American Community Survey Design and Methodology

Total Page:16

File Type:pdf, Size:1020Kb

American Community Survey Design and Methodology Design and Methodology Issued April 2009 American Community Survey ACS-DM1 U.S. Department of Commerce U S C E N S U S B U R E A U Economics and Statistics Administration U.S. CENSUS BUREAU Helping You Make Informed Decisions ACKNOWLEDGMENTS The updating of the May 2006 unedited version of this technical report was conducted under the direction of Susan Schechter, Chief, American Community Survey Office. Deborah H. Griffin, Special Assistant to the Chief, American Community Survey Office, provided overall management and coordination. The American Community Survey program is under the direction of Arnold A. Jackson, Associate Director for Decennial Census, and Daniel H. Weinberg, Assistant Director for American Community Survey and Decennial Census. Major contributing authors for this updated 2008 report include Herman A. Alvarado, Mark E. Asiala, Lawrence M. Bates, Judy G. Belton, Grace L. Clemons, Kenneth B. Dawson, Deborah H. Griffin, James E. Hartman, Steven P. Hefter, Douglas W. Hillmer, Jennifer L. Holland, Cynthia Davis Hollingsworth, Todd R. Hughes, Karen E. King, Debra L. U. Klein, Pamela M. Klein, Alfredo Navarro, Susan Schechter, Nicholas M. Spanos, John G. Stiller, Anthony G. Tersine, Jr., Nancy K. Torrieri, Kai T. Wu, and Matthew A. Zimolzak. The U. S. Census Bureau is also grateful to staff from Mathematica Policy Research, Inc., who provided valuable comments and revisions to an earlier draft of this report. Assisting in the production of this report were Cheryl V. Chambers, Destiny D. Cusick, Susan L. Hostetter, Clive Richmond, and Sue Wood. The May 2006 unedited version was produced through the efforts of a number of individuals, primarily Mark E. Asiala, Lisa Blumerman, Sharon K. Boyer, Maryann M. Chapin, Thomas M. Coughlin, Barbara N. Diskin, Donald P. Fischer, Brian Gregory, Deborah H. Griffin, Wendy Davis Hicks, Douglas W. Hillmer, David L. Hubble, Agnes Kee, Susan P. Love, Lawrence McGinn, Marc Meyer, Alfredo Navarro, Joan B. Peacock, David Raglin, Nicholas M. Spanos, and Lynn Weidman. Catherine M. Raymond, Christine E. Geter, Crystal Wade, and Linda Chen, of the Administrative and Customer Services Division (ACSD), Francis Grailand Hall, Chief, provided publications and printing management, graphics design and composition, and editorial review for the print and electronic media. Claudette E. Bennett, Assistant Division Chief, and Wanda Cevis, Chief, Publications Services Branch, provided general direction and production management. Design and Methodology Issued April 2009 American Community Survey ACS-DM1 U.S. Department of Commerce Gary Locke, Secretary Vacant, Deputy Secretary Economics and Statistics Administration Vacant, Under Secretary for Economic Affairs U.S. CENSUS BUREAU Thomas L. Mesenbourg, Acting Director SUGGE S TED CITATION U.S. CENSUS BUREAU Design and Methodology American Community Survey U.S. Government Printing Office, Washington, DC, 2009. Economics and Statistics Administration Vacant, Under Secretary for Economic Affairs U.S. CENSUS BUREAU Thomas L. Mesenbourg, Acting Director Thomas L. Mesenbourg, Deputy Director and Chief Operating Officer Arnold A. Jackson, Associate Director for Decennial Census Daniel H. Weinberg, Assistant Director for ACS and Decennial Census Foreword The American Community Survey—A Revolution in Data Collection The American Community Survey (ACS) is the cornerstone of the U.S. Census Bureau’s effort to keep pace with the nation’s ever-increasing demands for timely and relevant data about popula- tion and housing characteristics. The new survey provides current demographic, social, economic, and housing information about America’s communities every year—information that until now was only available once a decade. Implementation of the ACS is viewed by many as the single most important change in the way detailed decennial census information is collected since 1940, when the Census Bureau introduced statistical sampling as a way to collect ‘‘long-form’’ data from a sample of households. The ACS and the reengineering of the decennial census will affect data users and the public for decades to come. Beginning with the survey’s full implementation in 2005, the ACS has replaced the census long-form questionnaire that was sent to about one-in-six addresses in Census 2000. As with the long form, information from the ACS will be used to administer federal and state pro- grams and distribute more than $300 billion a year in federal funds. Obtaining more current data throughout the decade from the ACS will have long-lasting value for policy and decision-making across federal, state, local, and tribal governments, the private sector, and virtually every local community in the nation. The Beginning. In 1994, the Census Bureau started developing what became the ACS with the idea of continuously measuring the characteristics of population and housing, instead of collect- ing the data only once a decade with each decennial census. Testing started in four counties across the country and with encouraging results, the testing expanded to 31 test sites by 1999. Realizing that a continuous program would also be collecting information during a decennial cen- sus, the sample was increased to about 800,000 addresses in 2000 and continued its demonstra- tion period through 2004. This was a national sample that yielded results for the country, states, and most geographic areas with 250,000 or more population. Comparing the 2000 ACS data with the results from the Census 2000 long form proved that the idea of a monthly survey was feasible and would generate quality data. With some changes to the sample design and other methodologies, the ACS was fully implemented in 2005 with a sample of three million addresses each year. A sample also was implemented in Puerto Rico, where the sur- vey is known as the Puerto Rico Community Survey (PRCS). In 2006, a sample of group quarters facilities was included so that estimates from the ACS and the PRCS would reflect complete char- acteristics of all community residents. Annual results will be available for all areas by 2010. Currently, the ACS publishes single- year data for all areas with populations of 65,000 or more. Among the roughly 7,000 areas that meet this threshold are all states, all congressional districts, more than 700 counties, and more than 500 places. Areas with populations less than 65,000 will require the use of multiyear esti- mates to reach an appropriate sample size for data publication. In 2008, the Census Bureau will begin releasing 3-year estimates for areas with populations greater than 20,000. And, we plan to release the first 5-year estimates for all census tracts and block groups starting in 2010. These multiyear estimates will be updated annually, with data published for the largest areas in both 1-, 3-, and 5-year formats, and for those meeting the 3-year threshold in both 3- and 5-year formats. Of course, even the smallest communities will be able to obtain ACS data based on 5-year esti- mates annually. The 2008 release of the ACS Design and Methodology Report. This ACS Design and Methodology Report is an update of the first unedited version that was released in 2006. We released that draft version because of the need to provide data users with information about the first full sample year of the survey. The version released in 2006 provided design and methodol- ogy information for the 2005 ACS only. ACS Design and Methodology Foreword iii U.S. Census Bureau This version of the ACS Design and Methodology Report includes updated information reflecting survey changes, modifications, and improvements through the end of 2007. Many portions of each chapter have been revised. We hope that data users find this report helpful and that it will aid in improving the public’s understanding of the ACS statistical design and the methods it uses. Success of the Program. The ACS program has been successful in large part because of the innovation and dedication of many people who have worked so hard to bring it to this point in time. With this publication of the ACS Design and Methodology Report, many individuals—both past and current—deserve special congratulations. From those early beginnings with a handful of designers, survey methodologists, and technical experts, through full implementation, countless individuals have contributed to the survey’s successful implementation. All of the primary survey activities are designed and managed by the staff at Census Bureau head- quarters in Suitland, MD, who continually strive to improve the accuracy of the ACS estimates, streamline its operations, analyze its data, conduct important research and evaluation to achieve greater efficiencies and effectiveness, and serve as educational resources and experts for the countless data users who come to the Census Bureau in need of technical assistance and help. In addition, the Census Bureau’s field partners provide many of the critical day-to-day activities that are the hub of the ACS existence. The ACS, which is the largest household survey conducted by the federal government, could not be accomplished without the dedication and effort of staff at the Census Bureau’s National Processing Center (NPC) in Jeffersonville, IN; the Census Bureau tele- phone call centers in Jeffersonville, IN; Hagerstown, MD; and Tucson, AZ; and the thousands of field representatives across the country who collect ACS data. In addition, the ACS field operations are run by Census Bureau survey managers in the NPC, telephone call centers and the twelve Regional Offices, all of whom add immeasurably to the smooth and efficient running of a very complex and demanding survey operation. Finally, the ACS would not have achieved its success without the continued cooperation of mil- lions of Americans who willingly provide the data that are collected each year. The data they pro- vide are invaluable and contribute daily to the survey’s exceptional accomplishments. Sincere thanks are extended to each and every respondent who took the time and effort to participate in this worthwhile endeavor.
Recommended publications
  • 811D Ecollomic Statistics Adrllillistra!Tioll
    811d Ecollomic Statistics Adrllillistra!tioll BUREAU THE CENSUS • I n i • I Charles G. Langham Issued 1973 U.S. D OF COM ERCE Frederick B. Dent. Secretary Social Economic Statistics Edward D. Administrator BU OF THE CENSUS Vincent P. Barabba, Acting Director Vincent Director Associate Director for Economic Associate Director for Statistical Standards and 11/1",1"\"/1,, DATA USER SERVICES OFFICE Robert B. Chief ACKNOWLEDGMENTS This report was in the Data User Services Office Charles G. direction of Chief, Review and many persons the Bureau. Library of Congress Card No.: 13-600143 SUGGESTED CiTATION U.S. Bureau of the Census. The Economic Censuses of the United by Charles G. longham. Working Paper D.C., U.S. Government Printing Office, 1B13 For sale by Publication Oistribution Section. Social and Economic Statistics Administration, Washington, D.C. 20233. Price 50 cents. N Page Economic Censuses in the 19th Century . 1 The First "Economic Censuses" . 1 Economic Censuses Discontinued, Resumed, and Augmented . 1 Improvements in the 1850 Census . 2 The "Kennedy Report" and the Civil War . • . 3 Economic Censuses and the Industrial Revolution. 4 Economic Censuses Adjust to the Times: The Censuses of 1880, 1890, and 1900 .........................•.. , . 4 Economic Censuses in the 20th Century . 8 Enumerations on Specialized Economic Topics, 1902 to 1937 . 8 Censuses of Manufacturing and Mineral Industries, 1905 to 1920. 8 Wartime Data Needs and Biennial Censuses of Manufactures. 9 Economic Censuses and the Great Depression. 10 The War and Postwar Developments: Economic Censuses Discontinued, Resumed, and Rescheduled. 13 The 1954 Budget Crisis. 15 Postwar Developments in Economic Census Taking: The Computer, and" Administrative Records" .
    [Show full text]
  • E-Survey Methodology
    Chapter I E-Survey Methodology Karen J. Jansen The Pennsylvania State University, USA Kevin G. Corley Arizona State University, USA Bernard J. Jansen The Pennsylvania State University, USA ABSTRACT With computer network access nearly ubiquitous in much of the world, alternative means of data col- lection are being made available to researchers. Recent studies have explored various computer-based techniques (e.g., electronic mail and Internet surveys). However, exploitation of these techniques requires careful consideration of conceptual and methodological issues associated with their use. We identify and explore these issues by defining and developing a typology of “e-survey” techniques in organiza- tional research. We examine the strengths, weaknesses, and threats to reliability, validity, sampling, and generalizability of these approaches. We conclude with a consideration of emerging issues of security, privacy, and ethics associated with the design and implications of e-survey methodology. INTRODUCTION 1999; Oppermann, 1995; Saris, 1991). Although research over the past 15 years has been mixed on For the researcher considering the use of elec- the realization of these benefits (Kiesler & Sproull, tronic surveys, there is a rapidly growing body of 1986; Mehta & Sivadas, 1995; Sproull, 1986; Tse, literature addressing design issues and providing Tse, Yin, Ting, Yi, Yee, & Hong, 1995), for the laundry lists of costs and benefits associated with most part, researchers agree that faster response electronic survey techniques (c.f., Lazar & Preece, times and decreased costs are attainable benefits, 1999; Schmidt, 1997; Stanton, 1998). Perhaps the while response rates differ based on variables three most common reasons for choosing an e-sur- beyond administration mode alone.
    [Show full text]
  • 2019 TIGER/Line Shapefiles Technical Documentation
    TIGER/Line® Shapefiles 2019 Technical Documentation ™ Issued September 2019220192018 SUGGESTED CITATION FILES: 2019 TIGER/Line Shapefiles (machine- readable data files) / prepared by the U.S. Census Bureau, 2019 U.S. Department of Commerce Economic and Statistics Administration Wilbur Ross, Secretary TECHNICAL DOCUMENTATION: Karen Dunn Kelley, 2019 TIGER/Line Shapefiles Technical Under Secretary for Economic Affairs Documentation / prepared by the U.S. Census Bureau, 2019 U.S. Census Bureau Dr. Steven Dillingham, Albert Fontenot, Director Associate Director for Decennial Census Programs Dr. Ron Jarmin, Deputy Director and Chief Operating Officer GEOGRAPHY DIVISION Deirdre Dalpiaz Bishop, Chief Andrea G. Johnson, Michael R. Ratcliffe, Assistant Division Chief for Assistant Division Chief for Address and Spatial Data Updates Geographic Standards, Criteria, Research, and Quality Monique Eleby, Assistant Division Chief for Gregory F. Hanks, Jr., Geographic Program Management Deputy Division Chief and External Engagement Laura Waggoner, Assistant Division Chief for Geographic Data Collection and Products 1-0 Table of Contents 1. Introduction ...................................................................................................................... 1-1 1. Introduction 1.1 What is a Shapefile? A shapefile is a geospatial data format for use in geographic information system (GIS) software. Shapefiles spatially describe vector data such as points, lines, and polygons, representing, for instance, landmarks, roads, and lakes. The Environmental Systems Research Institute (Esri) created the format for use in their software, but the shapefile format works in additional Geographic Information System (GIS) software as well. 1.2 What are TIGER/Line Shapefiles? The TIGER/Line Shapefiles are the fully supported, core geographic product from the U.S. Census Bureau. They are extracts of selected geographic and cartographic information from the U.S.
    [Show full text]
  • 2020 Census Barriers, Attitudes, and Motivators Study Survey Report
    2020 Census Barriers, Attitudes, and Motivators Study Survey Report A New Design for the 21st Century January 24, 2019 Version 2.0 Prepared by Kyley McGeeney, Brian Kriz, Shawnna Mullenax, Laura Kail, Gina Walejko, Monica Vines, Nancy Bates, and Yazmín García Trejo 2020 Census Research | 2020 CBAMS Survey Report Page intentionally left blank. ii 2020 Census Research | 2020 CBAMS Survey Report Table of Contents List of Tables ................................................................................................................................... iv List of Figures .................................................................................................................................. iv Executive Summary ......................................................................................................................... 1 Introduction ............................................................................................................................. 3 Background .............................................................................................................................. 5 CBAMS I ......................................................................................................................................... 5 CBAMS II ........................................................................................................................................ 6 2020 CBAMS Survey Climate ........................................................................................................
    [Show full text]
  • THE CENSUS in U.S. HISTORY Library of Congress of Library
    Bill of Rights Constitutional Rights in Action Foundation FALL 2019 Volume 35 No1 THE CENSUS IN U.S. HISTORY Library of Congress of Library A census taker talks to a group of women, men, and children in 1870. The Constitution requires that a census be taken every ten After the 1910 census, the House set the total num- years. This means counting all persons, citizens and ber of House seats at 435. Since then, when Congress noncitizens alike, in the United States. In addition to reapportions itself after each census, those states gain- conducting a population count, the census has evolved to collect massive amounts of information on the growth and ing population may pick up more seats in the House at development of the nation. the expense of states declining in population that have to lose seats. Why Do We Have a Census? Who is counted in apportioning seats in the House? The original purpose of the census was to determine The Constitution originally included “the whole Number the number of representatives each state is entitled to in of free persons” plus indentured servants but excluded the U.S. House of Representatives. The apportionment “Indians not taxed.” What about slaves? The North and (distribution) of seats in the House depends on the pop- South argued about this at the Constitutional Conven- ulation of each state. Every state is guaranteed at least tion, finally agreeing to the three-fifths compromise. one seat. Slaves would be counted in each census, but only three- After the first census in 1790, the House decided a fifths of the count would be included in a state’s popu- state was allowed one representative for each approxi- lation for the purpose of House apportionment.
    [Show full text]
  • Survey Nonresponse Bias and the Coronavirus Pandemic∗
    Coronavirus Infects Surveys, Too: Survey Nonresponse Bias and the Coronavirus Pandemic∗ Jonathan Rothbaum U.S. Census Bureau† Adam Bee U.S. Census Bureau‡ May 3, 2021 Abstract Nonresponse rates have been increasing in household surveys over time, increasing the potential of nonresponse bias. We make two contributions to the literature on nonresponse bias. First, we expand the set of data sources used. We use information returns filings (such as W-2's and 1099 forms) to identify individuals in respondent and nonrespondent households in the Current Population Survey Annual Social and Eco- nomic Supplement (CPS ASEC). We link those individuals to income, demographic, and socioeconomic information available in administrative data and prior surveys and the decennial census. We show that survey nonresponse was unique during the pan- demic | nonresponse increased substantially and was more strongly associated with income than in prior years. Response patterns changed by education, Hispanic origin, and citizenship and nativity. Second, We adjust for nonrandom nonresponse using entropy balance weights { a computationally efficient method of adjusting weights to match to a high-dimensional vector of moment constraints. In the 2020 CPS ASEC, nonresponse biased income estimates up substantially, whereas in other years, we do not find evidence of nonresponse bias in income or poverty statistics. With the sur- vey weights, real median household income was $68,700 in 2019, up 6.8 percent from 2018. After adjusting for nonresponse bias during the pandemic, we estimate that real median household income in 2019 was 2.8 percent lower than the survey estimate at $66,790. ∗This report is released to inform interested parties of ongoing research and to encourage discussion.
    [Show full text]
  • 2021 RHFS Survey Methodology
    2021 RHFS Survey Methodology Survey Design For purposes of this document, the following definitions are provided: • Building—a separate physical structure identified by the respondent containing one or more units. • Property—one or more buildings owned by a single entity (person, group, leasing company, and so on). For example, an apartment complex may have several buildings but they are owned as one property. Target population: All rental housing properties in the United States, circa 2020. Sampling frame: The RHFS sample frame is a single frame based on a subset of the 2019 American Housing Survey (AHS) sample units. The RHFS frame included all 2019 AHS sample units that were identified as: 1. Rented or occupied without payment of rent. 2. Units that are owner occupied and listed as “for sale or rent”. 3. Vacant units for rent, for rent or sale, or rented but not yet occupied. By design, the RHFS sample frame excluded public housing and transient housing types (i.e. boat, RV, van, other). Public housing units are identified in the AHS through a match with the Department of Housing and Urban Development (HUD) administrative records. The RHFS frame is derived from the AHS sample, which is itself composed of housing units derived from the Census Bureau Master Address File. The AHS sample frame excludes group quarters housing. Group quarters are places where people live or stay in a group living arrangement. Examples include dormitories, residential treatment centers, skilled nursing facilities, correctional facilities, military barracks, group homes, and maritime or military vessels. As such, all of these types of group quarters housing facilities are, by design, excluded from the RHFS.
    [Show full text]
  • A Meta-Analysis of the Effect of Concurrent Web Options on Mail Survey Response Rates
    When More Gets You Less: A Meta-Analysis of the Effect of Concurrent Web Options on Mail Survey Response Rates Jenna Fulton and Rebecca Medway Joint Program in Survey Methodology, University of Maryland May 19, 2012 Background: Mixed-Mode Surveys • Growing use of mixed-mode surveys among practitioners • Potential benefits for cost, coverage, and response rate • One specific mixed-mode design – mail + Web – is often used in an attempt to increase response rates • Advantages: both are self-administered modes, likely have similar measurement error properties • Two strategies for administration: • “Sequential” mixed-mode • One mode in initial contacts, switch to other in later contacts • Benefits response rates relative to a mail survey • “Concurrent” mixed-mode • Both modes simultaneously in all contacts 2 Background: Mixed-Mode Surveys • Growing use of mixed-mode surveys among practitioners • Potential benefits for cost, coverage, and response rate • One specific mixed-mode design – mail + Web – is often used in an attempt to increase response rates • Advantages: both are self-administered modes, likely have similar measurement error properties • Two strategies for administration: • “Sequential” mixed-mode • One mode in initial contacts, switch to other in later contacts • Benefits response rates relative to a mail survey • “Concurrent” mixed-mode • Both modes simultaneously in all contacts 3 • Mixed effects on response rates relative to a mail survey Methods: Meta-Analysis • Given mixed results in literature, we conducted a meta- analysis
    [Show full text]
  • A Computationally Efficient Method for Selecting a Split Questionnaire Design
    Iowa State University Capstones, Theses and Creative Components Dissertations Spring 2019 A Computationally Efficient Method for Selecting a Split Questionnaire Design Matthew Stuart Follow this and additional works at: https://lib.dr.iastate.edu/creativecomponents Part of the Social Statistics Commons Recommended Citation Stuart, Matthew, "A Computationally Efficient Method for Selecting a Split Questionnaire Design" (2019). Creative Components. 252. https://lib.dr.iastate.edu/creativecomponents/252 This Creative Component is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Creative Components by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. A Computationally Efficient Method for Selecting a Split Questionnaire Design Matthew Stuart1, Cindy Yu1,∗ Department of Statistics Iowa State University Ames, IA 50011 Abstract Split questionnaire design (SQD) is a relatively new survey tool to reduce response burden and increase the quality of responses. Among a set of possible SQD choices, a design is considered as the best if it leads to the least amount of information loss quantified by the Kullback-Leibler divergence (KLD) distance. However, the calculation of the KLD distance requires computation of the distribution function for the observed data after integrating out all the missing variables in a particular SQD. For a typical survey questionnaire with a large number of categorical variables, this computation can become practically infeasible. Motivated by the Horvitz-Thompson estima- tor, we propose an approach to approximate the distribution function of the observed in much reduced computation time and lose little valuable information when comparing different choices of SQDs.
    [Show full text]
  • Chi Square Survey Questionnaire
    Chi Square Survey Questionnaire andCoptic polo-neck and monoclonal Guido catholicizes Petey slenderize while unidirectional her bottom tigerishness Guthrie doffs encased her lamplighter and cricks documentarily necromantically. and interlinedMattery disinterestedly,queenly. Garcon offerable enskied andhis balderdashesmanufactural. trivializes mushily or needily after Aamir ethylate and rainproof Chi Square test of any Contingency Table because Excel. Comparing frequencies Chi-Square tests Manny Gimond. Are independent of squared test have two tests are the. The Chi Squared Test is a statistical test that already often carried out connect the start of they intended geographical investigation. OpenStax Statistics CH11THE CHI-SQUARE Top Hat. There are classified according to chi square survey questionnaire. You can only includes a questionnaire can take these. ANOVA Regression and Chi-Square Educational Research. T-Tests & Survey Analysis SurveyMonkey. Aids victims followed the survey analysis has a given by using likert? Square test of questionnaires, surveys frequently scared to watch horror movies too small? In short terms with are regression tests t-test ANOVA chi square and. What you calculate a survey solution is two columns of questionnaires, surveys frequently than to download reports! Using Cross Tabulation and Chi-Square The Survey Says. The Chi-Square Test for Independence Department of. And you'll must plug the research into a chi-square test for independence. Table 4a reports the responses to questions 213 in framework study survey. What output it mean look the chi square beauty is high? Completing the survey and surveys frequently ask them? Chi square test is rejected: the survey in surveys, is the population, explain that minority male and choose your dv the population of.
    [Show full text]
  • 2017 National Population Projections: Methodology and Assumptions
    Methodology, Assumptions, and Inputs for the 2017 National Population Projections September 2018 Erratum Note: The 2017 National Population Projections were revised after their original release date to correct an error in infant mortality rates. The files were removed from the website on August 1, 2018 and an erratum note posted. The error incorrectly calculated infant mortality rates, which erroneously caused an increase in the number of deaths projected in the total population. Correcting the error in infant mortality results in a decrease in the number of deaths and a slight increase in the total projected population in the revised series. The error did not affect the other two components of population change in the projections series (fertility and migration). Major demographic trends, such as an aging population and an increase in racial and ethnic diversity, remain unchanged. Table of Contents Introduction .......................................................................................................................................................................2 Methods...............................................................................................................................................................................2 Base Population ...........................................................................................................................................................2 Fertility and Mortality Denominators...................................................................................................................3
    [Show full text]
  • Proposed 2020 Census Data Products Plan
    Proposed 2020 Census Data Products Plan National Advisory Committee Spring 2019 Meeting Jason Devine Cynthia Hollingsworth Population Division Decennial Census Management Division May 2-3, 2019 Background • The Census Bureau has a long history of protecting information provided by respondents • Over the decades, more and more granular census data have been published • Advances in data science, more powerful computers, and externally accessible ‘big data’ – which contain a lot of personal information – has increased the risk of identifying individuals from published statistics • To mitigate this risk, the Census Bureau is transitioning to a new disclosure avoidance method called differential privacy 2 Background (cont.) • Our goal for the 2020 Census data products is to meet data user needs while implementing the new disclosure avoidance method - however, there are some challenges • We currently do not have solutions for protecting tabulations based on complex variables (characteristics of people within households), and variables with many possible values (detailed race/ethnicity) • We need your help in understanding what the “must-have” tables are both in terms of detail and geography to allow us to focus efforts on researching potential solutions to meet critical needs 3 Stakeholder Feedback • The primary way we collected feedback was through a July 2018 Federal Register notice, its extension, and associated outreach • Approximately 1,200 comments were received • Comments provided examples detailing legal, programmatic, or statistical needs for specific tables and geographies within the decennial products • We are using the Federal Register comments to inform the development of the proposed suite of 2020 data products 4 Examples of Federal Register Comments Received • Sex and age data used by Michigan Department of Education for School Aid and dollars for Michigan's school-aged population are distributed based on the numbers of persons in age cohorts.
    [Show full text]