A Bibliography of Selected Statistical Methods and Development Related to Census 2000

Total Page:16

File Type:pdf, Size:1020Kb

A Bibliography of Selected Statistical Methods and Development Related to Census 2000 BUREAU OF THE CENSUS STATISTICAL RESEARCH DIVISION Statistical Research Report Series No. RR 2000/02 A BIBLIOGRAPHY OF SELECTED STATISTICAL METHODS AND DEVELOPMENT RELATED TO CENSUS 2000 Tommy Wright and Joyce Farmer Statistical Research Division Methodology and Standards Directorate U.S. Bureau of the Census Washington, D.C. 20233 Report Issued: August 10, 2000 OVERVIEW In an attempt to help facilitate the continuing national conversation on the supplemental use of sampling methods as a part of Census 2000, this report is offered and contains two main parts: PART I. SELECTED MOMENTS IN THE DEVELOPMENT OF PROBABILITY SAMPLING: Theory & Practice This part focuses briefly on probability sampling methodol- ogy while pointing to selected moments in its development as a serious tool for scientific inquiry. PART II. AN INDEX & THE LISTING OF AN ANNOTATED BIBLIOGRAPHY This part provides brief summaries of selected papers which might be of interest to anyone with an interest in beginning a technical background study of the methodology which helps form the foundation of the Census Bureau’s planned use of sampling and estimation to improve the count from Census 2000. This report is a major revision of the earlier issued report (Third Edition: May 1, 2000) under the same title. We are grateful to our colleagues: Hazel Beaton for her expert typing of this report as well as the three earlier drafts, Juanita Rasmann for editing the final draft, Don Malec for calling the 1786 Laplace paper to our attention, and Yves Thibaudeau who read the 1786 Laplace paper (in French) and verified for us that Laplace’s estimator can be viewed as similar to capture-recapture or dual system estimation. “This paper reports the results of research and analysis undertaken by Census Bureau staff. It has undergone a Census Bureau review more limited in scope than that given to official Census Bureau publications. This report is released to inform interested parties of ongoing research and to encourage discussion of work in progress.” i ii PART I characteristic, a sample, i.e., a subset or portion of the SELECTED MOMENTS population, may be selected to yield satisfactory information IN THE DEVELOPMENT OF regarding the particular population characteristic. The PROBABILITY SAMPLING: population characteristic is often a quantitative one. In such Theory & Practice cases, a statistic is computed using information collected from the small and more manageable sample, and its value Practically everyday we pick up the newspaper or tune in is used to estimate the unknown value of the population to a news broadcast and are bombarded with data in the form characteristic. Although we desire a sample that will of numbers, graphs, and tables. We see the results of a study provide a "good" estimate of the unknown value of the done by the Gallup Poll. Forest personnel can tell the number population characteristic, it is certainly conceivable that the of deer inhabiting a certain land area. Reliable estimates of sample information obtained could lead to a very incorrect world grain production can be made before harvest by use of estimate. satellite data. Nielson’s Ratings can tell the approximate number of people who watched television in a given week and Sampling and Nonsampling Errors the proportion who watched a particular program. Much of Error is the difference between the known value of the this information and more is made possible by an area of estimator from the sample and the (true but unknown) value statistics referred to as probability sampling or simply of the population characteristic. Error can occur due to sampling. sampling reasons and/or nonsampling reasons. Sampling error is the error that is caused by measuring only the Uses of Sampling sampling units instead of all of the population units. Sampling methods are used throughout the world by a variety Nonsampling error is the error that is caused by reasons of individuals, groups, and organizations, as well as by local, other than sampling. Examples of nonsampling errors state, and national governments. They are used successfully include failure to get responses from all of the sample units, in many fields including agriculture, business, defense, failure of the measuring device to operate properly, and economics, education, energy, environment, finance, health, failure to correctly process the sample data. industry, labor, natural resource management, demographics, and transportation. Some specific applications include taking It is well known that the magnitude of nonsampling opinion polls, election polls, and polls for rating TV programs; errors can far exceed the magnitude of sampling error in a surveying animal populations (particularly fish, deer, etc.) and given sample. Unfortunately while much has been written farms; taking a sample of buildings; taking air samples to about measuring and controlling sampling error, relatively monitor air quality; sampling to monitor traffic activity; little is known about the quantification and estimation of the sampling to estimate energy consumption; sampling to monitor magnitude of nonsampling errors. Current practice seeks to a nation’s economy; taking samples before marketing a new minimize both sampling and nonsampling errors. Small product; taking a sample for auditing or inventory purposes; samples where resources are used to implement high quality taking soil samples to measure radioactivity levels; sampling data collection methods which control and minimize to monitor employment; sampling to monitor education nonsampling errors along with efficient statistical tech- progress; and taking samples of products produced at a niques (sampling and estimation) that seek to minimize manufacturing plant to monitor output quality. sampling error is an attractive combination for success. "All scientific observation, whether statistical or not, is Probability Sampling based on sampling," says Stephan (1948).(1) "The earliest Probability sampling makes use of the laws of probability examples of sampling procedures are to be found in certain in the selection of the sample and in the construction of very ordinary human activities. The common practice of efficient estimators. With probability sampling, every taking a small part or portion for tasting or testing to deter- population unit has a known positive chance of being mine the characteristics of the whole precedes recorded selected for the sample. Probability sampling provides a history and is one of the roots from which sampling methodol- means for saying how good one believes an estimate is ogy stems..." relative to all the possible estimates from all of the possible samples. That is, probability allows us to extend results Population and Sample from the sample to the entire population. The current approach to sampling assumes a given finite collection of units, called a population. It is often the case A Census or a Sample that certain characteristics of the population are needed but When limited resources such as time and costs dictated that unknown. When examination of each and every unit in the a complete census was not possible, sampling has been an population is undesirable to know a particular population alternative. Historically, however, the application of iii sampling techniques has had its ups and downs, largely owing could begin by choosing districts, towns, parts of cities, to common misconceptions about sampling. streets, etc., to be followed by systematic, rather than probabilistic, choice of units (houses, families, individuals); The heart of these misconceptions seems to be a belief (2) there should be substantial sample sizes at all levels of that if one wants to know something about a given population, such a selection process; (3) the sample should be spread it is better to contact the entire population (a census) rather out in a variety of ways, primarily geographically, but in than only a sample of the population. As Kish (1979) (2) has other ways as well. For example, if a sample had a defi- pointed out, censuses, if done correctly, have the potential ciency of cattle farmers, he would add more of them. (5) advantage of providing precise, detailed, and credible information on all population units. On the other hand, 1896: Petersen presents a sampling methodology for samples have the advantage of providing richer, more com- estimating the size of a finite population. The Petersen plex, accurate, inexpensive, and timely information for a estimator provides the heuristic basis of most estimators of sample which can be extended to the entire population. wildlife population size; and from humble beginnings, a very large capture-recapture scientific literature has Indeed the joint judicious use of both sampling and developed. (6) census taking offers the best opportunity for greatest benefit. 1897: At a conference of Scandinavian statisticians held Sampling: Its Development in Stockholm, a conference resolution gives guarded Following are selected major moments in the theoretical and support for the representative method being promoted practical development of probability sampling methodology. by A.N. Kiaer. (5) 1802: P.S. Laplace uses sampling to estimate the popula- 1903: Randomization is proposed for use in sample tion of France as of September 22, 1802. Laplace persuaded selection. Lucien March, a French statistician, who, in the the French government to take a sample of the small adminis- discussion to Kiaer’s paper at the 1903 Berlin International trative districts known as communes and to count the total Statistical Institute meeting, was the first to introduce (with population y in the sample communes on September 22, 1802. caution) concepts related to the use of probability (i.e., From the known total number of registered births (birth randomization) in the selection of the sample.(5) registration was required) during the preceding year in these communes x and in the whole country 1906: Bowley presents a central limit theorem for random sampling. Arthur Lyon Bowley presents a paper y X, the ratio estimate X of the population of France could which seeks to give an empirical verification to a type of x be calculated.
Recommended publications
  • 811D Ecollomic Statistics Adrllillistra!Tioll
    811d Ecollomic Statistics Adrllillistra!tioll BUREAU THE CENSUS • I n i • I Charles G. Langham Issued 1973 U.S. D OF COM ERCE Frederick B. Dent. Secretary Social Economic Statistics Edward D. Administrator BU OF THE CENSUS Vincent P. Barabba, Acting Director Vincent Director Associate Director for Economic Associate Director for Statistical Standards and 11/1",1"\"/1,, DATA USER SERVICES OFFICE Robert B. Chief ACKNOWLEDGMENTS This report was in the Data User Services Office Charles G. direction of Chief, Review and many persons the Bureau. Library of Congress Card No.: 13-600143 SUGGESTED CiTATION U.S. Bureau of the Census. The Economic Censuses of the United by Charles G. longham. Working Paper D.C., U.S. Government Printing Office, 1B13 For sale by Publication Oistribution Section. Social and Economic Statistics Administration, Washington, D.C. 20233. Price 50 cents. N Page Economic Censuses in the 19th Century . 1 The First "Economic Censuses" . 1 Economic Censuses Discontinued, Resumed, and Augmented . 1 Improvements in the 1850 Census . 2 The "Kennedy Report" and the Civil War . • . 3 Economic Censuses and the Industrial Revolution. 4 Economic Censuses Adjust to the Times: The Censuses of 1880, 1890, and 1900 .........................•.. , . 4 Economic Censuses in the 20th Century . 8 Enumerations on Specialized Economic Topics, 1902 to 1937 . 8 Censuses of Manufacturing and Mineral Industries, 1905 to 1920. 8 Wartime Data Needs and Biennial Censuses of Manufactures. 9 Economic Censuses and the Great Depression. 10 The War and Postwar Developments: Economic Censuses Discontinued, Resumed, and Rescheduled. 13 The 1954 Budget Crisis. 15 Postwar Developments in Economic Census Taking: The Computer, and" Administrative Records" .
    [Show full text]
  • Computing the Oja Median in R: the Package Ojanp
    Computing the Oja Median in R: The Package OjaNP Daniel Fischer Karl Mosler Natural Resources Institute of Econometrics Institute Finland (Luke) and Statistics Finland University of Cologne and Germany School of Health Sciences University of Tampere Finland Jyrki M¨ott¨onen Klaus Nordhausen Department of Social Research Department of Mathematics University of Helsinki and Statistics Finland University of Turku Finland and School of Health Sciences University of Tampere Finland Oleksii Pokotylo Daniel Vogel Cologne Graduate School Institute of Complex Systems University of Cologne and Mathematical Biology Germany University of Aberdeen United Kingdom June 27, 2016 arXiv:1606.07620v1 [stat.CO] 24 Jun 2016 Abstract The Oja median is one of several extensions of the univariate median to the mul- tivariate case. It has many nice properties, but is computationally demanding. In this paper, we first review the properties of the Oja median and compare it to other multivariate medians. Afterwards we discuss four algorithms to compute the Oja median, which are implemented in our R-package OjaNP. Besides these 1 algorithms, the package contains also functions to compute Oja signs, Oja signed ranks, Oja ranks, and the related scatter concepts. To illustrate their use, the corresponding multivariate one- and C-sample location tests are implemented. Keywords Oja median, Oja signs, Oja signed ranks, Oja ranks, R,C,C++ 1 Introduction The univariate median is a popular location estimator. It is, however, not straightforward to generalize it to the multivariate case since no generalization known retains all properties of univariate estimator, and therefore different gen- eralizations emphasize different properties of the univariate median.
    [Show full text]
  • 2019 TIGER/Line Shapefiles Technical Documentation
    TIGER/Line® Shapefiles 2019 Technical Documentation ™ Issued September 2019220192018 SUGGESTED CITATION FILES: 2019 TIGER/Line Shapefiles (machine- readable data files) / prepared by the U.S. Census Bureau, 2019 U.S. Department of Commerce Economic and Statistics Administration Wilbur Ross, Secretary TECHNICAL DOCUMENTATION: Karen Dunn Kelley, 2019 TIGER/Line Shapefiles Technical Under Secretary for Economic Affairs Documentation / prepared by the U.S. Census Bureau, 2019 U.S. Census Bureau Dr. Steven Dillingham, Albert Fontenot, Director Associate Director for Decennial Census Programs Dr. Ron Jarmin, Deputy Director and Chief Operating Officer GEOGRAPHY DIVISION Deirdre Dalpiaz Bishop, Chief Andrea G. Johnson, Michael R. Ratcliffe, Assistant Division Chief for Assistant Division Chief for Address and Spatial Data Updates Geographic Standards, Criteria, Research, and Quality Monique Eleby, Assistant Division Chief for Gregory F. Hanks, Jr., Geographic Program Management Deputy Division Chief and External Engagement Laura Waggoner, Assistant Division Chief for Geographic Data Collection and Products 1-0 Table of Contents 1. Introduction ...................................................................................................................... 1-1 1. Introduction 1.1 What is a Shapefile? A shapefile is a geospatial data format for use in geographic information system (GIS) software. Shapefiles spatially describe vector data such as points, lines, and polygons, representing, for instance, landmarks, roads, and lakes. The Environmental Systems Research Institute (Esri) created the format for use in their software, but the shapefile format works in additional Geographic Information System (GIS) software as well. 1.2 What are TIGER/Line Shapefiles? The TIGER/Line Shapefiles are the fully supported, core geographic product from the U.S. Census Bureau. They are extracts of selected geographic and cartographic information from the U.S.
    [Show full text]
  • 2020 Census Barriers, Attitudes, and Motivators Study Survey Report
    2020 Census Barriers, Attitudes, and Motivators Study Survey Report A New Design for the 21st Century January 24, 2019 Version 2.0 Prepared by Kyley McGeeney, Brian Kriz, Shawnna Mullenax, Laura Kail, Gina Walejko, Monica Vines, Nancy Bates, and Yazmín García Trejo 2020 Census Research | 2020 CBAMS Survey Report Page intentionally left blank. ii 2020 Census Research | 2020 CBAMS Survey Report Table of Contents List of Tables ................................................................................................................................... iv List of Figures .................................................................................................................................. iv Executive Summary ......................................................................................................................... 1 Introduction ............................................................................................................................. 3 Background .............................................................................................................................. 5 CBAMS I ......................................................................................................................................... 5 CBAMS II ........................................................................................................................................ 6 2020 CBAMS Survey Climate ........................................................................................................
    [Show full text]
  • Annual Report of the Center for Statistical Research and Methodology Research and Methodology Directorate Fiscal Year 2017
    Annual Report of the Center for Statistical Research and Methodology Research and Methodology Directorate Fiscal Year 2017 Decennial Directorate Customers Demographic Directorate Customers Missing Data, Edit, Survey Sampling: and Imputation Estimation and CSRM Expertise Modeling for Collaboration Economic and Research Experimentation and Record Linkage Directorate Modeling Customers Small Area Simulation, Data Time Series and Estimation Visualization, and Seasonal Adjustment Modeling Field Directorate Customers Other Internal and External Customers ince August 1, 1933— S “… As the major figures from the American Statistical Association (ASA), Social Science Research Council, and new Roosevelt academic advisors discussed the statistical needs of the nation in the spring of 1933, it became clear that the new programs—in particular the National Recovery Administration—would require substantial amounts of data and coordination among statistical programs. Thus in June of 1933, the ASA and the Social Science Research Council officially created the Committee on Government Statistics and Information Services (COGSIS) to serve the statistical needs of the Agriculture, Commerce, Labor, and Interior departments … COGSIS set … goals in the field of federal statistics … (It) wanted new statistical programs—for example, to measure unemployment and address the needs of the unemployed … (It) wanted a coordinating agency to oversee all statistical programs, and (it) wanted to see statistical research and experimentation organized within the federal government … In August 1933 Stuart A. Rice, President of the ASA and acting chair of COGSIS, … (became) assistant director of the (Census) Bureau. Joseph Hill (who had been at the Census Bureau since 1900 and who provided the concepts and early theory for what is now the methodology for apportioning the seats in the U.S.
    [Show full text]
  • THE CENSUS in U.S. HISTORY Library of Congress of Library
    Bill of Rights Constitutional Rights in Action Foundation FALL 2019 Volume 35 No1 THE CENSUS IN U.S. HISTORY Library of Congress of Library A census taker talks to a group of women, men, and children in 1870. The Constitution requires that a census be taken every ten After the 1910 census, the House set the total num- years. This means counting all persons, citizens and ber of House seats at 435. Since then, when Congress noncitizens alike, in the United States. In addition to reapportions itself after each census, those states gain- conducting a population count, the census has evolved to collect massive amounts of information on the growth and ing population may pick up more seats in the House at development of the nation. the expense of states declining in population that have to lose seats. Why Do We Have a Census? Who is counted in apportioning seats in the House? The original purpose of the census was to determine The Constitution originally included “the whole Number the number of representatives each state is entitled to in of free persons” plus indentured servants but excluded the U.S. House of Representatives. The apportionment “Indians not taxed.” What about slaves? The North and (distribution) of seats in the House depends on the pop- South argued about this at the Constitutional Conven- ulation of each state. Every state is guaranteed at least tion, finally agreeing to the three-fifths compromise. one seat. Slaves would be counted in each census, but only three- After the first census in 1790, the House decided a fifths of the count would be included in a state’s popu- state was allowed one representative for each approxi- lation for the purpose of House apportionment.
    [Show full text]
  • Survey Nonresponse Bias and the Coronavirus Pandemic∗
    Coronavirus Infects Surveys, Too: Survey Nonresponse Bias and the Coronavirus Pandemic∗ Jonathan Rothbaum U.S. Census Bureau† Adam Bee U.S. Census Bureau‡ May 3, 2021 Abstract Nonresponse rates have been increasing in household surveys over time, increasing the potential of nonresponse bias. We make two contributions to the literature on nonresponse bias. First, we expand the set of data sources used. We use information returns filings (such as W-2's and 1099 forms) to identify individuals in respondent and nonrespondent households in the Current Population Survey Annual Social and Eco- nomic Supplement (CPS ASEC). We link those individuals to income, demographic, and socioeconomic information available in administrative data and prior surveys and the decennial census. We show that survey nonresponse was unique during the pan- demic | nonresponse increased substantially and was more strongly associated with income than in prior years. Response patterns changed by education, Hispanic origin, and citizenship and nativity. Second, We adjust for nonrandom nonresponse using entropy balance weights { a computationally efficient method of adjusting weights to match to a high-dimensional vector of moment constraints. In the 2020 CPS ASEC, nonresponse biased income estimates up substantially, whereas in other years, we do not find evidence of nonresponse bias in income or poverty statistics. With the sur- vey weights, real median household income was $68,700 in 2019, up 6.8 percent from 2018. After adjusting for nonresponse bias during the pandemic, we estimate that real median household income in 2019 was 2.8 percent lower than the survey estimate at $66,790. ∗This report is released to inform interested parties of ongoing research and to encourage discussion.
    [Show full text]
  • The Theory of the Design of Experiments
    The Theory of the Design of Experiments D.R. COX Honorary Fellow Nuffield College Oxford, UK AND N. REID Professor of Statistics University of Toronto, Canada CHAPMAN & HALL/CRC Boca Raton London New York Washington, D.C. C195X/disclaimer Page 1 Friday, April 28, 2000 10:59 AM Library of Congress Cataloging-in-Publication Data Cox, D. R. (David Roxbee) The theory of the design of experiments / D. R. Cox, N. Reid. p. cm. — (Monographs on statistics and applied probability ; 86) Includes bibliographical references and index. ISBN 1-58488-195-X (alk. paper) 1. Experimental design. I. Reid, N. II.Title. III. Series. QA279 .C73 2000 001.4 '34 —dc21 00-029529 CIP This book contains information obtained from authentic and highly regarded sources. Reprinted material is quoted with permission, and sources are indicated. A wide variety of references are listed. Reasonable efforts have been made to publish reliable data and information, but the author and the publisher cannot assume responsibility for the validity of all materials or for the consequences of their use. Neither this book nor any part may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopying, microfilming, and recording, or by any information storage or retrieval system, without prior permission in writing from the publisher. The consent of CRC Press LLC does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific permission must be obtained in writing from CRC Press LLC for such copying. Direct all inquiries to CRC Press LLC, 2000 N.W.
    [Show full text]
  • Statistical Theory
    Statistical Theory Prof. Gesine Reinert November 23, 2009 Aim: To review and extend the main ideas in Statistical Inference, both from a frequentist viewpoint and from a Bayesian viewpoint. This course serves not only as background to other courses, but also it will provide a basis for developing novel inference methods when faced with a new situation which includes uncertainty. Inference here includes estimating parameters and testing hypotheses. Overview • Part 1: Frequentist Statistics { Chapter 1: Likelihood, sufficiency and ancillarity. The Factoriza- tion Theorem. Exponential family models. { Chapter 2: Point estimation. When is an estimator a good estima- tor? Covering bias and variance, information, efficiency. Methods of estimation: Maximum likelihood estimation, nuisance parame- ters and profile likelihood; method of moments estimation. Bias and variance approximations via the delta method. { Chapter 3: Hypothesis testing. Pure significance tests, signifi- cance level. Simple hypotheses, Neyman-Pearson Lemma. Tests for composite hypotheses. Sample size calculation. Uniformly most powerful tests, Wald tests, score tests, generalised likelihood ratio tests. Multiple tests, combining independent tests. { Chapter 4: Interval estimation. Confidence sets and their con- nection with hypothesis tests. Approximate confidence intervals. Prediction sets. { Chapter 5: Asymptotic theory. Consistency. Asymptotic nor- mality of maximum likelihood estimates, score tests. Chi-square approximation for generalised likelihood ratio tests. Likelihood confidence regions. Pseudo-likelihood tests. • Part 2: Bayesian Statistics { Chapter 6: Background. Interpretations of probability; the Bayesian paradigm: prior distribution, posterior distribution, predictive distribution, credible intervals. Nuisance parameters are easy. 1 { Chapter 7: Bayesian models. Sufficiency, exchangeability. De Finetti's Theorem and its intepretation in Bayesian statistics. { Chapter 8: Prior distributions. Conjugate priors.
    [Show full text]
  • Chapter 5 Statistical Inference
    Chapter 5 Statistical Inference CHAPTER OUTLINE Section 1 Why Do We Need Statistics? Section 2 Inference Using a Probability Model Section 3 Statistical Models Section 4 Data Collection Section 5 Some Basic Inferences In this chapter, we begin our discussion of statistical inference. Probability theory is primarily concerned with calculating various quantities associated with a probability model. This requires that we know what the correct probability model is. In applica- tions, this is often not the case, and the best we can say is that the correct probability measure to use is in a set of possible probability measures. We refer to this collection as the statistical model. So, in a sense, our uncertainty has increased; not only do we have the uncertainty associated with an outcome or response as described by a probability measure, but now we are also uncertain about what the probability measure is. Statistical inference is concerned with making statements or inferences about char- acteristics of the true underlying probability measure. Of course, these inferences must be based on some kind of information; the statistical model makes up part of it. Another important part of the information will be given by an observed outcome or response, which we refer to as the data. Inferences then take the form of various statements about the true underlying probability measure from which the data were obtained. These take a variety of forms, which we refer to as types of inferences. The role of this chapter is to introduce the basic concepts and ideas of statistical inference. The most prominent approaches to inference are discussed in Chapters 6, 7, and 8.
    [Show full text]
  • Identification of Empirical Setting
    Chapter 1 Identification of Empirical Setting Purpose for Identifying the Empirical Setting The purpose of identifying the empirical setting of research is so that the researcher or research manager can appropriately address the following issues: 1) Understand the type of research to be conducted, and how it will affect the design of the research investigation, 2) Establish clearly the goals and objectives of the research, and how these translate into testable research hypothesis, 3) Identify the population from which inferences will be made, and the limitations of the sample drawn for purposes of the investigation, 4) To understand the type of data obtained in the investigation, and how this will affect the ensuing analyses, and 5) To understand the limits of statistical inference based on the type of analyses and data being conducted, and this will affect final study conclusions. Inputs/Assumptions Assumed of the Investigator It is assumed that several important steps in the process of conducting research have been completed before using this part of the manual. If any of these stages has not been completed, then the researcher or research manager should first consult Volume I of this manual. 1) A general understanding of research concepts introduced in Volume I including principles of scientific inquiry and the overall research process, 2) A research problem statement with accompanying research objectives, and 3) Data or a data collection plan. Outputs/Products of this Chapter After consulting this chapter, the researcher or research
    [Show full text]
  • Statistical Inference: Paradigms and Controversies in Historic Perspective
    Jostein Lillestøl, NHH 2014 Statistical inference: Paradigms and controversies in historic perspective 1. Five paradigms We will cover the following five lines of thought: 1. Early Bayesian inference and its revival Inverse probability – Non-informative priors – “Objective” Bayes (1763), Laplace (1774), Jeffreys (1931), Bernardo (1975) 2. Fisherian inference Evidence oriented – Likelihood – Fisher information - Necessity Fisher (1921 and later) 3. Neyman- Pearson inference Action oriented – Frequentist/Sample space – Objective Neyman (1933, 1937), Pearson (1933), Wald (1939), Lehmann (1950 and later) 4. Neo - Bayesian inference Coherent decisions - Subjective/personal De Finetti (1937), Savage (1951), Lindley (1953) 5. Likelihood inference Evidence based – likelihood profiles – likelihood ratios Barnard (1949), Birnbaum (1962), Edwards (1972) Classical inference as it has been practiced since the 1950’s is really none of these in its pure form. It is more like a pragmatic mix of 2 and 3, in particular with respect to testing of significance, pretending to be both action and evidence oriented, which is hard to fulfill in a consistent manner. To keep our minds on track we do not single out this as a separate paradigm, but will discuss this at the end. A main concern through the history of statistical inference has been to establish a sound scientific framework for the analysis of sampled data. Concepts were initially often vague and disputed, but even after their clarification, various schools of thought have at times been in strong opposition to each other. When we try to describe the approaches here, we will use the notions of today. All five paradigms of statistical inference are based on modeling the observed data x given some parameter or “state of the world” , which essentially corresponds to stating the conditional distribution f(x|(or making some assumptions about it).
    [Show full text]