“President” Landon and the 1936 Literary Digest Poll

Total Page:16

File Type:pdf, Size:1020Kb

“President” Landon and the 1936 Literary Digest Poll Downloaded from https://doi.org/10.1017/S014555320001035X Dominic Lusinchi https://www.cambridge.org/core “President” Landon and the 1936 Literary Digest Poll . IP address: Were Automobile and Telephone Owners to Blame? 170.106.40.139 The disastrous prediction of an Alf Landon victory in the 1936 presidential election , on by the Literary Digest poll is a landmark event in the history of American survey 01 Oct 2021 at 20:34:14 research in general and polling in particular. It marks both the demise of the straw poll, of which the Digest was the most conspicuous and well- regarded example, and the rise to prominence of the self- proclaimed “scientific” poll. Why did the Digest poll fail so miserably? One view has come to prevail over the years: because the Digest selected its sample primarily from telephone books and car registration lists and since these con- , subject to the Cambridge Core terms of use, available at tained, at the time, mostly well- to- do folks who would vote Republican, it is no wonder the magazine mistakenly predicted a Republican win. This “conventional explanation” has found its way into countless publications (scholarly and in the press) and college courses. It has been used to illustrate the disastrous effects of a poorly designed poll. But is it correct? Empirical evidence, in the form of a 1937 Gallup poll, shows that this “conventional explanation” is wrong, because voters with telephones and cars backed Franklin D. Roosevelt and because it was those who failed to participate in the poll (overwhelmingly supporters of Roosevelt) who were mainly responsible for the faulty prediction. Polling and survey research more generally are data collection technologies https://www.cambridge.org/core/terms widely used today in the social sciences. Many apprentices of these disci- plines will take a college course in research methods and be told what consti- tutes a good survey and how to recognize a bad one. Most likely, as an illus- Social Science History 36:1 (Spring 2012) DOI 10.1215/01455532-1461650 © 2012 by Social Science History Association . Downloaded from https://doi.org/10.1017/S014555320001035X 24 Social Science History tration of the latter category, the instructor will recount the story of the 1936 https://www.cambridge.org/core Literary Digest presidential poll: it predicted the wrong winner (Alf Lan- don, the Republican challenger), and, to top it all off, its forecast error was huge (nearly 20 points). They will also be told most probably that the reason for this fiasco was the magazine’s reliance on telephone directories and car registration lists to select its sample,1 thereby biasing it, in the midst of the Depression, in favor of the better- off portion of the electorate, which was bound to vote for the Republican nominee. IP address: Budding social scientists are not the only ones who will have heard this story. Just like baseball, polling is now entrenched in the American culture; 170.106.40.139 the general public is inundated with polls, especially during an election year. From time to time, this tale will appear in the press (e.g., Higgins 2010) and in other publications accessible to the lay public (e.g., Zunz 1998: 64). , on But is this account correct? And how did it come about? The answer 01 Oct 2021 at 20:34:14 to the first question will be the primary focus of this article. As for the sec- ond question, the explanation for what went wrong with the Digest poll was largely crafted by the new players in this arena: the so- called scientific poll takers (Archibald M. Crossley, George H. Gallup, and Elmo Roper). Indeed, the 1936 presidential election witnessed not only the demise of traditional , subject to the Cambridge Core terms of use, available at straw polls, chiefly as a result of the Digest poll’s massive blunder, but the emergence of “scientific” polling. The advent and development of these new polls was made possible by their use of novel sampling methods and by their relative success in predicting the outcome of the election but more specifi- cally by outperforming the Digest poll. Gallup, founder in 1935 of the American Institute of Public Opinion (AIPO), and his fellow pollsters (Crossley and Roper) had a major stake in this story. After their “triumph” over the Digest (Rusciano 2007: 315), they were particularly keen to explain the superiority of their approach and show why the old “methods” used by straw polls, of which the Digest was the most prestigious example, were fundamentally unsound. Crossley (1937: 27) put it most bluntly: “The Literary Digest method is outmoded.” In contrast, the https://www.cambridge.org/core/terms new “scientific” polling was capable, they believed, of rendering a true image of America. According to Roper (1940: 326), “Our purpose is to set up an America in microcosm.” The superior value of this methodology, as Gallup (1972a: 146) saw it, rested on what he called “scientific sampling proce- dures”2 that could yield a representative sample of the population of interest:3 specifically, it could produce a miniature version of the electorate.4 To pro- . Downloaded from https://doi.org/10.1017/S014555320001035X 1936 Literary Digest Poll 25 mote the new venture, Gallup had to demonstrate that straw polls in general, https://www.cambridge.org/core and the Digest in particular, were based on samples that were biased, that is, unrepresentative of the American electorate.5 All through his career Gallup reiterated his version of the failure of the 1936 Digest poll. His authority in the field of polling and survey research was so great that his interpretation has been repeated countless times to college students and others and has per- vaded textbooks and manuals on survey research methods and other schol- arly works as well as newspaper and magazine stories. Even in the face of . IP address: empirical evidence contradicting it (Squire 1988), this view persists. Building on previous investigations (Bryson 1976; Cahalan 1989; Caha- 170.106.40.139 lan and Meier 1939; Squire 1988), the focus of this article is to show that the available empirical evidence regarding the causes of this event does not support what has been called the “conventional explanation” (Erikson and , on Tedin 1981: 953). It will become clear that the Digest’s original list was no 01 Oct 2021 at 20:34:14 impediment to predict the correct winner. The reason for the magazine’s mistaken prognosis was that those who participated in the poll (the respon- dents) turned out heavily in favor of the Republican challenger, whereas most of the incumbent president’s supporters failed to return their straw ballots— this is referred to as nonresponse bias.6 , subject to the Cambridge Core terms of use, available at This article proceeds as follows. First, I describe how the conventional explanation emerged in the aftermath of the poll’s blunder and how pervasive this explanation is in the scholarly literature and in other publications that refer to this event. Then, I examine previous attempts to use the only avail- able empirical evidence about the Digest poll’s failure, and I reanalyze this evidence with techniques advocated by Adam J. Berinsky (2006) when survey results originate from quota- controlled polls of the 1930s. I also look at the response rate of various sociodemographic groups. Next, I discuss the short- comings of the Gallup poll of 1937 and see if they negate the conclusions of this reanalysis. Finally, I offer some general remarks about theDigest poll and propose lessons that can be drawn from it by social science researchers. https://www.cambridge.org/core/terms The Genesis of the Conventional Explanation and Its Diffusion A few days before the 1936 presidential election, in which the incumbent Roosevelt was being challenged by Landon, the Literary Digest, a weekly magazine with a surprisingly good track record in predicting the outcomes . Downloaded from https://doi.org/10.1017/S014555320001035X 26 Social Science History of previous electoral contests,7 published the results of its presidential poll: https://www.cambridge.org/core it predicted that Landon would defeat Roosevelt—the former would receive 54 percent of the total popular vote, while the incumbent president would get only 41 percent. In fact, Roosevelt obtained 61 percent of the vote (US Bureau of the Census 1970: 354). Although this severe blunder destroyed the credibility of the magazine, which went out of business two years later, it became a major milestone in the history of polling and survey research in the United States. IP address: What happened? The magazine mailed out in the neighborhood of 10 million “ballots” (postcards) based on a number of sources: the rosters of 170.106.40.139 clubs and associations, voter registration rolls, occupational records, clas- sified mail- order and city directories (Literary Digest 1936a: 3–4). But the bulk of its list was based on telephone directories and registers of automobile , on owners.8 Of the 10 million cards sent out, nearly 2.4 million were returned, 01 Oct 2021 at 20:34:14 for a response rate of roughly 24 percent. How is it possible that with a sample of that size the magazine blundered so badly? Much has been written about this event over the years, and many are those who have referred to it. For the purpose of this study, these commen- tators can be grouped into three categories. First, there are the protagonists: , subject to the Cambridge Core terms of use, available at those who lived through the event and had a direct stake in explaining what had gone wrong. These include the “scientific” pollsters (Crossley, Gallup, and Roper) who promoted the explanation that was to define the event for years to come.
Recommended publications
  • 13 Collecting Statistical Data
    13 Collecting Statistical Data 13.1 The Population 13.2 Sampling 13.3 Random Sampling 1.1 - 1 • Polls, studies, surveys and other data collecting tools collect data from a small part of a larger group so that we can learn something about the larger group. • This is a common and important goal of statistics: Learn about a large group by examining data from some of its members. 1.1 - 2 Data collections of observations (such as measurements, genders, survey responses) 1.1 - 3 Statistics is the science of planning studies and experiments, obtaining data, and then organizing, summarizing, presenting, analyzing, interpreting, and drawing conclusions based on the data 1.1 - 4 Population the complete collection of all individuals (scores, people, measurements, and so on) to be studied; the collection is complete in the sense that it includes all of the individuals to be studied 1.1 - 5 Census Collection of data from every member of a population Sample Subcollection of members selected from a population 1.1 - 6 A Survey • The practical alternative to a census is to collect data only from some members of the population and use that data to draw conclusions and make inferences about the entire population. • Statisticians call this approach a survey (or a poll when the data collection is done by asking questions). • The subgroup chosen to provide the data is called the sample, and the act of selecting a sample is called sampling. 1.1 - 7 A Survey • The first important step in a survey is to distinguish the population for which the survey applies (the target population) and the actual subset of the population from which the sample will be drawn, called the sampling frame.
    [Show full text]
  • Preface Chapter 1
    Notes Preface 1. Alfred Pearce Dennis, “Humanizing the Department of Commerce,” Saturday Evening Post, June 6, 1925, 8. 2. Herbert Hoover, Memoirs: The Cabinet and the Presidency, 1920–1930 (New York: Macmillan, 1952), 184. 3. Herbert Hoover, “The Larger Purposes of the Department of Commerce,” in “Republi- can National Committee, Brief Review of Activities and Policies of the Federal Executive Departments,” Bulletin No. 6, 1928, Herbert Hoover Papers, Campaign and Transition Period, Box 6, “Subject: Republican National Committee,” Hoover Presidential Library, West Branch, Iowa. 4. Herbert Hoover, “Responsibility of America for World Peace,” address before national con- vention of National League of Women Voters, Des Moines, Iowa, April 11, 1923, Bible no. 303, Hoover Presidential Library. 5. Bruce Bliven, “Hoover—And the Rest,” Independent, May 29, 1920, 275. Chapter 1 1. John W. Hallowell to Arthur (Hallowell?), November 21, 1918, Hoover Papers, Pre-Com- merce Period, Hoover Presidential Library, West Branch, Iowa, Box 6, “Hallowell, John W., 1917–1920”; Julius Barnes to Gertrude Barnes, November 27 and December 5, 1918, ibid., Box 2, “Barnes, Julius H., Nov. 27, 1918–Jan. 17, 1919”; Lewis Strauss, “Further Notes for Mr. Irwin,” ca. February 1928, Subject File, Lewis L. Strauss Papers, Hoover Presidential Library, West Branch, Iowa, Box 10, “Campaign of 1928: Campaign Literature, Speeches, etc., Press Releases, Speeches, etc., 1928 Feb.–Nov.”; Strauss, handwritten notes, December 1, 1918, ibid., Box 76, “Strauss, Lewis L., Diaries, 1917–19.” 2. The men who sailed with Hoover to Europe on the Olympic on November 18, 1918, were Julius Barnes, Frederick Chatfi eld, John Hallowell, Lewis Strauss, Robert Taft, and Alonzo Taylor.
    [Show full text]
  • George Gallup: Highlights of His Life and Work
    George Gallup: Highlights of His Life and Work PRE-BIOGRAPHY AND PRE-HISTORY Understanding the character of creative people and the significance of their achievements requires, in the first place, a study of their pre-biography, together with the pre-history of that branch of science, art or culture where their endeavors have taken place. In other words, the person’s biography is a product of both the genealogy and the environment in which his or her professional activity has taken place. The life and work of George Gallup in this sense are exemplary. In the first place, he belonged to a large family, whose members had vigorously participated in the development of the United States, and whose accomplishments and merits are recorded in the annals of the country. Secondly, even though the modern stage of public opinion research began with the pioneering work of George Gallup in 1935 and 1936, the study of electoral attitudes in the US had a long history prior to that. Accordingly, after examining the pre-biography of George Gallup, we will also review the pre-history of public opinion research. Tenth-Generation American For generations, the large Kollop family resided in Lotharingia (Lorraine). During the Middle Ages some of its descendants moved over to England, retaining the Gollop name. It is believed to have been forged from the German words Gott and Lobe, meaning respectively “God” and “praise”. Over time various spellings of the family name emerged: Gallop, Galloup, Galloupe, Gallupe, and Gollop, with the version prevalent in America becoming Gal- lup. A historical record has been preserved in England concerning John Gollop (born about 1440), who came ‘out of the North in the fifth year of the reign of Edward IV’ (1465).
    [Show full text]
  • Has Polling Enhanced Representation? Unearthing Evidence from the Literary Digest Issue Polls
    Studies in American Political Development, 21 (Spring 2007), 16–29. Has Polling Enhanced Representation? Unearthing Evidence from the Literary Digest Issue Polls David Karol, University of California, Berkeley How has representation changed over time in the Institutional reforms are not, however, the only United States? Has responsiveness to public opinion factors that can affect representation; technological waxed or waned among elected officials? What are change can also play a significant role. In fact, some the causes of such trends as we observe? Scholars scholars contend that the rise of scientific surveys have pursued these crucial questions in different since the 1930s has yielded more responsive govern- ways. Some explore earlier eras in search of the “elec- ment. According to this school of thought, polls toral connection”, i.e. the extent to which voters held provide recent cohorts of elected officials more accu- office-holders accountable for their actions and the rate assessments of public opinion than their prede- degree to which electoral concerns motivated poli- cessors enjoyed, which allows them to reflect their ticians’ behavior.1 Others explore the effects of insti- constituents’ views to a greater extent than the tutional changes such as the move to direct election politicians of yesteryear. Yet others doubt whether of senators or the “reapportionment revolution.”2 politicians were truly ignorant of public sentiment before the rise of the poll; nor is there much certainty regarding the level of current politicians’ understand- I thank Larry Bartels, Terri Bimes, Ben Bishin, Ben Fordham, ing of constituent opinion. Some also question John Geer, Brian Glenn, Susan Herbst, Mark Kayser, Brian Lawson, whether ignorance is at the root of elected officials’ Taeku Lee, Eileen McDonagh and Eric Plutzer for comments along frequent divergence from their constituents’ wishes.
    [Show full text]
  • AP Statistics
    AP Statistics 1 SAMPLING C H A P 11 The idea that the examination of a relatively small number of randomly selected individuals can furnish dependable information about the characteristics of a vast unseen universe is an idea so powerful that only familiarity makes it cease to be exciting. Helen Mary Walker (1891 - 1983) What is Sampling? 2 We want to know something about a population (a mean or proportion, for example), but it isn’t possible to go out and obtain the information from every member of the population (called a census). So we take a “random sample” from the population, obtain information about the sample, and then use the sample information to estimate what we would get if we could reach the entire population. What is Sampling? 3 The information we want to know about the population is called a parameter. The information we get from the sample is called a statistic. We use the statistic to provide an estimate of the parameter. Sample Population Notation 4 Name Statistic Parameter Meany (mu) Standard Deviations (sigma) Correlationr (rho) Regression coefficientb (beta) Proportionppˆ (pi -- but we will use ) Surveys 5 A survey is a common way to get information from a population of people. Follow your favorite news source and you will hear nearly every day the phrase “In a recent survey …” A survey generally consists of questions asked of a selected group (the sample) in order to obtain an estimate of the opinions of the population (U.S. voters, for example). Sampling Variability 6 Different samples will produce different results.
    [Show full text]
  • UC Santa Barbara UC Santa Barbara Electronic Theses and Dissertations
    UC Santa Barbara UC Santa Barbara Electronic Theses and Dissertations Title Campaign spheres in Latin America: How institutions affect digital media in presidential elections Permalink https://escholarship.org/uc/item/0026d2q5 Author Brandao, Francisco Publication Date 2017 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California UNIVERSITY OF CALIFORNIA Santa Barbara Campaign spheres in Latin America: How institutions affect digital media in presidential elections A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy in Political Science by Francisco de Assis Fernandes Brandão Junior Committee in charge: Professor Bruce Bimber, Chair Professor Kathleen Bruhn Professor Matto Mildenberger Professor Michael Stohl January 2018 The dissertation of Francisco Brandão Jr. is approved. ___________________________________________ Kathleen Bruhn ___________________________________________ Matto Mildenberger ___________________________________________ Michael Stohl ___________________________________________ Bruce Bimber, Committee Chair December 2017 Campaign spheres in Latin America: How institutions affect digital media in presidential elections Copyright © 2018 by Francisco de Assis Fernandes Brandão Junior iii Dedication In memory of the students who lost their lives in Isla Vista, on May 23 2014, praying that God brings peace and heals their families, friends, this great community and country. Let there be light! To my mother, who always wondered what would be her life if she hadn`t dropped Graduate School to raise her children. To my wife and daughter, who were together with me in this journey. iv Acknowledgements Some might think it is a truism when graduate students write on their dissertations how much they owe to their graduate advisors for the satisfaction of getting the job done.
    [Show full text]
  • Ericka Menchen-Trevino
    Ericka Menchen-Trevino 7/24/2020 c COPYRIGHT by Kurt Wirth July 23, 2020 ALL RIGHTS RESERVED PREDICTING CHANGES IN PUBLIC OPINION WITH TWITTER: WHAT SOCIAL MEDIA DATA CAN AND CAN'T TELL US ABOUT OPINION FORMATION by Kurt Wirth ABSTRACT With the advent of social media data, some researchers have claimed they have the potential to revolutionize the measurement of public opinion. Others have pointed to non-generalizable methods and other concerns to suggest that the role of social media data in the field is limited. Likewise, researchers remain split as to whether automated social media accounts, or bots, have the ability to influence con- versations larger than those with their direct audiences. This dissertation examines the relationship between public opinion as measured by random sample surveys, Twitter sentiment, and Twitter bot activity. Analyzing Twitter data on two topics, the president and the economy, as well as daily public polling data, this dissertation offers evidence that changes in Twitter sentiment of the president predict changes in public approval of the president fourteen days later. Likewise, it shows that changes in Twit- ter bot sentiment of both the president and the economy predict changes in overall Twitter sentiment on those topics between one and two days later. The methods also reveal a previously undiscovered phenomenon by which Twitter sentiment on a topic moves counter to polling approval of the topic at a seven-day interval. This dissertation also discusses the theoretical implications of various methods of calculating social media sentiment. Most importantly, its methods were pre-registered so as to maximize the generalizability of its findings and avoid data cherry-picking or overfitting.
    [Show full text]
  • A Rhetorical Study of the Public Speaking of Harold L. Ickes in the 1936 Presidential Campaign
    Louisiana State University LSU Digital Commons LSU Historical Dissertations and Theses Graduate School 1955 A Rhetorical Study of the Public Speaking of Harold L. Ickes in the 1936 Presidential Campaign. William Scott oblesN Louisiana State University and Agricultural & Mechanical College Follow this and additional works at: https://digitalcommons.lsu.edu/gradschool_disstheses Recommended Citation Nobles, William Scott, "A Rhetorical Study of the Public Speaking of Harold L. Ickes in the 1936 Presidential Campaign." (1955). LSU Historical Dissertations and Theses. 127. https://digitalcommons.lsu.edu/gradschool_disstheses/127 This Dissertation is brought to you for free and open access by the Graduate School at LSU Digital Commons. It has been accepted for inclusion in LSU Historical Dissertations and Theses by an authorized administrator of LSU Digital Commons. For more information, please contact [email protected]. A RHETORICAL STUDY OF THE PUBLIC SPEAKING OF HAROLD L. ICKES IN THE 1936 PRESIDENTIAL CAMPAIGN A Dissertation Submitted to the Graduate Faculty of the Louisiana State University and Agricultural and Mechanloal College in partial fulfillment of the requirements for the degree of Doctor of Philosophy in The Department of Speech by William Scott Nobles A., Southeastern Oklahoma State College, 19^7 M. A . , WeBtern Reserve University, 19^8 August, 1955 ACKNOWLEDGMENT The w r i t e r acknowledges the assistance of all those who h a v e contributed to the preparation of this dissertation. Particularly, he wishes to acknowledge his Indebtedness to his wife, Elaine, for her en­ couragement a n d assistance throughout the research and the preparation of the manuscript; to Professor Waldo W.
    [Show full text]
  • Social Media Analytics and the Role of Twitter in the 2014 South African General Election: a Case Study
    Social Media Analytics and the Role of Twitter in the 2014 South African General Election: A Case Study Asheen Singh (450623) Supervisor: Michael Mchunu A Dissertation submitted to the Faculty of Science, University of the Witwatersrand, Johannesburg, in fulfilment of the requirements for the degree of Master of Science. February 21, 2018 Declaration I declare that this Dissertation is my own, unaided work. It is being submitted for the Degree of Master of Science at the University of the Witwatersrand, Johannes- burg. It has not been submitted before for any degree or examination at any other University. Signature of candidate day of 2018 in 1 Please Note The views and opinions expressed in this document are those of the author and do not necessarily reflect the official policy or position of the University of the Wit- watersrand. The data used in this research was publicly available and collected through the Twitter API. The analysis of the data and the interpretation of the re- sults were carried out to the best of the authors ability, without any intention of favouring or being biased towards any political party. 2 Abstract Social network sites such as Twitter have created vibrant and diverse communities in which users express their opinions and views on a variety of topics such as poli- tics. Extensive research has been conducted in countries such as Ireland, Germany and the United States, in which text mining techniques have been used to obtain information from politically oriented tweets. The purpose of this research was to determine if text mining techniques can be used to uncover meaningful infor- mation from a corpus of political tweets collected during the 2014 South African General Election.
    [Show full text]
  • Polls Gone Wild
    Polls Gone Wild (From Probabilities: The Little Numbers that Rule Our Lives, Chapter 8.) Before the presidential election in 1936, the magazine Literary Digest pub- lished an opinion poll that predicted an easy win for Republican candidate Alf Landon over President Franklin D. Roosevelt: 57% for Landon and 43% for Roosevelt. The Digest had a solid reputation after having correctly pre- dicted the winner in each presidential election since 1916, and the poll was based on responses from 2.3 million people. Yes, you read correctly, two mil- lion three hundred thousand people! The election result? Roosevelt got 62% and Landon 38% (of those who voted for either of the two), which is one of the largest margins of victory in any presidential election. The Digest poll has gone down in history as the worst opinion poll, ever and the magazine went out of business shortly thereafter. How on earth could this happen? With 2.3 million people, the rule of thumb from above gives a margin of error of only 0.07%, so the predicted numbers ought to be almost certain. Did something happen that made people suddenly change their minds? No, the error is in the methods of the Digest. The reliability of the estimate as measured by the margin of error is only valid if we have a random sample, meaning that everybody is equally likely to be selected. In theory, if there had been a list of all voters and 2.3 million people had been chosen from this list, the prediction would have been very accurate.
    [Show full text]
  • Landon and the 1936 Literary Digest Poll
    Downloaded from https://doi.org/10.1017/S014555320001035X Dominic Lusinchi https://www.cambridge.org/core “President” Landon and the 1936 Literary Digest Poll . IP address: Were Automobile and Telephone Owners to Blame? 170.106.33.42 The disastrous prediction of an Alf Landon victory in the 1936 presidential election , on by the Literary Digest poll is a landmark event in the history of American survey 26 Sep 2021 at 14:16:30 research in general and polling in particular. It marks both the demise of the straw poll, of which the Digest was the most conspicuous and well- regarded example, and the rise to prominence of the self- proclaimed “scientific” poll. Why did the Digest poll fail so miserably? One view has come to prevail over the years: because the Digest selected its sample primarily from telephone books and car registration lists and since these con- , subject to the Cambridge Core terms of use, available at tained, at the time, mostly well- to- do folks who would vote Republican, it is no wonder the magazine mistakenly predicted a Republican win. This “conventional explanation” has found its way into countless publications (scholarly and in the press) and college courses. It has been used to illustrate the disastrous effects of a poorly designed poll. But is it correct? Empirical evidence, in the form of a 1937 Gallup poll, shows that this “conventional explanation” is wrong, because voters with telephones and cars backed Franklin D. Roosevelt and because it was those who failed to participate in the poll (overwhelmingly supporters of Roosevelt) who were mainly responsible for the faulty prediction.
    [Show full text]
  • Unit 17: Samples and Surveys
    Unit 17: Samples and Surveys Summary of Video Listen to a news broadcast, read a newspaper, open a magazine and you’ll find an item related to the results of a poll on some topic. But how do pollsters survey a population as large and diverse as the United States and wind up with a complete and unbiased picture of attitudes on a particular topic? Before selecting the sample to be surveyed, pollsters need to identify the population – the group of interest to the study. For example, that could be all registered voters, people in a particular age group, or households with a certain income. The characteristic of a population that we are interested in is called a parameter. We can’t determine the true value of a parameter unless we can examine the entire population, which isn’t usually possible. However, we can estimate the unknown parameter based on information collected from a sample, a subset of the population. Such an estimate is called a statistic. (Remember Parameters are for Populations and Statistics are for Samples.) The pollsters at the University of New Hampshire’s Survey Center conduct everything from academic research to political polls. They are experts at selecting samples to represent the attitudes and opinions of the whole population. For a public opinion survey they use random digit dialing to select households; they start with a random sample of households. However, another sampling stage is required because they talk to an individual at each house and not to everyone in the house. So, for example, they might ask to speak to the person in the house who has had the most recent birthday.
    [Show full text]