2018 Statistics Final Exam Academic Success Programme Columbia University Instructor: Mark England

Total Page:16

File Type:pdf, Size:1020Kb

2018 Statistics Final Exam Academic Success Programme Columbia University Instructor: Mark England 2018 Statistics Final Exam Academic Success Programme Columbia University Instructor: Mark England Name: UNI: You have two hours to finish this exam. § Calculators are necessary for this exam. § Always show your working and thought process throughout so that even if you don’t get the answer totally correct I can give you as many marks as possible. § It has been a pleasure teaching you all. Try your best and good luck! Page 1 Question 1 True or False for the following statements? Warning, in this question (and only this question) you will get minus points for wrong answers, so it is not worth guessing randomly. For this question you do not need to explain your answer. a) Correlation values can only be between 0 and 1. False. Correlation values are between -1 and 1. b) The Student’s t distribution gives a lower percentage of extreme events occurring compared to a normal distribution. False. The student’s t distribution gives a higher percentage of extreme events. c) The units of the standard deviation of � are the same as the units of �. True. d) For a normal distribution, roughly 95% of the data is contained within one standard deviation of the mean. False. It is within two standard deviations of the mean e) For a left skewed distribution, the mean is larger than the median. False. This is a right skewed distribution. f) If an event � is independent of �, then � (�) = �(�|�) is always true. True. g) For a given confidence level, increasing the sample size of a survey should increase the margin of error. False. Increasing the sample size should reduce the margin of error. Question 2 I have five cards, and I label them 1, 2, 3, 4, 5. I pick cards out at random without replacing them. a) What is the probability of picking an odd numbered card on my first turn? Page 2 3/5 b) I pick two cards out and deal them on the table. What is the chance that if I add the cards together, they will add up to 3? Hint: think about the order! - - The only two ways of doing this are 1 followed by 2 ( × ) and 2 followed by 1 . 0 - - - - - - 1 - ( × ). Therefore, it would be × + × = = . 0 . 0 . 0 12 -2 c) I pick five cards out and deal them out on the table. What is the chance of picking the five cards out in ascending order (1, 2, 3, 4, 5)? - - - - - - There is only one way to do this: × × × × = . 0 3 1 - -12 Question 3 We can do a study and find that the length of time people spend in Westside Market on Broadway is normally distributed with a mean of 21 minutes and a standard deviation of 6 minutes. a) What percentage of people spend over half an hour in the market? 3251- We can utilise z scores: � = = 1.5. Consulting the z score table, that would 6 be the 93.32 percentile so above this would be 6.68%. b) Is it more likely that someone spends (i) under 12 minutes or (ii) over 28 in the market? For this we could calculate z-score but it is easier to just to see which is further from the mean, as it is symmetric. 12 minutes is 9 minutes below the mean and 28 minutes is only 7 minutes over the mean. So, it is over 28 minutes. c) What percentage of people spend between 22 and 25 minutes in the market? 1151- First, we find how many spent below 22 minutes: � = = 0.17 which gives 6 56.75 percentile. Then, we find out how many spent above 25 minutes: � = 1.51- = 0.67 which is the 74.86 percentile. Therefore we can find the difference 6 between them 74.86 − 56.75 which gives 18.11%. Page 3 Question 4 In January 2017, the city of Emeryville, California, introduced a minimum wage increase from $9 dollars to $13.50 per hour. The hope was that increasing the minimum wage would give people a ‘living wage’ and reduce the number of people who have to work multiple jobs to make ends meet. They ask you to investigate whether the new minimum wage has resulted in a significant decline in the percentage of those working multiple jobs. They know that, before the minimum wage increase, the percentage of people who were working multiple jobs was 7.1%. 18 months after the minimum wage increase, the city surveyed 1400 people and found that 83 were working multiple jobs. Use hypothesis testing to decide whether the increase in the minimum wage has resulted in less people having to work multiple jobs. You may decide upon a suitable confidence level which you wish to use, as long as it is clearly stated. Null hypothesis: � = 0.071 Alternate hypothesis � < 0.071 2.2G-×2.H1H � = 1400, �̂ = 0.059, �� = F = 0.0069 -022 Using 95% confidence level, and normal distribution with mean of 0.071 and a standard error of 0.0069. 2.2.H52.2G- Using z scores: � = = −1.8. 2.226H If we are using a 1 sided test the z score we need to be smaller than is -1.65. -1.8 is further away from zero than -1.65 so we can reject the null hypothesis. Therefore we can accept the alternate hypothesis, at 95% confidence level, to say that the increase of the minimum wage has reduced the percentage of people working multiple jobs. Question 5 Jill Stein, the green party candidate for president, won 1.01% of the popular vote in the 2016 general election. I randomly talk to 1000 Americans from across the country who voted in the election. I want to enlist a soccer team of Jill Stein voters, and for that I need 11 people. (Assume that every Jill Stein voter I ask says yes to being on my team). What is the chance that I have enough players to fill my team? You will need to be careful with decimal places in this question. Page 4 2.2-2-×2.HIHH � = 0.0101, �� = F = 0.00316 and we are asking what is the -222 -- probability of sampling this distribution to get over �̂ ≥ = 0.011. -222 2.2--52.2-2- We use z scores: = 0.28. Consulting the z score table, this is the 61st 2.223-6 percentile so to get greater than this we get 39% chance. Question 6 We want to know whether CC2 did better than CC1 in the midterm. The mean for CC1 (24 students) was 36.5 and the standard deviation was 7.5. The mean for CC2 (24 students) was 38.3 and the standard deviation was 9.4. Using the formulae from the back of the exam and the appropriate table, find out whether the average mark of CC2 was significantly higher than CC1. You will need to propose a null and alternative hypothesis and then test them at a certain confidence level of your choosing to arrive at a conclusion. Null hypothesis: �- = �1 Alternate hypothesis �1 > �- (24 − 1) × 56.25 + (24 − 1)88.36 � = P = 8.50 O 46 |36..53I.3| �∗ = = 0.73, For a df of 46, 95% confidence from t table gives 2.01. 0.73<2.01 so cannot T I..F TU reject null hypothesis. Classes performed the same. Page 5 Question 7 We have two data series, yearly salary �, and yearly rent cost for apartment �. Both values are in thousands of dollars and are from New York City. We know that the means and standard deviations are described by �̅ = 57.9, �X = 20.1, �Y = 21.0, �Z = 9.2) with correlation coefficient � = 0.75. We will then carry out linear regression analysis on these two data sets where we aim to predict the y values from the x values. a) What will the equation of the line of best fit (least squares regression) be in the form �\ = �-� + �2? �Z 9.2 �- = � = 0.75 × = 0.343 �X 20.1 �2 = 21.0 − 57.9 × 0.343 = 1.14 �\ = 0.343� + 1.14 b) What does �- represent? �- represents the slope of the line, this is the amount rent goes up as income goes up. c) What is the �1 of the data? What does this value tell us about the data? �1 is 0.56 and says that 56% of variation in rent can be explained by the linear relationship with income. d) I want to make one final check that a linear relationship is really appropriate for the data. What is the best way to do that? Make a scatter plot of residuals. Check there is no shape or change in spread. The residuals should be random. e) I remove a very large outlier from the data. Will this have a large effect on the line of best fit? This will have a large effect. Outliers can have an outsized effect on calculating the line of best fit because the square of the residual will be large. Page 6 Question 8 In January of this year, Pew Research surveyed 1,200 people on the issue of marijuana legalisation. They found that 61% of respondents favoured legalisation. This is up from 31% in 2000. a) Come up with 95% confidence intervals for the true current opinion of the whole country. Explain the meaning of your results in words. 2.6-×2.3H �̂ = 0.61, �� = F = 0.014 -122 Roughly the confidence interval will be plus or minus two standard deviations Between [58.2%, 63.8%]. In words: We have 95% confidence that the true opinion of the population will lie between 58.2% and 63.8%.
Recommended publications
  • Lesson 7: Measuring Variability for Skewed Distributions (Interquartile Range)
    NYS COMMON CORE MATHEMATICS CURRICULUM Lesson 7 M2 ALGEBRA I Lesson 7: Measuring Variability for Skewed Distributions (Interquartile Range) Student Outcomes . Students explain why a median is a better description of a typical value for a skewed distribution. Students calculate the 5-number summary of a data set. Students construct a box plot based on the 5-number summary and calculate the interquartile range (IQR). Students interpret the IQR as a description of variability in the data. Students identify outliers in a data distribution. Lesson Notes Distributions that are not symmetrical pose some challenges in students’ thinking about center and variability. The observation that the distribution is not symmetrical is straightforward. The difficult part is to select a measure of center and a measure of variability around that center. In Lesson 3, students learned that, because the mean can be affected by unusual values in the data set, the median is a better description of a typical data value for a skewed distribution. This lesson addresses what measure of variability is appropriate for a skewed data distribution. Students construct a box plot of the data using the 5-number summary and describe variability using the interquartile range. Classwork Exploratory Challenge 1/Exercises 1–3 (10 minutes): Skewed Data and Their Measure of Center Verbally introduce the data set as described in the introductory paragraph and dot plot shown below. Exploratory Challenge 1/Exercises 1–3: Skewed Data and Their Measure of Center Consider the following scenario. A television game show, Fact or Fiction, was cancelled after nine shows. Many people watched the nine shows and were rather upset when it was taken off the air.
    [Show full text]
  • Normal Probability Distribution
    MATH2114 Week 8, Thursday The Normal Distribution Slides Adopted & modified from “Business Statistics: A First Course” 7th Ed, by Levine, Szabat and Stephan, 2016 Pearson Education Inc. Objective/Goals To compute probabilities from the normal distribution How to use the normal distribution to solve practical problems To use the normal probability plot to determine whether a set of data is approximately normally distributed Continuous Probability Distributions A continuous variable is a variable that can assume any value on a continuum (can assume an uncountable number of values) thickness of an item time required to complete a task temperature of a solution height, in inches These can potentially take on any value depending only on the ability to precisely and accurately measure The Normal Distribution Bell Shaped f(X) Symmetrical Mean=Median=Mode σ Location is determined by X the mean, μ μ Spread is determined by the standard deviation, σ Mean The random variable has an = Median infinite theoretical range: = Mode + to The Normal Distribution Density Function The formula for the normal probability density function is 2 1 (X μ) 1 f(X) e 2 2π Where e = the mathematical constant approximated by 2.71828 π = the mathematical constant approximated by 3.14159 μ = the population mean σ = the population standard deviation X = any value of the continuous variable The Normal Distribution Density Function By varying 휇 and 휎, we obtain different normal distributions Image taken from Wikipedia, released under public domain The
    [Show full text]
  • L-Moments Based Assessment of a Mixture Model for Frequency Analysis of Rainfall Extremes
    Advances in Geosciences, 2, 331–334, 2006 SRef-ID: 1680-7359/adgeo/2006-2-331 Advances in European Geosciences Union Geosciences © 2006 Author(s). This work is licensed under a Creative Commons License. L-moments based assessment of a mixture model for frequency analysis of rainfall extremes V. Tartaglia, E. Caporali, E. Cavigli, and A. Moro Dipartimento di Ingegneria Civile, Universita` degli Studi di Firenze, Italy Received: 2 December 2004 – Revised: 2 October 2005 – Accepted: 3 October 2005 – Published: 23 January 2006 Abstract. In the framework of the regional analysis of hy- events in Tuscany (central Italy), a statistical methodology drological extreme events, a probabilistic mixture model, for parameter estimation of a probabilistic mixture model is given by a convex combination of two Gumbel distributions, here developed. The use of a probabilistic mixture model in is proposed for rainfall extremes modelling. The papers con- the regional analysis of rainfall extreme events, comes from cerns with statistical methodology for parameter estimation. the hypothesis that rainfall phenomenon arises from two in- With this aim, theoretical expressions of the L-moments of dependent populations (Becchi et al., 2000; Caporali et al., the mixture model are here defined. The expressions have 2002). The first population describes the basic components been tested on the time series of the annual maximum height corresponding to the more frequent events of low magnitude, of daily rainfall recorded in Tuscany (central Italy). controlled by the local morpho-climatic conditions. The sec- ond population characterizes the outlier component of more intense and rare events, which is here considered due to the aggregation at the large scale of meteorological phenomena.
    [Show full text]
  • Estimation of Quantiles and the Interquartile Range in Complex Surveys Carol Ann Francisco Iowa State University
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1987 Estimation of quantiles and the interquartile range in complex surveys Carol Ann Francisco Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Statistics and Probability Commons Recommended Citation Francisco, Carol Ann, "Estimation of quantiles and the interquartile range in complex surveys " (1987). Retrospective Theses and Dissertations. 11680. https://lib.dr.iastate.edu/rtd/11680 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. INFORMATION TO USERS While the most advanced technology has been used to photograph and reproduce this manuscript, the quality of the reproduction is heavily dependent upon the quality of the material submitted. For example: • Manuscript pages may have indistinct print. In such cases, the best available copy has been filmed. • Manuscripts may not always be complete. In such cases, a note will indicate that it is not possible to obtain missing pages. • Copyrighted material may have been removed from the manuscript. In such cases, a note will indicate the deletion. Oversize materials (e.g., maps, drawings, and charts) are photographed by sectioning the original, beginning at the upper left-hand corner and continuing from left to right in equal sections with small overlaps. Each oversize page is also filmed as one exposure and is available, for an additional charge, as a standard 35mm slide or as a 17"x 23" black and white photographic print.
    [Show full text]
  • Hand-Book on STATISTICAL DISTRIBUTIONS for Experimentalists
    Internal Report SUF–PFY/96–01 Stockholm, 11 December 1996 1st revision, 31 October 1998 last modification 10 September 2007 Hand-book on STATISTICAL DISTRIBUTIONS for experimentalists by Christian Walck Particle Physics Group Fysikum University of Stockholm (e-mail: [email protected]) Contents 1 Introduction 1 1.1 Random Number Generation .............................. 1 2 Probability Density Functions 3 2.1 Introduction ........................................ 3 2.2 Moments ......................................... 3 2.2.1 Errors of Moments ................................ 4 2.3 Characteristic Function ................................. 4 2.4 Probability Generating Function ............................ 5 2.5 Cumulants ......................................... 6 2.6 Random Number Generation .............................. 7 2.6.1 Cumulative Technique .............................. 7 2.6.2 Accept-Reject technique ............................. 7 2.6.3 Composition Techniques ............................. 8 2.7 Multivariate Distributions ................................ 9 2.7.1 Multivariate Moments .............................. 9 2.7.2 Errors of Bivariate Moments .......................... 9 2.7.3 Joint Characteristic Function .......................... 10 2.7.4 Random Number Generation .......................... 11 3 Bernoulli Distribution 12 3.1 Introduction ........................................ 12 3.2 Relation to Other Distributions ............................. 12 4 Beta distribution 13 4.1 Introduction .......................................
    [Show full text]
  • Descriptive Statistics
    NCSS Statistical Software NCSS.com Chapter 200 Descriptive Statistics Introduction This procedure summarizes variables both statistically and graphically. Information about the location (center), spread (variability), and distribution is provided. The procedure provides a large variety of statistical information about a single variable. Kinds of Research Questions The use of this module for a single variable is generally appropriate for one of four purposes: numerical summary, data screening, outlier identification (which sometimes is incorporated into data screening), and distributional shape. We will briefly discuss each of these now. Numerical Descriptors The numerical descriptors of a sample are called statistics. These statistics may be categorized as location, spread, shape indicators, percentiles, and interval estimates. Location or Central Tendency One of the first impressions that we like to get from a variable is its general location. You might think of this as the center of the variable on the number line. The average (mean) is a common measure of location. When investigating the center of a variable, the main descriptors are the mean, median, mode, and the trimmed mean. Other averages, such as the geometric and harmonic mean, have specialized uses. We will now briefly compare these measures. If the data come from the normal distribution, the mean, median, mode, and the trimmed mean are all equal. If the mean and median are very different, most likely there are outliers in the data or the distribution is skewed. If this is the case, the median is probably a better measure of location. The mean is very sensitive to extreme values and can be seriously contaminated by just one observation.
    [Show full text]
  • Measures of Variability
    CHAPTER 4 Measures of Variability Chapter 3 introduced descriptive statistics, the purpose of which is to numerically describe or summarize data for a Chapter Outline variable. As we learned in Chapter 3, a set of data is often 4.1 An Example From the described in terms of the most typical, common, or frequently Research: How Many “Sometimes” occurring score by using a measure of central tendency such in an “Always”? as the mode, median, or mean. Measures of central ten- 4.2 Understanding Variability dency numerically describe two aspects of a distribution of scores, modality and symmetry, based ondistribute what the scores 4.3 The Range have in common with each other. However, researchers • Strengths and weaknesses of also describe the amount of differences among the scores of a the range variable in analyzing and reportingor their data. This chapter 4.4 The Interquartile Range introduces measures of variability, which are designed to rep- • Strengths and weaknesses of resent the amount of differences in a distribution of data. In the interquartile range this chapter, we will present and evaluate different measures 4.5 The Variance (s2) of variability in terms of their relative strengths and weak- • Definitional formula for the nesses. As in the previous chapters, our presentation will variance depend heavily on the findings and analysis from a published • Computational formula for the research study.post, variance • Why not use the absolute value of the deviation in 4.1 An Example From the calculating the variance? Research: How Many • Why divide by N – 1 rather than “Sometimes” in an “Always”? N in calculating the variance? 4.6 The Standard Deviation (s) copy,Throughout our daily lives, we are constantly asked to fill out • Definitional formula for the questionnaires and surveys.
    [Show full text]
  • Measures of Spread
    community project encouraging academics to share statistics support resources All stcp resources are released under a Creative Commons licence stcp-gilchristsamuels-6 The following resources are associated: Descriptive Statistics – Measures of Middle Value Descriptive Statistics – Measures of Spread Research question type: Most What kind of variables: Nominal, ordinal and interval/scale Common Applications: Most quantitative studies Descriptive statistics are a way of summarising your data into a few values, often for comparative purposes. Commonly they are presented in tables, although appropriate charts are also used. They include measures of middle values (or central tendency), and measures of spread (or dispersion) of your data. The following is a rough guide to appropriate measures for level of data: Data level Middle value (central tendency) Dispersion (spread) Nominal Mode Ordinal Mode, median Range, interquartile range (IQR) Interval/scale Median, mean Variance, standard deviation Range The range is defined as the difference between the maximum and the minimum values of a set of data. Example Find the range for: 2, 6, 3, 9, 5, 6, 2, 6. The maximum value of the data is 9, and the minimum value is 2. Hence the range is 9 - 2 = 7. Interquartile range (IQR) Quartiles split a dataset into four quarters when the values are written in ascending order. !!! The lower quartile is 25% of the way through a data set. This is the th value, where n is the ! number of data values. (Note: Quarter values are found by starting with the lower value and adding either ¼ or ¾ of the difference. For example, the 2¼th value is the 2nd value plus ¼ of the 3rd value minus the 2nd).
    [Show full text]
  • Chapter 4 – Analyzing Skewed Quantitative Data Introduction: in Chapter 3, We Focused on Analyzing Bell Shaped (Normal) Data, but Many Data Sets Are Not Bell Shaped
    Chapter 4 – Analyzing Skewed Quantitative Data Introduction: In chapter 3, we focused on analyzing bell shaped (normal) data, but many data sets are not bell shaped. How do we analyze quantitative data when it is not bell shaped? The main issue is that the mean and standard deviations are not accurate and should not be used in the analysis. Then what statistics should we use? We will be introducing a new kind of graph that is specially designed for analyzing skewed data. It is called the “box and whisker plot” or “box plot” for short. When data sets are not bell shaped, we will focus on the median, quartiles, interquartile range and boxplots to measure center and spread. Quartiles are more accurate because they are based on the order of the numbers instead of distances and so are not as effected by the skewed shape and extremely unusual values. 143 Section 4A – Review of Shapes and Centers, Creating Histograms and Dot Plots with Technology Let us start by reviewing how to create dot plots and histograms with technology and determine the shape of a quantitative data set. Here are the directions for making dot plots and histograms in Statcato. Making a dot plot in Statcato: Graph => Dot plot => Pick a column => OK Making a histogram in Statcato: Graph => Histogram => Pick a column => Chose number of bins => OK Example 1 Use the women’s weight data (in pounds) from the Health data and Statcato to create a dot plot, histogram and boxplot. What is the shape of the data set? Plugging in the women’s weight data into Statcato, we can use the directions above to create the following graphs.
    [Show full text]
  • Outing the Outliers – Tails of the Unexpected
    Presented at the 2016 International Training Symposium: www.iceaaonline.com/bristol2016 Outing the Outliers – Tails of the Unexpected Outing the Outliers – Tails of the Unexpected Cynics would say that we can prove anything we want with statistics as it is all down to … a word (or two) from the Wise? interpretation and misinterpretation. To some extent this is true, but in truth it is all down to “The great tragedy of Science: the slaying making a judgement call based on the balance of of a beautiful hypothesis by an ugly fact” probabilities. Thomas Henry Huxley (1825-1895) British Biologist The problem with using random samples in estimating is that for a small sample size, the values could be on the “extreme” side, relatively speaking – like throwing double six or double one with a pair of dice, and nothing else for three of four turns. The more random samples we have the less likely (statistically speaking) we are to have all extreme values. So, more is better if we can get it, but sometimes it is a question of “We would if we could, but we can’t so we don’t!” Now we could argue that Estimators don’t use random values (because it sounds like we’re just guessing); we base our estimates on the “actuals” we have collected for similar tasks or activities. However, in the context of estimating, any “actual” data is in effect random because the circumstances that created those “actuals” were all influenced by a myriad of random factors. Anyone who has ever looked at the “actuals” for a repetitive task will know that there are variations in those values.
    [Show full text]
  • Problem Max. Points Your Points 1-10 10 11 10 12 3 13 4 14 18 15 8 16 7 17 14 Total 75
    M 140 Test 1 B Name__________________ (1 point) SHOW YOUR WORK FOR FULL CREDIT! Problem Max. Points Your Points 1-10 10 11 10 12 3 13 4 14 18 15 8 16 7 17 14 Total 75 Multiple choice questions (1 point each) For questions 1-3 consider the following: The distribution of scores on a certain statistics test is strongly skewed to the right. 1. How would the mean and the median compare for this distribution? a) mean < median b) mean = median c) mean > median d) Can’t tell without more information When a distribution skewed to the right, the “tail” is on the right. Since the tail “pulls” the mean, the mean is greater than the median. 2. Which set of measures of center and spread are more appropriate for the distribution of scores? a) Mean and standard deviation b) Median and interquartile range c) Mean and interquartile range d) Median and standard deviation For skewed distributions the median and the interquartile range are the appropriate measures of center and spread because they are resistant to outliers. 3. What does this suggest about the difficulty of the test? a) It was an easy test b) It was a hard test c) It wasn’t too hard or too easy d) It is impossible to tell If the distribution is skewed to the right, then most of the values are to the left, toward the lower values. Thus, more students got lower scores, which suggests that the test was hard. 4. Which one of the following variables is NOT categorical? a) whether or not an individual has glasses b) weight of an elephant c) religion d) social security number All the other variables are categorical—the average doesn’t make sense for those.
    [Show full text]
  • For Questions 1-4, Consider the Following Data Set: 4 3 6 2 8 8
    Math 225 Quiz 2 Name:_________________ For questions 1-4, consider the following data set: 4 3 6 2 8 8 1. What is the median of the data set? a. 4 b. 5 c. 2 d. 4.5 2. If I change one of the 8’s in the data set to 18, how would that change the median? a. The median would be higher b. The median would be lower c. The median would remain the same 3. If I change one of the 8’s in the data set to 18, how would that change the mean? a. The mean would be higher b. The mean would be lower c. The mean would remain the same 4. What is the mode of the data set? a. there is no mode b. 8 c. 2 5. For the distribution in the picture a. mean ≈ median b. mean < median c. mean > median d. Can’t tell the relationship of the mean and the median without looking at the data. 6. Which one of these is NOT true about the standard deviation? a. The standard deviation is zero only if the observations are all the same. b. The standard deviation has the same unit as the data. c. The standard deviation is used as a measure of spread when the mean is use as the measure of center. d. The standard deviation is resistant to outliers. 1 7. What percent of the data lies between the first quartile, Q 1, and the Median? a. About 50% b. About 25% c.
    [Show full text]