Analysis of Variance Example: Friend Or Pet to • Compare Means for More Than 2 Groups

Total Page:16

File Type:pdf, Size:1020Kb

Analysis of Variance Example: Friend Or Pet to • Compare Means for More Than 2 Groups Announcements: • Special office hours for final exam: Friday, Jason will have 1-3 (instead of 1-2) Monday (just before exam) I will have 10-noon. Exam is Monday, 1:30 – 3:30. Analysis • Homework assigned today and Wed not due, but Chapter 16 solutions on website. of • Final exam review sheets posted on web (in list of lectures). Variance • Form and instructions for disputing points will be sent by email. Watch for it if you plan to do that. Fill out and bring to final exam to hand in. Homework: Chapter 16: 1, 7, 8, 16, 17 (not due) 1 ANOVA = Analysis of variance Example: Friend or Pet to • Compare means for more than 2 groups. Relieve Stress? • We have k independent samples and measure a quantitative variable on all units in all k • Randomized experiment using 45 women samples. volunteers who said they love dogs. • We want to know if the population means are • Each woman assigned to do a stressful task: all equal (null hypothesis) or if at least one is – 15 did it alone different (alternative hypothesis). – 15 did it with a good friend present • This is called one-way ANOVA because we – 15 did it with their dog present analyze variability in order to compare • Response variable = heart rate means. 3 4 16.1 Comparing Means Step 1 for the example (defining hypotheses) with an ANOVA F-Test H0: µ1 = µ2 = µ3 H0: µ1 = µ2 = … = µk Ha: The means are not all equal. Ha: The means are not all equal. Example: Parameters of interest are the population Step 2 (In general): mean heart rates for all such women if they were to do The test statistic is called an F-statistic. In words, it the task under the 3 conditions: is defined as: µ1: if doing the task alone. Variation among sample means µ if doing the task with good friend present. F = 2: Natural variation within groups µ3: if doing the task with dog present. 5 6 1 Variation among sample means Assumptions for the F-Test F = Natural variation within groups • Samples are independent random samples. • Distribution of response variable is a normal curve Variation among sample means is 0 if all within each population (but ok as long as large n). k sample means are equal and gets larger • Different populations may have different means. the more spread out they are. • All populations have same standard deviation, σ. If large enough => evidence at least one e.g. How k = 3 populations might look … population mean is different from others => reject null hypothesis. p-value found using an F-distribution (more later) 7 8 Conditions for Using the F-Test Notation for Summary Statistics k = number of groups • F-statistic can be used if data are not extremely xi , si, and ni are the mean, standard deviation, skewed, there are no extreme outliers, and group and sample size for the ith sample group standard deviations are not markedly different. N = total sample size (N = n1 + n2 + … + nk) • Tests based on F-statistic are valid for data with skewness or outliers if sample sizes are large. Example: Friends, Pets and Stress • A rough criterion for standard deviations is that Three different conditions => k = 3 the largest of the sample standard deviations n1 = n2 = n3 = 15; so N = 45 should not be more than twice as large as the xxx= 82.52, ==91.33, 73.48 smallest of the sample standard deviations. 123 sss123===9.24, 8.34, 9.97 9 10 Example, continued Conditions for Using the F-Test: Do friends, pets, or neither reduce stress more Does the example qualify? than when alone? Boxplots of sample data: • F-statistic can be used if data are not extremely Possible outlier 100 skewed, there are no extreme outliers. 9In the example there is one possible outlier 90 but with a sample of only 15 it is hard to tell. 80 • A rough criterion for standard deviations is that Heart rate Heart the largest of the sample standard deviations 70 should not be more than twice as large as the 60 smallest of the sample standard deviations. Alone Friend Pet 9That condition is clearly met. The sample Treatment groups 11 standard deviations s are very similar. 12 2 16.2 Details of the F Statistic Measuring variation between groups: for Analysis of Variance How far apart are the means? Fundamental concept: the variation among the data values in the overall sample can be separated into: Sum of squares for groups = SS Groups (1) differences between group means 2 (2) natural variation among observations within a group SS Groups = ∑ groups ni ()xi − x Total variation = Numerator of F-statistic = mean square for groups Variation between groups + Variation within groups SS Groups MS Groups = “ANOVA Table” displays this information in k −1 summary form, and gives F statistic and p-value. 13 14 Measuring variation within groups: Measuring Total Variation How variable are the individuals? Total sum of squares = SS Total = SSTO Sum of squared errors = SS Error SS Errors = ()()n −1 s 2 2 ∑groups i i SS Total = x − x ∑values ()ij Denominator of F-statistic = mean square error SS Error MSE = N − k SS Total = SS Groups + SS Error Pooled standard deviation: s p = MSE Measures internal variability within each group. 15 16 General Format of a Example: Stress, Friends and Pets One-Way ANOVA Table H0: µ1 = µ2 = µ3 Ha: The means are not all equal. One-way ANOVA: Heart rate versus treatment group Source DF SS MS F P Group 2 2387.7 1193.8 14.08 0.000 Error 42 3561.3 84.8 Total 44 5949.0 The F-statistic is 14.08 and p-value is 0.000... p-value so small => reject H0 and accept Ha. Conclude there are differences among population means. MORE LATER ABOUT FINDING P-VALUE. 17 18 3 Conclusion in Context: 95% Confidence Intervals •The population mean heart rates would differ if we for the Population Means subjected all people similar to our volunteers to the 3 conditions (alone, good friend, pet dog). In one-way analysis of variance, a confidence •Now we want to know which one(s) differ! interval for a population mean µι is Individual 95% confidence intervals (next slide for s p formula): x ± t* Individual 95% CIs For Mean Based on i Pooled StDev ni Group Mean --+---------+---------+---------+------- Alone 82.524 (------*------) Friend 91.325 (-----*------) where s p = MSE and t* is from Table A.2: Pet 73.483 (------*------) --+---------+---------+---------+------- t* is such that the confidence level is the probability 70.0 77.0 84.0 91.0 between -t* and t* in a t-distribution with df = N – k. 19 20 Multiple Comparisons Multiple Comparisons, continued Multiple comparisons: Problem is that each C.I. has Many statistical tests or C.I.s => increased risk 95% confidence, but we want overall 95% confidence. of making at least one type I error (erroneously Can do multiple C.I.s and/or tests at once. rejecting a null hypothesis). Several procedures Most common: all pairwise comparisons of means. to control the overall family type I error rate or overall family confidence level. Ways to make inferences about each pair of means: • Family error rate for set of significance tests is • Significance test to assess if two means probability of making one or more type I errors when significantly differ. more than one significance test is done. • Confidence interval for difference computed and • Family confidence level for procedure used to create a if 0 is not in the interval, there is a statistically set of confidence intervals is the proportion of times all significant difference. intervals in set capture their true parameter values. 21 22 Tukey method: Family confidence level of 0.95 The Family of F-Distributions • Skewed distributions with minimum value of 0. Confidence intervals for µi − µj (differences in means) • Specific F-distribution indicated by two parameters Difference is Alone subtracted from: called degrees of freedom: numerator degrees of Lower Center Upper freedom and denominator degrees of freedom. Friend 2.015 8.801 15.587 (Friend – Alone) Pet -15.827 -9.041 -2.255 (Pet – Alone) • In one-way ANOVA, Difference is Friend subtracted from: numerator df = k – 1, Lower Center Upper and Pet -24.628 -17.842 -11.056 (Pet – Friend) denominator df = N – k None of the confidence intervals cover 0! That • Looks similar to chi- indicates that all 3 population means differ. square distributions Ordering is Pet < Alone < Friend 23 24 4 Determining the p-Value Example: Stress, Friends and Pets Statistical Software reports the p-value in output. Reported F-statistic was F = 14.08 and p-value < 0.000 Table A.4 provides critical values (to find rejection region) for 1% and 5% significance levels. N = 15 women: num df = k – 1 = 3 – 1 = 2 • If the F-statistic is > than the 5% critical value, den df = N – k = 45 – 3 = 42 the p-value < 0.05. • If the F-statistic is > than the 1% critical value, the p-value < 0.01 . Table A.4 with df of (2, 42), closest available is (2, 40): • If the F-statistic is between the 1% and 5% critical The 5% critical value is 3.23. values, the p-value is between 0.01 and 0.05. The 1% critical value is 5.18 and The F-statistic was much larger, so p-value < 0.01. 25 26 Example 16.1 Seat Location and GPA Example 16.1 Seat Location and GPA (cont) Q: Do best students sit in the front of a classroom? • The boxplot showed two outliers in the group Data on seat location and GPA for n = 384 students; of students who typically sit in the middle of 88 sit in front, 218 in middle, 78 in back a classroom, but there are 218 students in that group so these outliers don’t have much influence on the results.
Recommended publications
  • Hypothesis Testing and Likelihood Ratio Tests
    Hypottthesiiis tttestttiiing and llliiikellliiihood ratttiiio tttesttts Y We will adopt the following model for observed data. The distribution of Y = (Y1, ..., Yn) is parameter considered known except for some paramett er ç, which may be a vector ç = (ç1, ..., çk); ç“Ç, the paramettter space. The parameter space will usually be an open set. If Y is a continuous random variable, its probabiiillliiittty densiiittty functttiiion (pdf) will de denoted f(yy;ç) . If Y is y probability mass function y Y y discrete then f(yy;ç) represents the probabii ll ii tt y mass functt ii on (pmf); f(yy;ç) = Pç(YY=yy). A stttatttiiistttiiicalll hypottthesiiis is a statement about the value of ç. We are interested in testing the null hypothesis H0: ç“Ç0 versus the alternative hypothesis H1: ç“Ç1. Where Ç0 and Ç1 ¶ Ç. hypothesis test Naturally Ç0 § Ç1 = ∅, but we need not have Ç0 ∞ Ç1 = Ç. A hypott hesii s tt estt is a procedure critical region for deciding between H0 and H1 based on the sample data. It is equivalent to a crii tt ii call regii on: a critical region is a set C ¶ Rn y such that if y = (y1, ..., yn) “ C, H0 is rejected. Typically C is expressed in terms of the value of some tttesttt stttatttiiistttiiic, a function of the sample data. For µ example, we might have C = {(y , ..., y ): y – 0 ≥ 3.324}. The number 3.324 here is called a 1 n s/ n µ criiitttiiicalll valllue of the test statistic Y – 0 . S/ n If y“C but ç“Ç 0, we have committed a Type I error.
    [Show full text]
  • Statistical Analysis 8: Two-Way Analysis of Variance (ANOVA)
    Statistical Analysis 8: Two-way analysis of variance (ANOVA) Research question type: Explaining a continuous variable with 2 categorical variables What kind of variables? Continuous (scale/interval/ratio) and 2 independent categorical variables (factors) Common Applications: Comparing means of a single variable at different levels of two conditions (factors) in scientific experiments. Example: The effective life (in hours) of batteries is compared by material type (1, 2 or 3) and operating temperature: Low (-10˚C), Medium (20˚C) or High (45˚C). Twelve batteries are randomly selected from each material type and are then randomly allocated to each temperature level. The resulting life of all 36 batteries is shown below: Table 1: Life (in hours) of batteries by material type and temperature Temperature (˚C) Low (-10˚C) Medium (20˚C) High (45˚C) 1 130, 155, 74, 180 34, 40, 80, 75 20, 70, 82, 58 2 150, 188, 159, 126 136, 122, 106, 115 25, 70, 58, 45 type Material 3 138, 110, 168, 160 174, 120, 150, 139 96, 104, 82, 60 Source: Montgomery (2001) Research question: Is there difference in mean life of the batteries for differing material type and operating temperature levels? In analysis of variance we compare the variability between the groups (how far apart are the means?) to the variability within the groups (how much natural variation is there in our measurements?). This is why it is called analysis of variance, abbreviated to ANOVA. This example has two factors (material type and temperature), each with 3 levels. Hypotheses: The 'null hypothesis' might be: H0: There is no difference in mean battery life for different combinations of material type and temperature level And an 'alternative hypothesis' might be: H1: There is a difference in mean battery life for different combinations of material type and temperature level If the alternative hypothesis is accepted, further analysis is performed to explore where the individual differences are.
    [Show full text]
  • 5. Dummy-Variable Regression and Analysis of Variance
    Sociology 740 John Fox Lecture Notes 5. Dummy-Variable Regression and Analysis of Variance Copyright © 2014 by John Fox Dummy-Variable Regression and Analysis of Variance 1 1. Introduction I One of the limitations of multiple-regression analysis is that it accommo- dates only quantitative explanatory variables. I Dummy-variable regressors can be used to incorporate qualitative explanatory variables into a linear model, substantially expanding the range of application of regression analysis. c 2014 by John Fox Sociology 740 ° Dummy-Variable Regression and Analysis of Variance 2 2. Goals: I To show how dummy regessors can be used to represent the categories of a qualitative explanatory variable in a regression model. I To introduce the concept of interaction between explanatory variables, and to show how interactions can be incorporated into a regression model by forming interaction regressors. I To introduce the principle of marginality, which serves as a guide to constructing and testing terms in complex linear models. I To show how incremental I -testsareemployedtotesttermsindummy regression models. I To show how analysis-of-variance models can be handled using dummy variables. c 2014 by John Fox Sociology 740 ° Dummy-Variable Regression and Analysis of Variance 3 3. A Dichotomous Explanatory Variable I The simplest case: one dichotomous and one quantitative explanatory variable. I Assumptions: Relationships are additive — the partial effect of each explanatory • variable is the same regardless of the specific value at which the other explanatory variable is held constant. The other assumptions of the regression model hold. • I The motivation for including a qualitative explanatory variable is the same as for including an additional quantitative explanatory variable: to account more fully for the response variable, by making the errors • smaller; and to avoid a biased assessment of the impact of an explanatory variable, • as a consequence of omitting another explanatory variables that is relatedtoit.
    [Show full text]
  • Analysis of Variance and Analysis of Variance and Design of Experiments of Experiments-I
    Analysis of Variance and Design of Experimentseriments--II MODULE ––IVIV LECTURE - 19 EXPERIMENTAL DESIGNS AND THEIR ANALYSIS Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur 2 Design of experiment means how to design an experiment in the sense that how the observations or measurements should be obtained to answer a qqyuery inavalid, efficient and economical way. The desigggning of experiment and the analysis of obtained data are inseparable. If the experiment is designed properly keeping in mind the question, then the data generated is valid and proper analysis of data provides the valid statistical inferences. If the experiment is not well designed, the validity of the statistical inferences is questionable and may be invalid. It is important to understand first the basic terminologies used in the experimental design. Experimental unit For conducting an experiment, the experimental material is divided into smaller parts and each part is referred to as experimental unit. The experimental unit is randomly assigned to a treatment. The phrase “randomly assigned” is very important in this definition. Experiment A way of getting an answer to a question which the experimenter wants to know. Treatment Different objects or procedures which are to be compared in an experiment are called treatments. Sampling unit The object that is measured in an experiment is called the sampling unit. This may be different from the experimental unit. 3 Factor A factor is a variable defining a categorization. A factor can be fixed or random in nature. • A factor is termed as fixed factor if all the levels of interest are included in the experiment.
    [Show full text]
  • This Is Dr. Chumney. the Focus of This Lecture Is Hypothesis Testing –Both What It Is, How Hypothesis Tests Are Used, and How to Conduct Hypothesis Tests
    TRANSCRIPT: This is Dr. Chumney. The focus of this lecture is hypothesis testing –both what it is, how hypothesis tests are used, and how to conduct hypothesis tests. 1 TRANSCRIPT: In this lecture, we will talk about both theoretical and applied concepts related to hypothesis testing. 2 TRANSCRIPT: Let’s being the lecture with a summary of the logic process that underlies hypothesis testing. 3 TRANSCRIPT: It is often impossible or otherwise not feasible to collect data on every individual within a population. Therefore, researchers rely on samples to help answer questions about populations. Hypothesis testing is a statistical procedure that allows researchers to use sample data to draw inferences about the population of interest. Hypothesis testing is one of the most commonly used inferential procedures. Hypothesis testing will combine many of the concepts we have already covered, including z‐scores, probability, and the distribution of sample means. To conduct a hypothesis test, we first state a hypothesis about a population, predict the characteristics of a sample of that population (that is, we predict that a sample will be representative of the population), obtain a sample, then collect data from that sample and analyze the data to see if it is consistent with our hypotheses. 4 TRANSCRIPT: The process of hypothesis testing begins by stating a hypothesis about the unknown population. Actually we state two opposing hypotheses. The first hypothesis we state –the most important one –is the null hypothesis. The null hypothesis states that the treatment has no effect. In general the null hypothesis states that there is no change, no difference, no effect, and otherwise no relationship between the independent and dependent variables.
    [Show full text]
  • The Scientific Method: Hypothesis Testing and Experimental Design
    Appendix I The Scientific Method The study of science is different from other disciplines in many ways. Perhaps the most important aspect of “hard” science is its adherence to the principle of the scientific method: the posing of questions and the use of rigorous methods to answer those questions. I. Our Friend, the Null Hypothesis As a science major, you are probably no stranger to curiosity. It is the beginning of all scientific discovery. As you walk through the campus arboretum, you might wonder, “Why are trees green?” As you observe your peers in social groups at the cafeteria, you might ask yourself, “What subtle kinds of body language are those people using to communicate?” As you read an article about a new drug which promises to be an effective treatment for male pattern baldness, you think, “But how do they know it will work?” Asking such questions is the first step towards hypothesis formation. A scientific investigator does not begin the study of a biological phenomenon in a vacuum. If an investigator observes something interesting, s/he first asks a question about it, and then uses inductive reasoning (from the specific to the general) to generate an hypothesis based upon a logical set of expectations. To test the hypothesis, the investigator systematically collects data, either with field observations or a series of carefully designed experiments. By analyzing the data, the investigator uses deductive reasoning (from the general to the specific) to state a second hypothesis (it may be the same as or different from the original) about the observations.
    [Show full text]
  • Hypothesis Testing – Examples and Case Studies
    Hypothesis Testing – Chapter 23 Examples and Case Studies Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 23.1 How Hypothesis Tests Are Reported in the News 1. Determine the null hypothesis and the alternative hypothesis. 2. Collect and summarize the data into a test statistic. 3. Use the test statistic to determine the p-value. 4. The result is statistically significant if the p-value is less than or equal to the level of significance. Often media only presents results of step 4. Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 2 23.2 Testing Hypotheses About Proportions and Means If the null and alternative hypotheses are expressed in terms of a population proportion, mean, or difference between two means and if the sample sizes are large … … the test statistic is simply the corresponding standardized score computed assuming the null hypothesis is true; and the p-value is found from a table of percentiles for standardized scores. Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 3 Example 2: Weight Loss for Diet vs Exercise Did dieters lose more fat than the exercisers? Diet Only: sample mean = 5.9 kg sample standard deviation = 4.1 kg sample size = n = 42 standard error = SEM1 = 4.1/ √42 = 0.633 Exercise Only: sample mean = 4.1 kg sample standard deviation = 3.7 kg sample size = n = 47 standard error = SEM2 = 3.7/ √47 = 0.540 measure of variability = [(0.633)2 + (0.540)2] = 0.83 Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 4 Example 2: Weight Loss for Diet vs Exercise Step 1.
    [Show full text]
  • THE ONE-SAMPLE Z TEST
    10 THE ONE-SAMPLE z TEST Only the Lonely Difficulty Scale ☺ ☺ ☺ (not too hard—this is the first chapter of this kind, but youdistribute know more than enough to master it) or WHAT YOU WILL LEARN IN THIS CHAPTERpost, • Deciding when the z test for one sample is appropriate to use • Computing the observed z value • Interpreting the z value • Understandingcopy, what the z value means • Understanding what effect size is and how to interpret it not INTRODUCTION TO THE Do ONE-SAMPLE z TEST Lack of sleep can cause all kinds of problems, from grouchiness to fatigue and, in rare cases, even death. So, you can imagine health care professionals’ interest in seeing that their patients get enough sleep. This is especially the case for patients 186 Copyright ©2020 by SAGE Publications, Inc. This work may not be reproduced or distributed in any form or by any means without express written permission of the publisher. Chapter 10 ■ The One-Sample z Test 187 who are ill and have a real need for the healing and rejuvenating qualities that sleep brings. Dr. Joseph Cappelleri and his colleagues looked at the sleep difficul- ties of patients with a particular illness, fibromyalgia, to evaluate the usefulness of the Medical Outcomes Study (MOS) Sleep Scale as a measure of sleep problems. Although other analyses were completed, including one that compared a treat- ment group and a control group with one another, the important analysis (for our discussion) was the comparison of participants’ MOS scores with national MOS norms. Such a comparison between a sample’s mean score (the MOS score for par- ticipants in this study) and a population’s mean score (the norms) necessitates the use of a one-sample z test.
    [Show full text]
  • Tests of Hypotheses Using Statistics
    Tests of Hypotheses Using Statistics Adam Massey¤and Steven J. Millery Mathematics Department Brown University Providence, RI 02912 Abstract We present the various methods of hypothesis testing that one typically encounters in a mathematical statistics course. The focus will be on conditions for using each test, the hypothesis tested by each test, and the appropriate (and inappropriate) ways of using each test. We conclude by summarizing the di®erent tests (what conditions must be met to use them, what the test statistic is, and what the critical region is). Contents 1 Types of Hypotheses and Test Statistics 2 1.1 Introduction . 2 1.2 Types of Hypotheses . 3 1.3 Types of Statistics . 3 2 z-Tests and t-Tests 5 2.1 Testing Means I: Large Sample Size or Known Variance . 5 2.2 Testing Means II: Small Sample Size and Unknown Variance . 9 3 Testing the Variance 12 4 Testing Proportions 13 4.1 Testing Proportions I: One Proportion . 13 4.2 Testing Proportions II: K Proportions . 15 4.3 Testing r £ c Contingency Tables . 17 4.4 Incomplete r £ c Contingency Tables Tables . 18 5 Normal Regression Analysis 19 6 Non-parametric Tests 21 6.1 Tests of Signs . 21 6.2 Tests of Ranked Signs . 22 6.3 Tests Based on Runs . 23 ¤E-mail: [email protected] yE-mail: [email protected] 1 7 Summary 26 7.1 z-tests . 26 7.2 t-tests . 27 7.3 Tests comparing means . 27 7.4 Variance Test . 28 7.5 Proportions . 28 7.6 Contingency Tables .
    [Show full text]
  • Understanding Statistical Hypothesis Testing: the Logic of Statistical Inference
    Review Understanding Statistical Hypothesis Testing: The Logic of Statistical Inference Frank Emmert-Streib 1,2,* and Matthias Dehmer 3,4,5 1 Predictive Society and Data Analytics Lab, Faculty of Information Technology and Communication Sciences, Tampere University, 33100 Tampere, Finland 2 Institute of Biosciences and Medical Technology, Tampere University, 33520 Tampere, Finland 3 Institute for Intelligent Production, Faculty for Management, University of Applied Sciences Upper Austria, Steyr Campus, 4040 Steyr, Austria 4 Department of Mechatronics and Biomedical Computer Science, University for Health Sciences, Medical Informatics and Technology (UMIT), 6060 Hall, Tyrol, Austria 5 College of Computer and Control Engineering, Nankai University, Tianjin 300000, China * Correspondence: [email protected]; Tel.: +358-50-301-5353 Received: 27 July 2019; Accepted: 9 August 2019; Published: 12 August 2019 Abstract: Statistical hypothesis testing is among the most misunderstood quantitative analysis methods from data science. Despite its seeming simplicity, it has complex interdependencies between its procedural components. In this paper, we discuss the underlying logic behind statistical hypothesis testing, the formal meaning of its components and their connections. Our presentation is applicable to all statistical hypothesis tests as generic backbone and, hence, useful across all application domains in data science and artificial intelligence. Keywords: hypothesis testing; machine learning; statistics; data science; statistical inference 1. Introduction We are living in an era that is characterized by the availability of big data. In order to emphasize the importance of this, data have been called the ‘oil of the 21st Century’ [1]. However, for dealing with the challenges posed by such data, advanced analysis methods are needed.
    [Show full text]
  • ANALYSIS of VARIANCE and MISSING OBSERVATIONS in COMPLETELY RANDOMIZED, RANDOMIZED BLOCKS and LATIN SQUARE DESIGNS a Thesis
    ANALYSIS OF VARIANCE AND MISSING OBSERVATIONS IN COMPLETELY RANDOMIZED, RANDOMIZED BLOCKS AND LATIN SQUARE DESIGNS A Thesis Presented to The Department of Mathematics Kansas State Teachers College, Emporia, Kansas In Partial Fulfillment of the Requirements for the Degree Master of Science by Kiritkumar K. Talati May 1972 c; , ACKNOWLEDGEMENTS Sincere thanks and gratitude is expressed to Dr. John Burger for his assistance, patience, and his prompt attention in all directions in preparing this paper. A special note of appreciation goes to Dr. Marion Emerson, Dr. Thomas Davis, Dr. George Poole, Dr. Darrell Wood, who rendered assistance in the research of this paper. TABLE OF CONTENTS CHAPTER I. INTRODUCTION • • • • • • • • • • • • • • • • 1 A. Preliminary Consideration. • • • • • • • 1 B. Purpose and Assumptions of the Analysis of Variance • • • • • • • • •• 1 C. Analysis of Covariance • • • • • • • • • 2 D. Definitions. • • • • • • • • • • • • • • ) E. Organization of Paper. • • • • • • • • • 4 II. COMPLETELY RANDOMIZED DESIGN • • • • • • • • 5 A. Description............... 5 B. Randomization. • • • • • • • • • • • • • 5 C. Problem and Computations •••••••• 6 D. Conclusion and Further Applications. •• 10 III. RANDOMIZED BLOCK DESIGN. • • • • • • • • • • 12 A. Description............... 12 B. Randomization. • • • • • • • • • • • • • 1) C. Problem and Statistical Analysis • • • • 1) D. Efficiency of Randomized Block Design as Compared to Completely Randomized Design. • • • • • • • • • •• 20 E. Missing Observations • • • • • • • • •• 21 F.
    [Show full text]
  • Chapter 8 Key Ideas Hypothesis (Null and Alternative)
    Chapter 8 Key Ideas Hypothesis (Null and Alternative), Hypothesis Test, Test Statistic, P-value Type I Error, Type II Error, Significance Level, Power Section 8-1: Overview Confidence Intervals (Chapter 7) are great for estimating the value of a parameter, but in some situations estimation is not the goal. For example, suppose that researchers somehow know that the average number of fast food meals eaten per week by families in the 1990s was 3.5. Furthermore, they are interested in seeing if families today eat more fast food than in the 1990s. In this case, researchers don’t need to estimate how much fast food families eat today. All they need to know is whether that average is larger than 3.5. For the purposes of this example, let µ denote the average number of fast food meals eaten per week by families today. The researchers’ question could then be formulated as follows: Is the average number of meals eaten today the same as in the 1990s ( µ = 3.5) or is it more ( µ > 3.5)? There are two different possibilities, and the correct answer is impossible to know, since µ is unknown. However, the researchers could observe the sample mean and determine which possibility is the most likely . This process is called hypothesis testing . Definition Hypothesis – A claim or statement about a property of the population (e.g. “ µ = 3.5” from above) Hypothesis Test – A standard procedure for testing a claim about a property of the population. To guide this discussion of hypothesis testing, remember the “Rare Event Rule”: If, under a given assumption, the probability of a particular observed event is exceptionally small, we conclude that the assumption is probably not correct.
    [Show full text]