A Test of Independence in Two-Way Contingency Tables Based on Maximal Correlation

Total Page:16

File Type:pdf, Size:1020Kb

A Test of Independence in Two-Way Contingency Tables Based on Maximal Correlation A TEST OF INDEPENDENCE IN TWO-WAY CONTINGENCY TABLES BASED ON MAXIMAL CORRELATION Deniz C. YenigÄun A Dissertation Submitted to the Graduate College of Bowling Green State University in partial ful¯llment of the requirements for the degree of DOCTOR OF PHILOSOPHY August 2007 Committee: G¶abor Sz¶ekely, Advisor Maria L. Rizzo, Co-Advisor Louisa Ha, Graduate Faculty Representative James Albert Craig L. Zirbel ii ABSTRACT G¶abor Sz¶ekely, Advisor Maximal correlation has several desirable properties as a measure of dependence, includ- ing the fact that it vanishes if and only if the variables are independent. Except for a few special cases, it is hard to evaluate maximal correlation explicitly. In this dissertation, we focus on two-dimensional contingency tables and discuss a procedure for estimating maxi- mal correlation, which we use for constructing a test of independence. For large samples, we present the asymptotic null distribution of the test statistic. For small samples or tables with sparseness, we use exact inferential methods, where we employ maximal correlation as the ordering criterion. We compare the maximal correlation test with other tests of independence by Monte Carlo simulations. When the underlying continuous variables are dependent but uncorre- lated, we point out some cases for which the new test is more powerful. iii ACKNOWLEDGEMENTS I would like to express my sincere appreciation to my advisor, G¶abor Sz¶ekely, and my co-advisor, Maria Rizzo, for their advice and help throughout this research. I thank to all the members of my committee, Craig Zirbel, Jim Albert, and Louisa Ha, for their time and advice. I also want to thank to the faculty members of the Department of Mathematics and Statistics, and the Department of Applied Statistics and Operations Research, for excellent instruction and help throughout my graduate studies at Bowling Green State University. Finally, I would like to thank to my ¯anc¶ee,GÄune»sErtan, and my family, for all their support and encouragement. iv Table of Contents CHAPTER 1: Introduction 1 1.1 Statement of the Problem ............................ 1 1.2 Summary and Objectives ............................. 2 CHAPTER 2: Contingency Table Analysis and Dependence Measures 4 2.1 Analysis of Contingency Tables ......................... 4 2.1.1 Preliminaries ............................... 5 2.1.2 Pearson and Likelihood Ratio Chi-Squared Tests of Independence . 7 2.1.3 Loglinear Models ............................. 8 2.1.4 Correspondence Analysis ......................... 10 2.1.5 Exact Tests ................................ 11 2.2 Measures of Dependence ............................. 12 2.2.1 Correlation Based Measures ....................... 13 2.2.2 Measures Based on Distribution and Density Functions . 17 2.2.3 Measures of Dependence for Cross Classi¯cations . 19 CHAPTER 3: Maximal Correlation Test of Independence 24 3.1 Maximal Correlation ............................... 25 3.1.1 De¯nition ................................. 25 3.1.2 Literature Review ............................. 26 3.1.3 Attainment ................................ 28 v 3.2 Maximal Correlation in Case of Contingency Tables . 32 3.2.1 Sample Maximal Correlation ....................... 36 3.3 Algebraic Form of Sample Maximal Correlation . 38 3.4 Maximal Correlation Test for Contingency Tables . 39 3.4.1 Large Sample Case ............................ 40 3.4.2 Small Sample Case ............................ 41 3.5 A Numerical Illustration ............................. 41 3.6 A Related Test of Independence: Correlation Ratio Test . 43 3.7 An Example: Lissajous Curve Case ....................... 45 CHAPTER 4: Empirical Results 49 2 4.1 The Null Distribution of nSn ........................... 49 4.2 Large Sample Behavior of Sn ........................... 51 4.3 Power Comparisons ................................ 52 4.3.1 Simulation Design ............................. 53 4.3.2 Empirical Signi¯cance .......................... 55 4.3.3 Results of Power Comparisons ...................... 55 4.4 Empirical Powers of Correlation Ratio Test and I-Test . 62 4.5 Summary of Power Comparisons ......................... 64 4.6 An Exploratory Study for the Lissajous Curve Case . 65 CHAPTER 5: Conclusions 69 BIBLIOGRAPHY 71 Appendices 75 CHAPTER A: Results of the Empirical Study 76 vi CHAPTER B: Algebraic form of Maximal Correlation 83 B.1 2 £ 2 Contingency Tables ............................. 83 B.2 3 £ 3 Contingency Tables ............................. 84 CHAPTER C: Selected R Code 87 C.1 R Function for Maximal Correlation ....................... 87 C.2 R Function for Correlation Ratio ........................ 88 C.3 R Code for Table A.10 .............................. 88 vii List of Figures 3.1 Plots of Lissajous Curves ............................. 46 2 4.1 Empirical Distribution of nSn .......................... 51 4.2 Empirical MSE of Sn ............................... 52 4.3 Empirical Power for Example 1 ......................... 56 4.4 Empirical Power for Example 2 ......................... 57 4.5 Empirical Power for Example 3, Case 1 ..................... 59 4.6 Empirical Power for Example 3, Case 2 ..................... 60 4.7 Empirical Power for Example 4 ......................... 61 4.8 Empirical Power for Example 5 ......................... 62 A.1 Several Lissajous Curves ............................. 82 viii List of Tables 2.1 The General Form of I £ J Contingency Tables ................ 5 2.2 The Joint Distribution of X and Y ....................... 7 3.1 Postulates vs. Dependence Measures ...................... 26 3.2 Hair and Eye Color of 264 Males ........................ 42 2 A.1 Critical values of nSn ............................... 76 A.2 Empirical Mean Squared Error of Sn ...................... 76 A.3 Empirical Signi¯cance .............................. 77 A.4 Empirical Power for Example 1 ......................... 77 A.5 Empirical Power for Example 2 ......................... 78 A.6 Empirical Power for Example 3, Case 1 ..................... 78 A.7 Empirical Power for Example 3, Case 2 ..................... 79 A.8 Empirical Power for Example 4 ......................... 79 A.9 Empirical Power for Example 5 ......................... 80 ¤ A.10 Empirical Power for In and Sn, Loglinear Case . 80 ¤ A.11 Empirical Power for In and Sn, Lissajuos Curve Case . 81 A.12 I, KX (Y ), and KY (X) for Lissajous Curve Case . 81 1 CHAPTER 1 Introduction A variety of data come in the form of cross-classi¯ed tables of frequency counts, referred to as contingency tables. After a century of great progress, the analysis of contingency tables is still an active ¯eld in statistics, mainly due to its important applications in biological and social sciences. For a comprehensive survey of the most important methods in the analysis of contingency tables, see Agresti (2002). A principal interest in many studies regarding contingency tables is to test if the variables are independent. Although many good tests are available, no single test is known to be optimal for all independence problems. An overview of several existing tests for independence in contingency tables follows in Chapter 2. In this dissertation, we construct a test of independence for two-dimensional contingency tables, based on maximal correlation. The details of this test are presented and the power performance is compared with other tests of independence. 1.1 Statement of the Problem Consider two categorical response variables X and Y having I and J levels respectively. Given a contingency table, we consider the problem of testing if X and Y are independent. The null hypothesis of statistical independence of X and Y is given by H0 : ¼ij = ¼i¢¼¢j for 2 i = 1; :::I, j = 1; :::J, where ¼ij denotes the probability that a randomly selected individual falls into category i of variable X and category j of variable Y , and the subscript \¢" denotes the sum over the index it replaces. Our tool for approaching this problem is maximal correlation, which is a convenient measure of dependence. Maximal correlation has several desirable properties, including the fact that it vanishes if and only if the variables are independent. However, except for a few special cases, it is hard to evaluate the maximal correlation explicitly. We discuss a procedure for estimating the maximal correlation for contingency tables, which we use for constructing a test of independence. 1.2 Summary and Objectives The main objective of this research is to introduce the maximal correlation test of indepen- dence for contingency tables, and evaluate the performance of this test. We describe the computation of maximal correlation for an observed contingency table, and give a detailed procedure on how to carry out the test of independence. Exact inferential methods are used for contingency tables with small sample size or sparseness. For large samples, the asymptotic null distribution of maximal correlation is used to carry out the test. We also evaluate this test of independence against a wide range of alternatives, and compare its power with two well-known tests of independence by an extensive empirical study. This dissertation has ¯ve chapters. We introduce the problem of interest in Chapter 1. Chapter 2 introduces the notation used in the analysis of contingency tables, and summarizes the well-known methods used in this ¯eld. This chapter also includes a review of several widely used measures of dependence, including the maximal correlation. In Chapter 3 we give further insight to the concept of maximal correlation,
Recommended publications
  • Contingency Tables Are Eaten by Large Birds of Prey
    Case Study Case Study Example 9.3 beginning on page 213 of the text describes an experiment in which fish are placed in a large tank for a period of time and some Contingency Tables are eaten by large birds of prey. The fish are categorized by their level of parasitic infection, either uninfected, lightly infected, or highly infected. It is to the parasites' advantage to be in a fish that is eaten, as this provides Bret Hanlon and Bret Larget an opportunity to infect the bird in the parasites' next stage of life. The observed proportions of fish eaten are quite different among the categories. Department of Statistics University of Wisconsin|Madison Uninfected Lightly Infected Highly Infected Total October 4{6, 2011 Eaten 1 10 37 48 Not eaten 49 35 9 93 Total 50 45 46 141 The proportions of eaten fish are, respectively, 1=50 = 0:02, 10=45 = 0:222, and 37=46 = 0:804. Contingency Tables 1 / 56 Contingency Tables Case Study Infected Fish and Predation 2 / 56 Stacked Bar Graph Graphing Tabled Counts Eaten Not eaten 50 40 A stacked bar graph shows: I the sample sizes in each sample; and I the number of observations of each type within each sample. 30 This plot makes it easy to compare sample sizes among samples and 20 counts within samples, but the comparison of estimates of conditional Frequency probabilities among samples is less clear. 10 0 Uninfected Lightly Infected Highly Infected Contingency Tables Case Study Graphics 3 / 56 Contingency Tables Case Study Graphics 4 / 56 Mosaic Plot Mosaic Plot Eaten Not eaten 1.0 0.8 A mosaic plot replaces absolute frequencies (counts) with relative frequencies within each sample.
    [Show full text]
  • Use of Chi-Square Statistics
    This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this site. Copyright 2008, The Johns Hopkins University and Marie Diener-West. All rights reserved. Use of these materials permitted only in accordance with license rights granted. Materials provided “AS IS”; no representations or warranties provided. User assumes all responsibility for use, and all liability related thereto, and must independently review all materials for accuracy and efficacy. May contain materials owned by others. User is responsible for obtaining permissions for use from third parties as needed. Use of the Chi-Square Statistic Marie Diener-West, PhD Johns Hopkins University Section A Use of the Chi-Square Statistic in a Test of Association Between a Risk Factor and a Disease The Chi-Square ( X2) Statistic Categorical data may be displayed in contingency tables The chi-square statistic compares the observed count in each table cell to the count which would be expected under the assumption of no association between the row and column classifications The chi-square statistic may be used to test the hypothesis of no association between two or more groups, populations, or criteria Observed counts are compared to expected counts 4 Displaying Data in a Contingency Table Criterion 2 Criterion 1 1 2 3 . C Total . 1 n11 n12 n13 n1c r1 2 n21 n22 n23 . n2c r2 . 3 n31 . r nr1 nrc rr Total c1 c2 cc n 5 Chi-Square Test Statistic
    [Show full text]
  • Knowledge and Human Capital As Sustainable Competitive Advantage in Human Resource Management
    sustainability Article Knowledge and Human Capital as Sustainable Competitive Advantage in Human Resource Management Miloš Hitka 1 , Alžbeta Kucharˇcíková 2, Peter Štarcho ˇn 3 , Žaneta Balážová 1,*, Michal Lukáˇc 4 and Zdenko Stacho 5 1 Faculty of Wood Sciences and Technology, Technical University in Zvolen, T. G. Masaryka 24, 960 01 Zvolen, Slovakia; [email protected] 2 Faculty of Management Science and Informatics, University of Žilina, Univerzitná 8215/1, 010 26 Žilina, Slovakia; [email protected] 3 Faculty of Management, Comenius University in Bratislava, Odbojárov 10, P.O. BOX 95, 82005 Bratislava, Slovakia; [email protected] 4 Faculty of Social Sciences, University of SS. Cyril and Methodius in Trnava, Buˇcianska4/A, 917 01 Trnava, Slovakia; [email protected] 5 Institut of Civil Society, University of SS. Cyril and Methodius in Trnava, Buˇcianska4/A, 917 01 Trnava, Slovakia; [email protected] * Correspondence: [email protected]; Tel.: +421-45-520-6189 Received: 2 August 2019; Accepted: 9 September 2019; Published: 12 September 2019 Abstract: The ability to do business successfully and to stay on the market is a unique feature of each company ensured by highly engaged and high-quality employees. Therefore, innovative leaders able to manage, motivate, and encourage other employees can be a great competitive advantage of an enterprise. Knowledge of important personality factors regarding leadership, incentives and stimulus, systematic assessment, and subsequent motivation factors are parts of human capital and essential conditions for effective development of its potential. Familiarity with various ways to motivate leaders and their implementation in practice are important for improving the work performance and reaching business goals.
    [Show full text]
  • Krishnan Correlation CUSB.Pdf
    SELF INSTRUCTIONAL STUDY MATERIAL PROGRAMME: M.A. DEVELOPMENT STUDIES COURSE: STATISTICAL METHODS FOR DEVELOPMENT RESEARCH (MADVS2005C04) UNIT III-PART A CORRELATION PROF.(Dr). KRISHNAN CHALIL 25, APRIL 2020 1 U N I T – 3 :C O R R E L A T I O N After reading this material, the learners are expected to : . Understand the meaning of correlation and regression . Learn the different types of correlation . Understand various measures of correlation., and, . Explain the regression line and equation 3.1. Introduction By now we have a clear idea about the behavior of single variables using different measures of Central tendency and dispersion. Here the data concerned with one variable is called ‗univariate data‘ and this type of analysis is called ‗univariate analysis‘. But, in nature, some variables are related. For example, there exists some relationships between height of father and height of son, price of a commodity and amount demanded, the yield of a plant and manure added, cost of living and wages etc. This is a a case of ‗bivariate data‘ and such analysis is called as ‗bivariate data analysis. Correlation is one type of bivariate statistics. Correlation is the relationship between two variables in which the changes in the values of one variable are followed by changes in the values of the other variable. 3.2. Some Definitions 1. ―When the relationship is of a quantitative nature, the approximate statistical tool for discovering and measuring the relationship and expressing it in a brief formula is known as correlation‖—(Craxton and Cowden) 2. ‗correlation is an analysis of the co-variation between two or more variables‖—(A.M Tuttle) 3.
    [Show full text]
  • Pearson-Fisher Chi-Square Statistic Revisited
    Information 2011 , 2, 528-545; doi:10.3390/info2030528 OPEN ACCESS information ISSN 2078-2489 www.mdpi.com/journal/information Communication Pearson-Fisher Chi-Square Statistic Revisited Sorana D. Bolboac ă 1, Lorentz Jäntschi 2,*, Adriana F. Sestra ş 2,3 , Radu E. Sestra ş 2 and Doru C. Pamfil 2 1 “Iuliu Ha ţieganu” University of Medicine and Pharmacy Cluj-Napoca, 6 Louis Pasteur, Cluj-Napoca 400349, Romania; E-Mail: [email protected] 2 University of Agricultural Sciences and Veterinary Medicine Cluj-Napoca, 3-5 M ănăş tur, Cluj-Napoca 400372, Romania; E-Mails: [email protected] (A.F.S.); [email protected] (R.E.S.); [email protected] (D.C.P.) 3 Fruit Research Station, 3-5 Horticultorilor, Cluj-Napoca 400454, Romania * Author to whom correspondence should be addressed; E-Mail: [email protected]; Tel: +4-0264-401-775; Fax: +4-0264-401-768. Received: 22 July 2011; in revised form: 20 August 2011 / Accepted: 8 September 2011 / Published: 15 September 2011 Abstract: The Chi-Square test (χ2 test) is a family of tests based on a series of assumptions and is frequently used in the statistical analysis of experimental data. The aim of our paper was to present solutions to common problems when applying the Chi-square tests for testing goodness-of-fit, homogeneity and independence. The main characteristics of these three tests are presented along with various problems related to their application. The main problems identified in the application of the goodness-of-fit test were as follows: defining the frequency classes, calculating the X2 statistic, and applying the χ2 test.
    [Show full text]
  • ED441031.Pdf
    DOCUMENT RESUME ED 441 031 TM 030 830 AUTHOR Fahoome, Gail; Sawilowsky, Shlomo S. TITLE Review of Twenty Nonparametric Statistics and Their Large Sample Approximations. PUB DATE 2000-04-00 NOTE 42p.; Paper presented at the Annual Meeting of the American Educational Research Association (New Orleans, LA, April 24-28, 2000). PUB TYPE Information Analyses (070)-- Numerical/Quantitative Data (110) Speeches/Meeting Papers (150) EDRS PRICE MF01/PCO2 Plus Postage. DESCRIPTORS Monte Carlo Methods; *Nonparametric Statistics; *Sample Size; *Statistical Distributions ABSTRACT Nonparametric procedures are often more powerful than classical tests for real world data, which are rarely normally distributed. However, there are difficulties in using these tests. Computational formulas are scattered tnrous-hout the literature, and there is a lack of avalialalicy of tables of critical values. This paper brings together the computational formulas for 20 commonly used nonparametric tests that have large-sample approximations for the critical value. Because there is no generally agreed upon lower limit for the sample size, Monte Carlo methods have been used to determine the smallest sample size that can be used with the large-sample approximations. The statistics reviewed include single-population tests, comparisons of two populations, comparisons of several populations, and tests of association. (Contains 4 tables and 59 references.)(Author/SLD) Reproductions supplied by EDRS are the best that can be made from the original document. Review of Twenty Nonparametric Statistics and Their Large Sample Approximations Gail Fahoome Shlomo S. Sawilowsky U.S. DEPARTMENT OF EDUCATION Mice of Educational Research and Improvement PERMISSION TO REPRODUCE AND UCATIONAL RESOURCES INFORMATION DISSEMINATE THIS MATERIAL HAS f BEEN GRANTED BY CENTER (ERIC) This document has been reproduced as received from the person or organization originating it.
    [Show full text]
  • Detecting Trends Using Spearman's Rank Correlation Coefficient
    Environmental Forensics 2001) 2, 359±362 doi:10.1006/enfo.2001.0061, available online at http://www.idealibrary.com on Detecting Trends Using Spearman's Rank Correlation Coecient Thomas D. Gauthier Sciences International Inc, 6150 State Road 70, East Bradenton, FL 34203, U.S.A. Received 23 August 2001, Revised manuscript accepted 1 October 2001) Spearman's rank correlation coecient is a useful tool for exploratory data analysis in environmental forensic investigations. In this application it is used to detect monotonic trends in chemical concentration with time or space. # 2001 AEHS Keywords: Spearman rank correlation coecient; trend analysis; soil; groundwater. Environmental forensic investigations can involve a The idea behind the rank correlation coecient is variety of statistical analysis techniques. A relatively simple. Each variable is ranked separately from lowest simple technique that can be used for exploratory data to highest e.g. 1, 2, 3, etc.) and the dierence between analysis is the Spearman rank correlation coecient. In ranks for each data pair is recorded. If the data are this technical note, I describe howthe Spearman rank correlated, then the sum of the square of the dierence correlation coecient can be used as a statistical tool between ranks will be small. The magnitude of the sum to detect monotonic trends in chemical concentrations is related to the signi®cance of the correlation. with time or space. The Spearman rank correlation coecient is calcu- Trend analysis can be useful for detecting chemical lated according to the following equation: ``footprints'' at a site or evaluating the eectiveness of Pn an installed or natural attenuation remedy.
    [Show full text]
  • 05 36534Nys130620 31
    Monte Carlos study on Power Rates of Some Heteroscedasticity detection Methods in Linear Regression Model with multicollinearity problem O.O. Alabi, Kayode Ayinde, O. E. Babalola, and H.A. Bello Department of Statistics, Federal University of Technology, P.M.B. 704, Akure, Ondo State, Nigeria Corresponding Author: O. O. Alabi, [email protected] Abstract: This paper examined the power rate exhibit by some heteroscedasticity detection methods in a linear regression model with multicollinearity problem. Violation of unequal error variance assumption in any linear regression model leads to the problem of heteroscedasticity, while violation of the assumption of non linear dependency between the exogenous variables leads to multicollinearity problem. Whenever these two problems exist one would faced with estimation and hypothesis problem. in order to overcome these hurdles, one needs to determine the best method of heteroscedasticity detection in other to avoid taking a wrong decision under hypothesis testing. This then leads us to the way and manner to determine the best heteroscedasticity detection method in a linear regression model with multicollinearity problem via power rate. In practices, variance of error terms are unequal and unknown in nature, but there is need to determine the presence or absence of this problem that do exist in unknown error term as a preliminary diagnosis on the set of data we are to analyze or perform hypothesis testing on. Although, there are several forms of heteroscedasticity and several detection methods of heteroscedasticity, but for any researcher to arrive at a reasonable and correct decision, best and consistent performed methods of heteroscedasticity detection under any forms or structured of heteroscedasticity must be determined.
    [Show full text]
  • The Effects of Simplifying Assumptions in Power Analysis
    University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Public Access Theses and Dissertations from Education and Human Sciences, College of the College of Education and Human Sciences (CEHS) 4-2011 The Effects of Simplifying Assumptions in Power Analysis Kevin A. Kupzyk University of Nebraska-Lincoln, [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/cehsdiss Part of the Educational Psychology Commons Kupzyk, Kevin A., "The Effects of Simplifying Assumptions in Power Analysis" (2011). Public Access Theses and Dissertations from the College of Education and Human Sciences. 106. https://digitalcommons.unl.edu/cehsdiss/106 This Article is brought to you for free and open access by the Education and Human Sciences, College of (CEHS) at DigitalCommons@University of Nebraska - Lincoln. It has been accepted for inclusion in Public Access Theses and Dissertations from the College of Education and Human Sciences by an authorized administrator of DigitalCommons@University of Nebraska - Lincoln. Kupzyk - i THE EFFECTS OF SIMPLIFYING ASSUMPTIONS IN POWER ANALYSIS by Kevin A. Kupzyk A DISSERTATION Presented to the Faculty of The Graduate College at the University of Nebraska In Partial Fulfillment of Requirements For the Degree of Doctor of Philosophy Major: Psychological Studies in Education Under the Supervision of Professor James A. Bovaird Lincoln, Nebraska April, 2011 Kupzyk - i THE EFFECTS OF SIMPLIFYING ASSUMPTIONS IN POWER ANALYSIS Kevin A. Kupzyk, Ph.D. University of Nebraska, 2011 Adviser: James A. Bovaird In experimental research, planning studies that have sufficient probability of detecting important effects is critical. Carrying out an experiment with an inadequate sample size may result in the inability to observe the effect of interest, wasting the resources spent on an experiment.
    [Show full text]
  • Measures of Association for Contingency Tables
    Newsom Psy 525/625 Categorical Data Analysis, Spring 2021 1 Measures of Association for Contingency Tables The Pearson chi-squared statistic and related significance tests provide only part of the story of contingency table results. Much more can be gleaned from contingency tables than just whether the results are different from what would be expected due to chance (Kline, 2013). For many data sets, the sample size will be large enough that even small departures from expected frequencies will be significant. And, for other data sets, we may have low power to detect significance. We therefore need to know more about the strength of the magnitude of the difference between the groups or the strength of the relationship between the two variables. Phi The most common measure of magnitude of effect for two binary variables is the phi coefficient. Phi can take on values between -1.0 and 1.0, with 0.0 representing complete independence and -1.0 or 1.0 representing a perfect association. In probability distribution terms, the joint probabilities for the cells will be equal to the product of their respective marginal probabilities, Pn( ij ) = Pn( i++) Pn( j ) , only if the two variables are independent. The formula for phi is often given in terms of a shortcut notation for the frequencies in the four cells, called the fourfold table. Azen and Walker Notation Fourfold table notation n11 n12 A B n21 n22 C D The equation for computing phi is a fairly simple function of the cell frequencies, with a cross- 1 multiplication and subtraction of the two sets of diagonal cells in the numerator.
    [Show full text]
  • Introduction to Hypothesis Testing
    Introduction to Hypothesis Testing OPRE 6301 Motivation . The purpose of hypothesis testing is to determine whether there is enough statistical evidence in favor of a certain belief, or hypothesis, about a parameter. Examples: Is there statistical evidence, from a random sample of potential customers, to support the hypothesis that more than 10% of the potential customers will pur- chase a new product? Is a new drug effective in curing a certain disease? A sample of patients is randomly selected. Half of them are given the drug while the other half are given a placebo. The conditions of the patients are then mea- sured and compared. These questions/hypotheses are similar in spirit to the discrimination example studied earlier. Below, we pro- vide a basic introduction to hypothesis testing. 1 Criminal Trials . The basic concepts in hypothesis testing are actually quite analogous to those in a criminal trial. Consider a person on trial for a “criminal” offense in the United States. Under the US system a jury (or sometimes just the judge) must decide if the person is innocent or guilty while in fact the person may be innocent or guilty. These combinations are summarized in the table below. Person is: Innocent Guilty Jury Says: Innocent No Error Error Guilty Error No Error Notice that there are two types of errors. Are both of these errors equally important? Or, is it as bad to decide that a guilty person is innocent and let them go free as it is to decide an innocent person is guilty and punish them for the crime? Or, is a jury supposed to be totally objective, not assuming that the person is either innocent or guilty and make their decision based on the weight of the evidence one way or another? 2 In a criminal trial, there actually is a favored assump- tion, an initial bias if you will.
    [Show full text]
  • Power of a Statistical Test
    Power of a Statistical Test By Smita Skrivanek, Principal Statistician, MoreSteam.com LLC What is the power of a test? The power of a statistical test gives the likelihood of rejecting the null hypothesis when the null hypothesis is false. Just as the significance level (alpha) of a test gives the probability that the null hypothesis will be rejected when it is actually true (a wrong decision), power quantifies the chance that the null hypothesis will be rejected when it is actually false (a correct decision). Thus, power is the ability of a test to correctly reject the null hypothesis. Why is it important? Although you can conduct a hypothesis test without it, calculating the power of a test beforehand will help you ensure that the sample size is large enough for the purpose of the test. Otherwise, the test may be inconclusive, leading to wasted resources. On rare occasions the power may be calculated after the test is performed, but this is not recommended except to determine an adequate sample size for a follow-up study (if a test failed to detect an effect, it was obviously underpowered – nothing new can be learned by calculating the power at this stage). How is it calculated? As an example, consider testing whether the average time per week spent watching TV is 4 hours versus the alternative that it is greater than 4 hours. We will calculate the power of the test for a specific value under the alternative hypothesis, say, 7 hours: The Null Hypothesis is H0: μ = 4 hours The Alternative Hypothesis is H1: μ = 6 hours Where μ = the average time per week spent watching TV.
    [Show full text]