Likelihood Ratio Test, Most Powerful Test, Uniformly Most Powerful Test

Total Page:16

File Type:pdf, Size:1020Kb

Likelihood Ratio Test, Most Powerful Test, Uniformly Most Powerful Test Likelihood Ratio Test, Most Powerful Test, Uniformly Most Powerful Test A. We first review the general definition of a hypothesis test. A hypothesis test is like a lawsuit: H 0 : the defendant is innocent versus H a : the defendant is guilty The truth H 0 : the defendant H a the defendant is innocent is guilty Jury’s H 0 Right decision Type II error Decision H a Type I error Right decision The significance level and the power of the test. α = P(Type I error) = P(Reject H 0 | H 0 ) ← signiicance level β = P(Type II error) = P(Fail to reject H 0 | H a ) Power = 1- β = P(Reject H 0 | H a ) B. Now we derive the likelihood ratio test for the usual two- sided hypotheses for a population mean. (It has a simple null hypothesis and a composite alternative hypothesis.) Example 1. Please derive the likelihood ratio test for H0: μ = μ0 versus Ha: μ ≠ μ0, when the population is normal and population variance σ2 is known. Solution: For a 2-sided test of H0: μ = μ0 versus Ha: μ ≠ μ0, when the population is normal and population variance σ2 is known, we have: 1 : 0 and : The likelihoods are: L L0 n 2 n 1 x 1 2 1 n 2 exp i 0 exp x i1 2 2 2 2 i1 i 0 2 2 2 2 There is no free parameter in L , thus Lˆ L . L L n 2 n 1 x 1 2 1 n 2 exp i exp x i1 2 2 2 2 i1 i 2 2 2 2 There is only one free parameter μ in L. Now we shall find the value of μ that maximizes the log likelihood n 1 n ln L ln2 2 x 2 . 2 2 2 i1 i d ln L 1 n By solving x 0 , we have ˆ x d 2 i1 i It is easy to verify that ˆ x indeed maximizes the loglikelihood, and thus the likelihood function. Therefore the likelihood ratio is: Lˆ L 0 ˆ L max L n 1 2 1 n 2 exp x L 2 2 2 2 i1 i 0 0 Lˆ n 1 2 1 n 2 exp x x 2 2 i1i 2 2 1 n 2 2 exp x x x 2 i1i 0 i 2 2 1 x 0 1 2 exp exp z0 2 / n 2 2 Therefore, the likelihood ratio test that will reject H0 when * is equivalent to the z-test that will reject H0 when Z0 c , where c can be determined by the significance level α as c z / 2 . C. MP Test, UMP Test, and the Neyman-Pearson Lemma Now considering some one-sided tests where we have : H0 : 0 versus H a : 0 or, H0 : 0 versus H a : 0 Given the composite null hypothesis for the second one-sided test, we need to expand our definition of the significance level as follows: For 퐻: 휃 ∈ 휔 versus 퐻: 휃 ∈ 휔 Let the random sample be denoted by 푿 = (푋, 푋,⋯, 푋) . Suppose the rejection region is 푪 such that: (*note: the rejection region can be defined in terms of the sample points directly instead of the test statistic values) We reject 퐻 if 푿 ∈ 푪 We fail to reject 퐻 if 푿 ∈ 푪 Definition: The size or significance level of the test is defined as: 휶 = 퐦퐚퐱 푷(푿 ∈ 푪) = 퐦퐚퐱 푷휽(푹풆풋풆풄풕 퐻|휃 ∈ 휔) ∈ Definition: The power of a test is given by ퟏ − 휷 = 휸(휃) = 푷휽(푹풆풋풆풄풕 퐻|휃 ∈ 휔) Definition: A most powerful test corresponds to the best rejection region 푪 such that in testing a simple null hypothesis 퐻: 휃 = 휃 versus a simple alternative hypothesis 퐻: 휃 = 휃 , this test has the largest possible test power (ퟏ − 휷) among all tests of size 휶. 3 Theorem (Neyman-Pearson Lemma). The likelihood ratio test for a simple null hypothesis 퐻: 휃 = 휃 versus a simple alternative hypothesis 퐻: 휃 = 휃 is a most powerful test. Jerzy Neyman (left) and Egon Pearson (right) Definition: A uniformly most powerful test in testing a simple null hypothesis 퐻: 휃 = 휃 versus a composite alternative hypothesis 퐻: 휃 ∈ 휔, or in testing a composite null hypothesis 퐻: 휃 ∈ 휔 versus a composite alternative 퐻: 휃 ∈ 휔 , is a size 휶 test such that this test has the largest possible test power 휸(휃) for every simple alternative hypothesis: 퐻: 휃 = 휃 ∈ 휔, among all such tests of size 휶. Now we derive the likelihood ratio test for the first one- sided hypotheses. ( ) Example 2. Let푓(푥 |휇) = 푒 , 푓표푟 푖 = 1,⋯, 푛, where √ 휎 is known. Derive the likelihood ratio test for the hypothesis 퐻: 휇 = 휇 푣푠. 퐻: 휇 > 휇, and show whether it is a UMP test or not. Solution: 흎 = {휇: 휇 = 휇} 훀 = {휇: 휇 ≥ 휇} Note: Now that we are not just dealing with the two-sided hypothesis, it is important to know that the more general definition of 흎 is the set of all unknown parameter values under 퐻, while 훀 is the set of all unknown parameter values under the union of 퐻 and 퐻. 4 Therefore we have: ⎧ 퐿 = 푓(푥 , 푥 ,⋯ , 푥 |휇 = 휇 ) = 푓(푥 |휇 = 휇 ) ⎪ ⎪ ( ) ⎪ 1 ∑ ( ) = 푒 = (2휋휎) 푒 ⎪ √2휋휎 ⎨ ⎪ 퐿 = 푓(푥, 푥,⋯ , 푥|휇 ≥ 휇) = 푓(푥|휇 ≥ 휇) ⎪ ⎪ ( ) 1 ∑ ( ) ⎪ = 푒 = (2휋휎) 푒 √2휋휎 ⎩ Since there is no free parameter in 퐿 , ∑() sup(퐿) = (2휋휎 ) 푒 . 휕푙표푔퐿 푛 ⎧ = (푥̅ − 휇) ⎪ 휕휇 휎 ∵ ,∴ sup(퐿) ⎨휕 푙표푔퐿 푛 ⎪ =− <0 ⎩ 휕휇 휎 ∑ ( ̅) (2휋휎) 푒 , 푖푓 푥̅ > 휇 = . ∑() (2휋휎 ) 푒 , 푖푓 푥̅ ≤ 휇 ( ̅ ) sup(퐿) ∴ 퐿푅 = = 푒 , 푖푓 푥̅ > 휇. sup(퐿 ) 1, 푖푓 푥̅ ≤ 휇 푋 − 휇 ∗ ∗ ∴ 푃(퐿푅 ≤ 푐 |퐻) = 훼 ⇔ 푃 푍 = ≥ −2푙표푔푐 퐻 휎⁄√푛 = 훼. ∴ Reject 퐻 in favor of 퐻 if 푍 = ≥ 푍 ⁄√ Furthermore, it is easy to show that for each 휇 > 휇 , the likelihood ratio test of 퐻: 휇 = 휇 푣푠. 퐻: 휇 = 휇 rejects the null hypothesis at the significance level α for 푍 = ≥ 푍 ⁄√ By the Neyman-Pearson Lemma, the likelihood ratio test is also the most powerful test. Now since for each 1 0 , the most powerful size α test of 퐻: 휇 = 휇 푣푠. 퐻: 휇 = 휇 rejects the null hypothesis for 푍 = ≥ 푍 . Since this same test function is most powerful ⁄√ for each 1 0 , this test is UMP for 퐻: 휇 = 휇 푣푠. 퐻: 휇 > 휇. 5 D. Monotone Likelihood Ratio, the Karlin-Rubin Theorem, and the Exponential Family. A condition under which the UMP tests exists is when the family of distributions being considered possesses a property called monotone likelihood ratio. Definition: Let 푓(풙|휃) be the joint pdf of the sample 푿 = (푋, 푋,⋯, 푋) . Then 푓(풙|휃) is said to have a monotone likelihood ratio in the statistic 푇(푿) if for any choice 휃 < 휃 of parameter values, the likelihood ratio 푓(풙|휃)/푓(풙|휃) depends on values of the data X only through the value of statistic 푇(푿) and, in addition, this ratio is a monotone function of 푇(푿). Theorem (Karlin-Rubin). Suppose that 푓(풙|휃) has an increasing monotone likelihood ratio for the statistic 푇(푿). Let 훼 and 푘 be chosen so that 휶 = 푷휽(푇(푿) ≥ 풌휶). Then 푪 = {풙: 푇(풙) ≥ 풌휶} is the rejection region (*also called ‘critical region’) for a UMP test for the one-sided tests of: 퐻: 휃 ≤ 휃 versus 퐻: 휃 > 휃. Theorem. For a Regular Exponential Family with only one unknown parameter θ, and a population p.d.f.: 푓(푥; 휃) = 푐(휃) ⋅ 푔(푥) ⋅ exp[푡(푥)푤(휃)] If 푤(휃) is a is an increasing function of 휃, then we have a monotone likelihood ratio in the statistic 푇푋 = ∑ 푡(푋). {*Recall 푇푋 = ∑ 푡(푋)} is also complete sufficient for 휃} Examples of a family with monotone likelihood ratio: For X1,, X n iid Exponential ( ), n X e 1 i n 1 n p(x| 1 ) i1 1 n exp (1 0 ) X i p(x| 0 ) 0 Xi 0 i1 0e i1 6 p(x| ) n 1 X For 1 0 , is an increasing function of i so the p(x| 0 ) i1 n family has monotone likelihood ratio in T(x ) X i . i1 Example 3. Let 푋, 푋,… 푋 ~푁(휇, 휎 ), where 휎 is known (1) Find the LRT for 퐻: 휇 ≤ 휇 푣푠 퐻: 휇 > 휇 (2) Show the test in (1) is a UMP test. Solution: 흎 = {휇: 휇 ≤ 휇} 훀 = {휇: −∞< 휇 <∞} Note: Now that we are not just dealing with the two-sided hypothesis, it is important to know that the more general definition of 흎 is the set of all unknown parameter values under 퐻, while 훀 is the set of all unknown parameter values under the union of 퐻 and 퐻. (1) The ratio of the likelihood is shown as the following, 퐿(휔) 휆(푥)= 퐿(Ω) 퐿(휔) is the maximum likelihood under 퐻: 휇 ≤ 휇 1 / ∑(푥 − 푥̅) ⎧ exp − , 푥̅ < 휇 ⎪ 2휋휎 2휎 퐿(휔) = ⎨ 1 / ∑(푥 − 휇 ) ⎪ exp − , 푥̅ ≥ 휇 ⎩ 2휋휎 2휎 퐿Ω is the maximum likelihood under 퐻 푢푛푖표푛 퐻 1 / ∑(푥 − 푥̅) 퐿Ω = exp − 2휋휎 2휎 The ratio of the maximized likelihood is, 1, 푥̅ < 휇 휆(푥) = (∑(푥 − 휇) − ∑(푥 − 푥̅) ) 푒푥푝 − , 푥̅ ≥ 휇 2휎 ⇒ 1, 푥̅ < 휇 휆(푥) = (̅) 푒 , 푥̅ ≥ 휇 7 By LRT, we reject null hypothesis if 휆(푥) < 푐. Generally, c is chosen to be less than 1.
Recommended publications
  • Hypothesis Testing and Likelihood Ratio Tests
    Hypottthesiiis tttestttiiing and llliiikellliiihood ratttiiio tttesttts Y We will adopt the following model for observed data. The distribution of Y = (Y1, ..., Yn) is parameter considered known except for some paramett er ç, which may be a vector ç = (ç1, ..., çk); ç“Ç, the paramettter space. The parameter space will usually be an open set. If Y is a continuous random variable, its probabiiillliiittty densiiittty functttiiion (pdf) will de denoted f(yy;ç) . If Y is y probability mass function y Y y discrete then f(yy;ç) represents the probabii ll ii tt y mass functt ii on (pmf); f(yy;ç) = Pç(YY=yy). A stttatttiiistttiiicalll hypottthesiiis is a statement about the value of ç. We are interested in testing the null hypothesis H0: ç“Ç0 versus the alternative hypothesis H1: ç“Ç1. Where Ç0 and Ç1 ¶ Ç. hypothesis test Naturally Ç0 § Ç1 = ∅, but we need not have Ç0 ∞ Ç1 = Ç. A hypott hesii s tt estt is a procedure critical region for deciding between H0 and H1 based on the sample data. It is equivalent to a crii tt ii call regii on: a critical region is a set C ¶ Rn y such that if y = (y1, ..., yn) “ C, H0 is rejected. Typically C is expressed in terms of the value of some tttesttt stttatttiiistttiiic, a function of the sample data. For µ example, we might have C = {(y , ..., y ): y – 0 ≥ 3.324}. The number 3.324 here is called a 1 n s/ n µ criiitttiiicalll valllue of the test statistic Y – 0 . S/ n If y“C but ç“Ç 0, we have committed a Type I error.
    [Show full text]
  • This Is Dr. Chumney. the Focus of This Lecture Is Hypothesis Testing –Both What It Is, How Hypothesis Tests Are Used, and How to Conduct Hypothesis Tests
    TRANSCRIPT: This is Dr. Chumney. The focus of this lecture is hypothesis testing –both what it is, how hypothesis tests are used, and how to conduct hypothesis tests. 1 TRANSCRIPT: In this lecture, we will talk about both theoretical and applied concepts related to hypothesis testing. 2 TRANSCRIPT: Let’s being the lecture with a summary of the logic process that underlies hypothesis testing. 3 TRANSCRIPT: It is often impossible or otherwise not feasible to collect data on every individual within a population. Therefore, researchers rely on samples to help answer questions about populations. Hypothesis testing is a statistical procedure that allows researchers to use sample data to draw inferences about the population of interest. Hypothesis testing is one of the most commonly used inferential procedures. Hypothesis testing will combine many of the concepts we have already covered, including z‐scores, probability, and the distribution of sample means. To conduct a hypothesis test, we first state a hypothesis about a population, predict the characteristics of a sample of that population (that is, we predict that a sample will be representative of the population), obtain a sample, then collect data from that sample and analyze the data to see if it is consistent with our hypotheses. 4 TRANSCRIPT: The process of hypothesis testing begins by stating a hypothesis about the unknown population. Actually we state two opposing hypotheses. The first hypothesis we state –the most important one –is the null hypothesis. The null hypothesis states that the treatment has no effect. In general the null hypothesis states that there is no change, no difference, no effect, and otherwise no relationship between the independent and dependent variables.
    [Show full text]
  • The Scientific Method: Hypothesis Testing and Experimental Design
    Appendix I The Scientific Method The study of science is different from other disciplines in many ways. Perhaps the most important aspect of “hard” science is its adherence to the principle of the scientific method: the posing of questions and the use of rigorous methods to answer those questions. I. Our Friend, the Null Hypothesis As a science major, you are probably no stranger to curiosity. It is the beginning of all scientific discovery. As you walk through the campus arboretum, you might wonder, “Why are trees green?” As you observe your peers in social groups at the cafeteria, you might ask yourself, “What subtle kinds of body language are those people using to communicate?” As you read an article about a new drug which promises to be an effective treatment for male pattern baldness, you think, “But how do they know it will work?” Asking such questions is the first step towards hypothesis formation. A scientific investigator does not begin the study of a biological phenomenon in a vacuum. If an investigator observes something interesting, s/he first asks a question about it, and then uses inductive reasoning (from the specific to the general) to generate an hypothesis based upon a logical set of expectations. To test the hypothesis, the investigator systematically collects data, either with field observations or a series of carefully designed experiments. By analyzing the data, the investigator uses deductive reasoning (from the general to the specific) to state a second hypothesis (it may be the same as or different from the original) about the observations.
    [Show full text]
  • Hypothesis Testing – Examples and Case Studies
    Hypothesis Testing – Chapter 23 Examples and Case Studies Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 23.1 How Hypothesis Tests Are Reported in the News 1. Determine the null hypothesis and the alternative hypothesis. 2. Collect and summarize the data into a test statistic. 3. Use the test statistic to determine the p-value. 4. The result is statistically significant if the p-value is less than or equal to the level of significance. Often media only presents results of step 4. Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 2 23.2 Testing Hypotheses About Proportions and Means If the null and alternative hypotheses are expressed in terms of a population proportion, mean, or difference between two means and if the sample sizes are large … … the test statistic is simply the corresponding standardized score computed assuming the null hypothesis is true; and the p-value is found from a table of percentiles for standardized scores. Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 3 Example 2: Weight Loss for Diet vs Exercise Did dieters lose more fat than the exercisers? Diet Only: sample mean = 5.9 kg sample standard deviation = 4.1 kg sample size = n = 42 standard error = SEM1 = 4.1/ √42 = 0.633 Exercise Only: sample mean = 4.1 kg sample standard deviation = 3.7 kg sample size = n = 47 standard error = SEM2 = 3.7/ √47 = 0.540 measure of variability = [(0.633)2 + (0.540)2] = 0.83 Copyright ©2005 Brooks/Cole, a division of Thomson Learning, Inc. 4 Example 2: Weight Loss for Diet vs Exercise Step 1.
    [Show full text]
  • Tests of Hypotheses Using Statistics
    Tests of Hypotheses Using Statistics Adam Massey¤and Steven J. Millery Mathematics Department Brown University Providence, RI 02912 Abstract We present the various methods of hypothesis testing that one typically encounters in a mathematical statistics course. The focus will be on conditions for using each test, the hypothesis tested by each test, and the appropriate (and inappropriate) ways of using each test. We conclude by summarizing the di®erent tests (what conditions must be met to use them, what the test statistic is, and what the critical region is). Contents 1 Types of Hypotheses and Test Statistics 2 1.1 Introduction . 2 1.2 Types of Hypotheses . 3 1.3 Types of Statistics . 3 2 z-Tests and t-Tests 5 2.1 Testing Means I: Large Sample Size or Known Variance . 5 2.2 Testing Means II: Small Sample Size and Unknown Variance . 9 3 Testing the Variance 12 4 Testing Proportions 13 4.1 Testing Proportions I: One Proportion . 13 4.2 Testing Proportions II: K Proportions . 15 4.3 Testing r £ c Contingency Tables . 17 4.4 Incomplete r £ c Contingency Tables Tables . 18 5 Normal Regression Analysis 19 6 Non-parametric Tests 21 6.1 Tests of Signs . 21 6.2 Tests of Ranked Signs . 22 6.3 Tests Based on Runs . 23 ¤E-mail: [email protected] yE-mail: [email protected] 1 7 Summary 26 7.1 z-tests . 26 7.2 t-tests . 27 7.3 Tests comparing means . 27 7.4 Variance Test . 28 7.5 Proportions . 28 7.6 Contingency Tables .
    [Show full text]
  • Understanding Statistical Hypothesis Testing: the Logic of Statistical Inference
    Review Understanding Statistical Hypothesis Testing: The Logic of Statistical Inference Frank Emmert-Streib 1,2,* and Matthias Dehmer 3,4,5 1 Predictive Society and Data Analytics Lab, Faculty of Information Technology and Communication Sciences, Tampere University, 33100 Tampere, Finland 2 Institute of Biosciences and Medical Technology, Tampere University, 33520 Tampere, Finland 3 Institute for Intelligent Production, Faculty for Management, University of Applied Sciences Upper Austria, Steyr Campus, 4040 Steyr, Austria 4 Department of Mechatronics and Biomedical Computer Science, University for Health Sciences, Medical Informatics and Technology (UMIT), 6060 Hall, Tyrol, Austria 5 College of Computer and Control Engineering, Nankai University, Tianjin 300000, China * Correspondence: [email protected]; Tel.: +358-50-301-5353 Received: 27 July 2019; Accepted: 9 August 2019; Published: 12 August 2019 Abstract: Statistical hypothesis testing is among the most misunderstood quantitative analysis methods from data science. Despite its seeming simplicity, it has complex interdependencies between its procedural components. In this paper, we discuss the underlying logic behind statistical hypothesis testing, the formal meaning of its components and their connections. Our presentation is applicable to all statistical hypothesis tests as generic backbone and, hence, useful across all application domains in data science and artificial intelligence. Keywords: hypothesis testing; machine learning; statistics; data science; statistical inference 1. Introduction We are living in an era that is characterized by the availability of big data. In order to emphasize the importance of this, data have been called the ‘oil of the 21st Century’ [1]. However, for dealing with the challenges posed by such data, advanced analysis methods are needed.
    [Show full text]
  • Chapter 8 Key Ideas Hypothesis (Null and Alternative)
    Chapter 8 Key Ideas Hypothesis (Null and Alternative), Hypothesis Test, Test Statistic, P-value Type I Error, Type II Error, Significance Level, Power Section 8-1: Overview Confidence Intervals (Chapter 7) are great for estimating the value of a parameter, but in some situations estimation is not the goal. For example, suppose that researchers somehow know that the average number of fast food meals eaten per week by families in the 1990s was 3.5. Furthermore, they are interested in seeing if families today eat more fast food than in the 1990s. In this case, researchers don’t need to estimate how much fast food families eat today. All they need to know is whether that average is larger than 3.5. For the purposes of this example, let µ denote the average number of fast food meals eaten per week by families today. The researchers’ question could then be formulated as follows: Is the average number of meals eaten today the same as in the 1990s ( µ = 3.5) or is it more ( µ > 3.5)? There are two different possibilities, and the correct answer is impossible to know, since µ is unknown. However, the researchers could observe the sample mean and determine which possibility is the most likely . This process is called hypothesis testing . Definition Hypothesis – A claim or statement about a property of the population (e.g. “ µ = 3.5” from above) Hypothesis Test – A standard procedure for testing a claim about a property of the population. To guide this discussion of hypothesis testing, remember the “Rare Event Rule”: If, under a given assumption, the probability of a particular observed event is exceptionally small, we conclude that the assumption is probably not correct.
    [Show full text]
  • Week 10: Causality with Measured Confounding
    Week 10: Causality with Measured Confounding Brandon Stewart1 Princeton November 28 and 30, 2016 1These slides are heavily influenced by Matt Blackwell, Jens Hainmueller, Erin Hartman, Kosuke Imai and Gary King. Stewart (Princeton) Week 10: Measured Confounding November 28 and 30, 2016 1 / 176 Where We've Been and Where We're Going... Last Week I regression diagnostics This Week I Monday: F experimental Ideal F identification with measured confounding I Wednesday: F regression estimation Next Week I identification with unmeasured confounding I instrumental variables Long Run I causality with measured confounding ! unmeasured confounding ! repeated data Questions? Stewart (Princeton) Week 10: Measured Confounding November 28 and 30, 2016 2 / 176 1 The Experimental Ideal 2 Assumption of No Unmeasured Confounding 3 Fun With Censorship 4 Regression Estimators 5 Agnostic Regression 6 Regression and Causality 7 Regression Under Heterogeneous Effects 8 Fun with Visualization, Replication and the NYT 9 Appendix Subclassification Identification under Random Assignment Estimation Under Random Assignment Blocking Stewart (Princeton) Week 10: Measured Confounding November 28 and 30, 2016 3 / 176 1 The Experimental Ideal 2 Assumption of No Unmeasured Confounding 3 Fun With Censorship 4 Regression Estimators 5 Agnostic Regression 6 Regression and Causality 7 Regression Under Heterogeneous Effects 8 Fun with Visualization, Replication and the NYT 9 Appendix Subclassification Identification under Random Assignment Estimation Under Random Assignment Blocking Stewart
    [Show full text]
  • Skedastic: Heteroskedasticity Diagnostics for Linear Regression
    Package ‘skedastic’ June 14, 2021 Type Package Title Heteroskedasticity Diagnostics for Linear Regression Models Version 1.0.3 Description Implements numerous methods for detecting heteroskedasticity (sometimes called heteroscedasticity) in the classical linear regression model. These include a test based on Anscombe (1961) <https://projecteuclid.org/euclid.bsmsp/1200512155>, Ramsey's (1969) BAMSET Test <doi:10.1111/j.2517-6161.1969.tb00796.x>, the tests of Bickel (1978) <doi:10.1214/aos/1176344124>, Breusch and Pagan (1979) <doi:10.2307/1911963> with and without the modification proposed by Koenker (1981) <doi:10.1016/0304-4076(81)90062-2>, Carapeto and Holt (2003) <doi:10.1080/0266476022000018475>, Cook and Weisberg (1983) <doi:10.1093/biomet/70.1.1> (including their graphical methods), Diblasi and Bowman (1997) <doi:10.1016/S0167-7152(96)00115-0>, Dufour, Khalaf, Bernard, and Genest (2004) <doi:10.1016/j.jeconom.2003.10.024>, Evans and King (1985) <doi:10.1016/0304-4076(85)90085-5> and Evans and King (1988) <doi:10.1016/0304-4076(88)90006-1>, Glejser (1969) <doi:10.1080/01621459.1969.10500976> as formulated by Mittelhammer, Judge and Miller (2000, ISBN: 0-521-62394-4), Godfrey and Orme (1999) <doi:10.1080/07474939908800438>, Goldfeld and Quandt (1965) <doi:10.1080/01621459.1965.10480811>, Harrison and McCabe (1979) <doi:10.1080/01621459.1979.10482544>, Harvey (1976) <doi:10.2307/1913974>, Honda (1989) <doi:10.1111/j.2517-6161.1989.tb01749.x>, Horn (1981) <doi:10.1080/03610928108828074>, Li and Yao (2019) <doi:10.1016/j.ecosta.2018.01.001>
    [Show full text]
  • Chapter 10 Heteroskedasticity
    Chapter 10 Heteroskedasticity In the multiple regression model yX , it is assumed that VI() 2 , i.e., 22 Var()i , Cov(ij ) 0, i j 1, 2,..., n . In this case, the diagonal elements of the covariance matrix of are the same indicating that the variance of each i is same and off-diagonal elements of the covariance matrix of are zero indicating that all disturbances are pairwise uncorrelated. This property of constancy of variance is termed as homoskedasticity and disturbances are called as homoskedastic disturbances. In many situations, this assumption may not be plausible, and the variances may not remain the same. The disturbances whose variances are not constant across the observations are called heteroskedastic disturbance, and this property is termed as heteroskedasticity. In this case 2 Var()ii , i 1,2,..., n and disturbances are pairwise uncorrelated. The covariance matrix of disturbances is 2 00 1 00 2 Vdiag( ) (22 , ,..., 2 )2 . 12 n 2 00 n Regression Analysis | Chapter 10 | Heteroskedasticity | Shalabh, IIT Kanpur 1 Graphically, the following pictures depict homoskedasticity and heteroskedasticity. Homoskedasticity Heteroskedasticity (Var(y) increases with x) Heteroskedasticity (Var(y) decreases with x) Examples: Suppose in a simple linear regression model, x denote the income and y denotes the expenditure on food. It is observed that as the income increases, the expenditure on food increases because of the choice and varieties in food increase, in general, up to a certain extent. So the variance of observations on y will not remain constant as income changes. The assumption of homoscedasticity implies that the consumption pattern of food will remain the same irrespective of the income of the person.
    [Show full text]
  • Introduction to Hypothesis Testing 3
    CHAPTER Introduction to Hypothesis 8 Testing 8.1 Inferential Statistics and Hypothesis Testing 8.2 Four Steps to LEARNING OBJECTIVES Hypothesis Testing After reading this chapter, you should be able to: 8.3 Hypothesis Testing and Sampling Distributions 8.4 Making a Decision: 1 Identify the four steps of hypothesis testing. Types of Error 8.5 Testing a Research 2 Define null hypothesis, alternative hypothesis, Hypothesis: Examples level of significance, test statistic, p value, and Using the z Test statistical significance. 8.6 Research in Focus: Directional Versus 3 Define Type I error and Type II error, and identify the Nondirectional Tests type of error that researchers control. 8.7 Measuring the Size of an Effect: Cohen’s d 4 Calculate the one-independent sample z test and 8.8 Effect Size, Power, and interpret the results. Sample Size 8.9 Additional Factors That 5 Distinguish between a one-tailed and two-tailed test, Increase Power and explain why a Type III error is possible only with 8.10 SPSS in Focus: one-tailed tests. A Preview for Chapters 9 to 18 6 Explain what effect size measures and compute a 8.11 APA in Focus: Reporting the Test Cohen’s d for the one-independent sample z test. Statistic and Effect Size 7 Define power and identify six factors that influence power. 8 Summarize the results of a one-independent sample z test in American Psychological Association (APA) format. 2 PART III: PROBABILITY AND THE FOUNDATIONS OF INFERENTIAL STATISTICS 8.1 INFERENTIAL STATISTICS AND HYPOTHESIS TESTING We use inferential statistics because it allows us to measure behavior in samples to learn more about the behavior in populations that are often too large or inaccessi­ ble.
    [Show full text]
  • Sample Quiz #4
    Statistics 13V NAME: Sample Quiz 4 Last six digits of Student ID#: Inserted answers in italics; multiple choice answers in bold. 1. A chi-square test of the relationship between personal perception of emotional health and marital status led to rejection of the null hypothesis, indicating that there is a relationship between these two variables. One conclusion that can be drawn is: A. Marriage leads to better emotional health. B. Better emotional health leads to marriage. C. The more emotionally healthy someone is, the more likely they are to be married. D. There are likely to be confounding variables related to both emotional health and marital status. Questions 2 to 5: A survey is done in which people are asked how often they exceed speed limits. The data are then categorized by age (Under 30, 30 and Over) and answer (Always, Not Always). The following computer output shows observed and expected counts, the chi-square statistic and the p-value. Expected counts are printed below observed counts Always Not Always Total Under 30 100 100 200 70.00 130.00 30 and Over 40 160 200 70.00 130.00 Total 140 260 400 Chi-Sq = 12.857 + 6.923 + 12.857 + 6.923 = 39.560 DF = 1, P-Value = 0.000 2. What is the relative risk of always exceeding the speed limit for people under 30 compared to always exceeding it for people 30 and over? Show your calculations. Risk for under 30 = 100/200 = .5 Risk for 30 and over = 40/200 = .2 Relative risk = .5/.2 = 2.5 3.
    [Show full text]