Heteroskedasticity in the Linear Model

Total Page:16

File Type:pdf, Size:1020Kb

Heteroskedasticity in the Linear Model Short Guides to Microeconometrics Kurt Schmidheiny Fall 2020 Unversit¨atBasel Heteroskedasticity in the Linear Model 1 Introduction This handout extends the handout on \The Multiple Linear Regression model" and refers to its definitions and assumptions in section 2. This handouts relaxes the homoscedasticity assumption (OLS4a) and shows how the parameters of the linear model are correctly estimated and tested when the error terms are heteroscedastic (OLS4b). 2 The Econometric Model Consider the multiple linear regression model for observation i = 1; :::; N 0 yi = xiβ + ui 0 where xi is a (K + 1)-dimensional row vector of K explanatory variables and a constant, β is a (K + 1)-dimensional column vector of parameters and ui is a scalar called the error term. Assume OLS1, OLS2, OLS3 and OLS4: Error Variance b) conditional heteroscedasticity: 2 2 2 V [uijxi] = σi = σ !i = σ !(xi) 2 0 and E[ui xixi] = QXΩX is p.d. and finite 2 where !(:) is a function constant across i. The decomposition of σi into 2 !i and σ is arbitrary but useful. Version: 29-10-2020, 23:35 Heteroskedasticity in the Linear Model 2 Note that under OLS2 (i.i.d. sample) the errors are unconditionally 2 homoscedastic, V [ui] = σ but allowed to be conditionally heteroscedas- 2 tic, V [uijxi] = σi . Assuming OLS2 and OLS3c provides that the errors are also not conditionally autocorrelated, i.e. 8i 6= j : Cov[ui; ujjxi; xj] = E[uiujjxi; xj] = E[uijxi] · E[ujjxj] = 0 . Also note that the condition- ing on xi is less restrictive than it may seem: if the conditional variance V [uijxi] depends on other exogenous variables (or functions of them), we can include these variables in xi and set the corresponding β parameters to zero. 0 The variance-covariance of the vector of error terms u = (u1; u2; :::; uN ) in the whole sample is therefore V [ujX] = E[uu0jX] = σ2Ω 0 2 1 0 1 σ1 0 ··· 0 !1 0 ··· 0 B 0 σ2 ··· 0 C B 0 ! ··· 0 C B 2 C 2 B 2 C = B . C = σ B . C B . .. C B . .. C @ . A @ . A 2 0 0 ··· σN 0 0 ··· !N 3 A Generic Case: Groupewise Heteroskedasticity Heteroskedasticity is sometimes a direct consequence of the construc- tion of the data. Consider the following linear regression model with homoscedastic errors 0 yi = xiβ + ui 2 with V [uijxi] = V [ui] = σ . Assume that instead of the individual observations yi and xi only the mean values yg and xg for g = 1; :::; G groups are observed. The error term in the resulting regression model 0 yg = xgβ + ug 3 Short Guides to Microeconometrics 2 2 is now conditionally heteroskedastic with V [ugjNg] = σg = σ =Ng, where Ng is a random variable with the number of observations in group g. Heteroskedasticity in the Linear Model 4 4 Estimation with OLS The parameter β can be estimated with the usual OLS estimator 0 −1 0 βbOLS = (X X) X y The OLS estimator of β remains unbiased under OLS1, OLS2, OLS3c, OLS4b, and OLS5 in small samples. Additionally assuming OLS3a, it is normally distributed in small samples. It is consistent and approximately normally distributed under OLS1, OLS2, OLS3d, OLS4a or OLS4b and OLS5, in large samples. However, the OLS estimator is not efficient any more. More importantly, the usual standard errors of the OLS estimator and tests (t-, F -, z-, Wald-) based on them are not valid any more. 5 Estimating the Variance of the OLS Estimator The small sample covariance matrix of βbOLS is under OLS4b 0 −1 0 2 0 −1 V [βbOLSjX] = (X X) X σ ΩX (X X) 2 0 −1 and differs from usual OLS where V [βbOLSjX] = σ (X X) . Conse- 2 0 −1 quently, the usual estimator Vb[βbOLSjX] = σc(X X) is biased. Usual small sample test procedures, such as the F - or t-Test, based on the usual estimator are therefore not valid. The OLS estimator is asymptotically normally distributed under OLS1, OLS2, OLS3d, and OLS5 p d −1 −1 N(βb − β) −! N 0;QXX QXΩX QXX 2 0 0 where QXΩX = E[ui xixi] and QXX = E[xixi]. The OLS estimator is therefore approximately normally distributed in large samples as A βb ∼ N β; Avar[βb] 5 Short Guides to Microeconometrics −1 −1 −1 where Avar[βb] = N QXX QXΩX QXX can be consistently estimated as " N # 0 −1 X 2 0 0 −1 Avar[ [βb] = (X X) ubi xixi (X X) i=1 with some additional assumptions on higher order moments of xi (see White 1980). This so-called White or Eicker-Huber-White estimator of the covari- ance matrix is a heteroskedasticity-consistent covariance matrix estimator that does not require any assumptions on the form of heteroscedasticity (though we assumed independence of the error terms in OLS2 ). Stan- dard errors based on the White estimator are often called robust. We can perform the usual z- and Wald-test for large samples using the White covariance estimator. Note: t- and F -Tests using the White covariance estimator are only asymptotically valid because the White covariance estimator is consistent but not unbiased. It is therefore more appropriate to use large sample tests (z, Wald). Bootstrapping (see the handout on \The Bootstrap") is an alternative method to estimate a heteroscedasticity robust covariance matrix. 6 Testing for Heteroskedasticity There are several tests for the assumption that the error term is ho- moskedastic. White (1980)'s test is general and does not presume a par- ticular form of heteroskedasticity. Unfortunately, little can be said about its power and it has poor small sample properties unless the number of 2 regressors is very small. If we have prior knowledge that the variance σi is a linear (in parameters) function of explanatory variables, the Breusch- Pagan (1979) test is more powerful. Koenker (1981) proposes a variant of the Breusch-Pagan test that does not assume normally distributed errors. Note: In practice we often do not test for heteroskedasticity but di- rectly report heteroskedasticity-robust standard errors. Heteroskedasticity in the Linear Model 6 7 Estimation with GLS/WLS when Ω is Known When Ω is known, β is efficiently estimated with generalized least squares (GLS) −1 0 −1 0 −1 βbGLS = X Ω X X Ω y: The GLS estimator simplifies in the case of heteroskedasticity to ~ 0 ~ −1 ~ 0 βbW LS = (X X) X y~ where 0 p 1 0 0 p 1 y1= !1 x1= !1 B . C ~ B . C y~ = B . C ; X = B . C @ A @ A p 0 p yN = !N xN = !N and is called weighted least squares (WLS) estimator. The WLS estimator of β can therefore be estimated by running OLS with the transformed variables. Note: the above transformation of the explanatory variables also ap- p plies to the constant, i.e.x ~i0 = 1= !i. The OLS regression using the transformed variables does not include an additional constant. The WLS estimator minimizes the sum of squared residuals weighted by 1=!i: N N X 0 2 X 1 0 2 S (α; β) = (~yi − x~iβ) = (yi − xiβ) ! min ! β i=1 i=1 i The WLS estimator of β is unbiased and efficient (under OLS1, OLS2, OLS3c, OLS4b, and OLS5 ) and normally distributed additionally assum- ing OLS3a (normality) in small samples. The WLS estimator of β is consistent, asymptotically efficient and ap- proximately normally distributed under OLS4b (conditional heteroscedas- ticity) and OLS2 , OLS1, OLS3d, and a modification of OLS5 7 Short Guides to Microeconometrics OLS5: Identifiability X0Ω−1X is p.d. and finite −1 0 E[!i xixi] = QXΩ−1X is p.d. and finite where !i = !(xi) is a function of xi. The covariance matrix of βbW LS can be estimated as 2 ~ 0 ~ −1 Avar[ [βbW LS] = Vb[βbW LS] = σc(X X) 0 0 where σc2 = ub~ u=b~ (N − K − 1) in small samples and σc2 = ub~ u=Nb~ in large ~ samples and ub~ =y ~ − XβbGLS. Usual tests (t-, F -) for small samples are valid (under OLS1, OLS2, OLS3a, OLS4b and OLS5 ; usual tests (z, Wald) for large samples are also valid (under OLS1, OLS2, OLS3d, OLS4b and OLS5 ). 8 Estimation with FGLS/FWLS when Ω is Unknown In practice, Ω is typically unknown. However, we can model the !i's as a function of the data and estimate this relationship. Feasible gener- alized least squares (FGLS) replaces !i by their predicted values !bi and calculates then βbF GLS as if !i were known. A useful model for the error variance is 2 0 σi = V [uijzi] = exp(ziδ) where zi are L + 1 variables that may belong to xi including a constant and δ is a vector of parameters. We can estimate the auxiliary regression 2 0 ubi = exp(ziδ) + νi 0 by nonlinear least squares (NLLS) where ubi = yi−xiβbOLS or alternatively, 2 0 log(ubi ) = ziδ + νi by ordinary least squares (OLS). In both cases, we use the predictions 0 !bi = exp(ziδb) Heteroskedasticity in the Linear Model 8 in the calculations for βbF GLS and Avar[ [βbF GLS]. The FGLS estimator is consistent and approximately normally dis- tributed in large samples under OLS1, OLS2 (fxi; zi; yig i.i.d.), OLS3d, 2 OLS4b, OLS5 and some additional more technical assumptions. If σi is correctly specified, βF GLS is asymptotically efficient and the usual tests (z, Wald) for large samples are valid; small samples tests are only asymp- 2 totically valid and nothing is gained from using them. If σi is not correctly specified, the usual covariance matrix is inconsistent and tests (z, Wald) invalid.
Recommended publications
  • Three Statistical Testing Procedures in Logistic Regression: Their Performance in Differential Item Functioning (DIF) Investigation
    Research Report Three Statistical Testing Procedures in Logistic Regression: Their Performance in Differential Item Functioning (DIF) Investigation Insu Paek December 2009 ETS RR-09-35 Listening. Learning. Leading.® Three Statistical Testing Procedures in Logistic Regression: Their Performance in Differential Item Functioning (DIF) Investigation Insu Paek ETS, Princeton, New Jersey December 2009 As part of its nonprofit mission, ETS conducts and disseminates the results of research to advance quality and equity in education and assessment for the benefit of ETS’s constituents and the field. To obtain a PDF or a print copy of a report, please visit: http://www.ets.org/research/contact.html Copyright © 2009 by Educational Testing Service. All rights reserved. ETS, the ETS logo, GRE, and LISTENING. LEARNING. LEADING. are registered trademarks of Educational Testing Service (ETS). SAT is a registered trademark of the College Board. Abstract Three statistical testing procedures well-known in the maximum likelihood approach are the Wald, likelihood ratio (LR), and score tests. Although well-known, the application of these three testing procedures in the logistic regression method to investigate differential item function (DIF) has not been rigorously made yet. Employing a variety of simulation conditions, this research (a) assessed the three tests’ performance for DIF detection and (b) compared DIF detection in different DIF testing modes (targeted vs. general DIF testing). Simulation results showed small differences between the three tests and different testing modes. However, targeted DIF testing consistently performed better than general DIF testing; the three tests differed more in performance in general DIF testing and nonuniform DIF conditions than in targeted DIF testing and uniform DIF conditions; and the LR and score tests consistently performed better than the Wald test.
    [Show full text]
  • Wald (And Score) Tests
    Wald (and Score) Tests 1 / 18 Vector of MLEs is Asymptotically Normal That is, Multivariate Normal This yields I Confidence intervals I Z-tests of H0 : θj = θ0 I Wald tests I Score Tests I Indirectly, the Likelihood Ratio tests 2 / 18 Under Regularity Conditions (Thank you, Mr. Wald) a.s. I θbn → θ √ d −1 I n(θbn − θ) → T ∼ Nk 0, I(θ) 1 −1 I So we say that θbn is asymptotically Nk θ, n I(θ) . I I(θ) is the Fisher Information in one observation. I A k × k matrix ∂2 I(θ) = E[− log f(Y ; θ)] ∂θi∂θj I The Fisher Information in the whole sample is nI(θ) 3 / 18 H0 : Cθ = h Suppose θ = (θ1, . θ7), and the null hypothesis is I θ1 = θ2 I θ6 = θ7 1 1 I 3 (θ1 + θ2 + θ3) = 3 (θ4 + θ5 + θ6) We can write null hypothesis in matrix form as θ1 θ2 1 −1 0 0 0 0 0 θ3 0 0 0 0 0 0 1 −1 θ4 = 0 1 1 1 −1 −1 −1 0 θ5 0 θ6 θ7 4 / 18 p Suppose H0 : Cθ = h is True, and Id(θ)n → I(θ) By Slutsky 6a (Continuous mapping), √ √ d −1 0 n(Cθbn − Cθ) = n(Cθbn − h) → CT ∼ Nk 0, CI(θ) C and −1 p −1 Id(θ)n → I(θ) . Then by Slutsky’s (6c) Stack Theorem, √ ! n(Cθbn − h) d CT → . −1 I(θ)−1 Id(θ)n Finally, by Slutsky 6a again, −1 0 0 −1 Wn = n(Cθb − h) (CId(θ)n C ) (Cθb − h) →d W = (CT − 0)0(CI(θ)−1C0)−1(CT − 0) ∼ χ2(r) 5 / 18 The Wald Test Statistic −1 0 0 −1 Wn = n(Cθbn − h) (CId(θ)n C ) (Cθbn − h) I Again, null hypothesis is H0 : Cθ = h I Matrix C is r × k, r ≤ k, rank r I All we need is a consistent estimator of I(θ) I I(θb) would do I But it’s inconvenient I Need to compute partial derivatives and expected values in ∂2 I(θ) = E[− log f(Y ; θ)] ∂θi∂θj 6 / 18 Observed Fisher Information I To find θbn, minimize the minus log likelihood.
    [Show full text]
  • Comparison of Wald, Score, and Likelihood Ratio Tests for Response Adaptive Designs
    Journal of Statistical Theory and Applications Volume 10, Number 4, 2011, pp. 553-569 ISSN 1538-7887 Comparison of Wald, Score, and Likelihood Ratio Tests for Response Adaptive Designs Yanqing Yi1∗and Xikui Wang2 1 Division of Community Health and Humanities, Faculty of Medicine, Memorial University of Newfoundland, St. Johns, Newfoundland, Canada A1B 3V6 2 Department of Statistics, University of Manitoba, Winnipeg, Manitoba, Canada R3T 2N2 Abstract Data collected from response adaptive designs are dependent. Traditional statistical methods need to be justified for the use in response adaptive designs. This paper gener- alizes the Rao's score test to response adaptive designs and introduces a generalized score statistic. Simulation is conducted to compare the statistical powers of the Wald, the score, the generalized score and the likelihood ratio statistics. The overall statistical power of the Wald statistic is better than the score, the generalized score and the likelihood ratio statistics for small to medium sample sizes. The score statistic does not show good sample properties for adaptive designs and the generalized score statistic is better than the score statistic under the adaptive designs considered. When the sample size becomes large, the statistical power is similar for the Wald, the sore, the generalized score and the likelihood ratio test statistics. MSC: 62L05, 62F03 Keywords and Phrases: Response adaptive design, likelihood ratio test, maximum likelihood estimation, Rao's score test, statistical power, the Wald test ∗Corresponding author. Fax: 1-709-777-7382. E-mail addresses: [email protected] (Yanqing Yi), xikui [email protected] (Xikui Wang) Y. Yi and X.
    [Show full text]
  • Heteroscedastic Errors
    Heteroscedastic Errors ◮ Sometimes plots and/or tests show that the error variances 2 σi = Var(ǫi ) depend on i ◮ Several standard approaches to fixing the problem, depending on the nature of the dependence. ◮ Weighted Least Squares. ◮ Transformation of the response. ◮ Generalized Linear Models. Richard Lockhart STAT 350: Heteroscedastic Errors and GLIM Weighted Least Squares ◮ Suppose variances are known except for a constant factor. 2 2 ◮ That is, σi = σ /wi . ◮ Use weighted least squares. (See Chapter 10 in the text.) ◮ This usually arises realistically in the following situations: ◮ Yi is an average of ni measurements where you know ni . Then wi = ni . 2 ◮ Plots suggest that σi might be proportional to some power of 2 γ γ some covariate: σi = kxi . Then wi = xi− . Richard Lockhart STAT 350: Heteroscedastic Errors and GLIM Variances depending on (mean of) Y ◮ Two standard approaches are available: ◮ Older approach is transformation. ◮ Newer approach is use of generalized linear model; see STAT 402. Richard Lockhart STAT 350: Heteroscedastic Errors and GLIM Transformation ◮ Compute Yi∗ = g(Yi ) for some function g like logarithm or square root. ◮ Then regress Yi∗ on the covariates. ◮ This approach sometimes works for skewed response variables like income; ◮ after transformation we occasionally find the errors are more nearly normal, more homoscedastic and that the model is simpler. ◮ See page 130ff and check under transformations and Box-Cox in the index. Richard Lockhart STAT 350: Heteroscedastic Errors and GLIM Generalized Linear Models ◮ Transformation uses the model T E(g(Yi )) = xi β while generalized linear models use T g(E(Yi )) = xi β ◮ Generally latter approach offers more flexibility.
    [Show full text]
  • Statistical Asymptotics Part II: First-Order Theory
    First-Order Asymptotic Theory Statistical Asymptotics Part II: First-Order Theory Andrew Wood School of Mathematical Sciences University of Nottingham APTS, April 15-19 2013 Andrew Wood Statistical Asymptotics Part II: First-Order Theory First-Order Asymptotic Theory Structure of the Chapter This chapter covers asymptotic normality and related results. Topics: MLEs, log-likelihood ratio statistics and their asymptotic distributions; M-estimators and their first-order asymptotic theory. Initially we focus on the case of the MLE of a scalar parameter θ. Then we study the case of the MLE of a vector θ, first without and then with nuisance parameters. Finally, we consider the more general setting of M-estimators. Andrew Wood Statistical Asymptotics Part II: First-Order Theory First-Order Asymptotic Theory Motivation Statistical inference typically requires approximations because exact answers are usually not available. Asymptotic theory provides useful approximations to densities or distribution functions. These approximations are based on results from probability theory. The theory underlying these approximation techniques is valid as some quantity, typically the sample size n [or more generally some measure of information], goes to infinity, but the approximations obtained are often accurate even for small sample sizes. Andrew Wood Statistical Asymptotics Part II: First-Order Theory First-Order Asymptotic Theory Test statistics Consider testing the null hypothesis H0 : θ = θ0, where θ0 is an arbitrary specified point in Ωθ. If desired, we may
    [Show full text]
  • Power Comparisons of the Mann-Whitney U and Permutation Tests
    Power Comparisons of the Mann-Whitney U and Permutation Tests Abstract: Though the Mann-Whitney U-test and permutation tests are often used in cases where distribution assumptions for the two-sample t-test for equal means are not met, it is not widely understood how the powers of the two tests compare. Our goal was to discover under what circumstances the Mann-Whitney test has greater power than the permutation test. The tests’ powers were compared under various conditions simulated from the Weibull distribution. Under most conditions, the permutation test provided greater power, especially with equal sample sizes and with unequal standard deviations. However, the Mann-Whitney test performed better with highly skewed data. Background and Significance: In many psychological, biological, and clinical trial settings, distributional differences among testing groups render parametric tests requiring normality, such as the z test and t test, unreliable. In these situations, nonparametric tests become necessary. Blair and Higgins (1980) illustrate the empirical invalidity of claims made in the mid-20th century that t and F tests used to detect differences in population means are highly insensitive to violations of distributional assumptions, and that non-parametric alternatives possess lower power. Through power testing, Blair and Higgins demonstrate that the Mann-Whitney test has much higher power relative to the t-test, particularly under small sample conditions. This seems to be true even when Welch’s approximation and pooled variances are used to “account” for violated t-test assumptions (Glass et al. 1972). With the proliferation of powerful computers, computationally intensive alternatives to the Mann-Whitney test have become possible.
    [Show full text]
  • Research Report Statistical Research Unit Goteborg University Sweden
    Research Report Statistical Research Unit Goteborg University Sweden Testing for multivariate heteroscedasticity Thomas Holgersson Ghazi Shukur Research Report 2003:1 ISSN 0349-8034 Mailing address: Fax Phone Home Page: Statistical Research Nat: 031-77312 74 Nat: 031-77310 00 http://www.stat.gu.se/stat Unit P.O. Box 660 Int: +4631 773 12 74 Int: +4631 773 1000 SE 405 30 G6teborg Sweden Testing for Multivariate Heteroscedasticity By H.E.T. Holgersson Ghazi Shukur Department of Statistics Jonkoping International GOteborg university Business school SE-405 30 GOteborg SE-55 111 Jonkoping Sweden Sweden Abstract: In this paper we propose a testing technique for multivariate heteroscedasticity, which is expressed as a test of linear restrictions in a multivariate regression model. Four test statistics with known asymptotical null distributions are suggested, namely the Wald (W), Lagrange Multiplier (LM), Likelihood Ratio (LR) and the multivariate Rao F-test. The critical values for the statistics are determined by their asymptotic null distributions, but also bootstrapped critical values are used. The size, power and robustness of the tests are examined in a Monte Carlo experiment. Our main findings are that all the tests limit their nominal sizes asymptotically, but some of them have superior small sample properties. These are the F, LM and bootstrapped versions of Wand LR tests. Keywords: heteroscedasticity, hypothesis test, bootstrap, multivariate analysis. I. Introduction In the last few decades a variety of methods has been proposed for testing for heteroscedasticity among the error terms in e.g. linear regression models. The assumption of homoscedasticity means that the disturbance variance should be constant (or homoscedastic) at each observation.
    [Show full text]
  • Econometrics-I-11.Pdf
    Econometrics I Professor William Greene Stern School of Business Department of Economics 11-1/78 Part 11: Hypothesis Testing - 2 Econometrics I Part 11 – Hypothesis Testing 11-2/78 Part 11: Hypothesis Testing - 2 Classical Hypothesis Testing We are interested in using the linear regression to support or cast doubt on the validity of a theory about the real world counterpart to our statistical model. The model is used to test hypotheses about the underlying data generating process. 11-3/78 Part 11: Hypothesis Testing - 2 Types of Tests Nested Models: Restriction on the parameters of a particular model y = 1 + 2x + 3T + , 3 = 0 (The “treatment” works; 3 0 .) Nonnested models: E.g., different RHS variables yt = 1 + 2xt + 3xt-1 + t yt = 1 + 2xt + 3yt-1 + wt (Lagged effects occur immediately or spread over time.) Specification tests: ~ N[0,2] vs. some other distribution (The “null” spec. is true or some other spec. is true.) 11-4/78 Part 11: Hypothesis Testing - 2 Hypothesis Testing Nested vs. nonnested specifications y=b1x+e vs. y=b1x+b2z+e: Nested y=bx+e vs. y=cz+u: Not nested y=bx+e vs. logy=clogx: Not nested y=bx+e; e ~ Normal vs. e ~ t[.]: Not nested Fixed vs. random effects: Not nested Logit vs. probit: Not nested x is (not) endogenous: Maybe nested. We’ll see … Parametric restrictions Linear: R-q = 0, R is JxK, J < K, full row rank General: r(,q) = 0, r = a vector of J functions, R(,q) = r(,q)/’. Use r(,q)=0 for linear and nonlinear cases 11-5/78 Part 11: Hypothesis Testing - 2 Broad Approaches Bayesian: Does not reach a firm conclusion.
    [Show full text]
  • T-Statistic Based Correlation and Heterogeneity Robust Inference
    t-Statistic Based Correlation and Heterogeneity Robust Inference Rustam IBRAGIMOV Economics Department, Harvard University, 1875 Cambridge Street, Cambridge, MA 02138 Ulrich K. MÜLLER Economics Department, Princeton University, Fisher Hall, Princeton, NJ 08544 ([email protected]) We develop a general approach to robust inference about a scalar parameter of interest when the data is potentially heterogeneous and correlated in a largely unknown way. The key ingredient is the following result of Bakirov and Székely (2005) concerning the small sample properties of the standard t-test: For a significance level of 5% or lower, the t-test remains conservative for underlying observations that are independent and Gaussian with heterogenous variances. One might thus conduct robust large sample inference as follows: partition the data into q ≥ 2 groups, estimate the model for each group, and conduct a standard t-test with the resulting q parameter estimators of interest. This results in valid and in some sense efficient inference when the groups are chosen in a way that ensures the parameter estimators to be asymptotically independent, unbiased and Gaussian of possibly different variances. We provide examples of how to apply this approach to time series, panel, clustered and spatially correlated data. KEY WORDS: Dependence; Fama–MacBeth method; Least favorable distribution; t-test; Variance es- timation. 1. INTRODUCTION property of the correlations. The key ingredient to the strategy is a result by Bakirov and Székely (2005) concerning the small Empirical analyses in economics often face the difficulty that sample properties of the usual t-test used for inference on the the data is correlated and heterogeneous in some unknown fash- mean of independent normal variables: For significance levels ion.
    [Show full text]
  • Package 'Lmtest'
    Package ‘lmtest’ September 9, 2020 Title Testing Linear Regression Models Version 0.9-38 Date 2020-09-09 Description A collection of tests, data sets, and examples for diagnostic checking in linear regression models. Furthermore, some generic tools for inference in parametric models are provided. LazyData yes Depends R (>= 3.0.0), stats, zoo Suggests car, strucchange, sandwich, dynlm, stats4, survival, AER Imports graphics License GPL-2 | GPL-3 NeedsCompilation yes Author Torsten Hothorn [aut] (<https://orcid.org/0000-0001-8301-0471>), Achim Zeileis [aut, cre] (<https://orcid.org/0000-0003-0918-3766>), Richard W. Farebrother [aut] (pan.f), Clint Cummins [aut] (pan.f), Giovanni Millo [ctb], David Mitchell [ctb] Maintainer Achim Zeileis <[email protected]> Repository CRAN Date/Publication 2020-09-09 05:30:11 UTC R topics documented: bgtest . .2 bondyield . .4 bptest . .6 ChickEgg . .8 coeftest . .9 coxtest . 11 currencysubstitution . 13 1 2 bgtest dwtest . 14 encomptest . 16 ftemp . 18 gqtest . 19 grangertest . 21 growthofmoney . 22 harvtest . 23 hmctest . 25 jocci . 26 jtest . 28 lrtest . 29 Mandible . 31 moneydemand . 31 petest . 33 raintest . 35 resettest . 36 unemployment . 38 USDistLag . 39 valueofstocks . 40 wages ............................................ 41 waldtest . 42 Index 46 bgtest Breusch-Godfrey Test Description bgtest performs the Breusch-Godfrey test for higher-order serial correlation. Usage bgtest(formula, order = 1, order.by = NULL, type = c("Chisq", "F"), data = list(), fill = 0) Arguments formula a symbolic description for the model to be tested (or a fitted "lm" object). order integer. maximal order of serial correlation to be tested. order.by Either a vector z or a formula with a single explanatory variable like ~ z.
    [Show full text]
  • Two-Way Heteroscedastic ANOVA When the Number of Levels Is Large
    Two-way Heteroscedastic ANOVA when the Number of Levels is Large Short Running Title: Two-way ANOVA Lan Wang 1 and Michael G. Akritas 2 Revision: January, 2005 Abstract We consider testing the main treatment e®ects and interaction e®ects in crossed two-way layout when one factor or both factors have large number of levels. Random errors are allowed to be nonnormal and heteroscedastic. In heteroscedastic case, we propose new test statistics. The asymptotic distributions of our test statistics are derived under both the null hypothesis and local alternatives. The sample size per treatment combination can either be ¯xed or tend to in¯nity. Numerical simulations indicate that the proposed procedures have good power properties and maintain approximately the nominal ®-level with small sample sizes. A real data set from a study evaluating forty varieties of winter wheat in a large-scale agricultural trial is analyzed. Key words: heteroscedasticity, large number of factor levels, unbalanced designs, quadratic forms, projection method, local alternatives 1385 Ford Hall, School of Statistics, University of Minnesota, 224 Church Street SE, Minneapolis, MN 55455. Email: [email protected], phone: (612) 625-7843, fax: (612) 624-8868. 2414 Thomas Building, Department of Statistics, the Penn State University, University Park, PA 16802. E-mail: [email protected], phone: (814) 865-3631, fax: (814) 863-7114. 1 Introduction In many experiments, data are collected in the form of a crossed two-way layout. If we let Xijk denote the k-th response associated with the i-th level of factor A and j -th level of factor B, then the classical two-way ANOVA model speci¯es that: Xijk = ¹ + ®i + ¯j + γij + ²ijk (1.1) where i = 1; : : : ; a; j = 1; : : : ; b; k = 1; 2; : : : ; nij, and to be identi¯able the parameters Pa Pb Pa Pb are restricted by conditions such as i=1 ®i = j=1 ¯j = i=1 γij = j=1 γij = 0.
    [Show full text]
  • Profiling Heteroscedasticity in Linear Regression Models
    358 The Canadian Journal of Statistics Vol. 43, No. 3, 2015, Pages 358–377 La revue canadienne de statistique Profiling heteroscedasticity in linear regression models Qian M. ZHOU1*, Peter X.-K. SONG2 and Mary E. THOMPSON3 1Department of Statistics and Actuarial Science, Simon Fraser University, Burnaby, British Columbia, Canada 2Department of Biostatistics, University of Michigan, Ann Arbor, Michigan, U.S.A. 3Department of Statistics and Actuarial Science, University of Waterloo, Waterloo, Ontario, Canada Key words and phrases: Heteroscedasticity; hybrid test; information ratio; linear regression models; model- based estimators; sandwich estimators; screening; weighted least squares. MSC 2010: Primary 62J05; secondary 62J20 Abstract: Diagnostics for heteroscedasticity in linear regression models have been intensively investigated in the literature. However, limited attention has been paid on how to identify covariates associated with heteroscedastic error variances. This problem is critical in correctly modelling the variance structure in weighted least squares estimation, which leads to improved estimation efficiency. We propose covariate- specific statistics based on information ratios formed as comparisons between the model-based and sandwich variance estimators. A two-step diagnostic procedure is established, first to detect heteroscedasticity in error variances, and then to identify covariates the error variance structure might depend on. This proposed method is generalized to accommodate practical complications, such as when covariates associated with the heteroscedastic variances might not be associated with the mean structure of the response variable, or when strong correlation is present amongst covariates. The performance of the proposed method is assessed via a simulation study and is illustrated through a data analysis in which we show the importance of correct identification of covariates associated with the variance structure in estimation and inference.
    [Show full text]