7 Detection and Modeling of Nonconstant Variance

Total Page:16

File Type:pdf, Size:1020Kb

7 Detection and Modeling of Nonconstant Variance CHAPTER 7 ST 762, M. DAVIDIAN 7 Detection and modeling of nonconstant variance 7.1 Introduction So far, we have focused on approaches to inference in mean-variance models of the form 2 2 E(Yj|xj)= f(xj, β), var(Yj|xj)= σ g (β, θ, xj) (7.1) under the assumption that we have already specified such a model. • Often, a model for the mean may be suggested by the nature of the response (e.g., binary or count), by subject-matter theoretical considerations (e.g., pharmacokinetics), or by the empirical evidence (e.g., models for assay response). • A model for variance may or may not be suggested by these features. When the response is binary, the form of the variance is indeed dictated by the Bernoulli distribution, while for data in the form of counts or proportions, for which the Poisson or binomial distributions may be appropriate, the form of the variance is again suggested. One may wish to consider the possibility of over- or underdispersion in these situations; this may reasonably be carried out by fitting a model that accommodates these features and determining if an improvement in fit is apparent using methods for inference on variance parameters we will discuss in Chapter 12. Alternatively, when the response is continuous (or approximately continuous), it is often the situation that there is not necessarily an obvious relevant distributional model. As we have discussed in some of the examples we have considered, several sources of variation may combine to produce patterns that are not well described by the kinds of variance models dictated by popular distributional assumptions such as the gamma or lognormal distributions. In fact, it may be unclear whether heterogeneity of variance is even an issue at all. In some applications, it is expected, and popular models may be available; in others, whether or not variance changes with the mean or covariate values may need to be deduced from the data. • In these situations, methods are required for detecting nonconstant variance, determining whether or not it changes smoothly across the range of the response or covariates, and identifying an appropriate model to characterize the change. To address these issues, both formal and informal approaches have been proposed: PAGE 155 CHAPTER 7 ST 762, M. DAVIDIAN • Graphical techniques. Both for detection and modeling, these often have a subjective flavor. In this chapter, we will focus on these procedures. • Formal hypothesis testing. Formal procedures are mainly used for detection. We will defer discus- sion of these until after we have covered the large-sample theoretical developments on which they are based. Because of the complexity of (7.1), no finite-sample, “exact” methods are available in general. COMMON THEME: Most graphical approaches are based on the OLS residuals ˆ rj = Yj − f(xj, βOLS) and functions thereof, or on related constructs. Our main focus will be on detection and modeling in situations where the response is continuous (or nearly continuous, such as in the case of moderate-to- large counts). A complementary treatment of some of the approaches we will discuss may be found in Carroll and Rup- pert (1988, Sections 2.7 and 2.8). Note that, in what follows, distributional statements are conditional on the xj. 7.2 Plots based on residuals We begin by first reviewing the basic rationale for the use of residuals as a tool for detecting nonconstant variance in regression. The “usual” residual plots described in a first course in linear regression analysis apply equally well in the nonlinear model situation. Specifically, one usually plots the rj or the “standardized” residuals rj/σˆOLS, where n 2 −1 2 σˆOLS = (n − p) rj , jX=1 versus one or more of the following: ˆ ˆ • Predicted values Yj = f(xj, βOLS) • Covariates (elements of xj) • log Yˆj in cases where many responses tend to be clustered in a very narrow range in order to “stretch things out” so that any patterns might be more readily discernible. We will see the value of this for some nonlinear models and designs later. PAGE 156 CHAPTER 7 ST 762, M. DAVIDIAN If the plot(s) exhibit an apparent pattern, with the magnitude of residuals changing with level of predicted value or covariate, this is taken as evidence of potential nonconstant variance. In particular, for the plot of residuals vs. predicted values or their logarithms, a “fan-shape” is accepted as evidence that variance increases smoothly with the level of the response (mean). More generally, any “nonhaphazard,” “systematic” pattern may well be evidence that variance does not remain constant over the range of the response. One must be careful, however. • A systematic pattern may also be the result of an ill-fitting mean model. The nature of the pattern must be critically assessed by the data analyst to determine a reasonable explanation for it given the particular mean model and circumstances. For example, for the indomethacin pharmacokinetic data in Examples 1.1 and 1.2, the model was the sum of two exponential terms. If a simple model containing only a single exponential term were fitted to these data, one would expect to see a systematic pattern in the residuals reflecting the lack of fit of this model. There is certainly subjectivity involved in this endeavor. • When responses are collected in time order, e.g., repeated measurements on the same individuals, one often plots the residuals against time to look for temporal patterns that may suggest possible serial correlation. Alternatively, more sophisticated plots for investigating this are available. We defer discussion of serial correlation until later chapters, as our current focus is on detecting and modeling nonconstant variance when the assumption of independence is reasonable. It is important to recognize, however, that this is an assumption that should be considered carefully in practice. MOTIVATION: The obvious motivation for the usual plots is that rj is a “proxy” for the true deviation Yj − f(xj, β). • If the data are normally (or at least symmetrically) distributed with constant variance, we would expect the rj to be roughly symmetrically distributed about 0 and to have approximate constant variance. • We would thus expect a “haphazard” pattern, with approximately equal numbers of positive and negative residuals with approximately the same magnitude across their entire range. • Even if the variance were nonconstant, if the data were at least normally or symmetrically dis- tributed, we would still expect approximately equal numbers of positive and negative residuals. PAGE 157 CHAPTER 7 ST 762, M. DAVIDIAN However, we would expect changing magnitude across the range. PROBLEMS WITH THE USUAL PLOTS: The OLS residuals rj may not have exactly the same ˆ properties as the true deviations because β is replaced by the OLS estimator βOLS. We will tackle this issue shortly. Some more immediate problems that may make the usual plots difficult to interpret are as follows: • The data may not be normally or even symmetrically distributed but may instead arise from a skewed (asymmetric) distribution. • The design (the settings of the xj) may be such that an unusual pattern of residuals may be due to something other than nonconstant variance. • Furthermore, although the usual plots may be sufficient for detection, they may not be very helpful for modeling of nonconstant variance. We thus consider refinements of the usual plots. REFINEMENT 1. A common idea is to base plots on transformations of (absolute) residuals or other residuals in order to account for sample size or asymmetry. A seminal reference for some of these ideas in the context of linear regression is Cook and Weisberg (1983). We have already discussed estimation of variance parameters based on transformations of absolute residuals, so it should come as no surprise that diagnostic plots would also be based on them. IDEA 1: “Visually double the sample size.” The usual plots may be difficult to interpret because the sample size is small. Under such conditions, a change in the placement of just a single residual in the plot can change the apparent pattern substantially. Thus, each observation may be very influential to the eye in gauging the pattern. 2 2 2 A simple remedy is to plot rj or rj /σˆOLS instead. In this plot, the magnitude but not the sign of the residuals is emphasized. Because the contribution of all residuals is positive, this has the effect of creating a “larger” sample size for the purpose of spotting changes in magnitude. Moreover, the visual influence of any single observation in dictating the pattern is reduced. Recall the data on the pharmacokinetics of indomethacin in Examples 1.1 and 1.2. Here, n = 11 concentration responses were collected over time on a single subject. PAGE 158 CHAPTER 7 ST 762, M. DAVIDIAN The data are plotted again in Figure 7.3 in Section 7.5; a usual residual plot was given in Figure 1.3 and exhibits a “fan-shaped” pattern that appears roughly symmetric about zero. Note that the residuals have been plotted against the logarithm of predicted values; because the response “tails off” rather quickly, there are many residuals at very small values of the response, so that residuals plotted against predicted values themselves are “bunched up” near zero, making the pattern difficult to assess. Figure 7.4 in Section 7.5 shows a plot of squared, standardized residuals against log predicted values and shows a “wedge shape” indicating the increase in magnitude across the range. One could substitute absolute residuals |rj| for squared ones and make similar plots. A purported advantage of squared over absolute residuals themselves is that squaring tends to highlight residuals that are “large” in magnitude and downplay those that are “small,” thus drawing attention to changes in magnitude over the range.
Recommended publications
  • Residuals, Part II
    Biostatistics 650 Mon, November 5 2001 Residuals, Part II Key terms External Studentization Outliers Added Variable Plot — Partial Regression Plot Partial Residual Plot — Component Plus Residual Plot Key ideas/results ¢ 1. An external estimate of ¡ comes from refitting the model without observation . Amazingly, it has an easy formula: £ ¦!¦ ¦ £¥¤§¦©¨ "# 2. Externally Studentized Residuals $ ¦ ¦©¨ ¦!¦)( £¥¤§¦&% ' Ordinary residuals standardized with £*¤§¦ . Also known as R-Student. 3. Residual Taxonomy Names Definition Distribution ¦!¦ ¦+¨-,¥¦ ,0¦ 1 Ordinary ¡ 5 /. &243 ¦768¤§¦©¨ ¦ ¦!¦ 1 ¦!¦ PRESS ¡ &243 9' 5 $ Studentized ¦©¨ ¦ £ ¦!¦ ¤0> % = Internally Studentized : ;<; $ $ Externally Studentized ¦+¨ ¦ £?¤§¦ ¦!¦ ¤0>¤A@ % R-Student = 4. Outliers are unusually large observations, due to an unmodeled shift or an (unmodeled) increase in variance. 5. Outliers are not necessarily bad points; they simply are not consistent with your model. They may posses valuable information about the inadequacies of your model. 1 PRESS Residuals & Studentized Residuals Recall that the PRESS residual has a easy computation form ¦ ¦©¨ PRESS ¦!¦ ( ¦!¦ It’s easy to show that this has variance ¡ , and hence a standardized PRESS residual is 9 ¦ ¦ ¦!¦ ¦ PRESS ¨ ¨ ¨ ¦ : £ ¦!¦ £ ¦!¦ £ ¦ ¦ % % % ' When we standardize a PRESS residual we get the studentized residual! This is very informa- ,¥¦ ¦ tive. We understand the PRESS residual to be the residual at ¦ if we had omitted from the 3 model. However, after adjusting for it’s variance, we get the same thing as a studentized residual. Hence the standardized residual can be interpreted as a standardized PRESS residual. Internal vs External Studentization ,*¦ ¦ The PRESS residuals remove the impact of point ¦ on the fit at . But the studentized 3 ¢ ¦ ¨ ¦ £?% ¦!¦ £ residual : can be corrupted by point by way of ; a large outlier will inflate the residual mean square, and hence £ .
    [Show full text]
  • Testing for Uncorrelated Residuals in Dynamic Count Models with an Application to Corporate Bankruptcy
    Munich Personal RePEc Archive Testing for Uncorrelated Residuals in Dynamic Count Models with an Application to Corporate Bankruptcy Sant’Anna, Pedro H. C. Universidad Carlos III de Madrid May 2013 Online at https://mpra.ub.uni-muenchen.de/48429/ MPRA Paper No. 48429, posted 19 Jul 2013 13:15 UTC Testing for Uncorrelated Residuals in Dynamic Count Models with an Application to Corporate Bankruptcy Pedro H. C. Sant’Anna Universidad Carlos III de Madrid, Getafe, 28903, Spain [email protected] Abstract This article proposes a new diagnostic test for dynamic count models, which is well suited for risk management. Our test proposal is of the Portmanteau-type test for lack of residual autocorrelation. Unlike previous proposals, the resulting test statistic is asymptotically pivotal when innovations are uncorrelated, but not necessarily iid nor a martingale difference. Moreover, the proposed test is able to detect local alter- natives converging to the null at the parametric rate T −1/2, with T the sample size. The finite sample performance of the test statistic is examined by means of a Monte Carlo experiment. Finally, using a dataset on U.S. corporate bankruptcies, we apply our test proposal to check if common risk models are correctly specified. Keywords: Time Series of counts; Residual autocorrelation function; Model checking; Credit risk management. 1 1. Introduction Credit risk affects virtually every financial contract. Therefore the measurement, pricing and management of credit risk have received much attention from economists, bank supervisors and regulators, and financial market practitioners. A widely used measure of credit risk is the probability of corporate default (PD).
    [Show full text]
  • Research Paper Commerce Statistics Inference in Nonlinear Regression Model with Heteroscedastic Errors B. Mahaboob Research Scho
    Volume-5, Issue-5, May - 2016 • ISSN No 2277 - 8160 IF : 3.62 | IC Value 70.36 Commerce Research Paper Statistics Inference in Nonlinear Regression Model With Heteroscedastic Errors B. Mahaboob Research Scholar, Department of Mathematics Dr. M. Ramesh Senior Data Scientist, Tech Mahindra, Hyderabad. Dr. Sk. Khadar Assistant Professor (Senior), Department of Mathematics, VIT, Vellore, Babu Tamilnadu. Academic Consultant, Department of CSE, SPMVV Engineering P. Srivyshnavi College, Tirupati. Prof. P. B.Sireesha Research Scholar, Department of Statistics, S.V.University, Tirupati. Balasiddamuni Rtd. Professor ABSTRACT An advent of modern computer science has made it possible for the applied mathematician to study the inferential aspects of an increasing number of nonlinear regression models in recent years. The inferential aspects for nonlinear regression models and the error assumptions are usually analogous to those made for linear regression models. The tests for the hypotheses on parameters of nonlinear regression models are usually based on nonlinear least squares estimates and the normal assumptions of error variables. A more common problem with data that are best fit by a nonlinear regression model than with data that can be adequately fit by the linear regression model is the problem of heteroscedasticity of errors. The problem of heteroscedasticity can be studied with reference to nonlinear regression models in a variety of ways. In the present study, an attempt has been made by developing inferential method for heteroscedastic nonlinear regression model. KEYWORDS : Intrinsically Nonlinear model; Violation of assumptions on Errors; Problem of Heteroscedasticity. I. INTRODUCTION In estimating the linear regression models, the least squares method of estimation is often applied to estimate the unknown parameters of the linear regression model and errors are usually assumed to be independent and identically distributed normal random variables with zero mean and unknown constant variance.
    [Show full text]
  • A Review on Empirical Likelihood Methods for Regression
    Noname manuscript No. (will be inserted by the editor) A Review on Empirical Likelihood Methods for Regression Song Xi Chen · Ingrid Van Keilegom Received: date / Accepted: date Abstract We provide a review on the empirical likelihood method for regression type inference problems. The regression models considered in this review include parametric, semiparametric and nonparametric models. Both missing data and censored data are accommodated. Keywords Censored data; empirical likelihood; missing data; nonparametric regression; parametric regression; semiparametric regression; Wilks’ theorem. 1 Introduction It has been twenty years since Art Owen published his seminal paper (Owen, 1988) that introduces the notion of empirical likelihood (EL). Since then, there has been a rich body of literature on the novel idea of formulating versions of nonparametric likelihood in various settings of statistical inference. There have been two major reviews on the empirical likelihood. The first review was given by Hall and La Scala (1990) in the early years of the EL method, which summarized some key properties of the method. The second one was the book by the inventor of the methodology (Owen, 2001), which provided a comprehensive overview up to that time. The body of empirical likelihood literature is increasing rapidly, and it would be a daunting task to review the entire field in one review paper like this one. We therefore decided to concentrate our review on regression due to its prominence in statistical S.X. Chen Department of Statistics, Iowa State University, Ames, Iowa 50011-1210, USA and Guanghua School of Management, Peking University, China Tel.: 1-515-2942729 Fax: 1-515-2944040 E-mail: [email protected]; [email protected] I.
    [Show full text]
  • Linear Models with R Julian J.Faraway Downloaded by [University of Toronto] at 16:20 23 May 2014 Texts in Statistical Science
    Downloaded by [University of Toronto] at 16:20 23 May 2014 CHAPMAN & HALL/CRC Texts in Statistical Science Series Series Editors Chris Chatfield, University of Bath, UK Martin Tanner, Northwestern University, USA Jim Zidek, University of British Columbia, Canada Analysis of Failure and Survival Data Peter J.Smith The Analysis and Interpretation of Multivariate Data for Social Scientists David J.Bartholomew, Fiona Steele, Irini Moustaki, and Jane Galbraith The Analysis of Time Series—An Introduction, Sixth Edition Chris Chatfield Applied Bayesian Forecasting and Time Series Analysis A.Pole, M.West and J.Harrison Applied Nonparametric Statistical Methods, Third Edition P.Sprent and N.C.Smeeton Applied Statistics—Handbook of GENSTAT Analysis E.J.Snell and H.Simpson Applied Statistics—Principles and Examples D.R.Cox and E.J.Snell Bayes and Empirical Bayes Methods for Data Analysis, Second Edition Bradley P.Carlin and Thomas A.Louis Downloaded by [University of Toronto] at 16:20 23 May 2014 Bayesian Data Analysis, Second Edition Andrew Gelman, John B.Carlin, Hal S.Stern, and Donald B.Rubin Beyond ANOVA—Basics of Applied Statistics R.G.Miller, Jr. Computer-Aided Multivariate Analysis, Third Edition A.A.Afifi and V.A.Clark A Course in Categorical Data Analysis T.Leonard A Course in Large Sample Theory T.S.Ferguson Data Driven Statistical Methods P.Sprent Decision Analysis—A Bayesian Approach J.Q.Smith Elementary Applications of Probability Theory, Second Edition H.C.Tuckwell Elements of Simulation B.J.T.Morgan Epidemiology—Study Design and Data
    [Show full text]
  • Analysis of the Spread of COVID-19 in the USA with a Spatio-Temporal Multivariate Time Series Model
    International Journal of Environmental Research and Public Health Article Analysis of the Spread of COVID-19 in the USA with a Spatio-Temporal Multivariate Time Series Model Rongxiang Rui 1 , Maozai Tian 2 , Man-Lai Tang 3,* , George To-Sum Ho 4 and Chun-Ho Wu 4 1 School of Statistics, Renmin University of China, Beijing 100872, China; [email protected] 2 College of Medical Engineering and Technology, Xinjiang Medical University, Ürümqi 830011, China; [email protected] 3 Department of Mathematics, Statistics and Insurance, Hang Seng University of Hong Kong, Hong Kong, China 4 Department of Supply Chain and Information Management, Hang Seng University of Hong Kong, Hong Kong, China; [email protected] (G.T.-S.H.); [email protected] (C.-H.W.) * Correspondence: [email protected] Abstract: With the rapid spread of the pandemic due to the coronavirus disease 2019 (COVID-19), the virus has already led to considerable mortality and morbidity worldwide, as well as having a severe impact on economic development. In this article, we analyze the state-level correlation between COVID-19 risk and weather/climate factors in the USA. For this purpose, we consider a spatio-temporal multivariate time series model under a hierarchical framework, which is especially suitable for envisioning the virus transmission tendency across a geographic area over time. Briefly, our model decomposes the COVID-19 risk into: (i) an autoregressive component that describes the within-state COVID-19 risk effect; (ii) a spatiotemporal component that describes the across-state COVID-19 risk effect; (iii) an exogenous component that includes other factors (e.g., weather/climate) that could envision future epidemic development risk; and (iv) an endemic component that captures the function of time and other predictors mainly for individual states.
    [Show full text]
  • Power Maximization and Size Control in Heteroskedasticity and Autocorrelation Robust Tests with Exponentiated Kernels
    POWER MAXIMIZATION AND SIZE CONTROL IN HETEROSKEDASTICITY AND AUTOCORRELATION ROBUST TESTS WITH EXPONENTIATED KERNELS By Yixiao Sun, Peter C. B. Phillips and Sainan Jin January 2010 COWLES FOUNDATION DISCUSSION PAPER NO. 1749 COWLES FOUNDATION FOR RESEARCH IN ECONOMICS YALE UNIVERSITY Box 208281 New Haven, Connecticut 06520-8281 Power Maximization and Size Control in Heteroskedasticity and Autocorrelation Robust Tests with Exponentiated Kernels Yixiao Sun Department of Economics University of California, San Diego Peter C. B. Phillips Yale University, University of Auckland, Singapore Management University & University of Southampton Sainan Jin School of Economics Singapore Management University, August 14, 2009 The authors gratefully acknowledge partial research support from NSF under Grant No. SES- 0752443 (Sun), SES 06-47086 (Phillips), and NSFC under grant No. 70501001 and 70601001 (Jin). ABSTRACT Using the power kernels of Phillips, Sun and Jin (2006, 2007), we examine the large sample asymptotic properties of the t-test for di¤erent choices of power parameter (). We show that the nonstandard …xed- limit distributions of the t-statistic provide more accurate approximations to the …nite sample distributions than the conventional large- limit distribution. We prove that the second-order corrected critical value based on an asymptotic expansion of the nonstandard limit distribution is also second- order correct under the large- asymptotics. As a further contribution, we propose a new practical procedure for selecting the test-optimal power parameter that addresses the central concern of hypothesis testing: the selected power parameter is test-optimal in the sense that it minimizes the type II error while controlling for the type I error. A plug-in procedure for implementing the test-optimal power parameter is suggested.
    [Show full text]
  • Optimal Bandwidth Selection in Heteroskedasticity#Autocorrelation
    Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing Yixiao Sun Department of Economics University of California, San Diego Peter C. B. Phillips Cowles Foundation, Yale University, University of Auckland & University of York Sainan Jin Guanghua School of Management Peking University January 5, 2011 Abstract The paper considers studentized tests in time series regressions with nonparamet- rically autocorrelated errors. The studentization is based on robust standard errors with truncation lag M = bT for some constant b (0; 1] and sample size T: It is shown that the nonstandard …xed-b limit distributions of2 such nonparametrically studentized tests provide more accurate approximations to the …nite sample distributions than the standard small-b limit distribution. We further show that, for typical economic time series, the optimal bandwidth that minimizes a weighted average of type I and type II errors is larger by an order of magnitude than the bandwidth that minimizes the asymptotic mean squared error of the corresponding long-run variance estimator. A plug-in procedure for implementing this optimal bandwidth is suggested and simula- tions (not reported here) con…rm that the new plug-in procedure works well in …nite samples. Keywords: Asymptotic expansion, bandwidth choice, kernel method, long-run variance, loss function, nonstandard asymptotics, robust standard error, Type I and Type II errors An earlier version of the paper is available as Sun, Phillips and Jin (2006). The present version is a substantially shorter paper and readers are referred to the earlier version for further details, discus- sion, proofs, and some asymptotic expansion results of independent interest. The authors thank Whitney Newey, two anonymous referees, and many seminar participants for comments on earlier versions.
    [Show full text]
  • Risk Segmentation Using Bayesian Quantile Regression with Natural Cubic Splines
    Open Access Austin Statistics A Austin Full Text Article Publishing Group Review Article Risk Segmentation Using Bayesian Quantile Regression with Natural Cubic Splines Xia M* Abstract Department of Statistics, Northern Illinois University, USA An insurance claims department is often interested in obtaining the possible *Corresponding author: Xia M, Division of Statistics, distribution of a claim for the purpose of risk management. Information such as Northern Illinois University, 300 Normal Road, DeKalb, the probability that a claim will exceed a certain amount is helpful for matching IL 60115, USA, Tel: 815-753-6795; Fax: 815-753-6776; the claim complexity with the specialty of claim adjusters. Using information Email: [email protected] available on the claim and the parties involved, we propose a Bayesian quantile regression model for the purpose of risk identification and segmentation in the Received: May 29, 2014; Accepted: July 11, 2014; claims department. Natural cubic splines are used in order to estimate a smooth Published: July 16, 2014 relationship between the expected quantiles and continuous explanatory variables such as the age of the claimant. A case study is conducted using the Medical Large Claims Experience Study data from the Society of Actuaries. For the claimant age factor that we study, we observe that the high-risk groups, such as the infants and elderly, exhibit a much higher risk in terms of high quantiles (such as the 99% and 99.5% percentiles) than that is revealed by the mean or the median. Particularly for the claims data where there are various characteristics available on the claimant and other parties involved, our model may reveal helpful information on the possible extremal risk that may be under looked in traditional claims modeling.
    [Show full text]
  • Gelman and Shalizi, 2013
    8 British Journal of Mathematical and Statistical Psychology (2013), 66, 8–38 © 2012 The British Psychological Society www.wileyonlinelibrary.com Philosophy and the practice of Bayesian statistics ∗ Andrew Gelman1 and Cosma Rohilla Shalizi2 1Department of Statistics and Department of Political Science, Columbia University, New York, USA 2Statistics Department, Carnegie Mellon University, Santa Fe Institute, Pittsburgh, USA A substantial school in the philosophy of science identifies Bayesian inference with inductive inference and even rationality as such, and seems to be strengthened by the rise and practical success of Bayesian statistics. We argue that the most successful forms of Bayesian statistics do not actually support that particular philosophy but rather accord much better with sophisticated forms of hypothetico-deductivism. We examine the actual role played by prior distributions in Bayesian models, and the crucial aspects of model checking and model revision, which fall outside the scope of Bayesian confirmation theory. We draw on the literature on the consistency of Bayesian updating and also on our experience of applied work in social science. Clarity about these matters should benefit not just philosophy of science, but also statistical practice. At best, the inductivist view has encouraged researchers to fit and compare models without checking them; at worst, theorists have actively discouraged practitioners from performing model checking because it does not fit into their framework. 1. The usual story – which we don’t like In so far as I have a coherent philosophy of statistics, I hope it is ‘robust’ enough to cope in principle with the whole of statistics, and sufficiently undogmatic not to imply that all those who may think rather differently from me are necessarily stupid.
    [Show full text]
  • Autocorrelation-Robust Inference*
    G. S. Maddala and C. R. Rao, eds., ok of Statistics, Vol. 15 © 1997 EIsevier Science B.V. AH rights reserved 11 Autocorrelation-Robust Inference* P. M. Robinson and C. Ve/asco 1. Introduction Time series data occur commonly in the natural and engineering sciences, eco­ nomics and many other fields of enquiry. A typical feature of such data is their apparent dependence across time, for example sometimes records close together in time are strongly correlated. Accounting for serial dependence can con­ siderably complicate statistical inference. The attempt to model the dependence parametrically, or even nonparametrically, can be difficult and computationally expensive. In sorne circumstances, the serial dependence is merely a nuisance feature, interest focusing on "static" aspects such as a location parameter, or a probability density function. Here, we can frequently base inference on point or function estimates that are natural ones to use in case of independence, and may well be optimal in that case. Such estimates will often be less efficient, at least asymp­ totically, than ones based on a comprehensive model that incorporates the serial dependence. But apart from their relative computational simplicity, they often remain consistent even in the presence of certain forms of dependence, and can be more reliable than the "efficient" ones, which can sometimes become inconsistent when the dependence is inappropriately dealt with, leading to statistical inferences that are invalid, even asymptotically. In this sense the estimates can be more "robust" than "efficient" ones, and the main focus in this paper is their use in statistical inference, though we also discuss attempts at more efficient inference.
    [Show full text]
  • Studentized Residuals Jacknife Residuals 3 Zambia 2 2 1 1 0 0 −1 −1 Jacknife Residuals −2 Studentized Residuals −2
    Chapter 7 Diagnostics Regression model building is often an iterative and interactive process. The first model we try may prove to be inadequate. Regression diagnostics are used to detect problems with the model and suggest improve- ments. This is a hands-on process. 7.1 Residuals and Leverage ¡ T 1 T We start with some basic diagnostic quantities - the residuals and the leverages. Recall that yˆ X X X ¢¤£ X y Hy where H is the hat-matrix. Now ¡ ¥ ¢ εˆ y ¥ yˆ I H y ¡ ¡ ¢ ¦ ¥ ¢ I ¥ H Xβ I H ε ¡ ¢ I ¥ H ε ¡ ¡ 2 2 ¢ ¥ ¢ So var εˆ var I ¥ H ε I H σ assuming that var ε σ I. We see that although the errors may have equal variance and be uncorrelated the residuals do not. ¡ 2 ¢ hi Hii are called leverages and are useful diagnostics. We see that var εˆi σ 1 ¥ hi so that a large leverage for hi will make var εˆi small — in other words the fit will be “forced” to be close to yi. The hi depends only on X — knowledge of y is required for a full interpretation. Some facts: ¨ © ∑hi p hi § 1 n i i ¨ An average value for hi is p ¨ n and a “rule of thumb” is that leverages of more than 2p n should be looked at more closely. Large values of hi are due to extreme values in X. hi corresponds to a Mahalanobis distance ¡ ¡ T 1 ¥ ¢ ¥ ¢ defined by X which is x x¯ Σˆ £ x x¯ where Σˆ is the estimated covariance of X.
    [Show full text]