Robust Bayesian General Linear Models ⁎ W.D
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Robust Statistics Part 3: Regression Analysis
Robust Statistics Part 3: Regression analysis Peter Rousseeuw LARS-IASC School, May 2019 Peter Rousseeuw Robust Statistics, Part 3: Regression LARS-IASC School, May 2019 p. 1 Linear regression Linear regression: Outline 1 Classical regression estimators 2 Classical outlier diagnostics 3 Regression M-estimators 4 The LTS estimator 5 Outlier detection 6 Regression S-estimators and MM-estimators 7 Regression with categorical predictors 8 Software Peter Rousseeuw Robust Statistics, Part 3: Regression LARS-IASC School, May 2019 p. 2 Linear regression Classical estimators The linear regression model The linear regression model says: yi = β0 + β1xi1 + ... + βpxip + εi ′ = xiβ + εi 2 ′ ′ with i.i.d. errors εi ∼ N(0,σ ), xi = (1,xi1,...,xip) and β =(β0,β1,...,βp) . ′ Denote the n × (p + 1) matrix containing the predictors xi as X =(x1,..., xn) , ′ ′ the vector of responses y =(y1,...,yn) and the error vector ε =(ε1,...,εn) . Then: y = Xβ + ε Any regression estimate βˆ yields fitted values yˆ = Xβˆ and residuals ri = ri(βˆ)= yi − yˆi . Peter Rousseeuw Robust Statistics, Part 3: Regression LARS-IASC School, May 2019 p. 3 Linear regression Classical estimators The least squares estimator Least squares estimator n ˆ 2 βLS = argmin ri (β) β i=1 X If X has full rank, then the solution is unique and given by ˆ ′ −1 ′ βLS =(X X) X y The usual unbiased estimator of the error variance is n 1 σˆ2 = r2(βˆ ) LS n − p − 1 i LS i=1 X Peter Rousseeuw Robust Statistics, Part 3: Regression LARS-IASC School, May 2019 p. 4 Linear regression Classical estimators Outliers in regression Different types of outliers: vertical outlier good leverage point • • y • • • regular data • ••• • •• ••• • • • • • • • • • bad leverage point • • •• • x Peter Rousseeuw Robust Statistics, Part 3: Regression LARS-IASC School, May 2019 p. -
Quantile Regression for Overdispersed Count Data: a Hierarchical Method Peter Congdon
Congdon Journal of Statistical Distributions and Applications (2017) 4:18 DOI 10.1186/s40488-017-0073-4 RESEARCH Open Access Quantile regression for overdispersed count data: a hierarchical method Peter Congdon Correspondence: [email protected] Abstract Queen Mary University of London, London, UK Generalized Poisson regression is commonly applied to overdispersed count data, and focused on modelling the conditional mean of the response. However, conditional mean regression models may be sensitive to response outliers and provide no information on other conditional distribution features of the response. We consider instead a hierarchical approach to quantile regression of overdispersed count data. This approach has the benefits of effective outlier detection and robust estimation in the presence of outliers, and in health applications, that quantile estimates can reflect risk factors. The technique is first illustrated with simulated overdispersed counts subject to contamination, such that estimates from conditional mean regression are adversely affected. A real application involves ambulatory care sensitive emergency admissions across 7518 English patient general practitioner (GP) practices. Predictors are GP practice deprivation, patient satisfaction with care and opening hours, and region. Impacts of deprivation are particularly important in policy terms as indicating effectiveness of efforts to reduce inequalities in care sensitive admissions. Hierarchical quantile count regression is used to develop profiles of central and extreme quantiles according to specified predictor combinations. Keywords: Quantile regression, Overdispersion, Poisson, Count data, Ambulatory sensitive, Median regression, Deprivation, Outliers 1. Background Extensions of Poisson regression are commonly applied to overdispersed count data, focused on modelling the conditional mean of the response. However, conditional mean regression models may be sensitive to response outliers. -
Robust Linear Regression: a Review and Comparison Arxiv:1404.6274
Robust Linear Regression: A Review and Comparison Chun Yu1, Weixin Yao1, and Xue Bai1 1Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802. Abstract Ordinary least-squares (OLS) estimators for a linear model are very sensitive to unusual values in the design space or outliers among y values. Even one single atypical value may have a large effect on the parameter estimates. This article aims to review and describe some available and popular robust techniques, including some recent developed ones, and compare them in terms of breakdown point and efficiency. In addition, we also use a simulation study and a real data application to compare the performance of existing robust methods under different scenarios. arXiv:1404.6274v1 [stat.ME] 24 Apr 2014 Key words: Breakdown point; Robust estimate; Linear Regression. 1 Introduction Linear regression has been one of the most important statistical data analysis tools. Given the independent and identically distributed (iid) observations (xi; yi), i = 1; : : : ; n, 1 in order to understand how the response yis are related to the covariates xis, we tradi- tionally assume the following linear regression model T yi = xi β + "i; (1.1) where β is an unknown p × 1 vector, and the "is are i.i.d. and independent of xi with E("i j xi) = 0. The most commonly used estimate for β is the ordinary least square (OLS) estimate which minimizes the sum of squared residuals n X T 2 (yi − xi β) : (1.2) i=1 However, it is well known that the OLS estimate is extremely sensitive to the outliers. -
Generalized Linear Models
Generalized Linear Models A generalized linear model (GLM) consists of three parts. i) The first part is a random variable giving the conditional distribution of a response Yi given the values of a set of covariates Xij. In the original work on GLM’sby Nelder and Wedderburn (1972) this random variable was a member of an exponential family, but later work has extended beyond this class of random variables. ii) The second part is a linear predictor, i = + 1Xi1 + 2Xi2 + + ··· kXik . iii) The third part is a smooth and invertible link function g(.) which transforms the expected value of the response variable, i = E(Yi) , and is equal to the linear predictor: g(i) = i = + 1Xi1 + 2Xi2 + + kXik. ··· As shown in Tables 15.1 and 15.2, both the general linear model that we have studied extensively and the logistic regression model from Chapter 14 are special cases of this model. One property of members of the exponential family of distributions is that the conditional variance of the response is a function of its mean, (), and possibly a dispersion parameter . The expressions for the variance functions for common members of the exponential family are shown in Table 15.2. Also, for each distribution there is a so-called canonical link function, which simplifies some of the GLM calculations, which is also shown in Table 15.2. Estimation and Testing for GLMs Parameter estimation in GLMs is conducted by the method of maximum likelihood. As with logistic regression models from the last chapter, the generalization of the residual sums of squares from the general linear model is the residual deviance, Dm 2(log Ls log Lm), where Lm is the maximized likelihood for the model of interest, and Ls is the maximized likelihood for a saturated model, which has one parameter per observation and fits the data as well as possible. -
A Guide to Robust Statistical Methods in Neuroscience
A GUIDE TO ROBUST STATISTICAL METHODS IN NEUROSCIENCE Authors: Rand R. Wilcox1∗, Guillaume A. Rousselet2 1. Dept. of Psychology, University of Southern California, Los Angeles, CA 90089-1061, USA 2. Institute of Neuroscience and Psychology, College of Medical, Veterinary and Life Sciences, University of Glasgow, 58 Hillhead Street, G12 8QB, Glasgow, UK ∗ Corresponding author: [email protected] ABSTRACT There is a vast array of new and improved methods for comparing groups and studying associations that offer the potential for substantially increasing power, providing improved control over the probability of a Type I error, and yielding a deeper and more nuanced understanding of data. These new techniques effectively deal with four insights into when and why conventional methods can be unsatisfactory. But for the non-statistician, the vast array of new and improved techniques for comparing groups and studying associations can seem daunting, simply because there are so many new methods that are now available. The paper briefly reviews when and why conventional methods can have relatively low power and yield misleading results. The main goal is to suggest some general guidelines regarding when, how and why certain modern techniques might be used. Keywords: Non-normality, heteroscedasticity, skewed distributions, outliers, curvature. 1 1 Introduction The typical introductory statistics course covers classic methods for comparing groups (e.g., Student's t-test, the ANOVA F test and the Wilcoxon{Mann{Whitney test) and studying associations (e.g., Pearson's correlation and least squares regression). The two-sample Stu- dent's t-test and the ANOVA F test assume that sampling is from normal distributions and that the population variances are identical, which is generally known as the homoscedastic- ity assumption. -
Sketching for M-Estimators: a Unified Approach to Robust Regression
Sketching for M-Estimators: A Unified Approach to Robust Regression Kenneth L. Clarkson∗ David P. Woodruffy Abstract ing algorithm often results, since a single pass over A We give algorithms for the M-estimators minx kAx − bkG, suffices to compute SA. n×d n n where A 2 R and b 2 R , and kykG for y 2 R is specified An important property of many of these sketching ≥0 P by a cost function G : R 7! R , with kykG ≡ i G(yi). constructions is that S is a subspace embedding, meaning d The M-estimators generalize `p regression, for which G(x) = that for all x 2 R , kSAxk ≈ kAxk. (Here the vector jxjp. We first show that the Huber measure can be computed norm is generally `p for some p.) For the regression up to relative error in O(nnz(A) log n + poly(d(log n)=")) d time, where nnz(A) denotes the number of non-zero entries problem of minimizing kAx − bk with respect to x 2 R , of the matrix A. Huber is arguably the most widely used for inputs A 2 Rn×d and b 2 Rn, a minor extension of M-estimator, enjoying the robustness properties of `1 as well the embedding condition implies S preserves the norm as the smoothness properties of ` . 2 of the residual vector Ax − b, that is kS(Ax − b)k ≈ We next develop algorithms for general M-estimators. kAx − bk, so that a vector x that makes kS(Ax − b)k We analyze the M-sketch, which is a variation of a sketch small will also make kAx − bk small. -
Generalized Linear Models Outline for Today
Generalized linear models Outline for today • What is a generalized linear model • Linear predictors and link functions • Example: estimate a proportion • Analysis of deviance • Example: fit dose-response data using logistic regression • Example: fit count data using a log-linear model • Advantages and assumptions of glm • Quasi-likelihood models when there is excessive variance Review: what is a (general) linear model A model of the following form: Y = β0 + β1X1 + β2 X 2 + ...+ error • Y is the response variable • The X ‘s are the explanatory variables • The β ‘s are the parameters of the linear equation • The errors are normally distributed with equal variance at all values of the X variables. • Uses least squares to fit model to data, estimate parameters • lm in R The predicted Y, symbolized here by µ, is modeled as µ = β0 + β1X1 + β2 X 2 + ... What is a generalized linear model A model whose predicted values are of the form g(µ) = β0 + β1X1 + β2 X 2 + ... • The model still include a linear predictor (to right of “=”) • g(µ) is the “link function” • Wide diversity of link functions accommodated • Non-normal distributions of errors OK (specified by link function) • Unequal error variances OK (specified by link function) • Uses maximum likelihood to estimate parameters • Uses log-likelihood ratio tests to test parameters • glm in R The two most common link functions Log • used for count data η η = logµ The inverse function is µ = e Logistic or logit • used for binary data µ eη η = log µ = 1− µ The inverse function is 1+ eη In all cases log refers to natural logarithm (base e). -
On Robust Regression with High-Dimensional Predictors
On robust regression with high-dimensional predictors Noureddine El Karoui ∗, Derek Bean ∗ , Peter Bickel ∗ , Chinghway Lim y, and Bin Yu ∗ ∗University of California, Berkeley, and yNational University of Singapore Submitted to Proceedings of the National Academy of Sciences of the United States of America We study regression M-estimates in the setting where p, the num- pointed out a surprising feature of the regime, p=n ! κ > 0 ber of covariates, and n, the number of observations, are both large for LSE; fitted values were not asymptotically Gaussian. He but p ≤ n. We find an exact stochastic representation for the dis- was unable to deal with this regime otherwise, see the discus- Pn 0 tribution of β = argmin p ρ(Y − X β) at fixed p and n b β2R i=1 i i sion on p.802 of [6]. under various assumptions on the objective function ρ and our sta- In this paper we intend to, in part heuristically and with tistical model. A scalar random variable whose deterministic limit \computer validation", analyze fully what happens in robust rρ(κ) can be studied when p=n ! κ > 0 plays a central role in this regression when p=n ! κ < 1. We do limit ourselves to Gaus- representation. Furthermore, we discover a non-linear system of two deterministic sian covariates but present grounds that the behavior holds equations that characterizes rρ(κ). Interestingly, the system shows much more generally. We also investigate the sensitivity of that rρ(κ) depends on ρ through proximal mappings of ρ as well as our results to the geometry of the design matrix. -
Robust Regression in Stata
The Stata Journal (yyyy) vv, Number ii, pp. 1–23 Robust Regression in Stata Vincenzo Verardi University of Namur (CRED) and Universit´eLibre de Bruxelles (ECARES and CKE) Rempart de la Vierge, 8, B-5000 Namur. E-mail: [email protected]. Vincenzo Verardi is Associated Researcher of the FNRS and gratefully acknowledges their finacial support Christophe Croux K.U.Leuven, Faculty of Business and Economics Naamsestraat 69. B-3000, Leuven E-mail: [email protected]. Abstract. In regression analysis, the presence of outliers in the data set can strongly distort the classical least squares estimator and lead to un- reliable results. To deal with this, several robust-to-outliers methods have been proposed in the statistical literature. In Stata, some of these methods are available through the commands rreg and qreg. Unfortunately, these methods only resist to some specific types of outliers and turn out to be ineffective under alternative scenarios. In this paper we present more ef- fective robust estimators that we implemented in Stata. We also present a graphical tool that allows recognizing the type of detected outliers. KEYWORDS: S-estimators, MM-estimators, Outliers, Robustness JEL CLASSIFICATION: C12, C21, C87 c yyyy StataCorp LP st0001 2 Robust Regression in Stata 1 Introduction The objective of linear regression analysis is to study how a dependent variable is linearly related to a set of regressors. In matrix notation, the linear regression model is given by: y = Xθ + ε (1) where, for a sample of size n, y is the (n × 1) vector containing the values for the dependent variable, X is the (n × p) matrix containing the values for the p regressors and ε is the (n × 1) vector containing the error terms. -
Robust Fitting of Parametric Models Based on M-Estimation
Robust Fitting of Parametric Models Based on M-Estimation Andreas Ruckstuhl ∗ IDP Institute of Data Analysis and Process Design ZHAW Zurich University of Applied Sciences in Winterthur Version 2018† ∗Email Address: [email protected]; Internet: http://www.idp.zhaw.ch †The author thanks Amanda Strong for their help in translating the original German version into English. Contents 1. Basic Concepts 1 1.1. The Regression Model and the Outlier Problem . ..... 1 1.2. MeasuringRobustness . 3 1.3. M-Estimation................................. 7 1.4. InferencewithM-Estimators . 10 2. Linear Regression 12 2.1. RegressionM-Estimation . 12 2.2. Example from Molecular Spectroscopy . 13 2.3. General Regression M-Estimation . 17 2.4. Robust Regression MM-Estimation . 19 2.5. Robust Inference and Variable Selection . ...... 23 3. Generalized Linear Models 29 3.1. UnifiedModelFormulation . 29 3.2. RobustEstimatingProcedure . 30 4. Multivariate Analysis 34 4.1. Robust Estimation of the Covariance Matrix . ..... 34 4.2. PrincipalComponentAnalysis . 39 4.3. Linear Discriminant Analysis . 42 5. Baseline Removal Using Robust Local Regression 45 5.1. AMotivatingExample.. .. .. .. .. .. .. .. 45 5.2. LocalRegression ............................... 45 5.3. BaselineRemoval............................... 49 6. Some Closing Comments 51 6.1. General .................................... 51 6.2. StatisticsPrograms. 51 6.3. Literature ................................... 52 A. Appendix 53 A.1. Some Thoughts on the Location Model . 53 A.2. Calculation of Regression M-Estimations . ....... 54 A.3. More Regression Estimators with High Breakdown Points ........ 56 Bibliography 58 Objectives The block course Robust Statistics in the post-graduated course (Weiterbildungslehrgang WBL) in applied statistics at the ETH Zürich should 1. introduce problems where robust procedures are advantageous, 2. explain the basic idea of robust methods for linear models, and 3. -
Stat 714 Linear Statistical Models
STAT 714 LINEAR STATISTICAL MODELS Fall, 2010 Lecture Notes Joshua M. Tebbs Department of Statistics The University of South Carolina TABLE OF CONTENTS STAT 714, J. TEBBS Contents 1 Examples of the General Linear Model 1 2 The Linear Least Squares Problem 13 2.1 Least squares estimation . 13 2.2 Geometric considerations . 15 2.3 Reparameterization . 25 3 Estimability and Least Squares Estimators 28 3.1 Introduction . 28 3.2 Estimability . 28 3.2.1 One-way ANOVA . 35 3.2.2 Two-way crossed ANOVA with no interaction . 37 3.2.3 Two-way crossed ANOVA with interaction . 39 3.3 Reparameterization . 40 3.4 Forcing least squares solutions using linear constraints . 46 4 The Gauss-Markov Model 54 4.1 Introduction . 54 4.2 The Gauss-Markov Theorem . 55 4.3 Estimation of σ2 in the GM model . 57 4.4 Implications of model selection . 60 4.4.1 Underfitting (Misspecification) . 60 4.4.2 Overfitting . 61 4.5 The Aitken model and generalized least squares . 63 5 Distributional Theory 68 5.1 Introduction . 68 i TABLE OF CONTENTS STAT 714, J. TEBBS 5.2 Multivariate normal distribution . 69 5.2.1 Probability density function . 69 5.2.2 Moment generating functions . 70 5.2.3 Properties . 72 5.2.4 Less-than-full-rank normal distributions . 73 5.2.5 Independence results . 74 5.2.6 Conditional distributions . 76 5.3 Noncentral χ2 distribution . 77 5.4 Noncentral F distribution . 79 5.5 Distributions of quadratic forms . 81 5.6 Independence of quadratic forms . 85 5.7 Cochran's Theorem . -
Robust Regression Examples
Chapter 9 Robust Regression Examples Chapter Table of Contents OVERVIEW ...................................177 FlowChartforLMS,LTS,andMVE......................179 EXAMPLES USING LMS AND LTS REGRESSION ............180 Example 9.1 LMS and LTS with Substantial Leverage Points: Hertzsprung- Russell Star Data . ......................180 Example9.2LMSandLTS:StacklossData..................184 Example 9.3 LMS and LTS Univariate (Location) Problem: Barnett and LewisData...........................193 EXAMPLES USING MVE REGRESSION ..................196 Example9.4BrainlogData...........................196 Example9.5MVE:StacklossData.......................201 EXAMPLES COMBINING ROBUST RESIDUALS AND ROBUST DIS- TANCES ................................208 Example9.6Hawkins-Bradu-KassData....................209 Example9.7StacklossData..........................218 REFERENCES ..................................220 176 Chapter 9. Robust Regression Examples SAS OnlineDoc: Version 8 Chapter 9 Robust Regression Examples Overview SAS/IML has three subroutines that can be used for outlier detection and robust re- gression. The Least Median of Squares (LMS) and Least Trimmed Squares (LTS) subroutines perform robust regression (sometimes called resistant regression). These subroutines are able to detect outliers and perform a least-squares regression on the remaining observations. The Minimum Volume Ellipsoid Estimation (MVE) sub- routine can be used to find the minimum volume ellipsoid estimator, which is the location and robust covariance matrix that can be used