Speaking Stata: Trimming to Taste
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Outlier Impact and Accommodation on Power Hongjing Liao Beijing Foreign Studies University, [email protected]
Journal of Modern Applied Statistical Methods Volume 16 | Issue 1 Article 15 5-1-2017 Outlier Impact and Accommodation on Power Hongjing Liao Beijing Foreign Studies University, [email protected] Yanju Li Western Carolina University, Cullowhee, NC, [email protected] Gordon P. Brooks Ohio University, [email protected] Follow this and additional works at: http://digitalcommons.wayne.edu/jmasm Part of the Applied Statistics Commons, Social and Behavioral Sciences Commons, and the Statistical Theory Commons Recommended Citation Liao, H., Yanju, Li., & Brooks, G. P. (2017). Outlier impact and accommodation on power. Journal of Modern Applied Statistical Methods, 16(1), 261-278. doi: 10.22237/jmasm/1493597640 This Regular Article is brought to you for free and open access by the Open Access Journals at DigitalCommons@WayneState. It has been accepted for inclusion in Journal of Modern Applied Statistical Methods by an authorized editor of DigitalCommons@WayneState. Outlier Impact and Accommodation on Power Cover Page Footnote This research was supported by School of English for Specific urP poses, Beijing Foreign Studies University, grant ZJ1513 This regular article is available in Journal of Modern Applied Statistical Methods: http://digitalcommons.wayne.edu/jmasm/vol16/ iss1/15 Journal of Modern Applied Statistical Methods Copyright © 2017 JMASM, Inc. May 2017, Vol. 16, No. 1, 261-278. ISSN 1538 − 9472 doi: 10.22237/jmasm/1493597640 Outlier Impact and Accommodation on Power Hongjing Liao Yanju Li Gordon P. Brooks Beijing Foreign Studies University Western Carolina University Ohio University Beijing, China Cullowhee, NC Athens, OH The outliers’ influence on power rates in ANOVA and Welch tests at various conditions was examined and compared with the effectiveness of nonparametric methods and Winsorizing in minimizing the impact of outliers. -
Regularization of Case-Specific Parameters for Robustness And
Regularization of Case-Specific Parameters for Robustness and Efficiency Yoonkyung Lee, Steven N. MacEachern, and Yoonsuh Jung The Ohio State University, USA Summary. Regularization methods allow one to handle a variety of inferential problems where there are more covariates than cases. This allows one to consider a potentially enormous number of covari- ates for a problem. We exploit the power of these techniques, supersaturating models by augmenting the “natural” covariates in the problem with an additional indicator for each case in the data set. We attach a penalty term for these case-specific indicators which is designed to produce a desired effect. For regression methods with squared error loss, an ℓ1 penalty produces a regression which is robust to outliers and high leverage cases; for quantile regression methods, an ℓ2 penalty decreases the variance of the fit enough to overcome an increase in bias. The paradigm thus allows us to robustify procedures which lack robustness and to increase the efficiency of procedures which are robust. We provide a general framework for the inclusion of case-specific parameters in regularization problems, describing the impact on the effective loss for a variety of regression and classification problems. We outline a computational strategy by which existing software can be modified to solve the augmented regularization problem, providing conditions under which such modification will converge to the opti- mum solution. We illustrate the benefits of including case-specific parameters in the context of mean regression and quantile regression through analysis of NHANES and linguistic data sets. Keywords: Case indicator; Large margin classifier; LASSO; Leverage point; Outlier; Penal- ized method; Quantile regression 1. -
Statistical Inference with Paired Observations and Independent Observations in Two Samples
Statistical inference with paired observations and independent observations in two samples Benjamin Fletcher Derrick A thesis submitted in partial fulfilment of the requirements of the University of the West of England, Bristol, for the degree of Doctor of Philosophy Applied Statistics Group, Engineering Design and Mathematics, Faculty of Environment and Technology, University of the West of England, Bristol February, 2020 Abstract A frequently asked question in quantitative research is how to compare two samples that include some combination of paired observations and unpaired observations. This scenario is referred to as ‘partially overlapping samples’. Most frequently the desired comparison is that of central location. Depend- ing on the context, the research question could be a comparison of means, distributions, proportions or variances. Approaches that discard either the paired observations or the independent observations are customary. Existing approaches evoke much criticism. Approaches that make use of all of the available data are becoming more prominent. Traditional and modern ap- proaches for the analyses for each of these research questions are reviewed. Novel solutions for each of the research questions are developed and explored using simulation. Results show that proposed tests which report a direct measurable difference between two groups provide the best solutions. These solutions advance traditional methods in this area that have remained largely unchanged for over 80 years. An R package is detailed to assist users to per- form these new tests in the presence of partially overlapping samples. Acknowledgments I pay tribute to my colleagues in the Mathematics Cluster and the Applied Statistics Group at the University of the West of England, Bristol. -
Applying Robust Methods to Operational Risk Modeling
Applying Robust Methods to Operational Risk Modeling Anna Chernobai∗ Svetlozar T. Rachev† University of California Universit¨at Karlsruhe, at Santa Barbara, USA Germany and University of California at Santa Barbara, USA March 7, 2006 Contents 1 Introduction 1 2 Actuarial Approach to Modeling Operational Risk 2 3 Why Robust Statistics? 4 3.1 Classical vs. Robust Methods . 4 3.2 Some Examples . 5 3.3 Overview of Literature on Robust Statistics . 6 3.4 Outlier Rejection Approach and Stress Tests: A Parallel . 8 4 Application to Operational Loss Data 9 5 Conclusions 10 6 Acknowledgements 12 ∗Email: [email protected]. Parts of the research were done when A.Chernobai was visiting Institute of Econometrics, Statistics, and Mathematical Finance, School of Economics and Business Engineering, University of Karlsruhe, Germany, in 2005. †Email: [email protected]. S.Rachev gratefully acknowledges research support by grants from Division of Mathematical, Life and Physical Sciences, College of Letters and Science, University of California, Santa Barbara, the German Research Foundation (DFG), and the German Academic Exchange Service (DAAD). 1 INTRODUCTION 1 1 Introduction In 2001, the Basel Committee of the Bank of International Settlements (BIS) released new regulatory capital guidelines on operational risk (BIS, 2001a,b), finalized in 2004 (BIS, 2004). The nature of operational risk, fundamentally different from that of market risk and credit risk, is highly bank-specific and calls for the development of complex quantitative and quali- tative solutions, new “know-how,” and setting additional standards for training bank person- nel. Despite significant progress in the operational risk management, numerous challenges remain, and the current development in the area is criticized by many as being “more art than science.” An actuarial type model dominates statistical models for operational risk under the Advanced Measurement Approach (see Section 2 for the discussion of the model; see also publications by BIS, 2001-2004). -
Influential Observations and Inference in Accounting Research
Influential Observations and Inference in Accounting Research Andrew J. Leone University Miami [email protected] Miguel Minutti-Meza University of Miami [email protected] Charles Wasley University of Rochester [email protected] October 2013 (First Draft: April 2012) Abstract The existence of potentially influential or outlier observations is ubiquitous in empirical accounting research. The purpose of this paper is to summarize the various methods used by accounting researchers to identify and control for influential observations; assess the effectiveness of such methods and of alternative methods from the statistics literature; and, provide guidance for future studies so they may systematically identify and mitigate the impact of influential observations. A survey of articles published in accounting journals shows considerable variation in the way researchers deal with influential observations prior to estimating a regression model. We demonstrate that the common approach of winsorizing each variable has only a modest impact on parameter estimates compared to “doing nothing”. We also demonstrate that truncation tends to bias coefficients toward zero. More generally, we show that both winsorizing and truncating do little to accommodate potential bias caused by unusual and infrequent events (i.e., data points that are not random errors). Alternatively, use of robust regression significantly reduces such bias and is preferable to winsorization or truncation when regression is used for hypothesis testing. In addition, robust regression methods are available in commonly-used statistical packages and they do not entail the ad hoc choice of winsorization or truncation rules, thus providing a convenient way to control for influential observations and enhance inter-study comparability. -
ELKI: a Large Open-Source Library for Data Analysis
ELKI: A large open-source library for data analysis ELKI Release 0.7.5 “Heidelberg” https://elki-project.github.io/ Erich Schubert Arthur Zimek Technische Universität Dortmund University of Southern Denmark 1 Introduction This paper documents the release of the ELKI data mining framework, version 0.7.5. ELKI is an open source (AGPLv3) data mining software written in Java. The focus of ELKI is research in algorithms, with an emphasis on unsupervised methods in cluster analysis and outlier detection. In order to achieve high performance and scalability, ELKI oers data index structures such as the R*-tree that can provide major performance gains. ELKI is designed to be easy to extend for researchers and students in this domain, and welcomes contributions of additional methods. ELKI aims at providing a large collection of highly parameterizable algorithms, in order to allow easy and fair evaluation and benchmarking of algorithms. We will rst outline the motivation for this release, the plans for the future, and then give a brief overview over the new functionality in this version. We also include an appendix presenting an overview on the overall implemented functionality. 2 ELKI 0.7.5: “Heidelberg” The majority of the work in this release was done at Heidelberg University, both by students (that for example contributed many unit tests as part of Java programming practicals) and by Dr. Erich Schubert. The last ocial release 0.7.1 is now three years old. Some changes in Java have broken this old release, and it does not run on Java 11 anymore (trivial xes are long available, but there has not been a released version) — three years where we have continuted to add functionality, improve APIs, x bugs, ...yet surprisingly many users still use the old version rather than building the latest development version from git (albeit conveniently available at Github, https://github.com/elki-project/elki). -
Quantitative Data Cleaning for Large Databases
Quantitative Data Cleaning for Large Databases Joseph M. Hellerstein∗ EECS Computer Science Division UC Berkeley http://db.cs.berkeley.edu/jmh February 27, 2008 1 Introduction Data collection has become a ubiquitous function of large organizations { not only for record keeping, but to support a variety of data analysis tasks that are critical to the organizational mission. Data analysis typically drives decision-making processes and efficiency optimizations, and in an increasing number of settings is the raison d'etre of entire agencies or firms. Despite the importance of data collection and analysis, data quality remains a pervasive and thorny problem in almost every large organization. The presence of incorrect or inconsistent data can significantly distort the results of analyses, often negating the potential benefits of information-driven approaches. As a result, there has been a variety of research over the last decades on various aspects of data cleaning: computational procedures to automatically or semi-automatically identify { and, when possible, correct { errors in large data sets. In this report, we survey data cleaning methods that focus on errors in quantitative at- tributes of large databases, though we also provide references to data cleaning methods for other types of attributes. The discussion is targeted at computer practitioners who manage large databases of quantitative information, and designers developing data entry and auditing tools for end users. Because of our focus on quantitative data, we take a statistical view of data quality, with an emphasis on intuitive outlier detection and exploratory data analysis methods based in robust statistics [Rousseeuw and Leroy, 1987, Hampel et al., 1986, Huber, 1981]. -
Influential Observations and Inference in Accounting Research
Influential Observations and Inference in Accounting Research Andrew J. Leone University Miami Miguel Minutti-Meza University of Miami Charles Wasley University of Rochester Very Preliminary April 2012 I. Introduction This study examines the statistical problems related to the presence of influential observations or outliers in financial data. The objectives are: (1) to summarize the various methods used to identify and adjust for influential observations in the accounting literature; (2) to assess the effectiveness of such methods in commonly used research designs; (3) to assess the effectiveness of new methods proposed by the statistics literature; and, (4) to provide guidance for future studies that need to identify and correct for the presence of influential observations. Although a precise definition is hard to come by, influential observations are generally considered to be data points that have a large impact on the calculated values of various estimates (e.g, mean, regression coefficients, standard errors, etc.), while outliers are data points that are located unusually far from the mean of a sample. Influential observations or outliers are not necessarily problematic, but researchers should be aware of their potential effect on statistical inferences.1 Belsley et al. (1980, p. 3) remark: “Unusual or influential data points, of course, are not necessarily bad data points; they may contain some of the most interesting sample information. They may also, however, be in error or result from circumstances different from those common to the remaining data.” This remark highlights an important issue for accounting researchers. Unusual data points may result either from erroneous data or from unusual events affecting a subset of the data. -
Robust Statistical Methods Using WRS2
Robust Statistical Methods Using WRS2 Patrick Mair Rand Wilcox Harvard University University of Southern California Abstract This vignette is a (slightly) modified version of Mair and Wilcox (2020), published in Behavior Research Methods. It introduces the R package WRS2 that implements various robust statistical methods. It elaborates on the basics of robust statistics by introducing robust location, dispersion, and correlation measures. The location and dispersion measures are then used in robust variants of independent and dependent samples t-tests and ANOVA, including between- within subject designs and quantile ANOVA. Further, robust ANCOVA as well as robust mediation models are introduced. The paper targets applied researchers; it is therefore kept rather non-technical and written in a tutorial style. Special emphasis is placed on applications in the social and behavioral sciences and illustrations of how to perform corresponding robust analyses in R. The R code for reproducing the results in the paper is given in the supplementary materials. Keywords: robust statistics, robust location measures, robust ANOVA, robust ANCOVA, robust mediation, robust correlation. 1. Introduction Classic inferential methods based on means (e.g., the ANOVA F -test) assume normality and homoscedasticity (equal variances). A fundamental issue is whether violating these two assumptions is a serious practical concern. Based on numerous articles summarized in Wilcox (2017), the answer is an unequivocal “yes”. Under general conditions they can have relatively poor power, they can yield inaccurate confidence intervals, and they can poorly characterize the extent groups differ. Even a small departure from normality can be a serious concern. Despite the central limit theorem, certain serious concerns persist even when dealing with large sample sizes. -
The Cambridge Dictionary of Statistics Fourth Edition
This page intentionally left blank THE CAMBRIDGE DICTIONARY OF STATISTICS FOURTH EDITION If you work with data and need easy access to clear, reliable definitions and explanations of modern statistical and statistics-related concepts, then look no further than this dictionary. Nearly 4000 terms are defined, covering medical, survey, theoretical and applied statistics, including computational and graphical aspects. Entries are provided for standard and specialized statistical software. In addition, short biographies of over 100 important statis- ticians are given. Definitions provide enough mathematical detail to clarify concepts and give standard formulae when these are helpful. The majority of definitions then give a reference to a book or article where the user can seek further or more specialized informa- tion, and many are accompanied by graphical material to aid understanding. B. S. EVERITT is Professor Emeritus of King’s College London. He is the author of almost 60 books on statistics and computing, including Medical Statistics from A to Z, also from Cambridge University Press. A. SKRONDAL is Senior Statistician in the Division of Epidemiology, Norwegian Institute of Public Health and Professor of Biostatistics in the Department of Mathematics, University of Oslo. Previous positions include Professor of Statistics and Director of The Methodology Institute at the London School of Economics. THE CAMBRIDGE DICTIONARY OF Statistics Fourth Edition B. S. EVERITT Institute of Psychiatry, King’s College London A. SKRONDAL Norwegian Institute of Public Health Department of Mathematics, University of Oslo CAMBRIDGE UNIVERSITY PRESS Cambridge, New York, Melbourne, Madrid, Cape Town, Singapore, São Paulo, Delhi, Dubai, Tokyo Cambridge University Press The Edinburgh Building, Cambridge CB2 8RU, UK Published in the United States of America by Cambridge University Press, New York www.cambridge.org Information on this title: www.cambridge.org/9780521766999 © B.