A-B-C of EDA 040127.Pdf

Total Page:16

File Type:pdf, Size:1020Kb

A-B-C of EDA 040127.Pdf A plications, &ics, and C puting* ' wm Applications, Basics, and Computing of Exploratory Data Analysis By Paul F. Velleman, Cornell University and David C. Hoaglin, Abt Associates, Inc. and Harvard University Previously published by Duxbury Press, Boston Copyright 2004 by Paul F. Velleman and David Hoaglin Republished by The Internet-First University Press This manuscript is among the initial offerings being published as part of a new approach to scholarly publishing. The manuscript is freely available from the Internet-First University Press repository within DSpace at Cornell University at http://dspace.library.cornell.edu/handle/1813/62 The online version of this work is available on an open access basis, without fees or restrictions on personal use. A professionally printed and bound version may be purchased through Cornell Business Services by contacting: [email protected] All mass reproduction, even for educational or not-for-profit use, requires permission and license. For more information, please contact [email protected]. We will provide a downloadable version of this document from the Internet- First University Press. Ithaca, N.Y. January, 2004 Applications, Basics, and Computing of Exploratory Data Analysis To John W. Tukey Contents Preface xiii Introduction xv Chapter 1 Stem-and-Leaf Displays 1.1 Stems and Leaves 2 1.2 Multiple Lines per Stem 7 1.3 Positive and Negative Values 11 1.4 Listing Apparent Strays 12 1.5 Histograms 13 1.6 Stem-and-Leaf Displays from the Computer 15 1.7 Algorithms I 16 f 1.8 Algorithms II 17 Chapter 2 Letter-Value Displays 41 2.1 Median, Hinges, and Other Summary Values 41 2.2 Letter Values 44 2.3 Displaying the Letter Values 46 Vll ABCs of EDA 2.4 Re-expression and the Ladder of Powers 48 2.5 Re-expressions for Symmetry: An Example 50 2.6 Comparing Spreads to the Gaussian Distribution 53 2.7 Letter Values from the Computer 55 2.8 Algorithms 55 2.9 Sorting 57 Chapter 3 Boxplots 65 3.1 Basic Purposes 65 3.2 The Skeletal Boxplot 66 3.3 Outliers 67 3.4 Making a Boxplot 69 3.5 Boxplots from the Computer 71 3.6 Comparing Batches 71 * 3.7 More Refined Comparisons: Notched Boxplots 73 3.8 Using the Programs 74 t 3.9 Algorithms 75 t 3.10 Implementation Details 78 t 3.11 Further Refinements in Display 78 * 3.12 Details of the Notched Boxplot 79 Chapter 4 x-y Plotting 93 4.1 x-y Plots 95 4.2 Computer Plots 96 4.3 Condensed Plots 96 4.4 Coded Plot Symbols 98 4.5 Condensed Plots and Stem-and-Leaf Displays 100 4.6 Bounds for Plots 104 4.7 Focusing Plots 105 4.8 Using the Programs 105 t 4.9 Algorithms 106 t 4.10 Alternatives 107 t 4.11 Details of the Programs 107 Chapter 5 Resistant Line 121 5.1 Slope and Intercept 121 5.2 Summary Points 123 Contents 5.3 Finding the Slope and the Intercept 125 5.4 Residuals 126 5.5 Polishing the Fit 127 5.6 Example: Breast Cancer Mortality versus Temperature 127 5.7 Outliers 134 5.8 Straightening Plots by Re-expression 135 5.9 Interpreting Fits to Re-expressed x-y Data 142 * 5.10 Resistant Lines and Least-Squares Regression 144 5.11 Resistant Lines from the Computer 144 t 5.12 Algorithms 145 Chapter 6 Smoothing Data 159 6.1 Data Sequences and Smooth Summaries 159 6.2 Elementary Smoothers 163 6.3 Compound Smoothers 170 6.4 Smoothing the Endpoints 173 6.5 Splitting and 3RSSH 177 6.6 Looking at the Rough 178 6.7 Smoothing and the Computer 181 t 6.8 Algorithms 182 Chapter 7 Coded Tables 201 7.1 Displaying Tables 203 7.2 Coded Tables from the Computer 203 7.3 Coded Tables and Boxplots 207 t 7.4 Algorithms 209 7.5 Details and Alternatives 212 Chapter 8 Median Polish 219 8.1 Two-Way Tables 219 8.2 A Model for Two-Way Tables 220 8.3 Residuals 223 8.4 Fitting an Additive Model by Median Polish 225 8.5 Re-expressing for Additivity 233 8.6 Median Polish from the Computer 240 * 8.7 Median Polish and ANOVA 241 ABCs of EDA * 8.8 Data Structure 241 t 8.9 Algorithms 242 Chapter 9 Rootograms 255 9.1 Histograms and the Area Principle 257 9.2 Comparisons and Residuals 262 9.3 Rootograms 263 9.4 Fitting a Gaussian Comparison Curve 267 9.5 Suspended Rootograms 274 9.6 Rootograms from the Computer 277 9.7 More on Double Roots 281 Appendix A Computer Graphics 293 A.I Terminology 293 A.2 Exploratory Displays 295 A.3 Resistant Scaling 295 A.4 Printer Plots 296 A.5 Display Details 297 Appendix B Utility Programs 301 B.I BASIC 301 B.2 FORTRAN 308 Appendix C Programming Conventions 319 C.I BASIC 319 C.2 FORTRAN 325 Appendix D Minitab Implementation 335 D.I Stem-and-Leaf Displays 337 D.2 Letter-Value Displays 337 D.3 Boxplots 338 D.4 Condensed Plotting 339 Contents D.5 Resistant Lines 340 D.6 Resistant Smoothing 341 D.7 Coded Tables 342 D.8 Median Polish 343 D.9 Suspended Rootograms 344 Index 347 The BASIC programs in this book are available in machine-readable form from CONDUIT, P.O. Box 388, Iowa City, Iowa 52244 (319)353-5789. The FORTRAN programs in this book are available in machine- readable form from CONDUIT and from International Mathematical & Statistical Libraries, Inc., 6th Floor, NBC Building, 7500 Bellaire Boulevard, Houston, Texas 77036 (713)772-1927. A version of the BASIC programs tailored for the Apple micro computer is available from CONDUIT. Preface Exploratory data analysis techniques have added a new dimension to the way that people approach data. Over the past ten years, we have continually been impressed by how easily they have enabled us, our colleagues, and our students to uncover features concealed among masses of numbers. Unfortunately, the diversity of these techniques has at times discouraged students and data analysts who may want to learn a few methods without studying the full collection of exploratory tools. In addition, the lack of precisely specified algorithms has meant that computer programs for these techniques have not been widely available. This software gap has delayed the spread of exploratory methods. We have selected nine exploratory techniques that we have found most often useful. Each of these forms the basis for a chapter, in which we • Lay the foundations for understanding the technique, • Describe useful variations, • Illustrate applications to real data, and • Provide computer programs in FORTRAN and BASIC. The choice of languages makes it very likely that at least one of the programs for each technique can be readily installed on whatever computer system is available, from personal microcomputers to the largest mainframe. • • • Xlll x|v ABCs of EDA Most of this book requires no college level mathematics and no more than an introduction to statistical concepts. It can serve as a supplementary text to introduce the ideas and techniques of exploratory data analysis into a beginning course in statistics. (In draft form we have used portions of the book in just this way.) Some chapters include advanced sections which assume some knowledge of statistics and are intended to relate the exploratory techniques to traditional statistical practice. These sections will be of greater interest to researchers who wish to use the methods and programs in their own data analysis. A reader who is primarily interested in computational aspects of exploratory data analysis will find both the essential details and many refinements in our programs. At the other extreme, a student who has no background in programming and no access to a computer should have no difficulty in learning the techniques and applying them by pencil and paper. Between these two extremes, the reader who has access to the Minitab statistical system can take immediate advantage of our programs because they have been incorporated into Minitab (Releases 81.1 and later). Acknowledgments We are deeply grateful to the colleagues and friends who encouraged and aided us while we were developing this book. John Tukey originally suggested that we provide computer software for exploratory data analysis; later he participated in formulating the new resistant-line algorithm in Chapter 5, and he gave us critical comments on the manuscript. Frederick Mosteller gave us steadfast encouragement and invaluable advice, helped us to aim our writing at a high standard, and made many of the arrangements that facilitated our collaboration. Cleo Youtz painstakingly worked through the manuscript and helped us to eliminate a number of errors, large and small. John Emerson, Kathy Godfrey, Colin Goodall, Arthur Klein, J. David Velleman, Stanley Wasserman, and Agelia Ypelaar read various drafts and contributed helpful suggestions. Stephen Peters, Barbara Ryan, Thomas Ryan, and Michael Stoto gave us critical comments on the programs. Jeffrey Birch, Lambert Koop- mans, Douglas Lea, Thomas Louis, and Thomas Ryan reviewed the manu- script and suggested improvements. Teresa Redmond typed the manuscript, and Evelyn Maybee and Marjorie Olson typed some earlier draft material. We also appreciate the support provided by the National Science Foundation through grant SOC75-15702 to Harvard University. Initial versions of some BASIC programs were developed on a Model 4051 on loan from Tektronix, Inc. Introduction One recent thrust in statistics, primarily through the efforts of John Tukey, has produced a wealth of novel and ingenious methods of data analysis. In his 1977 book, Exploratory Data Analysis, and elsewhere, Tukey has expounded a practical philosophy of data analysis which minimizes prior assumptions and thus allows the data to guide the choice of appropriate models.
Recommended publications
  • Birth Cohort Effects Among US-Born Adults Born in the 1980S: Foreshadowing Future Trends in US Obesity Prevalence
    International Journal of Obesity (2013) 37, 448–454 & 2013 Macmillan Publishers Limited All rights reserved 0307-0565/13 www.nature.com/ijo ORIGINAL ARTICLE Birth cohort effects among US-born adults born in the 1980s: foreshadowing future trends in US obesity prevalence WR Robinson1,2, KM Keyes3, RL Utz4, CL Martin1 and Y Yang2,5 BACKGROUND: Obesity prevalence stabilized in the US in the first decade of the 2000s. However, obesity prevalence may resume increasing if younger generations are more sensitive to the obesogenic environment than older generations. METHODS: We estimated cohort effects for obesity prevalence among young adults born in the 1980s. Using data collected from the National Health and Nutrition Examination Survey between 1971 and 2008, we calculated obesity for respondents aged between 2 and 74 years. We used the median polish approach to estimate smoothed age and period trends; residual non-linear deviations from age and period trends were regressed on cohort indicator variables to estimate birth cohort effects. RESULTS: After taking into account age effects and ubiquitous secular changes, cohorts born in the 1980s had increased propensity to obesity versus those born in the late 1960s. The cohort effects were 1.18 (95% CI: 1.01, 1.07) and 1.21 (95% CI: 1.02, 1.09) for the 1979–1983 and 1984–1988 birth cohorts, respectively. The effects were especially pronounced in Black males and females but appeared absent in White males. CONCLUSIONS: Our results indicate a generational divergence of obesity prevalence. Even if age-specific obesity prevalence stabilizes in those born before the 1980s, age-specific prevalence may continue to rise in the 1980s cohorts, culminating in record-high obesity prevalence as this generation enters its ages of peak obesity prevalence.
    [Show full text]
  • Median Polish with Covariate on Before and After Data
    IOSR Journal of Mathematics (IOSR-JM) e-ISSN: 2278-5728, p-ISSN: 2319-765X. Volume 12, Issue 3 Ver. IV (May. - Jun. 2016), PP 64-73 www.iosrjournals.org Median Polish with Covariate on Before and After Data Ajoge I., M.B Adam, Anwar F., and A.Y., Sadiq Abstract: The method of median polish with covariate is use for verifying the relationship between before and after treatment data. The relationship is base on the yield of grain crops for both before and after data in a classification of contingency table. The main effects such as grand, column and row effects were estimated using median polish algorithm. Logarithm transformation of data before applying median polish is done to obtain reasonable results as well as finding the relationship between before and after treatment data. The data to be transformed must be positive. The results of median polish with covariate were evaluated based on the computation and findings showed that the median polish with covariate indicated a relationship between before and after treatment data. Keywords: Before and after data, Robustness, Exploratory data analysis, Median polish with covariate, Logarithmic transformation I. Introduction This paper introduce the background of the problem by using median polish model for verifying the relationship between before and after treatment data in classification of contingency table. We first developed the algorithm of median polish, followed by sweeping of columns and rows medians. This computational procedure, as noted by Tukey (1977) operates iteratively by subtracting the median of each column from each observation in that column level, and then subtracting the median from each row from this updated table.
    [Show full text]
  • BIOINFORMATICS ORIGINAL PAPER Doi:10.1093/Bioinformatics/Btm145
    Vol. 23 no. 13 2007, pages 1648–1657 BIOINFORMATICS ORIGINAL PAPER doi:10.1093/bioinformatics/btm145 Data and text mining An efficient method for the detection and elimination of systematic error in high-throughput screening Vladimir Makarenkov1,*, Pablo Zentilli1, Dmytro Kevorkov1, Andrei Gagarin1, Nathalie Malo2,3 and Robert Nadon2,4 1Department d’informatique, Universite´ du Que´ bec a` Montreal, C.P.8888, s. Centre Ville, Montreal, QC, Canada, H3C 3P8, 2McGill University and Genome Quebec Innovation Centre, 740 Dr. Penfield Ave., Montreal, QC, Canada, H3A 1A4, 3Department of Epidemiology, Biostatistics, and Occupational Health, McGill University, 1020 Pine Av. West, Montreal, QC, Canada, H3A 1A4 and 4Department of Human Genetics, McGill University, 1205 Dr. Penfield Ave., N5/13, Montreal, QC, Canada, H3A 1B1 Received on December 7, 2006; revised on February 22, 2007; accepted on April 10, 2007 Advance Access publication April 26, 2007 Associate Editor: Jonathan Wren ABSTRACT experimental artefacts that might confound with important Motivation: High-throughput screening (HTS) is an early-stage biological or chemical effects. The description of several process in drug discovery which allows thousands of chemical methods for quality control and correction of HTS data can compounds to be tested in a single study. We report a method for be found in Brideau et al. (2003), Gagarin et al. (2006a), Gunter correcting HTS data prior to the hit selection process (i.e. selection et al. (2003), Heuer et al. (2003), Heyse (2002), Kevorkov and of active compounds). The proposed correction minimizes the Makarenkov (2005), Makarenkov et al. (2006), Malo et al. impact of systematic errors which may affect the hit selection in (2006) and Zhang et al.
    [Show full text]
  • Exploratory Data Analysis
    Copyright © 2011 SAGE Publications. Not for sale, reproduction, or distribution. 530 Data Analysis, Exploratory The ultimate quantitative extreme in textual data Klingemann, H.-D., Volkens, A., Bara, J., Budge, I., & analysis uses scaling procedures borrowed from McDonald, M. (2006). Mapping policy preferences II: item response theory methods developed originally Estimates for parties, electors, and governments in in psychometrics. Both Jon Slapin and Sven-Oliver Eastern Europe, European Union and OECD Proksch’s Poisson scaling model and Burt Monroe 1990–2003. Oxford, UK: Oxford University Press. and Ko Maeda’s similar scaling method assume Laver, M., Benoit, K., & Garry, J. (2003). Extracting that word frequencies are generated by a probabi- policy positions from political texts using words as listic function driven by the author’s position on data. American Political Science Review, 97(2), some latent scale of interest and can be used to 311–331. estimate those latent positions relative to the posi- Leites, N., Bernaut, E., & Garthoff, R. L. (1951). Politburo images of Stalin. World Politics, 3, 317–339. tions of other texts. Such methods may be applied Monroe, B., & Maeda, K. (2004). Talk’s cheap: Text- to word frequency matrixes constructed from texts based estimation of rhetorical ideal-points (Working with no human decision making of any kind. The Paper). Lansing: Michigan State University. disadvantage is that while the scaled estimates Slapin, J. B., & Proksch, S.-O. (2008). A scaling model resulting from the procedure represent relative dif- for estimating time-series party positions from texts. ferences between texts, they must be interpreted if a American Journal of Political Science, 52(3), 705–722.
    [Show full text]
  • Expression Summarization Interrogate for Each Gene, Called a Probe Set
    Expression Quantification: Affy Affymetrix Genechip is an oligonucleotide array consisting of a several perfect match (PM) and their corresponding mismatch (MM) probes that interrogate for a single gene. · PM is the exact complementary sequence of the target genetic sequence, composed of 25 base pairs · MM probe, which has the same sequence with exception that the middle base (13th) position has been reversed · There are roughly 11­20 PM/MM probe pairs that Expression summarization interrogate for each gene, called a probe set Mikhail Dozmorov Fall 2016 2/36 Affymetrix Expression Array Preprocessing Affymetrix Expression Array Preprocessing Background adjustment Normalization Remove intensity contributions from optical noise and cross­ Remove array effect, make array comparable hybridization 1. Constant or linear (MAS) · so the true measurements aren't affected by neighboring measurements 2. Rank invariant (dChip) 3. Quantile (RMA) 1. PM­MM 2. PM only 3. RMA 4. GC­RMA 3/36 4/36 Affymetrix Expression Array Preprocessing Expression Index estimates Summarization Summarization Combine probe intensities into one measure per gene · Reduce the 11­20 probe intensities on each array to a single number for gene expression. 1. MAS 4.0, MAS 5.0 · The goal is to produce a measure that will serve as an 2. Li­Wong (dChip) indicator of the level of expression of a transcript using 3. RMA the PM (and possibly MM values). 5/36 6/36 Expression Index estimates Expression Index estimates Single Chip Multiple Chip · MAS 4.0 (avgDiff): no longer recommended for use due · MBEI (Li­Wong): a multiplicative model (Model based to many flaws.
    [Show full text]
  • Median Polish Algorithm
    Introduction to Computational Chemical Biology Dr. Raimo Franke [email protected] Lecture Series Chemical Biology http://www.raimofranke.de/ Leibniz-Universität Hannover Where can you find me? • Dr. Raimo Franke Department of Chemical Biology (C-Building) Helmholtz Centre for Infection Research Inhoffenstr. 7, 38124 Braunschweig Tel.: 0531-6181-3415 Email: [email protected] Download of lecture slides: www.raimofranke.de • Research Topics - Metabolomics - Biostatistical Data Analysis of omics experiments (NGS (genome, RNAseq), Arrays, Metabolomics, Profiling) - Phenotypic Profiling of bioactive compounds with impedance measurements (xCelligence) (Peptide Synthesis, ABPP) I am looking for BSc, MSc and PhD students, feel free to contact me! Slide 2 | My journey… Slide 3 | Outline Primer on Statistical Methods Multi-parameter phenotypic profiling: using cellular effects to characterize bioactive compounds Network Pharmacology, Modeling of Signal Transduction Networks Paper Presentations: Proc Natl Acad Sci U S A. 2013 Feb 5;110(6):2336-41. doi: 10.1073/pnas.1218524110. Epub 2013 Jan 22. Antimicrobial drug resistance affects broad changes in metabolomic phenotype in addition to secondary metabolism. Derewacz DK, Goodwin CR, McNees CR, McLean JA, Bachmann BO. Slide 4 | Chemical Biology Arsenal of Methods In silico Target interactions Correlation signals Chemical Genetics & predictions Chemoproteomics Pharmacophore Biochemical assays Expression profiling Protein microarrays –based target SPR, ITC Proteomics Yeast-3-Hybrid prediction X-Ray, NMR Metabonomics Phage display Image analysis ABPP Impedance Affinity pulldown NCI60 panel Chemical probes Computational Methods in Chemical Biology • In silico –based target prediction: use molecular descriptors for in silico screening, molecular docking etc. (typical applications of Cheminformatics) • Biochemical Assays: SPR: curve fitting to determine kass und kdiss, Xray: model fitting, NMR: chemical shift prediction etc.
    [Show full text]
  • Introductory Methods for Area Data Analysis
    Analysis Methods for Area Data Introductory Methods for Values are associated with a fixed set of areal units covering the Area Data study region We assume a value has been observed for all areas Bailey and Gatrell Chapter 7 The areal units may take the form of a regular lattice or irregular units Lecture 17 November 11, 2003 Analysis Methods for Area Data Analysis Methods for Area Data Objectives For continuous data models we were more concerned with explanation of patterns in terms of locations § Not prediction - there are typically no unobserved values, the attribute is exhaustively measured. In this case explanation is in terms of covariates measured over the same units as well as in terms of the spatial § Model spatial patterns in the values associated with fixed arrangement of the areal units areas and determine possible explanations for such patterns Examples Relationship of disease rates and socio-economic variables 1 Analysis Methods for Area Data Analysis Methods for Area Data {Y (s), s Î R} Random variable Y indexed by locations Explore these attribute values in the context of global trend or first order variation and second order variation – the Random variable Y indexed by a fixed spatial arrangement of the set of areas {Y ( Ai ), ?i Î R} set of areal units § First order variation as variation in the of mean, mi of Yi A1 ÈL È An = R The set of areal units cover the study region R § Second order variation as variation in the COV(Yi , Yj ) Conceive of this sample as a sample from a super population – all realizations of the process over these areas that might ever occur.
    [Show full text]
  • Generalised Median Polish Based on Additive Generators
    Generalised Median Polish Based on Additive Generators Balasubramaniam Jayaram1 and Frank Klawonn2;3 Abstract Contingency tables often arise from collecting patient data and from lab experiments. A typical question to be answered based on a con- tingency table is whether the rows or the columns show a signi¯cant di®er- ence. Median Polish (MP) is fast becoming a prefered way to analyse con- tingency tables based on a simple additive model. Often, the data need to be transformed before applying the MP algorithm to get better results. A common transformation is the logarithm which essentially changes the un- derlying model to a multiplicative model. In this work, we propose a novel way of applying the MP algorithm with generalised transformations that still gives reasonable results. Our approach to the underlying model leads us to transformations that are similar to additive generators of some fuzzy logic connectives. In fact, we illustrate how to choose the best transformation that give meaningful results by proposing some modi¯ed additive generators of uninorms. In this way, MP is generalied from the simple additive model to more general nonlinear connectives. The recently proposed way of identifying a suitable power transformation based on IQRoQ plots [1] also plays a central role in this work. 1 Introduction Contingency tables often arise from collecting patient data and from lab experiments. The rows and columns of a contingency table correspond to two 1Department of Mathematics, Indian Institute of Technology Hyderabad, Yeddu- mailaram - 502 205, India [email protected] ¢2Department of Computer Science, Ostfalia University of Applied Sciences, Salzdahlumer Str.
    [Show full text]
  • A New Statistical Model for Analyzing Rating Scale Data Pertaining to Word Meaning
    Psychological Research DOI 10.1007/s00426-017-0864-8 ORIGINAL ARTICLE A new statistical model for analyzing rating scale data pertaining to word meaning 1,2,3 4 1,2,5 Felipe Munoz-Rubke • Karen Kafadar • Karin H. James Received: 28 October 2016 / Accepted: 31 March 2017 Ó Springer-Verlag Berlin Heidelberg 2017 Abstract The concrete-abstract categorization scheme has Imageability, and Multisensory. We analyzed the data guided several research programs. A popular way to clas- using both two-way and three-way MPA models. We also sify words into one of these categories is to calculate a calculated 95% CIs for the two-way models. Categorizing word’s mean value in a Concreteness or Imageability rating words with the Action scale revealed a continuum of word scale. However, this procedure has several limitations. For meaning for both nouns and verbs. The remaining scales instance, results can be highly distorted by outliers, ascribe produced dichotomous or stratified results for nouns, and differences among words when none may exist, and neglect continuous results for verbs. While the sample mean rating trends in participants. We suggest using an alterna- analysis generated continua irrespective of the rating scale, tive procedure to analyze rating scale data called median MPA differentiated among dichotomies and continua. We polish analysis (MPA). MPA is tolerant to outliers and conclude that MPA allowed us to better classify words by accounts for information in multiple dimensions, including discarding outliers, focusing on main trends, and consid- trends among participants. MPA performance can be ering the differences in rating criteria among participants.
    [Show full text]
  • Statistical Analysis of Some Gas Chromatography Measurements
    JOURNAL OF RESEARCHof the National Bureau of Standards Vol. 88, No. x, January-February 1983 Statistical Analysis of Some Gas Chromatography Measurements Karen Kafadar* and Keith R. Eberhardt* National Bureau of Standards, Washington, DC 20234 November 3, 1982 The National Bureau of Standards has certified Standard Reference Materials (SRMs) for the concentration of polychlorinated biphenyls (PCBs) in hydrocarbon matrices (transformer and motor oils). The certification of these SRMs involved measurements of extremely small concentrations of PCBs made by gas chromatography. Despite the high accuracy of the measurement techmique, the correlated data cannot he analyzed in a routine independent manner. A linear model for the measurements is described; its complexity encourages the use of simpler exploratory methods which reveal unexpected features and point the way towards obtaining valid statistical summaries of the data. Key words: exploratory analysis; linear models; median polish; robust estimates; statistical methods; uncertainty statement. 1. Introduction 2. The Data Exploratory methods in data analysis are used in many 2.1 Description of the SRM for PCBs in Oils fields of application. These methods are typically used on messy data because they are robust in nature; that is, they The National Bureau of Standards has certified are insensitive to unexpected departures from an assumed Standard Reference Materials (SRMs) for the model (e.g. outliers, non-normality). However, they can concentration of polychlorinated biphenyls (PCBs) in also provide valuable insight even for so-called "clean" hydrocarbon matrices, namely, transformer and motor data. This paper discusses an example of a measurement oils. PCBs are toxic contaminants; their chemical and process for which an exploratory approach can reveal thermal stability makes them commercially useful but also particularly interesting or unexpected trends even in leads to their persistence in the environment.
    [Show full text]
  • Using Approximations to Scale Exploratory Data Analysis In
    Using Approximations to Scale Exploratory Data Analysis in Datacub es Daniel Barbar a Xintao Wu George Mason University Information and Software Engineering Department Fairfax, VA 22303 dbarbara,[email protected] March 4, 1999 Pap er 188 Abstract Exploratory Data Analysis is a widely used technique to determine which factors have the most in uence on data values in a multi-way table, or which cells in the table can be considered anomalous with resp ect to the other cells. In particular, median p olish is a simple, yet robust metho d to p erform Exploratory Data Analysis. Median p olish is resistant to holes in the table cells that have no values, but it may require a lot of iterations through the data. This factor makes it dicult to apply median p olish to large multidimensional tables, since the I/O requirements may b e prohibitive. This pap er describ es a technique that uses median p olish over an approximation of a datacub e, easing the burden of I/O. The results obtained are tested for quality, using a variety of measures. The technique scales to large datacub es and proves to giveagood approximation of the results that would have b een obtained by median p olish in the original data. 1 Intro duction Exploratory Data Analysis EDA is a technique [13, 14, 22] that uncovers structure in data. EDA is p erformed without any a-priori hyp othesis in mind: rather it searches for \exceptions" of the data values relative to what those values would have b een if anticipated by an statistical mo del.
    [Show full text]
  • Universiti Putra Malaysia Median Polish Techniques
    UNIVERSITI PUTRA MALAYSIA MEDIAN POLISH TECHNIQUES FOR ANALYSING PAIRED DATA UPM IDRIS AJOGE COPYRIGHT © FS 2017 3 MEDIAN POLISH TECHNIQUES FOR ANALYSING PAIRED DATA UPM By IDRIS AJOGE COPYRIGHTThesis Submitted to the School of Graduate Studies, Universiti Putra Malaysia, in Fulfilment of the Requirements for the Degree of © Master of Science January 2017 UPM COPYRIGHT © COPYRIGHT All material contained within the thesis, including without limitation text, logos, icons, photographs and all other artwork, is copyright material of Universiti Putra Malaysia unless otherwise stated. Use may be made of any material contained within the thesis for non-commercial purposes from the copyright holder. Commercial use of material may only be made with the express, prior, written permission of Universiti Putra Malaysia. Copyright ©Universiti Putra Malaysia UPM COPYRIGHT © iii DEDICATIONS I would like to dedicate this dissertion work to • Almighty Allah, the benevolent and the merciful. • My respectful parents who have taught me a lot on the persistency in life. • My beloved wife for all her contribution, patience and understandings through- out my studies. She incredibly supported me and made it all possible for me. • My chidren for their patience and endurance and their love have always been my greatest inspiration. • My late younger brother Aminu S. Ajoge, whom I lost during the course ofUPM my study COPYRIGHT © Abstract of thesis presented to the Senate of Universiti Putra Malaysia in fulfilment of the requirement for the degree of Master of Science MEDIAN POLISH TECHNIQUES FOR ANALYSING PAIRED DATA By IDRIS AJOGE January 2017 Chairman: Mohd Bakri Adam, PhD UPM Faculty: Science Median polish is used as a data analysis technique for examining the significance of various factors in single or multi-way models.
    [Show full text]