A Comparison of Techniques for Assessing Central Tendency in Left

Total Page:16

File Type:pdf, Size:1020Kb

A Comparison of Techniques for Assessing Central Tendency in Left Chemosphere 80 (2010) 7–12 Contents lists available at ScienceDirect Chemosphere journal homepage: www.elsevier.com/locate/chemosphere A comparison of techniques for assessing central tendency in left-censored data using PCB and p,p0DDE contaminant concentrations from Michigan’s Bald Eagle Biosentinel Program Katherine F. Leith a,*, William W. Bowerman a, Michael R. Wierda a, Dave A. Best b, Teryl G. Grubb c, James G. Sikarske d a Department of Forestry and Natural Resources, Clemson University, Clemson, SC 29634, USA b United States Fish and Wildlife Service, East Lansing Field Office, East Lansing, MI 48823, USA c United States Forest Service, Rocky Mountain Research Station, Flagstaff, AZ 86001, USA d College of Veterinary Medicine, Small Animal Clinical Sciences, Michigan State University, East Lansing, MI 48825, USA article info abstract Article history: Monitoring of contaminants in the environment is an important part of understanding the fate of ecosys- Received 23 March 2010 tems after a chemical insult. Frequently, such monitoring efforts result in datasets with observations Accepted 30 March 2010 below the detection limit (DL) that are reported as ‘non-detect’ or ‘<DL’ and no value is provided. This Available online 22 April 2010 study explored the effects of non-detect data and their treatment on summary statistics. The data ana- lyzed in this paper are real-world data. They consist of both large (N = 234) and moderate (n = 12–64) Keywords: sample sizes with both good and marginal fit to an assumed distribution with a log-transformation. Sum- Kaplan–Meier mary statistics were calculated using (1) the ‘0.0001’ near-zero method of substitution, (2) substitution Substitution with ‘1/2 DL’, (3) multiple imputation, and (4) Kaplan–Meier estimation. Median was used for compar- Outliers à Multiple imputation ison. Several analytical options for datasets with non-detect observations are available. The general con- Non-detect sensus is that substitution methods ((1) and (2)) can produce biased summary statistics, especially as levels of substitution increase. Substitution methods continue to be used in research, likely because they are easy to implement. The objectives were to (1) assess the fit of lognormal distribution to the data, (2) compare and contrast the performance of four analytical treatments of left-censored data in terms of esti- mated geometric mean and standard error, and (3) make recommendations based on those results. Ó 2010 Elsevier Ltd. All rights reserved. 1. Introduction elli et al., 2005; Helsel, 2005a,b, 2006; Eastoe et al., 2006; Needham et al., 2007; Antweiler and Taylor, 2008). Specifically, it has been Monitoring contaminant levels in the environment is an impor- shown that the bias caused by substitution increases dramatically tant part of understanding the fate of ecosystems after a chemical as the percent of observations censored increases (Eastoe et al., insult. As levels decrease over time, limitations in analytical equip- 2006). In spite of this, various substitution methods continue to ment create a lower bound below which contaminant levels cannot be used in research, frequently with little regard for the proportion be accurately reported. This results in datasets with observations of observations censored. below the detection limit (DL) that are reported only as ‘non-de- In addition to censored observations, another complication tect’ or ‘<DL’ and no value is provided. This type of distribution is when analyzing environmental contaminant data is the possibility called ‘left-censored’, as the low-end observations that are un- that datasets display a right skew. This occurs when a few samples known generally occur near the origin of the x-axis in figures. Sev- show very high concentrations while the general tendency is for eral options for the analysis of these datasets have been concentrations to be lower. This distribution is common in envi- investigated leading to the conclusion that methods replacing all ronmental data and can frequently be accommodated by log-trans- non-detects with a single value (substitution methods) are fre- formation. The median has traditionally been an accepted measure quently inferior (Liu et al., 1997; Singh and Nocerino, 2002; Baccar- of central tendency for data which do not fit a normal distribution well, and this approach has been used in almost every field of sci- entific inquiry. Focus has shifted to newer approaches as more * Corresponding author. Address: Department of Forestry and Natural Resources, complex methods have been developed and computing power 261 Lehotsky Hall, Clemson University, Clemson, SC 29634, USA. Tel.: +1 864 656 has grown to make them feasible for the average researcher. While 5332 (O), +1 864 903 2728 (C); fax: +1 864 656 3304. E-mail addresses: [email protected], [email protected] (K.F. Leith). the median is still useful in that it is not based on indefensible 0045-6535/$ - see front matter Ó 2010 Elsevier Ltd. All rights reserved. doi:10.1016/j.chemosphere.2010.03.056 8 K.F. Leith et al. / Chemosphere 80 (2010) 7–12 assumptions about the shape of the distribution, it does not make plored the effects of non-detect data and their treatment on sum- use of all the information contained in a dataset. A final complexity mary statistics. The data analyzed in this paper represent both is added by the fact that lognormal data are frequently summa- large (N = 234) and moderate (n = 12–64) sample sizes with both rized using the geometric mean, which is particularly sensitive to good and marginal fit with a log-transformation. Summary statis- the choice of substitution value. tics were calculated using the current method of substitution with In 1997, the Michigan Department of Environmental Quality ‘0.0001’, the common method of substitution with ‘1/2 à DL’, MI, (MDEQ) implemented a Bald Eagle Biosentinel Program (BEBP) to and KM estimation. The median was also calculated for compari- monitor trends of a suite of organic pollutants under the Clean son with all four methods. The objectives were to (1) assess the Michigan Initiative (MDEQ, 1997). The data analyzed here are fit of lognormal distribution to the data, (2) compare and contrast 234 observations of total polychlorinated biphenol (PCB), which the performance of the four methods of non-detect handling in is a sum of all observed PCB congeners, and p,p0-dic- terms of estimated geometric mean, comparison to the median, hlorodiphenyldichloro-ethylene (p,p0DDE) concentrations found and standard error, and (3) make recommendations based on those in nestling bald eagle plasma samples from throughout the State results. of Michigan. Current statistical methods include tests of significant differences between regions at several geographic scales, and cal- 2. Methods culating descriptive statistics in the form of geometric means. Sub- stitution is currently used in cases of non-detect, or left-censored, Concentrations of total PCBs and p,p0DDE (lg/kg wet weight) in observations. plasma collected from nestling bald eagles across Michigan from This choice of methods for addressing the non-detects does not 1999 to 2003 were used in these analyses. In addition to analysis affect testing of significant differences between regions. The mon- as a single dataset representing the whole State of Michigan, data itoring program uses nonparametric Kruskal–Wallis and Wilcoxon were also classified geographically by subpopulation, based on the tests which are rank-based and make no assumptions about distri- classifications used in the BEBP. Subpopulations were defined by bution and so, are not sensitive to the problems of substitution first subdividing the state spatially into the categories of Great (Helsel, 2005b). However, summary statistics are reported as geo- Lakes and Inland breeding areas. Great Lakes breeding areas are de- metric means, which are affected by the choice of substituted va- fined as being within 8.0 km of Great Lakes shorelines and/or along lue. The BEBP currently substitutes the near-zero value of tributaries open to Great Lakes fish runs and inland breeding areas ‘0.0001’ for concentrations at non-detectable levels when analyz- are defined as being greater than 8.0 km from the Great Lakes ing contaminants that are summed, such as total PCBs. This near- shorelines and not along tributaries open to Great Lakes fish runs. zero value might appear to have little influence on the resulting These categories are then further subdivided into four Great Lakes calculations to those accustomed to arithmetic mean calculation and two Inland groups. The Great Lakes subpopulations consisted because the arithmetic mean is a function of addition, for which of Lake Superior (LS), Lake Michigan (LM), Lake Huron (LH), and ‘0’ is the identity value. Geometric means on the other hand are Lake Erie (LE). The Inland subpopulations consisted of Upper Pen- a function of multiplication, for which the identity value is ‘1’, insula (UP), and Lower Peninsula (LP) (Wierda, 2009). The data while near-zero values (like ‘0.0001’) have a drastic impact on analysis for this paper was generated using SASÒ software, Version the product. For contaminants that do not involve summation, a 9.1.2 of the SAS system for Windows. Copyright 2000–2004 SAS value of one-half of the quantification level is substituted for Institute Inc. SAS and all other SAS Institute Inc. product or service non-detects. names are registered trademarks or trademarks of SAS Institute The desire to summarize data more accurately has fueled recent Inc., Cary, NC, USA. comparisons of proposed analytical alternatives to substitution or simple median reporting (Helsel, 2005b). One alternative, maxi- mum likelihood estimation (MLE), forces the researcher to assume 2.1. Assessment of fit the shape of the underlying distribution, but is powerful if this assumption is correct. These MLE methods have been explored in Data were analyzed for significant departure from the lognor- a variety of environmental applications (Singh and Nocerino, mal distribution using the UNIVARIATE PROCEDURE with the op- 2002; Helsel, 2005b, 2006; Antweiler and Taylor, 2008; Jain et tions ‘normal’ and ‘plot’ activated (SAS Institute Inc., 2000–2004).
Recommended publications
  • Comparison of Harmonic, Geometric and Arithmetic Means for Change Detection in SAR Time Series Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas
    Comparison of Harmonic, Geometric and Arithmetic means for change detection in SAR time series Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas To cite this version: Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas. Comparison of Harmonic, Geometric and Arithmetic means for change detection in SAR time series. EUSAR. 9th European Conference on Synthetic Aperture Radar, 2012., Apr 2012, Germany. hal-00737524 HAL Id: hal-00737524 https://hal.archives-ouvertes.fr/hal-00737524 Submitted on 2 Oct 2012 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. EUSAR 2012 Comparison of Harmonic, Geometric and Arithmetic Means for Change Detection in SAR Time Series Guillaume Quin CEA, DAM, DIF, F-91297 Arpajon, France Béatrice Pinel-Puysségur CEA, DAM, DIF, F-91297 Arpajon, France Jean-Marie Nicolas Telecom ParisTech, CNRS LTCI, 75634 Paris Cedex 13, France Abstract The amplitude distribution in a SAR image can present a heavy tail. Indeed, very high–valued outliers can be observed. In this paper, we propose the usage of the Harmonic, Geometric and Arithmetic temporal means for amplitude statistical studies along time. In general, the arithmetic mean is used to compute the mean amplitude of time series.
    [Show full text]
  • University of Cincinnati
    UNIVERSITY OF CINCINNATI Date:___________________ I, _________________________________________________________, hereby submit this work as part of the requirements for the degree of: in: It is entitled: This work and its defense approved by: Chair: _______________________________ _______________________________ _______________________________ _______________________________ _______________________________ Gibbs Sampling and Expectation Maximization Methods for Estimation of Censored Values from Correlated Multivariate Distributions A dissertation submitted to the Division of Research and Advanced Studies of the University of Cincinnati in partial ful…llment of the requirements for the degree of DOCTORATE OF PHILOSOPHY (Ph.D.) in the Department of Mathematical Sciences of the McMicken College of Arts and Sciences May 2008 by Tina D. Hunter B.S. Industrial and Systems Engineering The Ohio State University, Columbus, Ohio, 1984 M.S. Aerospace Engineering University of Cincinnati, Cincinnati, Ohio, 1989 M.S. Statistics University of Cincinnati, Cincinnati, Ohio, 2003 Committee Chair: Dr. Siva Sivaganesan Abstract Statisticians are often called upon to analyze censored data. Environmental and toxicological data is often left-censored due to reporting practices for mea- surements that are below a statistically de…ned detection limit. Although there is an abundance of literature on univariate methods for analyzing this type of data, a great need still exists for multivariate methods that take into account possible correlation amongst variables. Two methods are developed here for that purpose. One is a Markov Chain Monte Carlo method that uses a Gibbs sampler to es- timate censored data values as well as distributional and regression parameters. The second is an expectation maximization (EM) algorithm that solves for the distributional parameters that maximize the complete likelihood function in the presence of censored data.
    [Show full text]
  • Chapter 4: Central Tendency and Dispersion
    Central Tendency and Dispersion 4 In this chapter, you can learn • how the values of the cases on a single variable can be summarized using measures of central tendency and measures of dispersion; • how the central tendency can be described using statistics such as the mode, median, and mean; • how the dispersion of scores on a variable can be described using statistics such as a percent distribution, minimum, maximum, range, and standard deviation along with a few others; and • how a variable’s level of measurement determines what measures of central tendency and dispersion to use. Schooling, Politics, and Life After Death Once again, we will use some questions about 1980 GSS young adults as opportunities to explain and demonstrate the statistics introduced in the chapter: • Among the 1980 GSS young adults, are there both believers and nonbelievers in a life after death? Which is the more common view? • On a seven-attribute political party allegiance variable anchored at one end by “strong Democrat” and at the other by “strong Republican,” what was the most Democratic attribute used by any of the 1980 GSS young adults? The most Republican attribute? If we put all 1980 95 96 PART II :: DESCRIPTIVE STATISTICS GSS young adults in order from the strongest Democrat to the strongest Republican, what is the political party affiliation of the person in the middle? • What was the average number of years of schooling completed at the time of the survey by 1980 GSS young adults? Were most of these twentysomethings pretty close to the average on
    [Show full text]
  • Central Tendency, Dispersion, Correlation and Regression Analysis Case Study for MBA Program
    Business Statistic Central Tendency, Dispersion, Correlation and Regression Analysis Case study for MBA program CHUOP Theot Therith 12/29/2010 Business Statistic SOLUTION 1. CENTRAL TENDENCY - The different measures of Central Tendency are: (1). Arithmetic Mean (AM) (2). Median (3). Mode (4). Geometric Mean (GM) (5). Harmonic Mean (HM) - The uses of different measures of Central Tendency are as following: Depends upon three considerations: 1. The concept of a typical value required by the problem. 2. The type of data available. 3. The special characteristics of the averages under consideration. • If it is required to get an average based on all values, the arithmetic mean or geometric mean or harmonic mean should be preferable over median or mode. • In case middle value is wanted, the median is the only choice. • To determine the most common value, mode is the appropriate one. • If the data contain extreme values, the use of arithmetic mean should be avoided. • In case of averaging ratios and percentages, the geometric mean and in case of averaging the rates, the harmonic mean should be preferred. • The frequency distributions in open with open-end classes prohibit the use of arithmetic mean or geometric mean or harmonic mean. Prepared by: CHUOP Theot Therith 1 Business Statistic • If the distribution is bell-shaped and symmetrical or flat-topped one with few extreme values, the arithmetic mean is the best choice, because it is less affected by sampling fluctuations. • If the distribution is sharply peaked, i.e., the data cluster markedly at the middle or if there are abnormally small or large values, the median has smaller sampling fluctuations than the arithmetic mean.
    [Show full text]
  • Incorporating a Geometric Mean Formula Into The
    Calculating the CPI Incorporating a geometric mean formula into the CPI Beginning in January 1999, a new geometric mean formula will replace the current Laspeyres formula in calculating most basic components of the Consumer Price Index; the new formula will better account for the economic substitution behavior of consumers 2 Kenneth V. Dalton, his article describes an important improve- bias” in the CPI. This upward bias was a techni- John S. Greenlees, ment in the calculation of the Consumer cal problem that tied the weight of a CPI sample and TPrice Index (CPI). The Bureau of Labor Sta- item to its expected price change. The flaw was Kenneth J. Stewart tistics plans to use a new geometric mean for- effectively eliminated by changes to the CPI mula for calculating most of the basic compo- sample rotation and substitution procedures and nents of the Consumer Price Index for all Urban to the functional form used to calculate changes Consumers (CPI-U) and the Consumer Price In- in the cost of shelter for homeowners. In 1997, a dex for Urban Wage Earners and Clerical Work- new approach to the measurement of consumer ers (CPI-W). This change will become effective prices for hospital services was introduced.3 Pric- with data for January 1999.1 ing procedures were revised, from pricing indi- The geometric mean formula will be used in vidual items (such as a unit of blood or a hospi- index categories that make up approximately 61 tal inpatient day) to pricing the combined sets of percent of total consumer spending represented goods and services provided on selected patient by the CPI-U.
    [Show full text]
  • Wavelet Operators and Multiplicative Observation Models
    Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series Abdourrahmane Atto, Emmanuel Trouvé, Jean-Marie Nicolas, Thu Trang Le To cite this version: Abdourrahmane Atto, Emmanuel Trouvé, Jean-Marie Nicolas, Thu Trang Le. Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series. 2016. hal-00950823v3 HAL Id: hal-00950823 https://hal.archives-ouvertes.fr/hal-00950823v3 Preprint submitted on 26 Jan 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 1 Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series Abdourrahmane M. Atto1;∗, Emmanuel Trouve1, Jean-Marie Nicolas2, Thu-Trang Le^1 Abstract|This paper first provides statistical prop- I. Introduction - Motivation erties of wavelet operators when the observation model IGHLY resolved data such as Synthetic Aperture can be seen as the product of a deterministic piece- Radar (SAR) image time series issued from new wise regular function (signal) and a stationary random H field (noise). This multiplicative observation model is generation sensors show minute details. Indeed, the evo- analyzed in two standard frameworks by considering lution of SAR imaging systems is such that in less than 2 either (1) a direct wavelet transform of the model decades: or (2) a log-transform of the model prior to wavelet • high resolution sensors can achieve metric resolution, decomposition.
    [Show full text]
  • Measures of Central Tendency for Censored Wage Data
    MEASURES OF CENTRAL TENDENCY FOR CENSORED WAGE DATA Sandra West, Diem-Tran Kratzke, and Shail Butani, Bureau of Labor Statistics Sandra West, 2 Massachusetts Ave. N.E. Washington, D.C. 20212 Key Words: Mean, Median, Pareto Distribution medians can lead to misleading results. This paper tested the linear interpolation method to estimate the median. I. INTRODUCTION First, the interval that contains the median was determined, The research for this paper began in connection with the and then linear interpolation is applied. The occupational need to measure the central tendency of hourly wage data minimum wage is used as the lower limit of the lower open from the Occupational Employment Statistics (OES) survey interval. If the median falls in the uppermost open interval, at the Bureau of Labor Statistics (BLS). The OES survey is all that can be said is that the median is equal to or above a Federal-State establishment survey of wage and salary the upper limit of hourly wage. workers designed to produce data on occupational B. MEANS employment for approximately 700 detailed occupations by A measure of central tendency with desirable properties is industry for the Nation, each State, and selected areas within the mean. In addition to the usual desirable properties of States. It provided employment data without any wage means, there is another nice feature that is proven by West information until recently. Wage data that are produced by (1985). It is that the percent difference between two means other Federal programs are limited in the level of is bounded relative to the percent difference between occupational, industrial, and geographic coverage.
    [Show full text]
  • Redalyc.Robust Analysis of the Central Tendency, Simple and Multiple
    International Journal of Psychological Research ISSN: 2011-2084 [email protected] Universidad de San Buenaventura Colombia Courvoisier, Delphine S.; Renaud, Olivier Robust analysis of the central tendency, simple and multiple regression and ANOVA: A step by step tutorial. International Journal of Psychological Research, vol. 3, núm. 1, 2010, pp. 78-87 Universidad de San Buenaventura Medellín, Colombia Available in: http://www.redalyc.org/articulo.oa?id=299023509005 How to cite Complete issue Scientific Information System More information about this article Network of Scientific Journals from Latin America, the Caribbean, Spain and Portugal Journal's homepage in redalyc.org Non-profit academic project, developed under the open access initiative International Journal of Psychological Research, 2010. Vol. 3. No. 1. Courvoisier, D.S., Renaud, O., (2010). Robust analysis of the central tendency, ISSN impresa (printed) 2011-2084 simple and multiple regression and ANOVA: A step by step tutorial. International ISSN electrónica (electronic) 2011-2079 Journal of Psychological Research, 3 (1), 78-87. Robust analysis of the central tendency, simple and multiple regression and ANOVA: A step by step tutorial. Análisis robusto de la tendencia central, regresión simple y múltiple, y ANOVA: un tutorial paso a paso. Delphine S. Courvoisier and Olivier Renaud University of Geneva ABSTRACT After much exertion and care to run an experiment in social science, the analysis of data should not be ruined by an improper analysis. Often, classical methods, like the mean, the usual simple and multiple linear regressions, and the ANOVA require normality and absence of outliers, which rarely occurs in data coming from experiments. To palliate to this problem, researchers often use some ad-hoc methods like the detection and deletion of outliers.
    [Show full text]
  • “Mean”? a Review of Interpreting and Calculating Different Types of Means and Standard Deviations
    pharmaceutics Review What Does It “Mean”? A Review of Interpreting and Calculating Different Types of Means and Standard Deviations Marilyn N. Martinez 1,* and Mary J. Bartholomew 2 1 Office of New Animal Drug Evaluation, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA 2 Office of Surveillance and Compliance, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA; [email protected] * Correspondence: [email protected]; Tel.: +1-240-3-402-0635 Academic Editors: Arlene McDowell and Neal Davies Received: 17 January 2017; Accepted: 5 April 2017; Published: 13 April 2017 Abstract: Typically, investigations are conducted with the goal of generating inferences about a population (humans or animal). Since it is not feasible to evaluate the entire population, the study is conducted using a randomly selected subset of that population. With the goal of using the results generated from that sample to provide inferences about the true population, it is important to consider the properties of the population distribution and how well they are represented by the sample (the subset of values). Consistent with that study objective, it is necessary to identify and use the most appropriate set of summary statistics to describe the study results. Inherent in that choice is the need to identify the specific question being asked and the assumptions associated with the data analysis. The estimate of a “mean” value is an example of a summary statistic that is sometimes reported without adequate consideration as to its implications or the underlying assumptions associated with the data being evaluated. When ignoring these critical considerations, the method of calculating the variance may be inconsistent with the type of mean being reported.
    [Show full text]
  • Expectation and Functions of Random Variables
    POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior of random variables, we first consider their expectation, which is also called mean value or expected value. The definition of expectation follows our intuition. Definition 1 Let X be a random variable and g be any function. 1. If X is discrete, then the expectation of g(X) is defined as, then X E[g(X)] = g(x)f(x), x∈X where f is the probability mass function of X and X is the support of X. 2. If X is continuous, then the expectation of g(X) is defined as, Z ∞ E[g(X)] = g(x)f(x) dx, −∞ where f is the probability density function of X. If E(X) = −∞ or E(X) = ∞ (i.e., E(|X|) = ∞), then we say the expectation E(X) does not exist. One sometimes write EX to emphasize that the expectation is taken with respect to a particular random variable X. For a continuous random variable, the expectation is sometimes written as, Z x E[g(X)] = g(x) d F (x). −∞ where F (x) is the distribution function of X. The expectation operator has inherits its properties from those of summation and integral. In particular, the following theorem shows that expectation preserves the inequality and is a linear operator. Theorem 1 (Expectation) Let X and Y be random variables with finite expectations. 1. If g(x) ≥ h(x) for all x ∈ R, then E[g(X)] ≥ E[h(X)].
    [Show full text]
  • Notes on Calculating Computer Performance
    Notes on Calculating Computer Performance Bruce Jacob and Trevor Mudge Advanced Computer Architecture Lab EECS Department, University of Michigan {blj,tnm}@umich.edu Abstract This report explains what it means to characterize the performance of a computer, and which methods are appro- priate and inappropriate for the task. The most widely used metric is the performance on the SPEC benchmark suite of programs; currently, the results of running the SPEC benchmark suite are compiled into a single number using the geometric mean. The primary reason for using the geometric mean is that it preserves values across normalization, but unfortunately, it does not preserve total run time, which is probably the figure of greatest interest when performances are being compared. Cycles per Instruction (CPI) is another widely used metric, but this method is invalid, even if comparing machines with identical clock speeds. Comparing CPI values to judge performance falls prey to the same prob- lems as averaging normalized values. In general, normalized values must not be averaged and instead of the geometric mean, either the harmonic or the arithmetic mean is the appropriate method for averaging a set running times. The arithmetic mean should be used to average times, and the harmonic mean should be used to average rates (1/time). A number of published SPECmarks are recomputed using these means to demonstrate the effect of choosing a favorable algorithm. 1.0 Performance and the Use of Means We want to summarize the performance of a computer; the easiest way uses a single number that can be compared against the numbers of other machines.
    [Show full text]
  • Robust Statistical Procedures for Testing the Equality of Central Tendency Parameters Under Skewed Distributions
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repository@USM ROBUST STATISTICAL PROCEDURES FOR TESTING THE EQUALITY OF CENTRAL TENDENCY PARAMETERS UNDER SKEWED DISTRIBUTIONS SHARIPAH SOAAD SYED YAHAYA UNIVERSITI SAINS MALAYSIA 2005 ii ROBUST STATISTICAL PROCEDURES FOR TESTING THE EQUALITY OF CENTRAL TENDENCY PARAMETERS UNDER SKEWED DISTRIBUTIONS by SHARIPAH SOAAD SYED YAHAYA Thesis submitted in fulfillment of the requirements for the degree of Doctor of Philosophy May 2005 ACKNOWLEDGEMENTS Firstly, my sincere appreciations to my supervisor Associate Professor Dr. Abdul Rahman Othman without whose guidance, support, patience, and encouragement, this study could not have materialized. I am indeed deeply indebted to him. My sincere thanks also to my co-supervisor, Associate Professor Dr. Sulaiman for his encouragement and support throughout this study. I would also like to thank Universiti Utara Malaysia (UUM) for sponsoring my study at Universiti Sains Malaysia. Thanks are due to Professor H.J. Keselman for according me access to his UNIX server to run the relevant programs; Professor R.R. Wilcox and Professor R.G. Staudte for promptly responding to my queries via emails; Pn. Azizah Ahmad and Pn. Ramlah Din for patiently helping me with some of the computer programs and En. Sofwan for proof reading the chapters of this thesis. I am deeply grateful to my parents and my wonderful daughter, Nur Azzalia Kamaruzaman for their inspiration, patience, enthusiasm and effort. I would also like to thank my brothers for their constant support. Finally, my grateful recognition are due to (in alphabetical order) Azilah and Nung, Bidin and Kak Syera, Izan, Linda, Min, Shikin, Yen, Zurni and to those who had directly or indirectly lend me their friendship, moral support and endless encouragement during my study.
    [Show full text]