The History of Histograms (Abridged)

Total Page:16

File Type:pdf, Size:1020Kb

The History of Histograms (Abridged) The History of Histograms (abridged) Yannis Ioannidis Department of Informatics and Telecommunications, University of Athens Panepistimioupolis, Informatics Buildings 157-84, Athens, Hellas (Greece) [email protected] word that was originally used in the Greek language1. The term `histogram' was coined by the famous statis- tician Karl Pearson2 to refer to a \common form of Abstract graphical representation". In the Oxford English Dic- tionary quotes from \Philosophical Transactions of the The history of histograms is long and rich, full Royal Society of London" Series A, Vol. CLXXXVI, of detailed information in every step. It in- (1895) p. 399, it is mentioned that \[The word `his- cludes the course of histograms in different togram' was] introduced by the writer in his lectures scientific fields, the successes and failures of on statistics as a term for a common form of graphical histograms in approximating and compressing representation, i.e., by columns marking as areas the information, their adoption by industry, and frequency corresponding to the range of their base.". solutions that have been given on a great va- Stigler identifies the lectures as the 1892 lectures on riety of histogram-related problems. In this the geometry of statistics [69]. paper and in the same spirit of the histogram The above quote suggests that histograms were used techniques themselves, we compress their en- long before they received their name, but their birth tire history (including their \future history" date is unclear. Bar charts (i.e., histograms with an as currently anticipated) in the given/fixed individual `base' element associated with each column) space budget, mostly recording details for the most likely predate histograms and this helps us put periods, events, and results with the highest a lower bound on the timing of their first appearance. (personally-biased) interest. In a limited set The oldest known bar chart appeared in a book by the of experiments, the semantic distance between Scottish political economist William Playfair3 titled the compressed and the full form of the history \The Commercial and Political Atlas (London 1786)" was found relatively small! and shows the imports and exports of Scotland to and from seventeen countries in 1781 [74]. Although Play- fair was skeptical of the usefulness of his invention, it was adopted by many in the following years, includ- 1 Prehistory ing for example, Florence Nightingale, who used them in 1859 to compare mortality in the peacetime army The word `histogram' is of Greek origin, as it is a com- to that of civilians and through those convinced the posite of the words `isto-s' (ιστos) (= `mast', also government to improve army hygiene. means `web' but this is not relevant to this discus- sion) and `gram-ma' (γραµµα) (= `something writ- From all the above, it is clear that histograms were ten'). Hence, it should be interpreted as a form of first conceived as a visual aid to statistical approxima- writing consisting of `masts', i.e., long shapes vertically tions. Even today this point is still emphasized in the standing, or something similar. It is not, however, a common conception of histograms: Webster's defines a 1To the contrary, the word `history' is indeed part of the Permission to copy without fee all or part of this material is Greek language (`istoria' - ιστoρια) and in use since the ancient granted provided that the copies are not made or distributed for times. Despite its similarity to `histogram', however, it appears direct commercial advantage, the VLDB copyright notice and to have a different etymology, one that is related to the original the title of the publication and its date appear, and notice is meaning of the word, which was `knowledge'. given that copying is by permission of the Very Large Data Base 2His claim to fame includes, among others, the chi-square test Endowment. To copy otherwise, or to republish, requires a fee for statistical significance and the term `standard deviation'. and/or special permission from the Endowment. 3In addition to the bar chart, Playfair is probably the fa- Proceedings of the 29th VLDB Conference, ther of the pie chart and other extremely intuitive and useful Berlin, Germany, 2003 visualizations that we use today. histogram as \a bar graph of a frequency distribution The data distribution of Xi is the set of pairs Ti = in which the widths of the bars are proportional to the f (vi(1); fi(1)); (vi(2); fi(2)); : : : ; (vi(Di); fi(Di)) g. classes into which the variable has been divided and The joint frequency f(k1; ::; kn) of the value combi- the heights of the bars are proportional to the class nation < v1(k1); ::; vn(kn) > is the number of tuples frequencies". Histograms, however, are extremely use- in R that contain vi(ki) in attribute Xi, for all i. The ful even when disassociated from their canonical visual joint data distribution T1;::;n of X1; ::; Xn is the entire representation and treated as purely mathematical ob- set of (value combination, joint frequency) pairs. jects capturing data distribution approximations. This In the sequel, for 1-dimensional cases, we use the is precisely how we approach them in this paper. above symbols without the subscript i. In the past few decades, histograms have been used in several fields of informatics. Besides databases, his- 2.2 Motivation for Histograms tograms have played a very important role primarily in image processing and computer vision. Given an Data distributions are very useful in database systems image (or a video) and a visual pixel parameter, a his- but are usually too large to be stored accurately, so togram captures for each possible value of the param- histograms come into play as an approximation mech- eter (Webster's \classes") the number of pixels that anism. The two most important applications of his- have this value (Webster's \frequencies"). Such a his- togram techniques in databases have been selectivity togram is a summary that is characteristic of the image estimation and approximate query answering within and can be very useful in several tasks: identifying sim- query optimization (for the former) or pre-execution ilar images, compressing the image, and others. Color user-level query feedback (for both). Our discussion histograms are the most common in the literature, e.g., below focuses exactly on these two, especially range- in the QBIC system [21], but several other parameters query selectivity estimation as this is the most pop- have been proposed as well, e.g., edge density, tex- ular issue in the literature. It should not be forgot- turedness, intensity gradient, etc. [61]. In general, his- ten, however, that histograms have proved to be useful tograms used in image processing and computer vision in the context of several other database problems as are accurate. For example, a color histogram contains well, e.g., load-balancing in parallel join query execu- a separate and precise count of pixels for each possi- tion [65], partition-based temporal join execution [68] ble distinct color in the image. The only element of and others. approximation might be in the number of bits used to represent different colors: fewer bits imply that several 2.3 Histograms actual colors are represented by one, which will be as- A histogram on an attribute X is constructed by parti- sociated with the number of pixels that have any of tioning the data distribution of X into β (≥ 1) mutu- the colors that are grouped together. Even this kind ally disjoint subsets called buckets and approximating of approximation is not common, however. the frequencies and values in each bucket in some com- In databases, histograms are used as a mechanism mon fashion. This definition leaves several degrees of for full-fledged compression and approximation of data freedom in designing specific histogram classes as there distributions. They first appeared in the literature and are several possible choices for each of the following in systems in the 1980's and have been studied exten- (mostly orthogonal) aspects of histograms [67]: sively since then at a continuously increasing rate. In Partition Rule: This is further analyzed into the this paper, we concentrate on the database notion of following characteristics: histograms, discuss the most important developments • Partition Class: This indicates if there are any on the topic so far, and outline several problems that restrictions on the buckets. Of great importance we believe are interesting and whose solution may fur- is the serial class, which requires that buckets are ther expand their applicability and usefulness. non-overlapping with respect to some parameter (the next characteristic), and its subclass end- 2 Histogram Definitions biased, which requires at most one non-singleton bucket. 2.1 Data Distributions • Sort Parameter: This is a parameter whose Consider a relation R with n numeric attributes Xi value for each element in the data distribution (i = 1::n). The value set Vi of attribute Xi is is derived from the corresponding attribute value the set of values of Xi that are present in R. Let and frequencies. All serial histograms require that Vi = f vi(k): 1 ≤ k ≤ Di g, where vi(k) < vi(j) when the sort parameter values in each bucket form a k < j. The spread si(k) of vi(k) is defined as contiguous range. Attribute value (V), frequency si(k) = vi(k + 1) − vi(k), for 1 ≤ k < Di. (We (F), and area (A) are examples of sort parameters take si(Di) = 1.) The frequency fi(k) of vi(k) is that have been discussed in the literature. the number of tuples in R with Xi = vi(k). The • Source Parameter: This captures the property area ai(k) of vi(k) is defined as ai(k) = fi(k) × si(k).
Recommended publications
  • Programming in Stata
    Programming and Post-Estimation • Bootstrapping • Monte Carlo • Post-Estimation Simulation (Clarify) • Extending Clarify to Other Models – Censored Probit Example What is Bootstrapping? • A computer-simulated nonparametric technique for making inferences about a population parameter based on sample statistics. • If the sample is a good approximation of the population, the sampling distribution of interest can be estimated by generating a large number of new samples from the original. • Useful when no analytic formula for the sampling distribution is available. B 2 ˆ B B b How do I do it? SE B = s(x )- s(x )/ B /(B -1) å[ åb =1 ] b=1 1. Obtain a Sample from the population of interest. Call this x = (x1, x2, … , xn). 2. Re-sample based on x by randomly sampling with replacement from it. 3. Generate many such samples, x1, x2, …, xB – each of length n. 4. Estimate the desired parameter in each sample, s(x1), s(x2), …, s(xB). 5. For instance the bootstrap estimate of the standard error is the standard deviation of the bootstrap replications. Example: Standard Error of a Sample Mean Canned in Stata use "I:\general\PRISM Programming\auto.dta", clear Type: sum mpg Example: Standard Error of a Sample Mean Canned in Stata Type: bootstrap "summarize mpg" r(mean), reps(1000) » 21.2973±1.96*.6790806 x B - x Example: Difference of Medians Test Type: sum length, detail Type: return list Example: Difference of Medians Test Example: Difference of Medians Test Example: Difference of Medians Test Type: bootstrap "mymedian" r(diff), reps(1000) The medians are not very different.
    [Show full text]
  • The Cascade Bayesian Approach: Prior Transformation for a Controlled Integration of Internal Data, External Data and Scenarios
    risks Article The Cascade Bayesian Approach: Prior Transformation for a Controlled Integration of Internal Data, External Data and Scenarios Bertrand K. Hassani 1,2,3,* and Alexis Renaudin 4 1 University College London Computer Science, 66-72 Gower Street, London WC1E 6EA, UK 2 LabEx ReFi, Université Paris 1 Panthéon-Sorbonne, CESCES, 106 bd de l’Hôpital, 75013 Paris, France 3 Capegemini Consulting, Tour Europlaza, 92400 Paris-La Défense, France 4 Aon Benfield, The Aon Centre, 122 Leadenhall Street, London EC3V 4AN, UK; [email protected] * Correspondence: [email protected]; Tel.: +44-(0)753-016-7299 Received: 14 March 2018; Accepted: 23 April 2018; Published: 27 April 2018 Abstract: According to the last proposals of the Basel Committee on Banking Supervision, banks or insurance companies under the advanced measurement approach (AMA) must use four different sources of information to assess their operational risk capital requirement. The fourth includes ’business environment and internal control factors’, i.e., qualitative criteria, whereas the three main quantitative sources available to banks for building the loss distribution are internal loss data, external loss data and scenario analysis. This paper proposes an innovative methodology to bring together these three different sources in the loss distribution approach (LDA) framework through a Bayesian strategy. The integration of the different elements is performed in two different steps to ensure an internal data-driven model is obtained. In the first step, scenarios are used to inform the prior distributions and external data inform the likelihood component of the posterior function. In the second step, the initial posterior function is used as the prior distribution and the internal loss data inform the likelihood component of the second posterior function.
    [Show full text]
  • What Is Bayesian Inference? Bayesian Inference Is at the Core of the Bayesian Approach, Which Is an Approach That Allows Us to Represent Uncertainty As a Probability
    Learn to Use Bayesian Inference in SPSS With Data From the National Child Measurement Programme (2016–2017) © 2019 SAGE Publications Ltd. All Rights Reserved. This PDF has been generated from SAGE Research Methods Datasets. SAGE SAGE Research Methods Datasets Part 2019 SAGE Publications, Ltd. All Rights Reserved. 2 Learn to Use Bayesian Inference in SPSS With Data From the National Child Measurement Programme (2016–2017) Student Guide Introduction This example dataset introduces Bayesian Inference. Bayesian statistics (the general name for all Bayesian-related topics, including inference) has become increasingly popular in recent years, due predominantly to the growth of evermore powerful and sophisticated statistical software. However, Bayesian statistics grew from the ideas of an English mathematician, Thomas Bayes, who lived and worked in the first half of the 18th century and have been refined and adapted by statisticians and mathematicians ever since. Despite its longevity, the Bayesian approach did not become mainstream: the Frequentist approach was and remains the dominant means to conduct statistical analysis. However, there is a renewed interest in Bayesian statistics, part prompted by software development and part by a growing critique of the limitations of the null hypothesis significance testing which dominates the Frequentist approach. This renewed interest can be seen in the incorporation of Bayesian analysis into mainstream statistical software, such as, IBM® SPSS® and in many major statistics text books. Bayesian Inference is at the heart of Bayesian statistics and is different from Frequentist approaches due to how it views probability. In the Frequentist approach, probability is the product of the frequency of random events occurring Page 2 of 19 Learn to Use Bayesian Inference in SPSS With Data From the National Child Measurement Programme (2016–2017) SAGE SAGE Research Methods Datasets Part 2019 SAGE Publications, Ltd.
    [Show full text]
  • Basic Econometrics / Statistics Statistical Distributions: Normal, T, Chi-Sq, & F
    Basic Econometrics / Statistics Statistical Distributions: Normal, T, Chi-Sq, & F Course : Basic Econometrics : HC43 / Statistics B.A. Hons Economics, Semester IV/ Semester III Delhi University Course Instructor: Siddharth Rathore Assistant Professor Economics Department, Gargi College Siddharth Rathore guj75845_appC.qxd 4/16/09 12:41 PM Page 461 APPENDIX C SOME IMPORTANT PROBABILITY DISTRIBUTIONS In Appendix B we noted that a random variable (r.v.) can be described by a few characteristics, or moments, of its probability function (PDF or PMF), such as the expected value and variance. This, however, presumes that we know the PDF of that r.v., which is a tall order since there are all kinds of random variables. In practice, however, some random variables occur so frequently that statisticians have determined their PDFs and documented their properties. For our purpose, we will consider only those PDFs that are of direct interest to us. But keep in mind that there are several other PDFs that statisticians have studied which can be found in any standard statistics textbook. In this appendix we will discuss the following four probability distributions: 1. The normal distribution 2. The t distribution 3. The chi-square (␹2 ) distribution 4. The F distribution These probability distributions are important in their own right, but for our purposes they are especially important because they help us to find out the probability distributions of estimators (or statistics), such as the sample mean and sample variance. Recall that estimators are random variables. Equipped with that knowledge, we will be able to draw inferences about their true population values.
    [Show full text]
  • Practical Meta-Analysis -- Lipsey & Wilson Overview
    Practical Meta-Analysis -- Lipsey & Wilson The Great Debate • 1952: Hans J. Eysenck concluded that there were no favorable Practical Meta-Analysis effects of psychotherapy, starting a raging debate • 20 years of evaluation research and hundreds of studies failed David B. Wilson to resolve the debate American Evaluation Association • 1978: To proved Eysenck wrong, Gene V. Glass statistically Orlando, Florida, October 3, 1999 aggregate the findings of 375 psychotherapy outcome studies • Glass (and colleague Smith) concluded that psychotherapy did indeed work • Glass called his method “meta-analysis” 1 2 The Emergence of Meta-Analysis The Logic of Meta-Analysis • Ideas behind meta-analysis predate Glass’ work by several • Traditional methods of review focus on statistical significance decades testing – R. A. Fisher (1944) • Significance testing is not well suited to this task • “When a number of quite independent tests of significance have been made, it sometimes happens that although few or none can be – highly dependent on sample size claimed individually as significant, yet the aggregate gives an – null finding does not carry to same “weight” as a significant impression that the probabilities are on the whole lower than would often have been obtained by chance” (p. 99). finding • Source of the idea of cumulating probability values • Meta-analysis changes the focus to the direction and – W. G. Cochran (1953) magnitude of the effects across studies • Discusses a method of averaging means across independent studies – Isn’t this what we are
    [Show full text]
  • Statistical Power in Meta-Analysis Jin Liu University of South Carolina - Columbia
    University of South Carolina Scholar Commons Theses and Dissertations 12-14-2015 Statistical Power in Meta-Analysis Jin Liu University of South Carolina - Columbia Follow this and additional works at: https://scholarcommons.sc.edu/etd Part of the Educational Psychology Commons Recommended Citation Liu, J.(2015). Statistical Power in Meta-Analysis. (Doctoral dissertation). Retrieved from https://scholarcommons.sc.edu/etd/3221 This Open Access Dissertation is brought to you by Scholar Commons. It has been accepted for inclusion in Theses and Dissertations by an authorized administrator of Scholar Commons. For more information, please contact [email protected]. STATISTICAL POWER IN META-ANALYSIS by Jin Liu Bachelor of Arts Chongqing University of Arts and Sciences, 2009 Master of Education University of South Carolina, 2012 Submitted in Partial Fulfillment of the Requirements For the Degree of Doctor of Philosophy in Educational Psychology and Research College of Education University of South Carolina 2015 Accepted by: Xiaofeng Liu, Major Professor Christine DiStefano, Committee Member Robert Johnson, Committee Member Brian Habing, Committee Member Lacy Ford, Senior Vice Provost and Dean of Graduate Studies © Copyright by Jin Liu, 2015 All Rights Reserved. ii ACKNOWLEDGEMENTS I would like to express my thanks to all committee members who continuously supported me in the dissertation work. My dissertation chair, Xiaofeng Liu, always inspired and trusted me during this study. I could not finish my dissertation at this time without his support. Brian Habing helped me with R coding and research design. I could not finish the coding process so quickly without his instruction. I appreciate the encouragement from Christine DiStefano and Robert Johnson.
    [Show full text]
  • Descriptive Statistics Frequency Distributions and Their Graphs
    Chapter 2 Descriptive Statistics § 2.1 Frequency Distributions and Their Graphs Frequency Distributions A frequency distribution is a table that shows classes or intervals of data with a count of the number in each class. The frequency f of a class is the number of data points in the class. Class Frequency, f 1 – 4 4 LowerUpper Class 5 – 8 5 Limits 9 – 12 3 Frequencies 13 – 16 4 17 – 20 2 Larson & Farber, Elementary Statistics: Picturing the World , 3e 3 1 Frequency Distributions The class width is the distance between lower (or upper) limits of consecutive classes. Class Frequency, f 1 – 4 4 5 – 1 = 4 5 – 8 5 9 – 5 = 4 9 – 12 3 13 – 9 = 4 13 – 16 4 17 – 13 = 4 17 – 20 2 The class width is 4. The range is the difference between the maximum and minimum data entries. Larson & Farber, Elementary Statistics: Picturing the World , 3e 4 Constructing a Frequency Distribution Guidelines 1. Decide on the number of classes to include. The number of classes should be between 5 and 20; otherwise, it may be difficult to detect any patterns. 2. Find the class width as follows. Determine the range of the data, divide the range by the number of classes, and round up to the next convenient number. 3. Find the class limits. You can use the minimum entry as the lower limit of the first class. To find the remaining lower limits, add the class width to the lower limit of the preceding class. Then find the upper class limits. 4.
    [Show full text]
  • Flood Frequency Analysis Using Copula with Mixed Marginal Distributions
    Flood frequency analysis using copula with mixed marginal distributions Project Report, Aug. 2007 Project Report August 2007 - 1 - Flood frequency analysis using copula with mixed marginal distributions Project Report, Aug. 2007 Prepared by and - 2 - Flood frequency analysis using copula with mixed marginal distributions Project Report, Aug. 2007 Abstract...................................................................................................................7 I. Introduction......................................................................................................8 II. Nonparametric method of estimating marginal distribution.................................14 II.1 Univariate kernel density estimation...........................................................14 II.2 Orthonormal series method.......................................................................17 III. The copula function.........................................................................................19 IV. Multivariate flood frequency analysis................................................................23 IV.1 Details of the case study............................................................................25 IV.2 Associations among flood characteristics.....................................................26 IV.3 Marginal distributions of flood characteristics..............................................29 IV.3.1 Parametric estimation.........................................................................30 IV.3.2 Nonparametric estimation...................................................................32
    [Show full text]
  • Formulas Used by the “Practical Meta-Analysis Effect Size Calculator”
    Formulas Used by the “Practical Meta-Analysis Effect Size Calculator” David B. Wilson George Mason University April 27, 2017 1 Introduction This document provides the equations used by the on-line Practical Meta- Analysis Effect Size Calculator, available at: http://www.campbellcollaboration.org/resources/effect_size _input.php and http://cebcp.org/practical-meta-analysis-effect-size-calculator The calculator is a companion to the book I co-authored with Mark Lipsey, ti- tled “Practical Meta-Analysis,” published by Sage in 2001 (ISBN-10: 9780761921684). This calculator computes four effect size types: 1. Standardized mean difference (d) 2. Correlation coefficient (r) 3. Odds ratio (OR) 4. Risk ratio (RR) In addition to computing effect sizes, the calculator computes the associated variance and 95% confidence interval. The variance can be used to generate the w = 1=v weight for inverse-variancep weighted meta-analysis ( ) or the standard error (se = v). 2 Notation Notation commonly used through this document is presented below. Other no- tation will be defined as needed. 1 Formulas for “Practical Meta–Analysis Online Calculator” d Standardized mean difference effect size n1 Sample size for group 1 n2 Sample size for group 2 N Total sample size s1 Standard deviation for group 1 s2 Standard deviation for group 2 s Full sample standard deviation se1 Standard error for the mean of group 1 se2 Standard error for the mean of group 2 sgain Gain score standard deviation X1 Mean for group 1 X2 Mean for group 2 Xij Mean for subgroup i of group j ∆1
    [Show full text]
  • Categorical Data Analysis
    Categorical Data Analysis Related topics/headings: Categorical data analysis; or, Nonparametric statistics; or, chi-square tests for the analysis of categorical data. OVERVIEW For our hypothesis testing so far, we have been using parametric statistical methods. Parametric methods (1) assume some knowledge about the characteristics of the parent population (e.g. normality) (2) require measurement equivalent to at least an interval scale (calculating a mean or a variance makes no sense otherwise). Frequently, however, there are research problems in which one wants to make direct inferences about two or more distributions, either by asking if a population distribution has some particular specifiable form, or by asking if two or more population distributions are identical. These questions occur most often when variables are qualitative in nature, making it impossible to carry out the usual inferences in terms of means or variances. For such problems, we use nonparametric methods. Nonparametric methods (1) do not depend on any assumptions about the parameters of the parent population (2) generally assume data are only measured at the nominal or ordinal level. There are two common types of hypothesis-testing problems that are addressed with nonparametric methods: (1) How well does a sample distribution correspond with a hypothetical population distribution? As you might guess, the best evidence one has about a population distribution is the sample distribution. The greater the discrepancy between the sample and theoretical distributions, the more we question the “goodness” of the theory. EX: Suppose we wanted to see whether the distribution of educational achievement had changed over the last 25 years. We might take as our null hypothesis that the distribution of educational achievement had not changed, and see how well our modern-day sample supported that theory.
    [Show full text]
  • 2. Graphing Distributions
    2. Graphing Distributions A. Qualitative Variables B. Quantitative Variables 1. Stem and Leaf Displays 2. Histograms 3. Frequency Polygons 4. Box Plots 5. Bar Charts 6. Line Graphs 7. Dot Plots C. Exercises Graphing data is the first and often most important step in data analysis. In this day of computers, researchers all too often see only the results of complex computer analyses without ever taking a close look at the data themselves. This is all the more unfortunate because computers can create many types of graphs quickly and easily. This chapter covers some classic types of graphs such bar charts that were invented by William Playfair in the 18th century as well as graphs such as box plots invented by John Tukey in the 20th century. 65 by David M. Lane Prerequisites • Chapter 1: Variables Learning Objectives 1. Create a frequency table 2. Determine when pie charts are valuable and when they are not 3. Create and interpret bar charts 4. Identify common graphical mistakes When Apple Computer introduced the iMac computer in August 1998, the company wanted to learn whether the iMac was expanding Apple’s market share. Was the iMac just attracting previous Macintosh owners? Or was it purchased by newcomers to the computer market and by previous Windows users who were switching over? To find out, 500 iMac customers were interviewed. Each customer was categorized as a previous Macintosh owner, a previous Windows owner, or a new computer purchaser. This section examines graphical methods for displaying the results of the interviews. We’ll learn some general lessons about how to graph data that fall into a small number of categories.
    [Show full text]
  • Chapter 16.Tst
    Chapter 16 Statistical Quality Control- Prof. Dr. Samir Safi TRUE/FALSE. Write 'T' if the statement is true and 'F' if the statement is false. 1) Statistical process control uses regression and other forecasting tools to help control processes. 1) 2) It is impossible to develop a process that has zero variability. 2) 3) If all of the control points on a control chart lie between the UCL and the LCL, the process is always 3) in control. 4) An x-bar chart would be appropriate to monitor the number of defects in a production lot. 4) 5) The central limit theorem provides the statistical foundation for control charts. 5) 6) If we are tracking quality of performance for a class of students, we should plot the individual 6) grades on an x-bar chart, and the pass/fail result on a p-chart. 7) A p-chart could be used to monitor the average weight of cereal boxes. 7) 8) If we are attempting to control the diameter of bowling bowls, we will find a p-chart to be quite 8) helpful. 9) A c-chart would be appropriate to monitor the number of weld defects on the steel plates of a 9) ship's hull. MULTIPLE CHOICE. Choose the one alternative that best completes the statement or answers the question. 1) Which of the following is not a popular definition of quality? 1) A) Quality is fitness for use. B) Quality is the totality of features and characteristics of a product or service that bears on its ability to satisfy stated or implied needs.
    [Show full text]