University of Cincinnati

Total Page:16

File Type:pdf, Size:1020Kb

University of Cincinnati UNIVERSITY OF CINCINNATI Date:___________________ I, _________________________________________________________, hereby submit this work as part of the requirements for the degree of: in: It is entitled: This work and its defense approved by: Chair: _______________________________ _______________________________ _______________________________ _______________________________ _______________________________ Gibbs Sampling and Expectation Maximization Methods for Estimation of Censored Values from Correlated Multivariate Distributions A dissertation submitted to the Division of Research and Advanced Studies of the University of Cincinnati in partial ful…llment of the requirements for the degree of DOCTORATE OF PHILOSOPHY (Ph.D.) in the Department of Mathematical Sciences of the McMicken College of Arts and Sciences May 2008 by Tina D. Hunter B.S. Industrial and Systems Engineering The Ohio State University, Columbus, Ohio, 1984 M.S. Aerospace Engineering University of Cincinnati, Cincinnati, Ohio, 1989 M.S. Statistics University of Cincinnati, Cincinnati, Ohio, 2003 Committee Chair: Dr. Siva Sivaganesan Abstract Statisticians are often called upon to analyze censored data. Environmental and toxicological data is often left-censored due to reporting practices for mea- surements that are below a statistically de…ned detection limit. Although there is an abundance of literature on univariate methods for analyzing this type of data, a great need still exists for multivariate methods that take into account possible correlation amongst variables. Two methods are developed here for that purpose. One is a Markov Chain Monte Carlo method that uses a Gibbs sampler to es- timate censored data values as well as distributional and regression parameters. The second is an expectation maximization (EM) algorithm that solves for the distributional parameters that maximize the complete likelihood function in the presence of censored data. Both methods are applied to bivariate normal data and compared to each other and to two commonly used simple substitution methods with respect to bias and mean squared error of the resulting parameter estimates. The EM method is the most consistent for estimating all distributional and regres- sion parameters across all levels of correlation and proportions of censoring. Both methods provide substantially better estimates of the correlation coe¢ cient than the univariate methods. iii iv To Bryce and Callie, with all my love. May you know the satisfaction and rewards of hard work and perseverance. v Acknowledgements I would like to thank Dr. Marc Mills and Dr. Bryan Boulanger, my mentors at the U.S. Environmental Protection Agency, for all of their time and support during my traineeship, without which this dissertation would not exist. They have been a pleasure to work with and have provided invaluable encouragement and feedback. I would also like to thank my entire committee, and especially my advisor, Dr. Siva Sivaganesan, for taking the time to review my ideas and writing, for helpful and constructive feedback, and for allowing me the freedom to follow my own ideas and interests. I am grateful for all of the friends who kept me sane along the entire road to this Ph.D., especially those traveling it with me. I have many fond memories of parties, travels, and long talks, all of which eased the stresses along the way. I am also grateful to the many caring professors in the Department of Mathematical Sciences. In particular, Don French has been a great friend as well as a talented teacher and generous tutor. Jim Deddens has always been especially encouraging and willing to help. Terri, Anita, Nancy, and Patti have all been very kind and helpful with a wide variety of administrative technicalities. Finally, I would like to express a very special thank you to my family. My parents have been wonderful cheerleaders and have helped me out in countless ways on the way to this Ph.D. Bryce and Callie have not only put up with me, but have even turned into amazing young adults along the way. This research was funded by the U.S. Environmental Protection Agency, Of- …ce of Research and Development, National Exposure Research Laboratory and National Risk Management Research Laboratory, through a Graduate Research Training Grant. vi Contents List of Tables x List of Figures xi Chapter 1. Introduction 1 1.1. General Problem Statement 1 1.2. The Speci…c Analysis that Inspired this Research 7 1.3. Simpli…ed Problem 12 Chapter 2. Literature Review and Analysis 14 2.1. Overview of Methods for Censored Data 14 2.2. Dropping Censored Values 15 2.3. Replacing Censored Values with Constants 17 2.4. Maximum Likelihood Methods 22 2.5. Regression Methods. 29 2.6. Other Methods 30 2.7. EPA Guidelines 31 2.8. Extensions to Multivariate Distributions 33 2.9. Summary 36 vii Chapter 3. Gibbs Sampling Theory and Method Development 39 3.1. Bayesian Inference 39 3.2. Markov Chain Monte Carlo 45 3.3. Gibbs Sampler Algorithm 48 Chapter 4. Expectation Maximization Theory and Method Development 57 4.1. General Theory of Expectation Maximization 58 4.2. Expectation Maximization for Multivariate Normal Data with Left Censoring 59 Chapter 5. Implementation and Analysis of Methods 71 5.1. Assumptions and Data Transformation 71 5.2. Evaluation of Methods 73 5.3. R Program 75 Chapter 6. Results 77 6.1. Mean, Variance, and Correlation Coe¢ cient in Normal Scale 78 6.2. Regression Parameters 101 6.3. Comparisons in Original Lognormal Scale 106 6.4. Results for Application of Methods to Lower Variance Distribution 107 6.5. Results for Larger Sample Sizes 112 6.6. Conclusions 115 Chapter 7. Discussion and Recommendations 119 viii 7.1. MCMC Method 119 7.2. EM Method 122 7.3. Future Extensions of MCMC and EM Methods 127 References 129 Appendix . R Code 133 ix List of Tables 1.1 Hormones Analyzed in E• uent Survey 9 1.2 Sample Correlation Coe¢ cients Above .5 10 1.3 Proportions of 46 Values Reported as <MDL 11 2.1 Comparison of Two Simple Substitution Methods for a Lognormal Distribution such that Ln(x) is Distributed N(0,1) 23 2.2 Comparison of Two Simple Substitution Methods for a Lognormal Distribution such that Ln(x) is Distributed N(0,1/2) 23 2.3 Comparison of Two Simple Substitution Methods for a Lognormal Distribution such that Ln(x) is Distributed N(0,2) 23 6.1 Con…dence Intervals for Bias Plots of Mean Estimates 79 6.2 E¤ects of Bias on Lognormal Parameters 107 7.1 Bias Introduced by Estimating Individual Variance Components Incorrectly125 7.2 Bias Reduction with Multiple Imputations 126 x List of Figures 2.1 Comparison of Tails for 10% Nondetects 21 2.2 Comparison of Tails for 30% Nondetects 22 6.1 Biases in Mean Estimates when = 0:1 80 6.2 MSEs of Mean Estimates when = 0:1 80 6.3 Biases in Mean Estimates when = 0:5 81 6.4 MSEs of Mean Estimates when = 0:5 81 6.5 Biases in Mean Estimates when = 0:9 82 6.6 MSEs of Mean Estimates when = 0:9 82 6.7 Biases in Mean Estimates when = 0:1 83 6.8 MSEs of Mean Estimates when = 0:1 83 6.9 Biases in Mean Estimates when = 0:5 84 6.10MSEs of Mean Estimates when = 0:5 84 6.11Biases in Mean Estimates when = 0:9 85 6.12MSEs of Mean Estimates when = 0:9 85 6.13Biases in Standard Deviation Estimates when = 0:1 88 xi 6.14MSEs of Standard Deviation Estimates when = 0:1 88 6.15Biases in Standard Deviation Estimates when = 0:5 89 6.16MSEs of Standard Deviation Estimates when = 0:5 89 6.17Biases in Standard Deviation Estimates when = 0:9 90 6.18MSEs of Standard Deviation Estimates when = 0:9 90 6.19Biases in Standard Deviation Estimates when = 0:1 91 6.20MSEs of Standard Deviation Estimates when = 0:1 91 6.21Biases in Standard Deviation Estimates when = 0:5 92 6.22MSEs of Standard Deviation Estimates when = 0:5 92 6.23Biases in Standard Deviation Estimates when = 0:9 93 6.24MSEs of Standard Deviation Estimates when = 0:9 93 6.25Biases in Estimates of when = 0:1 95 6.26MSEs of Estimates of when = 0:1 95 6.27Biases in Estimates of when = 0:5 96 6.28MSEs of Estimates of when = 0:5 96 6.29Biases in Estimates of when = 0:9 97 6.30MSEs of Estimates of when = 0:9 97 6.31Biases in Estimates of when = 0:1 98 6.32MSEs of Estimates of when = 0:1 98 xii 6.33Biases in Estimates of when = 0:5 99 6.34MSEs of Estimates of when = 0:5 99 6.35Biases in Estimates of when = 0:9 100 6.36MSEs of Estimates of when = 0:9 100 6.37Biases in Estimates of Intercept 0 when = 0:5 103 6.38MSEs of Estimates of Intercept 0 when = 0:5 103 6.39Biases in Estimates of Slope 1 when = 0:5 104 6.40MSEs of Estimates of Slope 1 when = 0:5 104 6.41Biases in Estimates of Slope 2 when = 0:5 105 6.42MSEs of Estimates of Slope 2 when = 0:5 105 6.43Biases in Estimates of Lognormal Mean when = 0:5 108 6.44MSEs of Estimates of Lognormal Mean when = 0:5 108 6.45Biases in Estimates of Lognormal Mean when = 0:5 109 6.46MSEs of Estimates of Lognormal Mean when = 0:5 109 6.47Biases in Estimates of Lognormal Variance when = 0:5 110 6.48MSEs of Estimates of Lognormal Variance when = 0:5 110 6.49Biases in Estimates of Lognormal Variance when = 0:5 111 6.50MSEs of Estimates of Lognormal Variance when = 0:5 111 6.51Biases in Estimates of Mean Mu when = 0:5 and Variances are 2 = 0:25113 xiii 6.52MSEs of Estimates of Mean Mu when = 0:5 and Variances are 2 = 0:25113 6.53Biases in Standard Deviation Estimates when = 0:5 and Variances are 2 = 0:25 114 6.54MSEs of Standard Deviation Estimates when = 0:5 and Variances are 2 = 0:25 114 6.55Biases in Mean Estimates for n = 100 when = 0:5 116 6.56MSEs of Mean Estimates for n = 100 when = 0:5 116 6.57Biases in Estimates of for n = 100 when = 0:5 117 6.58MSEs of Estimates of for n = 100 when = 0:5 117 7.1 Autocorrelation vs.
Recommended publications
  • Comparison of Harmonic, Geometric and Arithmetic Means for Change Detection in SAR Time Series Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas
    Comparison of Harmonic, Geometric and Arithmetic means for change detection in SAR time series Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas To cite this version: Guillaume Quin, Béatrice Pinel-Puysségur, Jean-Marie Nicolas. Comparison of Harmonic, Geometric and Arithmetic means for change detection in SAR time series. EUSAR. 9th European Conference on Synthetic Aperture Radar, 2012., Apr 2012, Germany. hal-00737524 HAL Id: hal-00737524 https://hal.archives-ouvertes.fr/hal-00737524 Submitted on 2 Oct 2012 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. EUSAR 2012 Comparison of Harmonic, Geometric and Arithmetic Means for Change Detection in SAR Time Series Guillaume Quin CEA, DAM, DIF, F-91297 Arpajon, France Béatrice Pinel-Puysségur CEA, DAM, DIF, F-91297 Arpajon, France Jean-Marie Nicolas Telecom ParisTech, CNRS LTCI, 75634 Paris Cedex 13, France Abstract The amplitude distribution in a SAR image can present a heavy tail. Indeed, very high–valued outliers can be observed. In this paper, we propose the usage of the Harmonic, Geometric and Arithmetic temporal means for amplitude statistical studies along time. In general, the arithmetic mean is used to compute the mean amplitude of time series.
    [Show full text]
  • Incorporating a Geometric Mean Formula Into The
    Calculating the CPI Incorporating a geometric mean formula into the CPI Beginning in January 1999, a new geometric mean formula will replace the current Laspeyres formula in calculating most basic components of the Consumer Price Index; the new formula will better account for the economic substitution behavior of consumers 2 Kenneth V. Dalton, his article describes an important improve- bias” in the CPI. This upward bias was a techni- John S. Greenlees, ment in the calculation of the Consumer cal problem that tied the weight of a CPI sample and TPrice Index (CPI). The Bureau of Labor Sta- item to its expected price change. The flaw was Kenneth J. Stewart tistics plans to use a new geometric mean for- effectively eliminated by changes to the CPI mula for calculating most of the basic compo- sample rotation and substitution procedures and nents of the Consumer Price Index for all Urban to the functional form used to calculate changes Consumers (CPI-U) and the Consumer Price In- in the cost of shelter for homeowners. In 1997, a dex for Urban Wage Earners and Clerical Work- new approach to the measurement of consumer ers (CPI-W). This change will become effective prices for hospital services was introduced.3 Pric- with data for January 1999.1 ing procedures were revised, from pricing indi- The geometric mean formula will be used in vidual items (such as a unit of blood or a hospi- index categories that make up approximately 61 tal inpatient day) to pricing the combined sets of percent of total consumer spending represented goods and services provided on selected patient by the CPI-U.
    [Show full text]
  • Wavelet Operators and Multiplicative Observation Models
    Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series Abdourrahmane Atto, Emmanuel Trouvé, Jean-Marie Nicolas, Thu Trang Le To cite this version: Abdourrahmane Atto, Emmanuel Trouvé, Jean-Marie Nicolas, Thu Trang Le. Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series. 2016. hal-00950823v3 HAL Id: hal-00950823 https://hal.archives-ouvertes.fr/hal-00950823v3 Preprint submitted on 26 Jan 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. 1 Wavelet Operators and Multiplicative Observation Models - Application to Change-Enhanced Regularization of SAR Image Time Series Abdourrahmane M. Atto1;∗, Emmanuel Trouve1, Jean-Marie Nicolas2, Thu-Trang Le^1 Abstract|This paper first provides statistical prop- I. Introduction - Motivation erties of wavelet operators when the observation model IGHLY resolved data such as Synthetic Aperture can be seen as the product of a deterministic piece- Radar (SAR) image time series issued from new wise regular function (signal) and a stationary random H field (noise). This multiplicative observation model is generation sensors show minute details. Indeed, the evo- analyzed in two standard frameworks by considering lution of SAR imaging systems is such that in less than 2 either (1) a direct wavelet transform of the model decades: or (2) a log-transform of the model prior to wavelet • high resolution sensors can achieve metric resolution, decomposition.
    [Show full text]
  • “Mean”? a Review of Interpreting and Calculating Different Types of Means and Standard Deviations
    pharmaceutics Review What Does It “Mean”? A Review of Interpreting and Calculating Different Types of Means and Standard Deviations Marilyn N. Martinez 1,* and Mary J. Bartholomew 2 1 Office of New Animal Drug Evaluation, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA 2 Office of Surveillance and Compliance, Center for Veterinary Medicine, US FDA, Rockville, MD 20855, USA; [email protected] * Correspondence: [email protected]; Tel.: +1-240-3-402-0635 Academic Editors: Arlene McDowell and Neal Davies Received: 17 January 2017; Accepted: 5 April 2017; Published: 13 April 2017 Abstract: Typically, investigations are conducted with the goal of generating inferences about a population (humans or animal). Since it is not feasible to evaluate the entire population, the study is conducted using a randomly selected subset of that population. With the goal of using the results generated from that sample to provide inferences about the true population, it is important to consider the properties of the population distribution and how well they are represented by the sample (the subset of values). Consistent with that study objective, it is necessary to identify and use the most appropriate set of summary statistics to describe the study results. Inherent in that choice is the need to identify the specific question being asked and the assumptions associated with the data analysis. The estimate of a “mean” value is an example of a summary statistic that is sometimes reported without adequate consideration as to its implications or the underlying assumptions associated with the data being evaluated. When ignoring these critical considerations, the method of calculating the variance may be inconsistent with the type of mean being reported.
    [Show full text]
  • Expectation and Functions of Random Variables
    POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior of random variables, we first consider their expectation, which is also called mean value or expected value. The definition of expectation follows our intuition. Definition 1 Let X be a random variable and g be any function. 1. If X is discrete, then the expectation of g(X) is defined as, then X E[g(X)] = g(x)f(x), x∈X where f is the probability mass function of X and X is the support of X. 2. If X is continuous, then the expectation of g(X) is defined as, Z ∞ E[g(X)] = g(x)f(x) dx, −∞ where f is the probability density function of X. If E(X) = −∞ or E(X) = ∞ (i.e., E(|X|) = ∞), then we say the expectation E(X) does not exist. One sometimes write EX to emphasize that the expectation is taken with respect to a particular random variable X. For a continuous random variable, the expectation is sometimes written as, Z x E[g(X)] = g(x) d F (x). −∞ where F (x) is the distribution function of X. The expectation operator has inherits its properties from those of summation and integral. In particular, the following theorem shows that expectation preserves the inequality and is a linear operator. Theorem 1 (Expectation) Let X and Y be random variables with finite expectations. 1. If g(x) ≥ h(x) for all x ∈ R, then E[g(X)] ≥ E[h(X)].
    [Show full text]
  • Notes on Calculating Computer Performance
    Notes on Calculating Computer Performance Bruce Jacob and Trevor Mudge Advanced Computer Architecture Lab EECS Department, University of Michigan {blj,tnm}@umich.edu Abstract This report explains what it means to characterize the performance of a computer, and which methods are appro- priate and inappropriate for the task. The most widely used metric is the performance on the SPEC benchmark suite of programs; currently, the results of running the SPEC benchmark suite are compiled into a single number using the geometric mean. The primary reason for using the geometric mean is that it preserves values across normalization, but unfortunately, it does not preserve total run time, which is probably the figure of greatest interest when performances are being compared. Cycles per Instruction (CPI) is another widely used metric, but this method is invalid, even if comparing machines with identical clock speeds. Comparing CPI values to judge performance falls prey to the same prob- lems as averaging normalized values. In general, normalized values must not be averaged and instead of the geometric mean, either the harmonic or the arithmetic mean is the appropriate method for averaging a set running times. The arithmetic mean should be used to average times, and the harmonic mean should be used to average rates (1/time). A number of published SPECmarks are recomputed using these means to demonstrate the effect of choosing a favorable algorithm. 1.0 Performance and the Use of Means We want to summarize the performance of a computer; the easiest way uses a single number that can be compared against the numbers of other machines.
    [Show full text]
  • E5df876f5e0f28178c70af790e47
    Journal of Applied Pharmaceutical Science Vol. 8(07), pp 001-009, July, 2018 Available online at http://www.japsonline.com DOI: 10.7324/JAPS.2018.8701 ISSN 2231-3354 Application of Chemometrics for the simultaneous estimation of stigmasterol and β-sitosterol in Manasamitra Vatakam-an ayurvedic herbomineral formulation using HPLC-PDA method Srikalyani Vemuri1, Mohan Kumar Ramasamy1, Pandiyan Rajakanu1, Rajappan Chandra Satish Kumar2, Ilango Kalliappan1,3* 1Division of Analytical Chemistry, Interdisciplinary Institute of Indian System of Medicine (IIISM), SRM Institute of Science and Technology, Kattanku- lathur-603 203, Kancheepuram (Dt), Tamil Nadu, India. 2Clinical Trial and Research Unit (Metabolic Ward), Interdisciplinary Institute of Indian System of Medicine (IIISM), SRM Institute of Science and Tech- nology, Kattankulathur-603 203, Kancheepuram (Dt), Tamil Nadu, India. 3Department of Pharmaceutical Chemistry, SRM College of Pharmacy, SRM Institute of Science and Technology, Kattankulathur-603 203, Kancheepuram (Dt), Tamil Nadu, India. ARTICLE INFO ABSTRACT Article history: A new research method has been developed to approach multiresponse optimization for simultaneously optimizing Received on: 08/05/2018 a large number of experimental factors. LC Chromatogram was optimized using Phenomenex RP C18 column (250 x Accepted on: 15/06/2018 4.6 mm; 5 µm); mobile phase was surged at isocratic mode with a flow of 1.0 mL/min using methanol and acetonitrile Available online: 30/07/2018 (95:5% v/v) at the detection max of 208 nm with the retention time of 16.3 and 18.1 min for Stigmasterol and β-Sitosterol respectively. Amount of Stigmasterol and β-Sitosterol was quantified and found to be 51.0 and 56.3 µg/ mg respectively.
    [Show full text]
  • Geometric Mean
    Project AMP Dr. Antonio Quesada – Director, Project AMP Exploring Geometric Mean Lesson Summary: The students will explore the Geometric Mean through the use of Cabrii II software or TI – 92 Calculators and inquiry based activities. Keywords: Geometric Mean, Ratios NCTM Standards: 1. Analyze characteristics and properties of two-dimensional geometric shapes and develop mathematical arguments about geometric relationships (p.310). 2. Apply appropriate techniques, tools, and formulas to determine measurements (p.322). Learning Objectives: 1. Students will be able to construct a right triangle with an altitude drawn from the right angle. 2. Students will be able to use the geometric mean to find missing measures for parts of a right triangle. Materials Needed: 1. Cabri II Geometry Software 2. Pencil, Paper 3. Lab Handout Procedures/Review: 1. Attention Grabber: Begin by asking the students, “What are some ways we can find missing lengths of a right triangle?” (The students may already know how to apply the Pythagorean Theorem, or the formulas for special right triangles, or maybe even how to use trigonometric ratios). The answer you will be showing them is how to use the geometric mean. 2. Students can be grouped in teams of two or, if enough computers are available, they may work individually. 3. Assessment will be based on the students’ completion of the lab worksheet. 4. A review of ratios and proportions may be necessary prior to instruction. Project AMP Dr. Antonio Quesada – Director, Project AMP Exploring Geometric Mean Team Members: ____________________ ____________________ File Name: ____________________ Lab Goal: To analyze the relationship between different sides and segments of right triangles.
    [Show full text]
  • Better Than Average: Calculating Geometric Means Using SAS Kimberly Roenfeldt Henry M
    Paper 56-2018 Better than Average: Calculating Geometric Means Using SAS Kimberly Roenfeldt Henry M. Jackson Foundation for the Advancement of Military Medicine ABSTRACT Geometric means are a robust and precise way to visualize the central tendency of a data set, particularly when examining skewed data or comparing ratios. Measures of central tendency are predominantly presented as arithmetic means or medians that are relatively simple to calculate and interpret, but may be inaccurate in representing data that are not strictly normal. Geometric means represent the best of both worlds, providing estimates that take into account all the observations in a data set without being influenced by the extremes. They can be applied by SAS® programmers working in multiple industries including business, finance, health care, and research. Examples are varied and include examining compounded interest rates or returns on investments, assessing population changes in longitudinal data, or investigating lognormal data such lab assay results, biological concentrations, or decay rates. Fortunately, there are multiple methods in SAS that can be used to calculate geometric means including the GEOMEAN() function, the geomean keyword in PROC SURVEYMEANS, as well as manual data manipulations such as log transformation combined with PROC MEANS and exponentiation. This paper will explain the utility of geometric means and provide examples for using SAS to calculate geometric means and their measures of variability for your data. INTRODUCTION What is a geometric mean? Geometric means are a type of “average”, or measure of central tendency in a distribution of data points, in the same group as the median, mode, or arithmetic mean.
    [Show full text]
  • Superlets: Time-Frequency Super-Resolution Using Wavelet Sets
    bioRxiv preprint doi: https://doi.org/10.1101/583732; this version posted September 8, 2019. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. Superlets: time-frequency super-resolution using wavelet sets Vasile V. Moca1, Adriana Nagy-Dăbâcan1, Harald Bârzan1,2, Raul C. Murean1,* 1Transylvanian Institute of Neuroscience, Department of Experimental and Theoretical Neuroscience, Pta. Timotei Cipariu 9/20, 400191 Cluj-Napoca, Romania 2Technical University of Cluj-Napoca, Basis of Electronics Department, Str. G. Baritiu 26-28, 400027 Cluj-Napoca, Romania * Correspondence to: Dr. Raul C. Mureşan Transylvanian Institute of Neuroscience, Department of Experimental and Theoretical Neuroscience, Pta. Timotei Cipariu 9/20, 400191 Cluj-Napoca, Romania Tel: +40364730730 Fax: +4036470731 E-mail: [email protected] 1 bioRxiv preprint doi: https://doi.org/10.1101/583732; this version posted September 8, 2019. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. Abstract Time-frequency analysis is ubiquitous in many fields of science. Due to the Heisenberg-Gabor uncertainty principle, a single measurement cannot estimate precisely the location of a finite oscillation in both time and frequency. Classical spectral estimators, like the short-time Fourier transform (STFT) or the continuous-wavelet transform (CWT) optimize either temporal or frequency resolution, or find a tradeoff that is suboptimal in both dimensions.
    [Show full text]
  • A Geometric Mean-Based DEMATEL Model for Evaluating the Critical Challenges of Spare Parts Planning 121
    A Geometric Mean-based DEMATEL Model for Evaluating the Critical Challenges of Spare Parts Planning 121 A Geometric Mean-based DEMATEL Model for Evaluating the Critical Challenges of Spare Parts Planning Sung-Shun Weng1, Kai-Ying Chen2, Chih-Yuan Li2 1 Department of Information and Finance Management, National Taipei University of Technology, Taiwan 2 Department of Industrial Engineering and Management, National Taipei University of Technology, Taiwan [email protected], [email protected], [email protected]* Abstract meet service-level requirements is the main challenge of SPP. Other researchers, including the present We present a method for evaluating the critical authors, contend that both factors are challenges in SPP. challenges of spare parts planning (SPP) to help planners The distributed nature of the spare parts, relatively develop appropriate SPP strategies. The purpose of this large number of stock keeping units, varying demand study is to propose a framework that clarifies key factors patterns, and lead time variation are various factors that shaping the understanding of critical challenges of SPP. increase the complexity of SPP optimization [3]. Through a literature review, 16 critical challenges of the The present study investigates the probable SPP criteria of planning are identified. Field study data existence of a geometric mean-based decision making are acquired through questionnaires given to logistics trial and evaluation laboratory (DEMATEL) model for managers at a well-known computer company. After the evaluating the critical challenges of SPP. Our data are analyzed through principal component analysis structured method can provide valuable information to (PCA), the geometric mean is calculated, and we then evaluate the relationships among the critical challenges evaluate these critical challenges of the SPP criteria using of SPP, which is why we explore this topic using an the decision making trial and evaluation laboratory objective and systematic analysis.
    [Show full text]
  • The Precision of the Arithmetic Mean, Geometric Mean And
    1 The precision of the arithmetic mean, geometric mean and percentiles for citation data: An experimental simulation modelling approach1 Mike Thelwall, Statistical Cybermetrics Research Group, School of Mathematics and Computer Science, University of Wolverhampton, Wulfruna Street, Wolverhampton, UK. When comparing the citation impact of nations, departments or other groups of researchers within individual fields, three approaches have been proposed: arithmetic means, geometric means, and percentage in the top X%. This article compares the precision of these statistics using 97 trillion experimentally simulated citation counts from 6875 sets of different parameters (although all having the same scale parameter) based upon the discretised lognormal distribution with limits from 1000 repetitions for each parameter set. The results show that the geometric mean is the most precise, closely followed by the percentage of a country’s articles in the top 50% most cited articles for a field, year and document type. Thus the geometric mean citation count is recommended for future citation-based comparisons between nations. The percentage of a country’s articles in the top 1% most cited is a particularly imprecise indicator and is not recommended for international comparisons based on individual fields. Moreover, whereas standard confidence interval formulae for the geometric mean appear to be accurate, confidence interval formulae are less accurate and consistent for percentile indicators. These recommendations assume that the scale parameters of the samples are the same but the choice of indicator is complex and partly conceptual if they are not. Keywords: scientometrics; citation analysis; research evaluation; geometric mean; percentile indicators; MNCS 1. Introduction The European Union and some countries and regions produce periodic reports that compare their scientific performance with the world average or with appropriate comparators (EC, 2007; Elsevier, 2013; NIFU, 2014; NISTEP, 2014; NSF, 2014; Salmi, 2015).
    [Show full text]