Don't Have to Worry About Multiple Comparisons
Total Page:16
File Type:pdf, Size:1020Kb

Load more
Recommended publications
-
Multiple Testing
The p-value Familywise Error Rate Fisher's Method False Discovery Rate Chapter 8 Hypothesis Testing Multiple Testing 1 / 20 The p-value Familywise Error Rate Fisher's Method False Discovery Rate Outline The p-value Familywise Error Rate Fisher's Method Benjamini-Hochberg Procedure False Discovery Rate 2 / 20 The p-value Familywise Error Rate Fisher's Method False Discovery Rate The p-value In 2016, the American Statistical Association set for itself a task to make a statement on p-values. They note that it is all too easy to set a test, create a test statistic and compute a p-value. Proper statistical practice is much more than this and includes • appropriately chosen techniques based on a thorough understanding of the phenomena under study, • adequate visual and numerical summaries of the data, • properly conducted analyses whose logic and quantitative approaches are clearly explained, • correct interpretation of statistical results in context, and • reproducibility of results via a thorough reporting. Expressing a p-value is one of many approaches to summarize the results of a statistical investigation. 3 / 20 The p-value Familywise Error Rate Fisher's Method False Discovery Rate The p-value The statement's six principles, many of which address misconceptions and misuse of the p-value, are the following: 1. P-values can indicate how incompatible the data are with a specified statistical model. 2. P-values do not measure the probability that the studied hypothesis is true, or the probability that the data were produced by random chance alone. 3. Scientific conclusions and business or policy decisions should not be based only on whether a p-value passes a specific threshold. -
©2018 Oxford University Press
©2018 PART E OxfordChallenges in Statistics University Press mot43560_ch22_201-213.indd 201 07/31/17 11:56 AM ©2018 Oxford University Press ©2018 CHAPTER 22 Multiple Comparisons Concepts If you torture your data long enough, they will tell you what- ever you want to hear. Oxford MILLS (1993) oping with multiple comparisons is one of the biggest challenges Cin data analysis. If you calculate many P values, some are likely to be small just by random chance. Therefore, it is impossible to inter- pret small P values without knowing how many comparisons were made. This chapter explains three approaches to coping with multiple compari- sons. Chapter 23 will show that the problem of multiple comparisons is pervasive, and Chapter 40 will explain special strategies for dealing with multiple comparisons after ANOVA.University THE PROBLEM OF MULTIPLE COMPARISONS The problem of multiple comparisons is easy to understand If you make two independent comparisons, what is the chance that one or both comparisons will result in a statistically significant conclusion just by chance? It is easier to answer the opposite question. Assuming both null hypotheses are true, what is the chance that both comparisons will not be statistically significant? The answer is the chance that the first comparison will not be statistically significant (0.95) times the chance that the second one will not be statistically significant (also 0.95), which equals 0.9025, or about 90%. That leaves about a 10% chance of obtaining at least one statistically significant conclusion by chance. It is easy to generalize that logic to more comparisons. -
The Problem of Multiple Testing and Its Solutions for Genom-Wide Studies
The problem of multiple testing and its solutions for genom-wide studies How to cite: Gyorffy B, Gyorffy A, Tulassay Z:. The problem of multiple testing and its solutions for genom-wide studies. Orv Hetil , 2005;146(12):559-563 ABSTRACT The problem of multiple testing and its solutions for genome-wide studies. Even if there is no real change, the traditional p = 0.05 can cause 5% of the investigated tests being reported significant. Multiple testing corrections have been developed to solve this problem. Here the authors describe the one-step (Bonferroni), multi-step (step-down and step-up) and graphical methods. However, sometimes a correction for multiple testing creates more problems, than it solves: the universal null hypothesis is of little interest, the exact number of investigations to be adjusted for can not determined and the probability of type II error increases. For these reasons the authors suggest not to perform multiple testing corrections routinely. The calculation of the false discovery rate is a new method for genome-wide studies. Here the p value is substituted by the q value, which also shows the level of significance. The q value belonging to a measurement is the proportion of false positive measurements when we accept it as significant. The authors propose using the q value instead of the p value in genome-wide studies. Free keywords: multiple testing, Bonferroni-correction, one-step, multi-step, false discovery rate, q-value List of abbreviations: FDR: false discovery rate FWER: family familywise error rate SNP: Single Nucleotide Polymorphism PFP: proportion of false positives 1 INTRODUCTION A common feature in all of the 'omics studies is the inspection of a large number of simultaneous measurements in a small number of samples. -
Discussion: on Methods Controlling the False Discovery Rate
Sankhy¯a: The Indian Journal of Statistics 2008, Volume 70, Part 2, pp. 169-176 c 2008, Indian Statistical Institute Discussion: On Methods Controlling the False Discovery Rate Joseph P. Romano Stanford University, USA Azeem M. Shaikh University of Chicago, USA Michael Wolf University of Zurich, Switzerland 1 Introduction It is a pleasure to acknowledge another insightful article by Sarkar. By developing clever expressions for the FDP, FDR, and FNR, he manages to prove fundamental properties of stepdown and stepup methods. It is particularly important that the theory is sufficiently developed so as to apply to what Sarkar calls adaptive BH methods. Here, the goal is to improve upon the Benjamini Hochberg procedure by incorporating a data-dependent estimate of the number of true null hypotheses. Theoretical justification of such methods is vital and Sarkar’s analysis is useful for this purpose. A perhaps more ambitious task is to develop methods which implicitly or explicitly estimate the joint dependence structure of the test statistics (or p-values). The focus of our discussion is to show how resampling methods can be used to construct stepdown procedures which control the FDR or other general measures of error control. The main benefit of using the boot- strap or subsampling is the ability to estimate the joint distribution of the test statistics, and thereby offer the potential of improving upon methods based on the marginal distributions of test statistics. The procedure below is a generalization of one we developed for FDR control, and the utility of the bootstrap is that it can apply to essentially arbitrary measures of error control, such as the pairwise FDR of Sarkar, the k-FWER, or the tail prob- abilities of the false discovery proportion. -
A Primer for Analyzing Nested Data: Multilevel Modeling in SPSS Using
REL 2015–046 The National Center for Education Evaluation and Regional Assistance (NCEE) conducts unbiased large-scale evaluations of education programs and practices supported by federal funds; provides research-based technical assistance to educators and policymakers; and supports the synthesis and the widespread dissemination of the results of research and evaluation throughout the United States. December 2014 This report was prepared for the Institute of Education Sciences (IES) under contract ED-IES-12-C-0009 by Regional Educational Laboratory Northeast & Islands administered by Education Development Center, Inc. (EDC). The content of the publication does not necessarily reflect the views or policies of IES or the U.S. Department of Education nor does mention of trade names, commercial products, or organizations imply endorsement by the U.S. Government. This REL report is in the public domain. While permission to reprint this publication is not necessary, it should be cited as: O’Dwyer, L. M., and Parker, C. E. (2014). A primer for analyzing nested data: multilevel mod eling in SPSS using an example from a REL study (REL 2015–046). Washington, DC: U.S. Department of Education, Institute of Education Sciences, National Center for Educa tion Evaluation and Regional Assistance, Regional Educational Laboratory Northeast & Islands. Retrieved from http://ies.ed.gov/ncee/edlabs. This report is available on the Regional Educational Laboratory website at http://ies.ed.gov/ ncee/edlabs. Summary Researchers often study how students’ academic outcomes are associated with the charac teristics of their classrooms, schools, and districts. They also study subgroups of students such as English language learner students and students in special education. -
A Bayesian Multilevel Model for Time Series Applied to Learning in Experimental Auctions
Linköpings universitet | Institutionen för datavetenskap Kandidatuppsats, 15 hp | Statistik Vårterminen 2016 | LIU-IDA/STAT-G--16/003—SE A Bayesian Multilevel Model for Time Series Applied to Learning in Experimental Auctions Torrin Danner Handledare: Bertil Wegmann Examinator: Linda Wänström Linköpings universitet SE-581 83 Linköping, Sweden 013-28 10 00, www.liu.se Abstract Establishing what variables affect learning rates in experimental auctions can be valuable in determining how competitive bidders in auctions learn. This study aims to be a foray into this field. The differences, both absolute and actual, between participant bids and optimal bids are evaluated in terms of the effects from a variety of variables such as age, sex, etc. An optimal bid in the context of an auction is the best bid a participant can place to win the auction without paying more than the value of the item, thus maximizing their revenues. This study focuses on how two opponent types, humans and computers, affect the rate at which participants learn to optimize their winnings. A Bayesian multilevel model for time series is used to model the learning rate of actual bids from participants in an experimental auction study. The variables examined at the first level were auction type, signal, round, interaction effects between auction type and signal and interaction effects between auction type and round. At a 90% credibility interval, the true value of the mean for the intercept and all slopes falls within an interval that also includes 0. Therefore, none of the variables are deemed to be likely to influence the model. The variables on the second level were age, IQ, sex and answers from a short quiz about how participants felt when they won or lost auctions. -
Inference on the Limiting False Discovery Rate and the P-Value Threshold Parameter Assuming Weak Dependence Between Gene Expression Levels Within Subject
Inference on the limiting false discovery rate and the p-value threshold parameter assuming weak dependence between gene expression levels within subject Glenn Heller1 and Jing Qin2 1Department of Epidemiology and Biostatistics Memorial Sloan-Kettering Cancer Center New York, NY 10021, USA 2Biostatistics Research Branch National Institute of Allergy and Infectious Diseases Bethesda, Maryland 20892, USA Corresponding author: Glenn Heller email: [email protected] phone: 646 735 8112 fax: 646 735 0010 Running Head: FDR analysis with dependent data 1 Summary. An objective of microarray data analysis is to identify gene expressions that are associated with a disease related outcome. For each gene, a test statistic is computed to determine if an association exists, and this statistic generates a marginal p-value. In an effort to pool this information across genes, a p-value density function is derived. The p-value density is modeled as a mixture of a uniform (0,1) density and a scaled ratio of normal densities derived from the asymptotic normality of the test statistic. The p-values are assumed to be weakly dependent and a quasi-likelihood is used to estimate the parameters in the mixture density. The quasi-likelihood and the weak dependence assumption enables estimation and asymptotic inference on the false discovery rate for a given rejection region, and its inverse, the p-value threshold parameter for a fixed false discovery rate. A false discovery rate analysis on a local- ized prostate cancer data set is used to illustrate the methodology. Simulations are performed to assess the performance of this methodology. Keywords: Asymptotic normal test statistic, confidence interval, microarray, p-value mixture model, quasi-likelihood, weak dependence. -
Heterocedasticity-Adjusted Ranking and Thresholding for Large-Scale Multiple Testing
Heterocedasticity-Adjusted Ranking and Thresholding for Large-Scale Multiple Testing Luella Fu1, Bowen Gang2, Gareth M. James3, and Wenguang Sun3 Abstract Standardization has been a widely adopted practice in multiple testing, for it takes into account the variability in sampling and makes the test statistics comparable across different study units. However, despite conventional wisdom to the contrary, we show that there can be a significant loss in information from basing hypothesis tests on stan- dardized statistics rather than the full data. We develop a new class of heteroscedasticity{ adjusted ranking and thresholding (HART) rules that aim to improve existing methods by simultaneously exploiting commonalities and adjusting heterogeneities among the study units. The main idea of HART is to bypass standardization by directly incor- porating both the summary statistic and its variance into the testing procedure. A key message is that the variance structure of the alternative distribution, which is subsumed under standardized statistics, is highly informative and can be exploited to achieve higher power. The proposed HART procedure is shown to be asymptotically valid and optimal for false discovery rate (FDR) control. Our simulation results demon- strate that HART achieves substantial power gain over existing methods at the same FDR level. We illustrate the implementation through a microarray analysis of myeloma. Keywords: covariate-assisted inference; data processing and information loss; false dis- covery rate; heteroscedasticity; multiple testing with side information; structured multiple testing 1Department of Mathematics, San Francisco State University. 2Department of Statistics, Fudan University. 3Department of Data Sciences and Operations, University of Southern California. The research of Wen- guang Sun was supported in part by NSF grant DMS-1712983. -
Introduction to Multilevel Modeling
1 INTRODUCTION TO MULTILEVEL MODELING OVERVIEW Multilevel modeling goes back over half a century to when social scientists becamedistribute attuned to the fact that what was true at the group level was not necessarily true at the individual level. The classic example was percentage literate and percentage African American. Using data aggregated to the state level for the 50 American states in the 1930s as units ofor observation, Robinson (1950) found that there was a strong correlation of race with illiteracy. When using individual- level data, however, the correlation was observed to be far lower. This difference arose from the fact that the Southern states, which had many African Americans, also had many illiterates of all races. Robinson described potential misinterpretations based on aggregated data, such as state-level data, as a type of “ecological fallacy” (see also Clancy, Berger, & Magliozzi, 2003). In the classic article by Davis, Spaeth, and Huson (1961),post, the same problem was put in terms of within-group versus between-group effects, corresponding to individual-level and group-level effects. A central function of multilevel modeling is to separate within-group individual effects from between-group aggregate effects. Multilevel modeling (MLM) is appropriate whenever there is clustering of the outcome variable by a categorical variable such that error is no longer independent as required by ordinary least squares (OLS) regression but rather error is correlated. Clustering of level 1 outcome scores within levels formed by a levelcopy, 2 clustering variable (e.g., employee ratings clustering by work unit) means that OLS estimates will be too high for some units and too low for others. -
Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods
H OH metabolites OH Article Comprehensive Comparative Analysis of Local False Discovery Rate Control Methods Shin June Kim, Youngjae Oh and Jaesik Jeong * Department of Mathematics and Statistics, Chonnam National University, Gwangju 61186, Korea; [email protected] (S.J.K.); [email protected] (Y.O.) * Correspondence: [email protected]; Tel.: +82-062-530-3442 Abstract: Due to the advance in technology, the type of data is getting more complicated and large-scale. To analyze such complex data, more advanced technique is required. In case of omics data from two different groups, it is interesting to find significant biomarkers between two groups while controlling error rate such as false discovery rate (FDR). Over the last few decades, a lot of methods that control local false discovery rate have been developed, ranging from one-dimensional to k-dimensional FDR procedure. For comparison study, we select three of them, which have unique and significant properties: Efron’s approach, Ploner’s approach, and Kim’s approach in chronological order. The first approach is one-dimensional approach while the other two are two-dimensional ones. Furthermore, we consider two more variants of Ploner’s approach. We compare the performance of those methods on both simulated and real data. Keywords: biomarker; familywise error rate; false discovery rate; large scale inference 1. Introduction Due to the advent of advanced high-throughput technologies, a large amount of raw data have been produced and various methods that can appropriately preprocess Citation: Kim, S.J.; Oh, Y.; Jeong, J. such data have been developed. After various preprocessing steps, statistical methods Comprehensive Comparative are applied to the preprocessed, yet large-scale data. -
Asymptotic Results on Adaptive False Discovery Rate Controlling Procedures Based on Kernel Estimators
JournalofMachineLearningResearch14(2013)1423-1459 Submitted 4/08; Revised 8/12; Published 5/13 Asymptotic Results on Adaptive False Discovery Rate Controlling Procedures Based on Kernel Estimators Pierre Neuvial∗ [email protected] Laboratoire Statistique et Génome Université d’Évry Val d’Essonne UMR CNRS 8071 – USC INRA 23 boulevard de France 91 037 Évry, France Editor: Olivier Teytaud Abstract The False Discovery Rate (FDR) is a commonly used type I error rate in multiple testing problems. It is defined as the expected False Discovery Proportion (FDP), that is, the expected fraction of false positives among rejected hypotheses. When the hypotheses are independent, the Benjamini- Hochberg procedure achieves FDR control at any pre-specified level. By construction, FDR control offers no guarantee in terms of power, or type II error. A number of alternative procedures have been developed, including plug-in procedures that aim at gaining power by incorporating an estimate of the proportion of true null hypotheses. In this paper, we study the asymptotic behavior of a class of plug-in procedures based on kernel estimators of the density of the p-values, as the number m of tested hypotheses grows to infinity. In a setting where the hypotheses tested are independent, we prove that these procedures are asymptotically more powerful in two respects: (i) a tighter asymptotic FDR control for any target FDR level and (ii) a broader range of target levels yielding positive asymptotic power. We also show that this increased asymptotic power comes at the price of slower, non-parametric convergence rates k/(2k+1) for the FDP. -
An Introduction to Bayesian Multilevel (Hierarchical) Modelling Using
An Intro duction to Bayesian Multilevel Hierarchical Mo delling using MLwiN by William Browne and Harvey Goldstein Centre for Multilevel Mo delling Institute of Education London May Summary What Is Multilevel Mo delling Variance comp onents and Random slop es re gression mo dels Comparisons between frequentist and Bayesian approaches Mo del comparison in multilevel mo delling Binary resp onses multilevel logistic regres sion mo dels Comparisons between various MCMC meth ods Areas of current research Junior Scho ol Project JSP Dataset The JSP dataset Mortimore et al was a longitudinal study of around primary scho ol children from scho ols in the Inner Lon don Education Authority Sample of interest here consists of pupils from scho ols Main outcome of intake is a score out of in year in mathematics MATH Main predictor is a score in year again out of on a mathematics test MATH Other predictors are child gender and fathers o ccupation status manualnon manual Why Multilevel mo delling Could t a simple linear regression of MATH and MATH on the full dataset Could t separate regressions of MATH and MATH for each scho ol indep endently The rst approach ignores scho ol eects dif ferences whilst the second ignores similari ties between the scho ols Solution A multilevel or hierarchical or ran dom eects mo del treats both the students and scho ols as random variables and can be seen as a compromise between the two single level approaches The predicted regression lines for each scho ol pro duced by a multilevel mo del are a weighted average of the individual scho ol lines and the full dataset regression line A mo del with both random scho ol intercepts and slop es is called a random slop es regression mo del Random Slopes Regression (RSR) Model MLwiN (Rasbash et al.