Confidence Interval Estimation for Continuous Outcomes in Cluster Randomization Trials

Total Page:16

File Type:pdf, Size:1020Kb

Confidence Interval Estimation for Continuous Outcomes in Cluster Randomization Trials Western University Scholarship@Western Electronic Thesis and Dissertation Repository 4-1-2011 12:00 AM Confidence Interval Estimation for Continuous Outcomes in Cluster Randomization Trials Julia Taleban The University of Western Ontario Supervisor Dr. Guang Yong Zou The University of Western Ontario Graduate Program in Biostatistics A thesis submitted in partial fulfillment of the equirr ements for the degree in Doctor of Philosophy © Julia Taleban 2011 Follow this and additional works at: https://ir.lib.uwo.ca/etd Part of the Biostatistics Commons Recommended Citation Taleban, Julia, "Confidence Interval Estimation for Continuous Outcomes in Cluster Randomization Trials" (2011). Electronic Thesis and Dissertation Repository. 121. https://ir.lib.uwo.ca/etd/121 This Dissertation/Thesis is brought to you for free and open access by Scholarship@Western. It has been accepted for inclusion in Electronic Thesis and Dissertation Repository by an authorized administrator of Scholarship@Western. For more information, please contact [email protected]. CONFIDENCE INTERVAL ESTIMATION FOR CONTINUOUS OUTCOMES IN CLUSTER RANDOMIZATION TRIALS (Spine title: Confidence Intervals in Cluster Randomization Trials) (Thesis format: Monograph) by Julia Taleban Graduate Program in Epidemiology & Biostatistics A thesis submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy School of Graduate and Postdoctoral Studies University of Western Ontario London, Ontario April 27, 2011 c Julia Taleban, 2011 THE UNIVERSITY OF WESTERN ONTARIO SCHOOL OF GRADUATE AND POSTDOCTORAL STUDIES CERTIFICATE OF EXAMINATION Supervisor Examining Board aaaaaaaaaaaaaaaaaaaaaa aaaaaaaaaaaaaaaaaaaaaa Dr. Guangyong Zou Dr. Neil Klar Co-Supervisor aaaaaaaaaaaaaaaaaaaaaa Dr. Duncan Murdoch aaaaaaaaaaaaaaaaaaaaaa Dr. John Koval aaaaaaaaaaaaaaaaaaaaaa Dr. Serge Provost Committee member aaaaaaaaaaaaaaaaaaaaaa Dr. Wendy Lou aaaaaaaaaaaaaaaaaaaaaa Dr. Allan Donner The thesis by Julia Taleban Entitled Confidence Interval Estimation for Continuous Outcomes in Cluster Randomization Trials is accepted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Date: aaaaaaaaaaaaaaaaaa aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa Chair of the Thesis Examination Board ii ABSTRACT Cluster randomization trials are experiments where intact social units (e.g. hos- pitals, schools, communities, and families) are randomized to the arms of the trial rather than individuals. The popularity of this design among health researchers is partially due to reduced contamination of treatment effects and convenience. How- ever, the advantages of cluster randomization trials come with a price. Due to the dependence of individuals within a cluster, cluster randomization trials suffer reduced statistical efficiency and often require a complex analysis of study outcomes. The primary purpose of this thesis is to propose new confidence intervals for effect measures commonly of interest for continuous outcomes arising from cluster random- ization trials. Specifically, we construct new confidence intervals for the difference between two normal means, the difference between two lognormal means, and the exceedance probability. The proposed confidence intervals, which use the method of variance estimates recovery (MOVER), do not make certain assumptions that existing procedures make on the data. For instance, symmetry is not forced when the sampling distribution of the parameter estimate is skewed and the assumption of homoscedasticity is not made. Furthermore, the MOVER results in simple confidence interval procedures rather than complex simulation-based methods which currently exist. Simulation studies are used to investigate the small sample properties of the MOVER as compared with existing procedures. Unbalanced cluster sizes are sim- ulated, with an average range of 50 to 200 individuals per cluster and 6 to 24 clusters per arm. The effects of various degrees of dependence between individuals within the same cluster are also investigated. When comparing the empirical coverage, tail errors, and median widths of confi- dence interval procedures, the MOVER has coverage close to the nominal, relatively balanced tail errors, and narrow widths as compared to existing procedure for the iii majority of the parameter combinations investigated. Existing data from cluster ran- domization trials are then used to illustrate each of the methods. Keywords: cluster randomization trials, confidence intervals, normal mean, lognor- mal mean, exceedance probability, method of variance estimates recovery, generalized confidence interval procedure, Wald method. iv ACKNOWLEDGMENTS The completion of this thesis would not have been possible without the help and support of many people. First, I would like to thank my Ph.D. supervisor, Dr. Guang Yong Zou, who provided me with invaluable guidance and insight. I would also like to thank my co-supervisor, Dr. John Koval, whose encouragement and suggestions have been a great help, and Dr. Allan Donner for the time he took to discuss the drafts of the thesis and his helpful suggestions. I am grateful to Dr. Thomas Marrie and Dr. Alan Montgomery for the use of their datasets while illustrating various confidence interval methods in Chapter 5. I am indebted to my parents and friends for their moral support and encourage- ment. My fiance Luc has shown constant support and endless patience, and for that I am thankful. This work has been financially supported in part by the Ontario Graduate Schol- arship in Science and Technology and by the Ontario Graduate Scholarship from the Ontario Ministry of Training, Colleges, and Universities. v TABLE OF CONTENTS Certificate of Examination ii Abstract iii Acknowledgments v List of Tables x List of Figures xiii Chapter 1 Introduction 1 1.1 Randomizedcontrolledtrials. 1 1.2 Clusterrandomizationtrials . 1 1.3 Notation.................................. 5 1.4 Brief summary of methods for cluster randomization trials with con- tinuousdata................................ 8 Cluster-levelanalyses... .. .. .. .. ... .. .. .. .. ... .. 8 Individuallevelanalyses . 9 Current methods for a difference between two normal means . 12 1.5 Alternative effect measures in cluster randomization trials ...... 15 Thestandardizedmeandifference . 15 Thedifferencebetweentwolognormalmeans . 21 1.6 Scopeofthethesis ............................ 24 1.7 Objectives................................. 26 1.8 Organizationofthethesis .. .. .. .. ... .. .. .. .. ... .. 27 Chapter 2 Fundamentals of confidence interval estimation 28 2.1 Introduction................................ 28 2.2 Definitionofaconfidenceinterval . 31 2.3 Confidence interval estimation for a single parameter . ....... 32 Theinversionprinciple . 32 Thetransformationprinciple. 33 2.4 Wald-type confidence intervals and the delta method . ..... 33 Asingleparameter ............................ 34 Afunctionofmultipleparameters . 34 Properties of Wald-type confidence intervals . 35 vi 2.5 Confidence intervals for a function of multiple parameters....... 37 The MOVER for a linear combination of parameters . 38 The MOVER for the ratio of two independent parameters . 45 PropertiesoftheMOVER .. .. .. .. ... .. .. .. .. ... .. 47 Previousapplications . 49 Chapter 3 Confidence interval estimation for effect measures in clus- ter randomization trials 53 3.1 Adifferencebetweentwonormalmeans. 54 TheMOVER ............................... 54 Alternativeconfidenceintervals . 55 Waldconfidenceinterval . 55 Cluster-adjusted confidence interval . 56 Generalizedconfidenceinterval . 57 3.2 Adifferencebetweentwolognormalmeans . 59 TheMOVERforasinglemean . 59 The MOVER for a difference between two lognormal means . 62 Alternativeconfidenceintervals . 62 Wald confidence interval and the delta method . 62 Generalizedconfidenceintervals . 64 3.3 Theexceedanceprobability . 65 TheMOVER ............................... 65 Alternativeconfidenceintervals . 66 Wald confidence interval and the delta method . 66 Generalizedconfidenceinterval . 68 Chapter 4 Simulation study of confidence interval procedures 70 4.1 Introduction................................ 70 4.2 Objectives................................. 71 4.3 Methods.................................. 72 Parametercombinations . 72 Datageneration.............................. 75 Clustersizes............................ 75 Correlatednormaldata. 77 Correlatedlognormaldata . 78 Computer software for data generation . 79 Methodsofcomparison.. .. .. .. .. ... .. .. .. .. ... .. 79 4.4 Results................................... 80 Thedifferencebetweentwonormalmeans . 80 Confidenceintervalcoverage . 81 Tailerrors............................. 87 Medianwidth........................... 87 vii Thedifferencebetweentwolognormalmeans . 88 Confidenceintervalcoverage . 88 Tailerrors............................. 89 Medianwidth........................... 95 Theexceedanceprobability . 95 Confidenceintervalcoverage . 95 Tailerrors............................. 97 Medianwidth ........................... 108 4.5 Discussion................................. 108 Thedifferencebetweentwonormalmeans . 108 Confidence interval coverage . 109 Confidenceintervaltailerrors . 110 Confidenceintervalwidths . 111 Thedifferencebetweentwolognormalmeans . 111 Confidence interval coverage . 112 Confidenceintervaltailerrors . 113 Confidenceintervalwidths . 114 Theexceedanceprobability . 114 Confidence interval coverage . 115 Confidenceintervaltailerrors . 115 Confidenceintervalwidths . 116 4.6 Overallconclusions . ... .. .. .. .. ... .. .. .. .. ... .. 116 Chapter 5
Recommended publications
  • Confidence Interval Estimation in System Dynamics Models: Bootstrapping Vs
    Confidence Interval Estimation in System Dynamics Models: Bootstrapping vs. Likelihood Ratio Method Gokhan Dogan* MIT Sloan School of Management, 30 Wadsworth Street, E53-364, Cambridge, Massachusetts 02142 [email protected] Abstract In this paper we discuss confidence interval estimation for system dynamics models. Confidence interval estimation is important because without confidence intervals, we cannot determine whether an estimated parameter value is significantly different from 0 or any other value, and therefore we cannot determine how much confidence to place in the estimate. We compare two methods for confidence interval estimation. The first, the “likelihood ratio method,” is based on maximum likelihood estimation. This method has been used in the system dynamics literature and is built in to some popular software packages. It is computationally efficient but requires strong assumptions about the model and data. These assumptions are frequently violated by the autocorrelation, endogeneity of explanatory variables and heteroskedasticity properties of dynamic models. The second method is called “bootstrapping.” Bootstrapping requires more computation but does not impose strong assumptions on the model or data. We describe the methods and illustrate them with a series of applications from actual modeling projects. Considering the empirical results presented in the paper and the fact that the bootstrapping method requires less assumptions, we suggest that bootstrapping is a better tool for confidence interval estimation in system dynamics
    [Show full text]
  • WHAT DID FISHER MEAN by an ESTIMATE? 3 Ideas but Is in Conflict with His Ideology of Statistical Inference
    Submitted to the Annals of Applied Probability WHAT DID FISHER MEAN BY AN ESTIMATE? By Esa Uusipaikka∗ University of Turku Fisher’s Method of Maximum Likelihood is shown to be a proce- dure for the construction of likelihood intervals or regions, instead of a procedure of point estimation. Based on Fisher’s articles and books it is justified that by estimation Fisher meant the construction of likelihood intervals or regions from appropriate likelihood function and that an estimate is a statistic, that is, a function from a sample space to a parameter space such that the likelihood function obtained from the sampling distribution of the statistic at the observed value of the statistic is used to construct likelihood intervals or regions. Thus Problem of Estimation is how to choose the ’best’ estimate. Fisher’s solution for the problem of estimation is Maximum Likeli- hood Estimate (MLE). Fisher’s Theory of Statistical Estimation is a chain of ideas used to justify MLE as the solution of the problem of estimation. The construction of confidence intervals by the delta method from the asymptotic normal distribution of MLE is based on Fisher’s ideas, but is against his ’logic of statistical inference’. Instead the construc- tion of confidence intervals from the profile likelihood function of a given interest function of the parameter vector is considered as a solution more in line with Fisher’s ’ideology’. A new method of cal- culation of profile likelihood-based confidence intervals for general smooth interest functions in general statistical models is considered. 1. Introduction. ’Collected Papers of R.A.
    [Show full text]
  • A Framework for Sample Efficient Interval Estimation with Control
    A Framework for Sample Efficient Interval Estimation with Control Variates Shengjia Zhao Christopher Yeh Stefano Ermon Stanford University Stanford University Stanford University Abstract timates, e.g. for each individual district or county, meaning that we need to draw a sufficient number of We consider the problem of estimating confi- samples for every district or county. Another diffi- dence intervals for the mean of a random vari- culty can arise when we need high accuracy (i.e. a able, where the goal is to produce the smallest small confidence interval), since confidence intervals possible interval for a given number of sam- producedp by typical concentration inequalities have size ples. While minimax optimal algorithms are O(1= number of samples). In other words, to reduce known for this problem in the general case, the size of a confidence interval by a factor of 10, we improved performance is possible under addi- need 100 times more samples. tional assumptions. In particular, we design This dilemma is generally unavoidable because concen- an estimation algorithm to take advantage of tration inequalities such as Chernoff or Chebychev are side information in the form of a control vari- minimax optimal: there exist distributions for which ate, leveraging order statistics. Under certain these inequalities cannot be improved. No-free-lunch conditions on the quality of the control vari- results (Van der Vaart, 2000) imply that any alter- ates, we show improved asymptotic efficiency native estimation algorithm that performs better (i.e. compared to existing estimation algorithms. outputs a confidence interval with smaller size) on some Empirically, we demonstrate superior perfor- problems, must perform worse on other problems.
    [Show full text]
  • STATS 305 Notes1
    STATS 305 Notes1 Art Owen2 Autumn 2013 1The class notes were beautifully scribed by Eric Min. He has kindly allowed his notes to be placed online for stat 305 students. Reading these at leasure, you will spot a few errors and omissions due to the hurried nature of scribing and probably my handwriting too. Reading them ahead of class will help you understand the material as the class proceeds. 2Department of Statistics, Stanford University. 0.0: Chapter 0: 2 Contents 1 Overview 9 1.1 The Math of Applied Statistics . .9 1.2 The Linear Model . .9 1.2.1 Other Extensions . 10 1.3 Linearity . 10 1.4 Beyond Simple Linearity . 11 1.4.1 Polynomial Regression . 12 1.4.2 Two Groups . 12 1.4.3 k Groups . 13 1.4.4 Different Slopes . 13 1.4.5 Two-Phase Regression . 14 1.4.6 Periodic Functions . 14 1.4.7 Haar Wavelets . 15 1.4.8 Multiphase Regression . 15 1.5 Concluding Remarks . 16 2 Setting Up the Linear Model 17 2.1 Linear Model Notation . 17 2.2 Two Potential Models . 18 2.2.1 Regression Model . 18 2.2.2 Correlation Model . 18 2.3 TheLinear Model . 18 2.4 Math Review . 19 2.4.1 Quadratic Forms . 20 3 The Normal Distribution 23 3.1 Friends of N (0; 1)...................................... 23 3.1.1 χ2 .......................................... 23 3.1.2 t-distribution . 23 3.1.3 F -distribution . 24 3.2 The Multivariate Normal . 24 3.2.1 Linear Transformations . 25 3.2.2 Normal Quadratic Forms .
    [Show full text]
  • Likelihood Confidence Intervals When Only Ranges Are Available
    Article Likelihood Confidence Intervals When Only Ranges are Available Szilárd Nemes Institute of Clinical Sciences, Sahlgrenska Academy, University of Gothenburg, Gothenburg 405 30, Sweden; [email protected] Received: 22 December 2018; Accepted: 3 February 2019; Published: 6 February 2019 Abstract: Research papers represent an important and rich source of comparative data. The change is to extract the information of interest. Herein, we look at the possibilities to construct confidence intervals for sample averages when only ranges are available with maximum likelihood estimation with order statistics (MLEOS). Using Monte Carlo simulation, we looked at the confidence interval coverage characteristics for likelihood ratio and Wald-type approximate 95% confidence intervals. We saw indication that the likelihood ratio interval had better coverage and narrower intervals. For single parameter distributions, MLEOS is directly applicable. For location-scale distribution is recommended that the variance (or combination of it) to be estimated using standard formulas and used as a plug-in. Keywords: range, likelihood, order statistics, coverage 1. Introduction One of the tasks statisticians face is extracting and possibly inferring biologically/clinically relevant information from published papers. This aspect of applied statistics is well developed, and one can choose to form many easy to use and performant algorithms that aid problem solving. Often, these algorithms aim to aid statisticians/practitioners to extract variability of different measures or biomarkers that is needed for power calculation and research design [1,2]. While these algorithms are efficient and easy to use, they mostly are not probabilistic in nature, thus they do not offer means for statistical inference. Yet another field of applied statistics that aims to help practitioners in extracting relevant information when only partial data is available propose a probabilistic approach with order statistics.
    [Show full text]
  • Pivotal Quantities with Arbitrary Small Skewness Arxiv:1605.05985V1 [Stat
    Pivotal Quantities with Arbitrary Small Skewness Masoud M. Nasari∗ School of Mathematics and Statistics of Carleton University Ottawa, ON, Canada Abstract In this paper we present randomization methods to enhance the accuracy of the central limit theorem (CLT) based inferences about the population mean µ. We introduce a broad class of randomized versions of the Student t- statistic, the classical pivot for µ, that continue to possess the pivotal property for µ and their skewness can be made arbitrarily small, for each fixed sam- ple size n. Consequently, these randomized pivots admit CLTs with smaller errors. The randomization framework in this paper also provides an explicit relation between the precision of the CLTs for the randomized pivots and the volume of their associated confidence regions for the mean for both univariate and multivariate data. This property allows regulating the trade-off between the accuracy and the volume of the randomized confidence regions discussed in this paper. 1 Introduction The CLT is an essential tool for inferring on parameters of interest in a nonpara- metric framework. The strength of the CLT stems from the fact that, as the sample size increases, the usually unknown sampling distribution of a pivot, a function of arXiv:1605.05985v1 [stat.ME] 19 May 2016 the data and an associated parameter, approaches the standard normal distribution. This, in turn, validates approximating the percentiles of the sampling distribution of the pivot by those of the normal distribution, in both univariate and multivariate cases. The CLT is an approximation method whose validity relies on large enough sam- ples.
    [Show full text]
  • 32. Statistics 1 32
    32. Statistics 1 32. STATISTICS Revised September 2007 by G. Cowan (RHUL). This chapter gives an overview of statistical methods used in High Energy Physics. In statistics, we are interested in using a given sample of data to make inferences about a probabilistic model, e.g., to assess the model’s validity or to determine the values of its parameters. There are two main approaches to statistical inference, which we may call frequentist and Bayesian. In frequentist statistics, probability is interpreted as the frequency of the outcome of a repeatable experiment. The most important tools in this framework are parameter estimation, covered in Section 32.1, and statistical tests, discussed in Section 32.2. Frequentist confidence intervals, which are constructed so as to cover the true value of a parameter with a specified probability, are treated in Section 32.3.2. Note that in frequentist statistics one does not define a probability for a hypothesis or for a parameter. Frequentist statistics provides the usual tools for reporting objectively the outcome of an experiment without needing to incorporate prior beliefs concerning the parameter being measured or the theory being tested. As such they are used for reporting most measurements and their statistical uncertainties in High Energy Physics. In Bayesian statistics, the interpretation of probability is more general and includes degree of belief (called subjective probability). One can then speak of a probability density function (p.d.f.) for a parameter, which expresses one’s state of knowledge about where its true value lies. Bayesian methods allow for a natural way to input additional information, such as physical boundaries and subjective information; in fact they require as input the prior p.d.f.
    [Show full text]
  • Stat 3701 Lecture Notes: Bootstrap Charles J
    Stat 3701 Lecture Notes: Bootstrap Charles J. Geyer April 17, 2017 1 License This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License (http: //creativecommons.org/licenses/by-sa/4.0/). 2 R • The version of R used to make this document is 3.3.3. • The version of the rmarkdown package used to make this document is 1.4. • The version of the knitr package used to make this document is 1.15.1. • The version of the bootstrap package used to make this document is 2017.2. 3 Relevant and Irrelevant Simulation 3.1 Irrelevant Most statisticians think a statistics paper isn’t really a statistics paper or a statistics talk isn’t really a statistics talk if it doesn’t have simulations demonstrating that the methods proposed work great (at least in some toy problems). IMHO, this is nonsense. Simulations of the kind most statisticians do prove nothing. The toy problems used are often very special and do not stress the methods at all. In fact, they may be (consciously or unconsciously) chosen to make the methods look good. In scientific experiments, we know how to use randomization, blinding, and other techniques to avoid biasing the results. Analogous things are never AFAIK done with simulations. When all of the toy problems simulated are very different from the statistical model you intend to use for your data, what could the simulation study possibly tell you that is relevant? Nothing. Hence, for short, your humble author calls all of these millions of simulation studies statisticians have done irrelevant simulation.
    [Show full text]
  • Interval Estimation Statistics (OA3102)
    Module 5: Interval Estimation Statistics (OA3102) Professor Ron Fricker Naval Postgraduate School Monterey, California Reading assignment: WM&S chapter 8.5-8.9 Revision: 1-12 1 Goals for this Module • Interval estimation – i.e., confidence intervals – Terminology – Pivotal method for creating confidence intervals • Types of intervals – Large-sample confidence intervals – One-sided vs. two-sided intervals – Small-sample confidence intervals for the mean, differences in two means – Confidence interval for the variance • Sample size calculations Revision: 1-12 2 Interval Estimation • Instead of estimating a parameter with a single number, estimate it with an interval • Ideally, interval will have two properties: – It will contain the target parameter q – It will be relatively narrow • But, as we will see, since interval endpoints are a function of the data, – They will be variable – So we cannot be sure q will fall in the interval Revision: 1-12 3 Objective for Interval Estimation • So, we can’t be sure that the interval contains q, but we will be able to calculate the probability the interval contains q • Interval estimation objective: Find an interval estimator capable of generating narrow intervals with a high probability of enclosing q Revision: 1-12 4 Why Interval Estimation? • As before, we want to use a sample to infer something about a larger population • However, samples are variable – We’d get different values with each new sample – So our point estimates are variable • Point estimates do not give any information about how far
    [Show full text]
  • Confidence Intervals in Analysis and Reporting of Clinical Trials Guangbin Peng, Eli Lilly and Company, Indianapolis, IN
    Confidence Intervals in Analysis and Reporting of Clinical Trials Guangbin Peng, Eli Lilly and Company, Indianapolis, IN ABSTRACT for more frequent use of confidence Regulatory agencies around the world intervals (Simon, 1993). have recommended reporting confidence There is a close relationship between intervals for treatment differences along confidence intervals and significance with the results of significance tests. tests (Hahn and Meeker, 1991); in fact, a SAS provides easy and convenient ways confidence interval can often be used to to produce confidence intervals using test a hypothesis. If the 100(1-α)% procedures such as PROC GLM and confidence interval for the mean PROC UNIVARIATE in conjunction treatment difference in a clinical trial with ODS (output delivery system). In does not contain zero, there is evidence this paper, I will discuss the relationship to indicate a treatment difference at the between significance tests and 100 α % significance level. This strategy confidence intervals, summarize the is equivalent to the hypothesis test that types of confidence intervals used in rejects the null hypothesis of no mean clinical study reports, and provide treatment difference at the level of α. examples from clinical trials to illustrate Compared to p-values, confidence the computation of distribution- intervals are generally more informative. dependent confidence intervals for the They provide quantitative bounds that mean treatment difference and express the uncertainty inherent in distribution-free confidence intervals for estimation, instead of merely an accept the median response within each or reject statement. The length of a treatment group using SAS. confidence interval depends on the sample size; this influence of sample INTRODUCTION size is evident from observing the length Confidence interval estimation and of the interval, while this is not the case significance testing (hypothesis testing) for a significance test.
    [Show full text]
  • Interval Estimation for Drop-The-Losers Designs
    Biometrika (2010), 97,2,pp. 405–418 doi: 10.1093/biomet/asq003 C 2010 Biometrika Trust Advance Access publication 14 April 2010 Printed in Great Britain Interval estimation for drop-the-losers designs BY SAMUEL S. WU Department of Epidemiology and Health Policy Research, University of Florida, Gainesville, Florida 32610, U.S.A. [email protected]fl.edu WEIZHEN WANG Downloaded from Department of Mathematics and Statistics, Wright State University, Dayton, Ohio 45435, U.S.A. [email protected] AND MARK C. K. YANG Department of Statistics, University of Florida, Gainesville, Florida 32610, U.S.A. http://biomet.oxfordjournals.org [email protected]fl.edu SUMMARY In the first stage of a two-stage, drop-the-losers design, a candidate for the best treatment is selected. At the second stage, additional observations are collected to decide whether the candidate is actually better than the control. The design also allows the investigator to stop the trial for ethical reasons at the end of the first stage if there is already strong evidence of futility or superiority. Two types of tests have recently been developed, one based on the combined means and the other based on the combined p-values, but corresponding inter- at University of Florida on May 27, 2010 val estimators are unavailable except in special cases. The problem is that, in most cases, the interval estimators depend on the mean configuration of all treatments in the first stage, which is unknown. In this paper, we prove a basic stochastic ordering lemma that enables us to bridge the gap between hypothesis testing and interval estimation.
    [Show full text]
  • Interval Estimation, Point Estimation, and Null Hypothesis Significance Testing Calibrated by an Estimated Posterior Probability of the Null Hypothesis David R
    Interval estimation, point estimation, and null hypothesis significance testing calibrated by an estimated posterior probability of the null hypothesis David R. Bickel To cite this version: David R. Bickel. Interval estimation, point estimation, and null hypothesis significance testing cali- brated by an estimated posterior probability of the null hypothesis. 2020. hal-02496126 HAL Id: hal-02496126 https://hal.archives-ouvertes.fr/hal-02496126 Preprint submitted on 2 Mar 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Interval estimation, point estimation, and null hypothesis significance testing calibrated by an estimated posterior probability of the null hypothesis March 2, 2020 David R. Bickel Ottawa Institute of Systems Biology Department of Biochemistry, Microbiology and Immunology Department of Mathematics and Statistics University of Ottawa 451 Smyth Road Ottawa, Ontario, K1H 8M5 +01 (613) 562-5800, ext. 8670 [email protected] Abstract Much of the blame for failed attempts to replicate reports of scientific findings has been placed on ubiquitous and persistent misinterpretations of the p value. An increasingly popular solution is to transform a two-sided p value to a lower bound on a Bayes factor. Another solution is to interpret a one-sided p value as an approximate posterior probability.
    [Show full text]