Statistical Analysis and Optimal Design for Cluster Randomized Trials
Total Page:16
File Type:pdf, Size:1020Kb
Load more
										Recommended publications
									
								- 
												  Choosing the SampleCHAPTER IV CHOOSING THE SAMPLE This chapter is written for survey coordinators and technical resource persons. It will enable you to: U Understand the basic concepts of sampling. U Calculate the required sample size for national and subnational estimates. U Determine the number of clusters to be used. U Choose a sampling scheme. UNDERSTANDING THE BASIC CONCEPTS OF SAMPLING In the context of multiple-indicator surveys, sampling is a process for selecting respondents from a population. In our case, the respondents will usually be the mothers, or caretakers, of children in each household visited,1 who will answer all of the questions in the Child Health modules. The Water and Sanitation and Salt Iodization modules refer to the whole household and may be answered by any adult. Questions in these modules are asked even where there are no children under the age of 15 years. In principle, our survey could cover all households in the population. If all mothers being interviewed could provide perfect answers, we could measure all indicators with complete accuracy. However, interviewing all mothers would be time-consuming, expensive and wasteful. It is therefore necessary to interview a sample of these women to obtain estimates of the actual indicators. The difference between the estimate and the actual indicator is called sampling error. Sampling errors are caused by the fact that a sample&and not the entire population&is surveyed. Sampling error can be minimized by taking certain precautions: 3 Choose your sample of respondents in an unbiased way. 3 Select a large enough sample for your estimates to be precise.
- 
												![Harmonizing Fully Optimal Designs with Classic Randomization in Fixed Trial Experiments Arxiv:1810.08389V1 [Stat.ME] 19 Oct 20](https://docslib.b-cdn.net/cover/7263/harmonizing-fully-optimal-designs-with-classic-randomization-in-fixed-trial-experiments-arxiv-1810-08389v1-stat-me-19-oct-20-167263.webp)  Harmonizing Fully Optimal Designs with Classic Randomization in Fixed Trial Experiments Arxiv:1810.08389V1 [Stat.ME] 19 Oct 20Harmonizing Fully Optimal Designs with Classic Randomization in Fixed Trial Experiments Adam Kapelner, Department of Mathematics, Queens College, CUNY, Abba M. Krieger, Department of Statistics, The Wharton School of the University of Pennsylvania, Uri Shalit and David Azriel Faculty of Industrial Engineering and Management, The Technion October 22, 2018 Abstract There is a movement in design of experiments away from the classic randomization put forward by Fisher, Cochran and others to one based on optimization. In fixed-sample trials comparing two groups, measurements of subjects are known in advance and subjects can be divided optimally into two groups based on a criterion of homogeneity or \imbalance" between the two groups. These designs are far from random. This paper seeks to understand the benefits and the costs over classic randomization in the context of different performance criterions such as Efron's worst-case analysis. In the criterion that we motivate, randomization beats optimization. However, the optimal arXiv:1810.08389v1 [stat.ME] 19 Oct 2018 design is shown to lie between these two extremes. Much-needed further work will provide a procedure to find this optimal designs in different scenarios in practice. Until then, it is best to randomize. Keywords: randomization, experimental design, optimization, restricted randomization 1 1 Introduction In this short survey, we wish to investigate performance differences between completely random experimental designs and non-random designs that optimize for observed covariate imbalance. We demonstrate that depending on how we wish to evaluate our estimator, the optimal strategy will change. We motivate a performance criterion that when applied, does not crown either as the better choice, but a design that is a harmony between the two of them.
- 
												  Optimal Design of Experiments Session 4: Some TheoryOptimal design of experiments Session 4: Some theory Peter Goos 1 / 40 Optimal design theory Ï continuous or approximate optimal designs Ï implicitly assume an infinitely large number of observations are available Ï is mathematically convenient Ï exact or discrete designs Ï finite number of observations Ï fewer theoretical results 2 / 40 Continuous versus exact designs continuous Ï ½ ¾ x1 x2 ... xh Ï » Æ w1 w2 ... wh Ï x1,x2,...,xh: design points or support points w ,w ,...,w : weights (w 0, P w 1) Ï 1 2 h i ¸ i i Æ Ï h: number of different points exact Ï ½ ¾ x1 x2 ... xh Ï » Æ n1 n2 ... nh Ï n1,n2,...,nh: (integer) numbers of observations at x1,...,xn P n n Ï i i Æ Ï h: number of different points 3 / 40 Information matrix Ï all criteria to select a design are based on information matrix Ï model matrix 2 3 2 T 3 1 1 1 1 1 1 f (x1) ¡ ¡ Å Å Å T 61 1 1 1 1 17 6f (x2)7 6 Å ¡ ¡ Å Å 7 6 T 7 X 61 1 1 1 1 17 6f (x3)7 6 ¡ Å ¡ Å Å 7 6 7 Æ 6 . 7 Æ 6 . 7 4 . 5 4 . 5 T 100000 f (xn) " " " " "2 "2 I x1 x2 x1x2 x1 x2 4 / 40 Information matrix Ï (total) information matrix 1 1 Xn M XT X f(x )fT (x ) 2 2 i i Æ σ Æ σ i 1 Æ Ï per observation information matrix 1 T f(xi)f (xi) σ2 5 / 40 Information matrix industrial example 2 3 11 0 0 0 6 6 6 0 6 0 0 0 07 6 7 1 6 0 0 6 0 0 07 ¡XT X¢ 6 7 2 6 7 σ Æ 6 0 0 0 4 0 07 6 7 4 6 0 0 0 6 45 6 0 0 0 4 6 6 / 40 Information matrix Ï exact designs h X T M ni f(xi)f (xi) Æ i 1 Æ where h = number of different points ni = number of replications of point i Ï continuous designs h X T M wi f(xi)f (xi) Æ i 1 Æ 7 / 40 D-optimality criterion Ï seeks designs that minimize variance-covariance matrix of ¯ˆ ¯ 2 T 1¯ Ï .
- 
												  Computing Effect Sizes for Clustered Randomized TrialsComputing Effect Sizes for Clustered Randomized Trials Terri Pigott, C2 Methods Editor & Co-Chair Professor, Loyola University Chicago [email protected] The Campbell Collaboration www.campbellcollaboration.org Computing effect sizes in clustered trials • In an experimental study, we are interested in the difference in performance between the treatment and control group • In this case, we use the standardized mean difference, given by YYTC− d = gg Control group sp mean Treatment group mean Pooled sample standard deviation Campbell Collaboration Colloquium – August 2011 www.campbellcollaboration.org 1 Variance of the standardized mean difference NNTC+ d2 Sd2 ()=+ NNTC2( N T+ N C ) where NT is the sample size for the treatment group, and NC is the sample size for the control group Campbell Collaboration Colloquium – August 2011 www.campbellcollaboration.org TREATMENT GROUP CONTROL GROUP TT T CC C YY,,..., Y YY12,,..., YC 12 NT N Overall Trt Mean Overall Cntl Mean T Y C Yg g S 2 2 Trt SCntl 2 S pooled Campbell Collaboration Colloquium – August 2011 www.campbellcollaboration.org 2 In cluster randomized trials, SMD more complex • In cluster randomized trials, we have clusters such as schools or clinics randomized to treatment and control • We have at least two means: mean performance for each cluster, and the overall group mean • We also have several components of variance – the within- cluster variance, the variance between cluster means, and the total variance • Next slide is an illustration Campbell Collaboration Colloquium – August 2011 www.campbellcollaboration.org TREATMENT GROUP CONTROL GROUP Cntl Cluster mC Trt Cluster 1 Trt Cluster mT Cntl Cluster 1 TT T T CC C C YY,...ggg Y ,..., Y YY11,..
- 
												  Evaluating Probability Sampling Strategies for Estimating Redd Counts: an Example with Chinook Salmon (Oncorhynchus Tshawytscha)1814 Evaluating probability sampling strategies for estimating redd counts: an example with Chinook salmon (Oncorhynchus tshawytscha) Jean-Yves Courbois, Stephen L. Katz, Daniel J. Isaak, E. Ashley Steel, Russell F. Thurow, A. Michelle Wargo Rub, Tony Olsen, and Chris E. Jordan Abstract: Precise, unbiased estimates of population size are an essential tool for fisheries management. For a wide variety of salmonid fishes, redd counts from a sample of reaches are commonly used to monitor annual trends in abundance. Using a 9-year time series of georeferenced censuses of Chinook salmon (Oncorhynchus tshawytscha) redds from central Idaho, USA, we evaluated a wide range of common sampling strategies for estimating the total abundance of redds. We evaluated two sampling-unit sizes (200 and 1000 m reaches), three sample proportions (0.05, 0.10, and 0.29), and six sampling strat- egies (index sampling, simple random sampling, systematic sampling, stratified sampling, adaptive cluster sampling, and a spatially balanced design). We evaluated the strategies based on their accuracy (confidence interval coverage), precision (relative standard error), and cost (based on travel time). Accuracy increased with increasing number of redds, increasing sample size, and smaller sampling units. The total number of redds in the watershed and budgetary constraints influenced which strategies were most precise and effective. For years with very few redds (<0.15 reddsÁkm–1), a stratified sampling strategy and inexpensive strategies were most efficient, whereas for years with more redds (0.15–2.9 reddsÁkm–1), either of two more expensive systematic strategies were most precise. Re´sume´ : La gestion des peˆches requiert comme outils essentiels des estimations pre´cises et non fausse´es de la taille des populations.
- 
												  Advanced Power Analysis WorkshopAdvanced Power Analysis Workshop www.nicebread.de PD Dr. Felix Schönbrodt & Dr. Stella Bollmann www.researchtransparency.org Ludwig-Maximilians-Universität München Twitter: @nicebread303 •Part I: General concepts of power analysis •Part II: Hands-on: Repeated measures ANOVA and multiple regression •Part III: Power analysis in multilevel models •Part IV: Tailored design analyses by simulations in 2 Part I: General concepts of power analysis •What is “statistical power”? •Why power is important •From power analysis to design analysis: Planning for precision (and other stuff) •How to determine the expected/minimally interesting effect size 3 What is statistical power? A 2x2 classification matrix Reality: Reality: Effect present No effect present Test indicates: True Positive False Positive Effect present Test indicates: False Negative True Negative No effect present 5 https://effectsizefaq.files.wordpress.com/2010/05/type-i-and-type-ii-errors.jpg 6 https://effectsizefaq.files.wordpress.com/2010/05/type-i-and-type-ii-errors.jpg 7 A priori power analysis: We assume that the effect exists in reality Reality: Reality: Effect present No effect present Power = α=5% Test indicates: 1- β p < .05 True Positive False Positive Test indicates: β=20% p > .05 False Negative True Negative 8 Calibrate your power feeling total n Two-sample t test (between design), d = 0.5 128 (64 each group) One-sample t test (within design), d = 0.5 34 Correlation: r = .21 173 Difference between two correlations, 428 r₁ = .15, r₂ = .40 ➙ q = 0.273 ANOVA, 2x2 Design: Interaction effect, f = 0.21 180 (45 each group) All a priori power analyses with α = 5%, β = 20% and two-tailed 9 total n Two-sample t test (between design), d = 0.5 128 (64 each group) One-sample t test (within design), d = 0.5 34 Correlation: r = .21 173 Difference between two correlations, 428 r₁ = .15, r₂ = .40 ➙ q = 0.273 ANOVA, 2x2 Design: Interaction effect, f = 0.21 180 (45 each group) All a priori power analyses with α = 5%, β = 20% and two-tailed 10 The power ofwithin-SS designs Thepower 11 May, K., & Hittner, J.
- 
												  Cluster SamplingDay 5 sampling - clustering SAMPLE POPULATION SAMPLING: IS ESTIMATING THE CHARACTERISTICS OF THE WHOLE POPULATION USING INFORMATION COLLECTED FROM A SAMPLE GROUP. The sampling process comprises several stages: •Defining the population of concern •Specifying a sampling frame, a set of items or events possible to measure •Specifying a sampling method for selecting items or events from the frame •Determining the sample size •Implementing the sampling plan •Sampling and data collecting 2 Simple random sampling 3 In a simple random sample (SRS) of a given size, all such subsets of the frame are given an equal probability. In particular, the variance between individual results within the sample is a good indicator of variance in the overall population, which makes it relatively easy to estimate the accuracy of results. SRS can be vulnerable to sampling error because the randomness of the selection may result in a sample that doesn't reflect the makeup of the population. Systematic sampling 4 Systematic sampling (also known as interval sampling) relies on arranging the study population according to some ordering scheme and then selecting elements at regular intervals through that ordered list Systematic sampling involves a random start and then proceeds with the selection of every kth element from then onwards. In this case, k=(population size/sample size). It is important that the starting point is not automatically the first in the list, but is instead randomly chosen from within the first to the kth element in the list. STRATIFIED SAMPLING 5 WHEN THE POPULATION EMBRACES A NUMBER OF DISTINCT CATEGORIES, THE FRAME CAN BE ORGANIZED BY THESE CATEGORIES INTO SEPARATE "STRATA." EACH STRATUM IS THEN SAMPLED AS AN INDEPENDENT SUB-POPULATION, OUT OF WHICH INDIVIDUAL ELEMENTS CAN BE RANDOMLY SELECTED Cluster sampling Sometimes it is more cost-effective to select respondents in groups ('clusters') Quota sampling Minimax sampling Accidental sampling Voluntary Sampling ….
- 
												  Approximation Algorithms for D-Optimal DesignApproximation Algorithms for D-optimal Design Mohit Singh School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, GA 30332, [email protected]. Weijun Xie Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA 24061, [email protected]. Experimental design is a classical statistics problem and its aim is to estimate an unknown m-dimensional vector β from linear measurements where a Gaussian noise is introduced in each measurement. For the com- binatorial experimental design problem, the goal is to pick k out of the given n experiments so as to make the most accurate estimate of the unknown parameters, denoted as βb. In this paper, we will study one of the most robust measures of error estimation - D-optimality criterion, which corresponds to minimizing the volume of the confidence ellipsoid for the estimation error β − βb. The problem gives rise to two natural vari- ants depending on whether repetitions of experiments are allowed or not. We first propose an approximation 1 algorithm with a e -approximation for the D-optimal design problem with and without repetitions, giving the first constant factor approximation for the problem. We then analyze another sampling approximation algo- 4m 12 1 rithm and prove that it is (1 − )-approximation if k ≥ + 2 log( ) for any 2 (0; 1). Finally, for D-optimal design with repetitions, we study a different algorithm proposed by literature and show that it can improve this asymptotic approximation ratio. Key words: D-optimal Design; approximation algorithm; determinant; derandomization. History: 1. Introduction Experimental design is a classical problem in statistics [5, 17, 18, 23, 31] and recently has also been applied to machine learning [2, 39].
- 
												  A Review on Optimal Experimental DesignA Review On Optimal Experimental Design Noha A. Youssef London School Of Economics 1 Introduction Finding an optimal experimental design is considered one of the most impor- tant topics in the context of the experimental design. Optimal design is the design that achieves some targets of our interest. The Bayesian and the Non- Bayesian approaches have introduced some criteria that coincide with the target of the experiment based on some speci¯c utility or loss functions. The choice between the Bayesian and Non-Bayesian approaches depends on the availability of the prior information and also the computational di±culties that might be encountered when using any of them. This report aims mainly to provide a short summary on optimal experi- mental design focusing more on Bayesian optimal one. Therefore a background about the Bayesian analysis was given in Section 2. Optimality criteria for non-Bayesian design of experiments are reviewed in Section 3 . Section 4 il- lustrates how the Bayesian analysis is employed in the design of experiments. The remaining sections of this report give a brief view of the paper written by (Chalenor & Verdinelli 1995). An illustration for the Bayesian optimality criteria for normal linear model associated with di®erent objectives is given in Section 5. Also, some problems related to Bayesian optimal designs for nor- mal linear model are discussed in Section 5. Section 6 presents some ideas for Bayesian optimal design for one way and two way analysis of variance. Section 7 discusses the problems associated with nonlinear models and presents some ideas for solving these problems.
- 
												  Introduction to Survey Sampling and Analysis Procedures (Chapter)SAS/STAT® 9.3 User’s Guide Introduction to Survey Sampling and Analysis Procedures (Chapter) SAS® Documentation This document is an individual chapter from SAS/STAT® 9.3 User’s Guide. The correct bibliographic citation for the complete manual is as follows: SAS Institute Inc. 2011. SAS/STAT® 9.3 User’s Guide. Cary, NC: SAS Institute Inc. Copyright © 2011, SAS Institute Inc., Cary, NC, USA All rights reserved. Produced in the United States of America. For a Web download or e-book: Your use of this publication shall be governed by the terms established by the vendor at the time you acquire this publication. The scanning, uploading, and distribution of this book via the Internet or any other means without the permission of the publisher is illegal and punishable by law. Please purchase only authorized electronic editions and do not participate in or encourage electronic piracy of copyrighted materials. Your support of others’ rights is appreciated. U.S. Government Restricted Rights Notice: Use, duplication, or disclosure of this software and related documentation by the U.S. government is subject to the Agreement with SAS Institute and the restrictions set forth in FAR 52.227-19, Commercial Computer Software-Restricted Rights (June 1987). SAS Institute Inc., SAS Campus Drive, Cary, North Carolina 27513. 1st electronic book, July 2011 SAS® Publishing provides a complete selection of books and electronic products to help customers use SAS software to its fullest potential. For more information about our e-books, e-learning products, CDs, and hard-copy books, visit the SAS Publishing Web site at support.sas.com/publishing or call 1-800-727-3228.
- 
												![Sampling and Household Listing Manual [DHSM4]](https://docslib.b-cdn.net/cover/5729/sampling-and-household-listing-manual-dhsm4-1365729.webp)  Sampling and Household Listing Manual [DHSM4]SAMPLING AND HOUSEHOLD LISTING MANuaL Demographic and Health Surveys Methodology This document is part of the Demographic and Health Survey’s DHS Toolkit of methodology for the MEASURE DHS Phase III project, implemented from 2008-2013. This publication was produced for review by the United States Agency for International Development (USAID). It was prepared by MEASURE DHS/ICF International. [THIS PAGE IS INTENTIONALLY BLANK] Demographic and Health Survey Sampling and Household Listing Manual ICF International Calverton, Maryland USA September 2012 MEASURE DHS is a five-year project to assist institutions in collecting and analyzing data needed to plan, monitor, and evaluate population, health, and nutrition programs. MEASURE DHS is funded by the U.S. Agency for International Development (USAID). The project is implemented by ICF International in Calverton, Maryland, in partnership with the Johns Hopkins Bloomberg School of Public Health/Center for Communication Programs, the Program for Appropriate Technology in Health (PATH), Futures Institute, Camris International, and Blue Raster. The main objectives of the MEASURE DHS program are to: 1) provide improved information through appropriate data collection, analysis, and evaluation; 2) improve coordination and partnerships in data collection at the international and country levels; 3) increase host-country institutionalization of data collection capacity; 4) improve data collection and analysis tools and methodologies; and 5) improve the dissemination and utilization of data. For information about the Demographic and Health Surveys (DHS) program, write to DHS, ICF International, 11785 Beltsville Drive, Suite 300, Calverton, MD 20705, U.S.A. (Telephone: 301-572- 0200; fax: 301-572-0999; e-mail: [email protected]; Internet: http://www.measuredhs.com).
- 
												  Taylor's Power Law and Fixed-Precision Sampling87 ARTICLE Taylor’s power law and fixed-precision sampling: application to abundance of fish sampled by gillnets in an African lake Meng Xu, Jeppe Kolding, and Joel E. Cohen Abstract: Taylor’s power law (TPL) describes the variance of population abundance as a power-law function of the mean abundance for a single or a group of species. Using consistently sampled long-term (1958–2001) multimesh capture data of Lake Kariba in Africa, we showed that TPL robustly described the relationship between the temporal mean and the temporal variance of the captured fish assemblage abundance (regardless of species), separately when abundance was measured by numbers of individuals and by aggregate weight. The strong correlation between the mean of abundance and the variance of abundance was not altered after adding other abiotic or biotic variables into the TPL model. We analytically connected the parameters of TPL when abundance was measured separately by the aggregate weight and by the aggregate number, using a weight–number scaling relationship. We utilized TPL to find the number of samples required for fixed-precision sampling and compared the number of samples when sampling was performed with a single gillnet mesh size and with multiple mesh sizes. These results facilitate optimizing the sampling design to estimate fish assemblage abundance with specified precision, as needed in stock management and conservation. Résumé : La loi de puissance de Taylor (LPT) décrit la variance de l’abondance d’une population comme étant une fonction de puissance de l’abondance moyenne pour une espèce ou un groupe d’espèces. En utilisant des données de capture sur une longue période (1958–2001) obtenues de manière cohérente avec des filets de mailles variées dans le lac Kariba en Afrique, nous avons démontré que la LPT décrit de manière robuste la relation entre la moyenne temporelle et la variance temporelle de l’abondance de l’assemblage de poissons capturés (peu importe l’espèce), que l’abondance soit mesurée sur la base du nombre d’individus ou de la masse cumulative.