HST 190: Introduction to Biostatistics

Total Page:16

File Type:pdf, Size:1020Kb

HST 190: Introduction to Biostatistics HST 190: Introduction to Biostatistics Lecture 3: Two-sample comparisons, hypothesis testing and power/sample size calculations 1 HST 190: Intro to Biostatistics Estimating two-sample comparisons • In Lecture 2, we introduced: § probability distributions § sampling from a population § estimation of the mean of some random variable § quantification of uncertainty via standard errors and confidence intervals • Today, we focus on a common setting in public health and medical research, specifically the comparison of two groups • Consider, for example, the question: “What is the difference in average birthweight between children born at a specific hospital in a disadvantaged area and children born at a specific hospital in a wealthy area?” 2 HST 190: Intro to Biostatistics • This question moves beyond the estimation scenario we discussed in Lecture 2 § that is, interest lies in more than a single mean � • Now, we want to compare two unknown means § each corresponding to some (sub-)population • How we achieve this, specifically estimation and the quantification of uncertainty, depends on how the data we have at our disposal was collected. 3 HST 190: Intro to Biostatistics Paired samples • Solish et al (2010, Clinical Ophthalmology) reported on a study of medical therapy for uncontrolled glaucoma § � = 55 patients § bimatoprost 0.03% once daily in one eye § either travoprost 0.004% or latanoprost 0.005% once daily in the fellow eye § measured compared intraocular pressure (IOP) at baseline and at 4- 6 weeks follow-up • This is an example of a paired study § � is number of pairs not number of total observations 4 HST 190: Intro to Biostatistics • Let � denote the change in IOP between baseline and follow-up • Measurements in each of the two eyes: § �$$, �&$, … , �)$ § �$&, �&&, … , �)& • Solish et al report estimated mean reductions in IOP of: § �$̅ = 2.7 mmHg among bimatoprost-treated eyes § �̅&= 1.7 mmHg among travoprost-treated eyes • Suggests that treatment with bimatoprost is associated with greater improvements • Q: How can we use the methods we’ve been exposed to so far to formally assess this comparison? 5 HST 190: Intro to Biostatistics • Define a new random variable � to be the difference in the IOP change between the two eyes: § �- = �-& – �-$ • Thus, we have exploited the special nature of the paired data to turn the two-sample problem into a one-sample § E D = Δ & § ��� � = �7 • We can now proceed just as in Lecture 2: $ § use �̅ = ∑) � sample mean of the differences to estimate Δ ) -:$ - ? $ & § use � = ∑) � − �̅ to estimate the standard deviation < )=$ -:$ - ̅ ̅ ? o note, this means that the standard error of � is sd � = �</ � <C=D § use ? ~�)=$ to derive 100(1 − �)% confidence interval for Δ EF/ ) � ̅ < � ± � P ⋅ ? )=$,$=& � 6 HST 190: Intro to Biostatistics Unpaired samples • In many settings there may not be an obvious path for creating “pairs” on which to base the two-sample comparison • As an example, Maciejewski et al (2016, JAMA) report on weight loss following bariatric surgery: § 1787 veterans who underwent RYGB § 5305 non-surgical patients from the same VA populationResearch Original Investigation Bariatric Surgery and Long-term Weight Loss Durability Statistical Analysis Figure 1. Differences in Estimated Weight Changes Among Patients Undergoing Roux-en-Y Gastric Bypass (RYGB) and Nonsurgical Matches The statistical analysis was developed to estimate mean tra- jectories of the percentage of weight change and estimates at Patients undergoing RYGB specific postsurgical years. Penalized spline mixed-effects 10 Nonsurgical matches models were used, with piecewise linear functions with pre- specified knots included as fixed and random effects at the 0 population-average and subject-specific levels.31,32 This ap- proach allowed us to estimate overall mean trajectories and –10 individual trajectories with enough flexibility to reflect weight 30 27 24 22 21 (29-30) (25-29) (20-28) (16-28) (11-31) change fluctuations during an individual’s entire follow-up –20 (eFigure 3 in the Supplement illustrates 75 cases). Because of the extent of weight change in the 2 years after –30 surgery,our primary model specification for the overall mean in- Weight Change From Baseline, % Change From Weight cluded a linear term for time and knots at 3 months, 6 months, –40 and 1, 2, 3, 5, 7,and9 years. All time terms were interacted with 0 1 2 3 4 5 6 7 8 9 10 11 atreatmentindicatortoallowdifferentialestimatesforpatients Years Since Baseline undergoing RYGB and nonsurgical matches using HPMIXED in SAS statistical software, version 9.4(SAS Institute Inc). The model 7 Estimated values, differences, and 95% CIs (shown in parentheses) were HST 190: Intro to Biostatistics generated from a penalized spline mixed-effects model (7092 patients: 5305 also included individual-level random effects with knots at 6 nonsurgical matches and 1787 patients undergoing RYGB). Numbers and arrows months and 1, 2, 3, 5, 7,and9 years;asubject-levelrandomslope in the center of the figure represent the differences and 95% CIs of the to account for correlation among patients’ repeated measures; differences between nonsurgical matches and patients undergoing RYGB at and a VA geographic area–level random slope to account for cor- years 1, 3, 5, 7, and 10. The sample for whom there was follow-up weight data for each year are as follows: year 1, n = 6894 patients (5131 nonsurgical matches relation between each pair matched within VA geographic area and 1763 patients undergoing RYGB); year 3, n = 6301 (4629 nonsurgical and patients matching more than once. Estimate statements were matches and 1672 patients undergoing RYGB); year 5, n = 5172 (3748 used to generate the predicted percentage of weight change at nonsurgical matches and 1424 patients undergoing RYGB); year 7, n = 3942 (2806 nonsurgical matches and 1136 patients undergoing RYGB); and year 10, 1, 3, 5, 7 and 10 years of follow-up for the patients undergoing n = 1847 (1274 nonsurgical matches and 573 patients undergoing RYGB). RYGBcomparedwiththenonsurgicalmatches.Themodel’sem- pirical Bayes estimates were used to calculate patients’ predicted weights at 1,3,5,7,and10 yearsonlyforthosewithaweightmea- surement within a 12-month interval of (6 months before to 6 (n = 61), who only had inpatient weights after surgery (n = 1), months after) and at least 1 measurement after each follow-up who lacked postsurgical weights (n = 6), or who did not have year,whichwereusedtoclassifyeachpatientashavinglostless a weight measurement within 5 years after surgery (n = 2 pa- than 5%, 5%ormore,10%ormore,20%ormore,and30%or tients undergoing RYGB). This resulted in a final sample of 1785 more of their baseline weight. The 12-month interval requirement patients undergoing RYGB, 246 undergoing AGB, and 379 un- modestly reduced the number of patients for whom individual dergoing SG (eFigure 2 in the Supplement). weights could be predicted. For example, 573 surgical patients were followed up for at least 10 years, but only 564 of these 573 Definition and Cleaning of Weight Outcome had a weight measurement between 9.5and 10.5years. Patients’ Follow-up weight data were obtained from measurements re- predicted weights at 10 years of follow-up were used to calcu- corded in the electronic health records during outpatient vis- late mean weight loss and excess body weight loss from baseline. its from January 1, 2000,throughDecember31, 2014.Thepri- Asimilarmodelwithknotsspecifiedat3 months, 6 months, mary outcome was percentage change in weight at follow-up and 1,2,and3yearswasusedtocompareweightchangethrough compared with baseline, which is less confounded by base- 4yearsoffollow-upacrosssurgicalprocedures,whichhadshorter line BMI than other commonly reported measures (eg, per- follow-up because of the recent introduction of the SG procedure centage of excess weight loss).30 Baseline weight data on the and study end date of December 31,2014 (eFigure 2in the Supple- surgery day was available for 78.6%ofpatientsundergoing ment). To understand the effect of baseline differences in patients RYGB but only 2.2%ofnonsurgicalmatches,andthemean(SD) who underwent each surgical procedure, patients’ predicted per- number of days for the baseline weight measurement before centage of weight loss at 4 years was regressed on surgery type surgery was 7.3 (24.0) for patients undergoing RYGB and 59.8 while adjusting for baseline BMI, comorbidity score, demograph- (48.6) for nonsurgical matches. ics, and diabetes status (eTable in the Supplement). Postsurgical weight measurements are highly variable and have a nonlinear trend, so we developed a multistep outlier- detection algorithm that examined the SD of consecutively Results measured weights (eAppendix in the Supplement). An algo- rithm to identify multiple weight measures on the same day, Characteristics of Patients Undergoing RYGB or weights that deviated from clinically plausible trends over and Nonsurgical Matches time, excluded a small proportion (4201 of 89 757 [4.7%]) of Patients undergoing RYGB had a mean (SD) age of 52.1 (8.5) weight measurements. yearsandnonsurgicalmatcheshadamean(SD)ageof 52.2 (8.4) E4 JAMA Surgery Published online August 31, 2016 (Reprinted) jamasurgery.com Copyright 2016 American Medical Association. All rights reserved. Downloaded From: http://archsurg.jamanetwork.com/ by a Department of Veterans Affairs User on 08/31/2016 • More generally, consider two independent (unpaired) samples of size �$ and �& • Suppose: & § �$$, �&$, … , �)R$ ~ ������ �$, �$ & § �$&, �&&, … ,
Recommended publications
  • Hypothesis Testing and Likelihood Ratio Tests
    Hypottthesiiis tttestttiiing and llliiikellliiihood ratttiiio tttesttts Y We will adopt the following model for observed data. The distribution of Y = (Y1, ..., Yn) is parameter considered known except for some paramett er ç, which may be a vector ç = (ç1, ..., çk); ç“Ç, the paramettter space. The parameter space will usually be an open set. If Y is a continuous random variable, its probabiiillliiittty densiiittty functttiiion (pdf) will de denoted f(yy;ç) . If Y is y probability mass function y Y y discrete then f(yy;ç) represents the probabii ll ii tt y mass functt ii on (pmf); f(yy;ç) = Pç(YY=yy). A stttatttiiistttiiicalll hypottthesiiis is a statement about the value of ç. We are interested in testing the null hypothesis H0: ç“Ç0 versus the alternative hypothesis H1: ç“Ç1. Where Ç0 and Ç1 ¶ Ç. hypothesis test Naturally Ç0 § Ç1 = ∅, but we need not have Ç0 ∞ Ç1 = Ç. A hypott hesii s tt estt is a procedure critical region for deciding between H0 and H1 based on the sample data. It is equivalent to a crii tt ii call regii on: a critical region is a set C ¶ Rn y such that if y = (y1, ..., yn) “ C, H0 is rejected. Typically C is expressed in terms of the value of some tttesttt stttatttiiistttiiic, a function of the sample data. For µ example, we might have C = {(y , ..., y ): y – 0 ≥ 3.324}. The number 3.324 here is called a 1 n s/ n µ criiitttiiicalll valllue of the test statistic Y – 0 . S/ n If y“C but ç“Ç 0, we have committed a Type I error.
    [Show full text]
  • Data 8 Final Stats Review
    Data 8 Final Stats review I. Hypothesis Testing Purpose: To answer a question about a process or the world by testing two hypotheses, a null and an alternative. Usually the null hypothesis makes a statement that “the world/process works this way”, and the alternative hypothesis says “the world/process does not work that way”. Examples: Null: “The customer was not cheating-his chances of winning and losing were like random tosses of a fair coin-50% chance of winning, 50% of losing. Any variation from what we expect is due to chance variation.” Alternative: “The customer was cheating-his chances of winning were something other than 50%”. Pro tip: You must be very precise about chances in your hypotheses. Hypotheses such as “the customer cheated” or “Their chances of winning were normal” are vague and might be considered incorrect, because you don’t state the exact chances associated with the events. Pro tip: Null hypothesis should also explain differences in the data. For example, if your hypothesis stated that the coin was fair, then why did you get 70 heads out of 100 flips? Since it’s possible to get that many (though not expected), your null hypothesis should also contain a statement along the lines of “Any difference in outcome from what we expect is due to chance variation”. Steps: 1) Precisely state your null and alternative hypotheses. 2) Decide on a test statistic (think of it as a general formula) to help you either reject or fail to reject the null hypothesis. • If your data is categorical, a good test statistic might be the Total Variation Distance (TVD) between your sample and the distribution it was drawn from.
    [Show full text]
  • Use of Statistical Tables
    TUTORIAL | SCOPE USE OF STATISTICAL TABLES Lucy Radford, Jenny V Freeman and Stephen J Walters introduce three important statistical distributions: the standard Normal, t and Chi-squared distributions PREVIOUS TUTORIALS HAVE LOOKED at hypothesis testing1 and basic statistical tests.2–4 As part of the process of statistical hypothesis testing, a test statistic is calculated and compared to a hypothesised critical value and this is used to obtain a P- value. This P-value is then used to decide whether the study results are statistically significant or not. It will explain how statistical tables are used to link test statistics to P-values. This tutorial introduces tables for three important statistical distributions (the TABLE 1. Extract from two-tailed standard Normal, t and Chi-squared standard Normal table. Values distributions) and explains how to use tabulated are P-values corresponding them with the help of some simple to particular cut-offs and are for z examples. values calculated to two decimal places. STANDARD NORMAL DISTRIBUTION TABLE 1 The Normal distribution is widely used in statistics and has been discussed in z 0.00 0.01 0.02 0.03 0.050.04 0.05 0.06 0.07 0.08 0.09 detail previously.5 As the mean of a Normally distributed variable can take 0.00 1.0000 0.9920 0.9840 0.9761 0.9681 0.9601 0.9522 0.9442 0.9362 0.9283 any value (−∞ to ∞) and the standard 0.10 0.9203 0.9124 0.9045 0.8966 0.8887 0.8808 0.8729 0.8650 0.8572 0.8493 deviation any positive value (0 to ∞), 0.20 0.8415 0.8337 0.8259 0.8181 0.8103 0.8206 0.7949 0.7872 0.7795 0.7718 there are an infinite number of possible 0.30 0.7642 0.7566 0.7490 0.7414 0.7339 0.7263 0.7188 0.7114 0.7039 0.6965 Normal distributions.
    [Show full text]
  • Receiver Operating Characteristic (ROC) Curve: Practical Review for Radiologists
    Receiver Operating Characteristic (ROC) Curve: Practical Review for Radiologists Seong Ho Park, MD1 The receiver operating characteristic (ROC) curve, which is defined as a plot of Jin Mo Goo, MD1 test sensitivity as the y coordinate versus its 1-specificity or false positive rate Chan-Hee Jo, PhD2 (FPR) as the x coordinate, is an effective method of evaluating the performance of diagnostic tests. The purpose of this article is to provide a nonmathematical introduction to ROC analysis. Important concepts involved in the correct use and interpretation of this analysis, such as smooth and empirical ROC curves, para- metric and nonparametric methods, the area under the ROC curve and its 95% confidence interval, the sensitivity at a particular FPR, and the use of a partial area under the ROC curve are discussed. Various considerations concerning the collection of data in radiological ROC studies are briefly discussed. An introduc- tion to the software frequently used for performing ROC analyses is also present- ed. he receiver operating characteristic (ROC) curve, which is defined as a Index terms: plot of test sensitivity as the y coordinate versus its 1-specificity or false Diagnostic radiology positive rate (FPR) as the x coordinate, is an effective method of evaluat- Receiver operating characteristic T (ROC) curve ing the quality or performance of diagnostic tests, and is widely used in radiology to Software reviews evaluate the performance of many radiological tests. Although one does not necessar- Statistical analysis ily need to understand the complicated mathematical equations and theories of ROC analysis, understanding the key concepts of ROC analysis is a prerequisite for the correct use and interpretation of the results that it provides.
    [Show full text]
  • 8.5 Testing a Claim About a Standard Deviation Or Variance
    8.5 Testing a Claim about a Standard Deviation or Variance Testing Claims about a Population Standard Deviation or a Population Variance ² Uses the chi-squared distribution from section 7-4 → Requirements: 1. The sample is a simple random sample 2. The population has a normal distribution (n −1)s 2 → Test Statistic for Testing a Claim about or ²: 2 = 2 where n = sample size s = sample standard deviation σ = population standard deviation s2 = sample variance σ2 = population variance → P-values and Critical Values: Use table A-4 with df = n – 1 for the number of degrees of freedom *Remember that table A-4 is based on cumulative areas from the right → Properties of the Chi-Square Distribution: 1. All values of 2 are nonnegative and the distribution is not symmetric 2. There is a different 2 distribution for each number of degrees of freedom 3. The critical values are found in table A-4 (based on cumulative areas from the right) --locate the row corresponding to the appropriate number of degrees of freedom (df = n – 1) --the significance level is used to determine the correct column --Right-tailed test: Because the area to the right of the critical value is 0.05, locate 0.05 at the top of table A-4 --Left-tailed test: With a left-tailed area of 0.05, the area to the right of the critical value is 0.95 so locate 0.95 at the top of table A-4 --Two-tailed test: Divide the significance level of 0.05 between the left and right tails, so the areas to the right of the two critical values are 0.975 and 0.025.
    [Show full text]
  • A Study of Non-Central Skew T Distributions and Their Applications in Data Analysis and Change Point Detection
    A STUDY OF NON-CENTRAL SKEW T DISTRIBUTIONS AND THEIR APPLICATIONS IN DATA ANALYSIS AND CHANGE POINT DETECTION Abeer M. Hasan A Dissertation Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY August 2013 Committee: Arjun K. Gupta, Co-advisor Wei Ning, Advisor Mark Earley, Graduate Faculty Representative Junfeng Shang. Copyright c August 2013 Abeer M. Hasan All rights reserved iii ABSTRACT Arjun K. Gupta, Co-advisor Wei Ning, Advisor Over the past three decades there has been a growing interest in searching for distribution families that are suitable to analyze skewed data with excess kurtosis. The search started by numerous papers on the skew normal distribution. Multivariate t distributions started to catch attention shortly after the development of the multivariate skew normal distribution. Many researchers proposed alternative methods to generalize the univariate t distribution to the multivariate case. Recently, skew t distribution started to become popular in research. Skew t distributions provide more flexibility and better ability to accommodate long-tailed data than skew normal distributions. In this dissertation, a new non-central skew t distribution is studied and its theoretical properties are explored. Applications of the proposed non-central skew t distribution in data analysis and model comparisons are studied. An extension of our distribution to the multivariate case is presented and properties of the multivariate non-central skew t distri- bution are discussed. We also discuss the distribution of quadratic forms of the non-central skew t distribution. In the last chapter, the change point problem of the non-central skew t distribution is discussed under different settings.
    [Show full text]
  • Epidemiology and Biostatistics (EPBI) 1
    Epidemiology and Biostatistics (EPBI) 1 Epidemiology and Biostatistics (EPBI) Courses EPBI 2219. Biostatistics and Public Health. 3 Credit Hours. This course is designed to provide students with a solid background in applied biostatistics in the field of public health. Specifically, the course includes an introduction to the application of biostatistics and a discussion of key statistical tests. Appropriate techniques to measure the extent of disease, the development of disease, and comparisons between groups in terms of the extent and development of disease are discussed. Techniques for summarizing data collected in samples are presented along with limited discussion of probability theory. Procedures for estimation and hypothesis testing are presented for means, for proportions, and for comparisons of means and proportions in two or more groups. Multivariable statistical methods are introduced but not covered extensively in this undergraduate course. Public Health majors, minors or students studying in the Public Health concentration must complete this course with a C or better. Level Registration Restrictions: May not be enrolled in one of the following Levels: Graduate. Repeatability: This course may not be repeated for additional credits. EPBI 2301. Public Health without Borders. 3 Credit Hours. Public Health without Borders is a course that will introduce you to the world of disease detectives to solve public health challenges in glocal (i.e., global and local) communities. You will learn about conducting disease investigations to support public health actions relevant to affected populations. You will discover what it takes to become a field epidemiologist through hands-on activities focused on promoting health and preventing disease in diverse populations across the globe.
    [Show full text]
  • Biostatistics (EHCA 5367)
    The University of Texas at Tyler Executive Health Care Administration MPA Program Spring 2019 COURSE NUMBER EHCA 5367 COURSE TITLE Biostatistics INSTRUCTOR Thomas Ross, PhD INSTRUCTOR INFORMATION email - [email protected] – this one is checked once a week [email protected] – this one is checked Monday thru Friday Phone - 828.262.7479 REQUIRED TEXT: Wayne Daniel and Chad Cross, Biostatistics, 10th ed., Wiley, 2013, ISBN: 978-1- 118-30279-8 COURSE DESCRIPTION This course is designed to teach students the application of statistical methods to the decision-making process in health services administration. Emphasis is placed on understanding the principles in the meaning and use of statistical analyses. Topics discussed include a review of descriptive statistics, the standard normal distribution, sampling distributions, t-tests, ANOVA, simple and multiple regression, and chi-square analysis. Students will use Microsoft Excel or other appropriate computer software in the completion of assignments. COURSE LEARNING OBJECTIVES Upon completion of this course, the student will be able to: 1. Use descriptive statistics and graphs to describe data 2. Understand basic principles of probability, sampling distributions, and binomial, Poisson, and normal distributions 3. Formulate research questions and hypotheses 4. Run and interpret t-tests and analyses of variance (ANOVA) 5. Run and interpret regression analyses 6. Run and interpret chi-square analyses GRADING POLICY Students will be graded on homework, midterm, and final exam, see weights below. ATTENDANCE/MAKE UP POLICY All students are expected to attend all of the on-campus sessions. Students are expected to participate in all online discussions in a substantive manner. If circumstances arise that a student is unable to attend a portion of the on-campus session or any of the online discussions, special arrangements must be made with the instructor in advance for make-up activity.
    [Show full text]
  • Biostatistics
    Biostatistics As one of the nation’s premier centers of biostatistical research, the Department of Biostatistics is dedicated to addressing high priority public health and biomedical issues and driving scientific discovery through specialized analytic techniques and catalyzing interdisciplinary research. Our world-renowned experts are engaged in developing novel statistical methods to break new ground in current research frontiers. The Department has become a leader in mission the analysis and interpretation of complex data in genetics, brain science, clinical trials, and personalized medicine. Our mission is to improve disease prevention, In these and other areas, our faculty play key roles in diagnosis, treatment, and the public’s health by public health, mental health, social science, and biomedical developing new theoretical results and applied research as collaborators and conveners. Current research statistical methods, collaborating with scientists in is wide-ranging. designing informative experiments and analyzing their data to weigh evidence and obtain valid Examples of our work: findings, and educating the next generation of biostatisticians through an innovative and relevant • Researching efficient statistical methods using large-scale curriculum, hands-on experience, and exposure biomarker data for disease risk prediction, informing to the latest in research findings and methods. clinical trial designs, discovery of personalized treatment regimes, and guiding new and more effective interventions history • Analyzing data to
    [Show full text]
  • Biostatistics (BIOSTAT) 1
    Biostatistics (BIOSTAT) 1 This course covers practical aspects of conducting a population- BIOSTATISTICS (BIOSTAT) based research study. Concepts include determining a study budget, setting a timeline, identifying study team members, setting a strategy BIOSTAT 301-0 Introduction to Epidemiology (1 Unit) for recruitment and retention, developing a data collection protocol This course introduces epidemiology and its uses for population health and monitoring data collection to ensure quality control and quality research. Concepts include measures of disease occurrence, common assurance. Students will demonstrate these skills by engaging in a sources and types of data, important study designs, sources of error in quarter-long group project to draft a Manual of Operations for a new epidemiologic studies and epidemiologic methods. "mock" population study. BIOSTAT 302-0 Introduction to Biostatistics (1 Unit) BIOSTAT 429-0 Systematic Review and Meta-Analysis in the Medical This course introduces principles of biostatistics and applications Sciences (1 Unit) of statistical methods in health and medical research. Concepts This course covers statistical methods for meta-analysis. Concepts include descriptive statistics, basic probability, probability distributions, include fixed-effects and random-effects models, measures of estimation, hypothesis testing, correlation and simple linear regression. heterogeneity, prediction intervals, meta regression, power assessment, BIOSTAT 303-0 Probability (1 Unit) subgroup analysis and assessment of publication
    [Show full text]
  • 05 36534Nys130620 31
    Monte Carlos study on Power Rates of Some Heteroscedasticity detection Methods in Linear Regression Model with multicollinearity problem O.O. Alabi, Kayode Ayinde, O. E. Babalola, and H.A. Bello Department of Statistics, Federal University of Technology, P.M.B. 704, Akure, Ondo State, Nigeria Corresponding Author: O. O. Alabi, [email protected] Abstract: This paper examined the power rate exhibit by some heteroscedasticity detection methods in a linear regression model with multicollinearity problem. Violation of unequal error variance assumption in any linear regression model leads to the problem of heteroscedasticity, while violation of the assumption of non linear dependency between the exogenous variables leads to multicollinearity problem. Whenever these two problems exist one would faced with estimation and hypothesis problem. in order to overcome these hurdles, one needs to determine the best method of heteroscedasticity detection in other to avoid taking a wrong decision under hypothesis testing. This then leads us to the way and manner to determine the best heteroscedasticity detection method in a linear regression model with multicollinearity problem via power rate. In practices, variance of error terms are unequal and unknown in nature, but there is need to determine the presence or absence of this problem that do exist in unknown error term as a preliminary diagnosis on the set of data we are to analyze or perform hypothesis testing on. Although, there are several forms of heteroscedasticity and several detection methods of heteroscedasticity, but for any researcher to arrive at a reasonable and correct decision, best and consistent performed methods of heteroscedasticity detection under any forms or structured of heteroscedasticity must be determined.
    [Show full text]
  • The Effects of Simplifying Assumptions in Power Analysis
    University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Public Access Theses and Dissertations from Education and Human Sciences, College of the College of Education and Human Sciences (CEHS) 4-2011 The Effects of Simplifying Assumptions in Power Analysis Kevin A. Kupzyk University of Nebraska-Lincoln, [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/cehsdiss Part of the Educational Psychology Commons Kupzyk, Kevin A., "The Effects of Simplifying Assumptions in Power Analysis" (2011). Public Access Theses and Dissertations from the College of Education and Human Sciences. 106. https://digitalcommons.unl.edu/cehsdiss/106 This Article is brought to you for free and open access by the Education and Human Sciences, College of (CEHS) at DigitalCommons@University of Nebraska - Lincoln. It has been accepted for inclusion in Public Access Theses and Dissertations from the College of Education and Human Sciences by an authorized administrator of DigitalCommons@University of Nebraska - Lincoln. Kupzyk - i THE EFFECTS OF SIMPLIFYING ASSUMPTIONS IN POWER ANALYSIS by Kevin A. Kupzyk A DISSERTATION Presented to the Faculty of The Graduate College at the University of Nebraska In Partial Fulfillment of Requirements For the Degree of Doctor of Philosophy Major: Psychological Studies in Education Under the Supervision of Professor James A. Bovaird Lincoln, Nebraska April, 2011 Kupzyk - i THE EFFECTS OF SIMPLIFYING ASSUMPTIONS IN POWER ANALYSIS Kevin A. Kupzyk, Ph.D. University of Nebraska, 2011 Adviser: James A. Bovaird In experimental research, planning studies that have sufficient probability of detecting important effects is critical. Carrying out an experiment with an inadequate sample size may result in the inability to observe the effect of interest, wasting the resources spent on an experiment.
    [Show full text]