Chapter 10 10-1

Total Page:16

File Type:pdf, Size:1020Kb

Chapter 10 10-1 Chapter 10 10-1 Learning Objectives In this chapter, you learn how to use hypothesis testing for comparing the difference between: • The means of two independent populations • The means of two related populations • The proportions of two independent Chapter 10 populations Two-Sample Tests • The variances of two independent populations by testing the ratio of the two variances Prof. Shuguang Liu Chap 10-1 Prof. Shuguang Liu Chap 10-2 Two-Sample Tests Difference Between Two Means DCOVA DCOVA Two-Sample Tests Population means, Goal: Test hypothesis or form independent * a confidence interval for the samples difference between two Population Population population means, µ – µ Means, Means, Population Population 1 2 Proportions Variances Independent Related σ1 and σ2 unknown, Samples Samples assumed equal The point estimate for the difference is Examples: Group 1 vs. Same group Proportion 1 vs. Variance 1 vs. Group 2 before vs. after Proportion 2 Variance 2 X1 – X2 treatment σ1 and σ2 unknown, not assumed equal Prof. Shuguang Liu Chap 10-3 Prof. Shuguang Liu Chap 10-4 Difference Between Two Means: Hypothesis Tests for Independent Samples Two Population Means DCOVA DCOVA § Different data sources Two Population Means, Independent Samples Population• Unrelated means, independent• Independent * Lower-tail test: Upper-tail test: Two-tail test: samplesu Sample selected from one population has no effect on the sample selected from the other population H0: µ1 ≥ µ2 H0: µ1 ≤ µ2 H0: µ1 = µ2 H1: µ1 < µ2 H1: µ1 > µ2 H1: µ1 ≠ µ2 Use S to estimate unknown i.e., i.e., i.e., p σ and σ unknown, 1 2 σ. Use a Pooled-Variance t H : µ – µ ≤ 0 H : µ – µ = 0 assumed equal H0: µ1 – µ2 ≥ 0 0 1 2 0 1 2 test. H : µ – µ > 0 H : µ – µ ≠ 0 H1: µ1 – µ2 < 0 1 1 2 1 1 2 Use S and S to estimate σ1 and σ2 unknown, 1 2 not assumed equal unknown σ1 and σ2. Use a Separate-Variance t test. Prof. Shuguang Liu Chap 10-5 Prof. Shuguang Liu Chap 10-6 Chapter 10 10-2 Hypothesis tests for µ1 - µ2 with σ1 and Hypothesis tests for µ1 – µ2 σ2 unknown and assumed equal DCOVA DCOVA Two Population Means, Independent Samples Population means, Assumptions: Lower-tail test: Upper-tail test: Two-tail test: independent § Samples are randomly and H : µ – µ 0 H : µ – µ ≤ 0 H : µ – µ = 0 samples 0 1 2 ≥ 0 1 2 0 1 2 independently drawn H1: µ1 – µ2 < 0 H1: µ1 – µ2 > 0 H1: µ1 – µ2 ≠ 0 § Populations are normally σ and σ unknown, α α α/2 α/2 1 2 * distributed or both sample assumed equal sizes are at least 30 -tα tα -tα/2 tα/2 § Population variances are unknown but assumed equal Reject H0 if tSTAT < -tα Reject H0 if tSTAT > tα Reject H0 if tSTAT < -tα/2 σ1 and σ2 unknown, or tSTAT > t /2 α not assumed equal Prof. Shuguang Liu Chap 10-7 Prof. Shuguang Liu Chap 10-8 Hypothesis tests for µ - µ with σ and 1 2 1 Confidence interval for µ1 - µ2 with σ1 σ2 unknown and assumed equal and σ2 unknown and assumed equal (continued) DCOVA DCOVA • The pooled variance is: Population means, 2 2 Population means, 2 (n1 − 1)S1 + (n2 − 1)S2 independent S = independent samples p (n 1) (n 1) samples 1 − + 2 − The confidence interval for • The test statistic is: µ1 – µ2 is: σ and σ unknown, 1 2 * σ1 and σ2 unknown, assumed equal * (X1 − X2 )− (µ1 − µ2 ) assumed equal tSTAT = 2 ⎛ 1 1 ⎞ X1 X2 t S ⎜ ⎟ ( − ) ± α/2 p ⎜ + ⎟ 2 ⎛ 1 1 ⎞ n n S ⎝ 1 2 ⎠ p ⎜ + ⎟ ⎝ n1 n2 ⎠ σ and σ unknown, σ and σ unknown, 1 2 1 2 Where tα/2 has d.f. = n1 + n2 – 2 not assumed equal • Where tSTAT has d.f. = (n1 + n2 – 2) not assumed equal Prof. Shuguang Liu Chap 10-9 Prof. Shuguang Liu Chap 10-10 Pooled-Variance t Test Example: Pooled-Variance t Test Example Calculating the Test Statistic You are a financial analyst for a brokerage firm. Is thereDCOV a A (continued) difference in dividend yield between stocks listed on the H0: µ1 - µ2 = 0 i.e. (µ1 = µ2) DCOVA NYSE & NASDAQ? You collect the following data: H1: µ1 - µ2 ≠ 0 i.e. (µ1 ≠ µ2) NYSE NASDAQ Number 21 25 The test statistic is: Sample mean 3.27 2.53 Sample std dev 1.30 1.16 (X1 − X2 )− (µ1 − µ2 ) (3.27 − 2.53)− 0 tSTAT = = = 2.040 ⎛ 1 1 ⎞ ⎛ 1 1 ⎞ S2 ⎜ + ⎟ 1.5021⎜ + ⎟ p ⎜ ⎟ 21 25 Assuming both populations are ⎝ n1 n2 ⎠ ⎝ ⎠ approximately normal with equal variances, is 2 2 2 2 2 (n1 −1)S1 + (n2 −1)S2 (21 −1)1.30 + (25 −1)1.16 there a difference in mean S = = = 1.5021 P (n −1) + (n −1) (21-1) + (25 −1) yield (α = 0.05)? 1 2 Prof. Shuguang Liu Chap 10-11 Prof. Shuguang Liu Chap 10-12 Chapter 10 10-3 Pooled-Variance t Test Example: Excel Pooled-Variance t test Comparing NYSE & Hypothesis Test Solution NASDAQ DCOVA H0: µ1 - µ2 = 0 i.e. (µ1 = µ2) DCOVA Reject H0 Reject H0 Pooled-Variance t Test for the Difference Between Two Means H1: µ1 - µ2 ≠ 0 i.e. (µ1 ≠ µ2) (assumes equal populaon variances) Data Hypothesized Difference 0 α = 0.05 Level of Significance 0.05 Decision: .025 .025 Populaon 1 Sample df = 21 + 25 - 2 = 44 Sample Size 21 =COUNT(DATA!$A:$A) Sample Mean 3.27 =AVERAGE(DATA!$A:$A) Reject H0 at α = 0.05 Critical Values: t = ± 2.0154 Sample Standard Deviaon 1.3 =STDEV(DATA!$A:$A) -2.0154 0 2.0154 Populaon 2 Sample t Sample Size 25 =COUNT(DATA!$B:$B) Conclusion: Sample Mean 2.53 =AVERAGE(DATA!$B:$B) Sample Standard Deviaon 1.16 =STDEV(DATA!$B:$B) Test Statistic: 2.040 There is evidence of a Intermediate Calculaons Populaon 1 Sample Degrees of Freedom 20 =B7 - 1 difference in means. Populaon 2 Sample Degrees of Freedom 24 =B11 - 1 Decision: Total Degrees of Freedom 44 =B16 + B17 3.27 − 2.53 Pooled Variance 1.502 =((B16 * B9^2) + (B17 * B13^2)) / B18 Reject H at α = 0.05 Standard Error 0.363 =SQRT(B19 * (1/B7 + 1/B11)) tSTAT = = 2.040 0 Difference in Sample Means 0.74 =B8 - B12 ⎛ 1 1 ⎞ t Test Stas=c 2.040 =(B21 - B4) / B20 1.5021 ⎜ + ⎟ Conclusion: Two-Tail Test 21 25 Lower Cri=cal Value -2.015 =-TINV(B5, B18) ⎝ ⎠ There is evidence of a Upper Cri=cal Value 2.015 =TINV(B5, B18) p-value 0.047 =TDIST(ABS(B22),B18,2) Reject the null hypothesis =IF(B27<B5,"Reject the null hypothesis", difference in means. "Do not reject the null hypothesis") Prof. Shuguang Liu Chap 10-13 Prof. Shuguang Liu Chap 10-14 Minitab Pooled-Variance t test Comparing NYSE & Pooled-Variance t Test Example: NASDAQ Confidence Interval for µ1 - µ2 DCOVA DCOVA Two-Sample T-Test and CI Since we rejected H0 can we be 95% confident that µNYSE Sample N Mean StDev SE Mean 1 21 3.27 1.30 0.28 > µNASDAQ? 2 25 2.53 1.16 0.23 Difference = mu (1) - mu (2) 95% Confidence Interval for µNYSE - µNASDAQ Estimate for difference: 0.740 95% CI for difference: (0.009, 1.471) T-Test of difference = 0 (vs not =): T-Value = 2.04 P-Value = 0.047 DF = 44 2 ⎛ 1 1 ⎞ Both use Pooled StDev = 1.2256 X1 − X2 ± t /2 Sp ⎜ + ⎟ = 0.74 ± 2.0154× 0.3628 = (0.009, 1.471) ( ) α ⎜ ⎟ ⎝ n1 n 2 ⎠ Decision: Reject H0 at α = 0.05 Since 0 is less than the entire interval, we can be 95% Conclusion: confident that µ > µ There is evidence of a NYSE NASDAQ difference in means. Prof. Shuguang Liu Chap 10-15 Prof. Shuguang Liu Chap 10-16 Hypothesis tests for µ1 - µ2 with σ1 and Hypothesis tests for µ1 - µ2 with σ1 and σ2 unknown, not assumed equal σ2 unknown and not assumed equal (continued) DCOVA DCOVA The test statistic is: Population means, Assumptions: Population means, independent independent X1 X2 µ µ § Samples are randomly and ( − )− ( 1 − 2 ) samples samples tSTAT = independently drawn S2 S2 2 1 2 ⎛ S 2 S 2 ⎞ ⎜ 1 + 2 ⎟ + ⎜ n n ⎟ ⎝ 1 2 ⎠ n1 n2 ν = 2 2 § Populations are normally ⎛ S 2 ⎞ ⎛ S 2 ⎞ σ1 and σ2 unknown, ⎜ 1 ⎟ ⎜ 2 ⎟ σ1 and σ2 unknown, ⎜ ⎟ ⎜ ⎟ n1 n2 distributed or both sample assumed equal ⎝ ⎠ + ⎝ ⎠ assumed equal t n1 − 1 hasn2 − 1 d.f. ν = sizes are at least 30 STAT § Population variances are unknown and cannot be σ and σ unknown, σ and σ unknown, 1 2 assumed to be equal 1 2 not assumed equal * not assumed equal * Prof. Shuguang Liu Chap 10-17 Prof. Shuguang Liu Chap 10-18 Chapter 10 10-4 Related Populations Related Populations The Paired Difference Test The Paired Difference Test DCOVA DCOVA th (continued) Tests Means of 2 Related Populations The i paired difference is Di , where – Paired or matched samples Related – Repeated measures (before/after) Related Di = X1i - X2i samples – Use difference between paired values: samples n The point estimate for the D paired difference ∑ i Di = X1i - X2i D = i=1 § Eliminates Variation Among Subjects population mean µD is D : n § Assumptions: n The sample standard (D D)2 • Both Populations Are Normally Distributed ∑ i − deviation is SD S = i=1 • Or, if not Normal, use large samples D n − 1 n is the number of pairs in the paired sample Prof.
Recommended publications
  • T-Testing (Single)
    t-Testing (Single) Young W. Lim 2019-07-13 Sat Young W. Lim t-Testing (Single) 2019-07-13 Sat 1 / 24 Outline 1 Based on 2 t Testing Young W. Lim t-Testing (Single) 2019-07-13 Sat 2 / 24 Based on "Understanding Statistics in the Behavioral Sciences" R. R. Pagano I, the copyright holder of this work, hereby publish it under the following licenses: GNU head Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the section entitled GNU Free Documentation License. CC BY SA This file is licensed under the Creative Commons Attribution ShareAlike 3.0 Unported License. In short: you are free to share and make derivative works of the file under the conditions that you appropriately attribute it, and that you distribute it only under a license compatible with this one. Young W. Lim t-Testing (Single) 2019-07-13 Sat 3 / 24 Student t-Test (1) The t-test is any statistical hypothesis test in which the test statistic follows a Student’s t-distribution under the null hypothesis. Young W. Lim t-Testing (Single) 2019-07-13 Sat 4 / 24 Student t-Test (2) A t-test is most commonly applied when the test statistic would follow a normal distribution if the value of a scaling term in the test statistic were known.
    [Show full text]
  • Nonparametric Statistics
    14 Nonparametric Statistics • These methods required that the data be normally CONTENTS distributed or that the sampling distributions of the 14.1 Introduction: Distribution-Free Tests relevant statistics be normally distributed. 14.2 Single-Population Inferences Where We're Going 14.3 Comparing Two Populations: • Develop the need for inferential techniques that Independent Samples require fewer or less stringent assumptions than the 14.4 Comparing Two Populations: Paired methods of Chapters 7 – 10 and 11 (14.1) Difference Experiment • Introduce nonparametric tests that are based on ranks 14.5 Comparing Three or More Populations: (i.e., on an ordering of the sample measurements Completely Randomized Design according to their relative magnitudes) (14.2–14.7) • Present a nonparametric test about the central 14.6 Comparing Three or More Populations: tendency of a single population (14.2) Randomized Block Design • Present a nonparametric test for comparing two 14.7 Rank Correlation populations with independent samples (14.3) • Present a nonparametric test for comparing two Where We've Been populations with paired samples (14.4) • Presented methods for making inferences about • Present a nonparametric test for comparing three or means ( Chapters 7 – 10 ) and for making inferences more populations using a designed experiment about the correlation between two quantitative (14.5–14.6) variables ( Chapter 11 ) • Present a nonparametric test for rank correlation (14.7) 14-1 Copyright (c) 2013 Pearson Education, Inc M14_MCCL6947_12_AIE_C14.indd 14-1 10/1/11 9:59 AM Statistics IN Action How Vulnerable Are New Hampshire Wells to Groundwater Contamination? Methyl tert- butyl ether (commonly known as MTBE) is a the measuring device, the MTBE val- volatile, flammable, colorless liquid manufactured by the chemi- ues for these wells are recorded as .2 cal reaction of methanol and isobutylene.
    [Show full text]
  • Hypothesis Testing-I SUBJECT FORENSIC SCIENCE
    SUBJECT FORENSIC SCIENCE Paper No. and Title PAPER No.15: Forensic Psychology Module No. and Title MODULE No.31: Hypothesis Testing-1 Module Tag FSC_P15_M31 FORENSIC SCIENCE PAPER No.15 : Forensic Psychology MODULE No.31: Hypothesis Testing-I TABLE OF CONTENTS 1. Learning Outcomes 2. Introduction 3. Single Sample Parametric Test 3.1 Single sample Z test 3.2 Single sample t-test 4. Single Sample Non-Parametric Test 4.1 Wilcoxon Signed-Rank Test 5. Summary FORENSIC SCIENCE PAPER No.15 : Forensic Psychology MODULE No.31: Hypothesis Testing-I 1. Learning Outcomes After studying this module, you will be able to- Know about parametric and non-parametric method. Understand one sample Z and t test. Understand Wilcoxon Signed-Rank Test 2. Introduction Parametric Methods These methods are categorized on the basis the population we are studying. Parametric methods are those for which the population is almost normal, or we can estimate using a normal distribution after we invoke the central limit theorem. Eventually the categorization of a method as non-parametric depends upon the conventions that are made about a population. A few parametric methods include: Confidence interval for a population mean, with known standard deviation. Confidence interval for a population mean with unknown standard deviation Confidence interval for a population variance. Confidence interval for the difference of two means with unknown standard deviation. Non-Parametric Methods Non-parametric methods are statistical techniques for which any assumption of normality for the population is not required unlike parametric methods. Certainly the methods do not have any dependency on the population of interest.
    [Show full text]
  • The Statistics Tutor's Quick Guide to Commonly Used Statistical Tests
    statstutor community project encouraging academics to share statistics support resources All stcp resources are released under a Creative Commons licence Stcp-marshallowen-7 The Statistics Tutor’s Quick Guide to Commonly Used Statistical Tests © Ellen Marshall, University of Sheffield Reviewer: Jean Russell University of Sheffield www.statstutor.ac.uk Contents CONTENTS ..................................................................................................................................................................................2 INTRODUCTION ............................................................................................................................................................................4 TIPS FOR TUTORING .......................................................................................................................................................................5 SECTION 1 GENERAL INFORMATION ............................................................................................................................................6 DATA TYPES .................................................................................................................................................................................7 SUMMARY STATISTICS ....................................................................................................................................................................7 Summary of descriptive and graphical statistics ..................................................................................................................8
    [Show full text]
  • 10: Inference from Small Samples
    10: Inference from Small Samples 10.1 Refer to Table 4, Appendix I, indexing df along the left or right margin and tα across the top. a t.05 = 2.015 with 5 df b t.025 = 2.306 with 8 df c t.10 = 1.330 with 18 df c t.025 ≈ 1.96 with 30 df 10.2 The value Pt(>ta )=ais the tabled entry for a particular number of degrees of freedom. a For a two-tailed test with α = .01, the critical value for the rejection region cuts off α 2 = .005 in the two tails of the t distribution shown below, so that t.005 = 3.055 . The null hypothesis H0 will be rejected if t > 3.055 or t <−3.055 (which you can also write as t > 3.055 ). b For a right-tailed test, the critical value that separates the rejection and nonrejection regions for a right tailed test based on a t-statistic will be a value of t (called tα ) such that Pt( >=tα ) α =.05 and df = 16 . That is, t.05 = 1.746 . The null hypothesis H0 will be rejected if t > 1.746 . c For a two-tailed test with α 2 = .025 and df = 25 , H0 will be rejected if t > 2.060 . d For a left-tailed test withα = .01and df = 7 , H0 will be rejected if t < −2.998 . 10.3 a The p-value for a two-tailed test is defined as pP-value =>( t2.43) =2P(t>2.43) so that 1 Pt()>=2.43 p-value 2 Refer to Table 4, Appendix I, with df = 12 .
    [Show full text]
  • Correction Factors for Freshwater Dissolved and Total Nutrients for 1998 and 2006 Instrument and Method Changes by the King County Environmental Laboratory
    Correction Factors for Freshwater Dissolved and Total Nutrients for 1998 and 2006 Instrument and Method Changes by the King County Environmental Laboratory March 2016 Alternate Formats Available Correction Factors for Freshwater Dissolved and Total Nutrients for 1998 and 2006 Instrument and Method Changes by the King County Environmental Laboratory Submitted by: Timothy Clark Science and Technical Support Section King County Water and Land Resources Division Department of Natural Resources and Parks Correction Factors for Freshwater Dissolved and Total Nutrients Acknowledgements I would like to express my gratitude to the current and past employees of the King County Environmental Laboratory; their constant drive to improve their methods of analysis is inspiring and a tremendous asset to the Division. My thanks to Sally Abella, Debra Bouchard, Curtis DeGasperi, and Jim Simmonds for their review of this document. Citation King County. 2016. Correction Factors for Dissolved and Total Nutrients for 1998 and 2006 Instrument and Method Changes by the King County Environmental Laboratory. Prepared by Timothy Clark, Water and Land Resources Division. Seattle, Washington. King County Science and Technical Support Section i March 2016 Correction Factors for Freshwater Dissolved and Total Nutrients Table of Contents Executive Summary............................................................................................................................................. vi Acronyms and Abbreviations ........................................................................................................................
    [Show full text]
  • [PSS] Glossary
    Glossary 2 × 2 contingency table. A 2 × 2 contingency table is used to describe the association between a binary independent variable and a binary response variable of interest. acceptance region. In hypothesis testing, an acceptance region is a set of sample values for which the null hypothesis cannot be rejected or can be accepted. It is the complement of the rejection region. actual alpha, actual significance level. This is an attained or observed significance level. allocation ratio. This ratio n2=n1 represents the number of subjects in the comparison, experimental group relative to the number of subjects in the reference, control group. Also see[ PSS] unbalanced designs. alpha. Alpha, α, denotes the significance level. alternative hypothesis. In hypothesis testing, the alternative hypothesis represents the counterpoint to which the null hypothesis is compared. When the parameter being tested is a scalar, the alternative hypothesis can be either one sided or two sided. alternative value, alternative parameter. This value of the parameter of interest under the alternative hypothesis is fixed by the investigator in a power and sample-size analysis. For example, alternative mean value and alternative mean refer to a value of the mean parameter under the alternative hypothesis. analysis of variance, ANOVA. This is a class of statistical models that studies differences between means from multiple populations by partitioning the variance of the continuous outcome into independent sources of variation due to effects of interest and random variation. The test statistic is then formed as a ratio of the expected variation due to the effects of interest to the expected random variation.
    [Show full text]
  • Reporting Mann Whitney U Test
    Reporting Mann Whitney U Test medially.palmatelyApproved asButch boozier never Dimitri dissert clobbers so nowhere her kiddles or garbles balloting any hautboiscloudlessly. proverbially. Equiangular Vasily Brewer pirouetting dabbling Mann-Whitney U Test SPSS UPDATED YouTube. Mann-Whitney U Test AO1 AO2 PSYCHOLOGY WIZARD. Mann-Whitney U Test in SPSS Statistics Setup Procedure. With the MannWhitney test several options on what do report alter the effect. Thus when conducting this test aside from reporting the p-value the. Why use Mann Whitney U test instead of t test? 10 Rank score tests BMJcom. To conclude that i know who gave us that longer than large samples of reporting mann whitney u test as predictors of statistical methods are kindly thanking you? 22 A Mann-Whitney test indicated that the slices of pizza eaten was greater for football Mdn 9 than for basketball players Mdn 5. If that assumption does not pluck the nonparametric Mann-Whitney test is living better for drawing. In reporting the results of a MannWhitney test it with important the state aid measure with the central tendencies of sight two groups means or medians since the MannWhitney is an ordinal test medians are usually recommended The provided of U The sample sizes The significance level. The paired t test repeated-measures ANOVA Wilcoxon signed-rank test and. The Wilcoxon rank-sum test is a nonparametric alternative to point two- sample t-test which is based solely on chain order in known the observations from for two. Mann Whitney Wilcoxon TEST EXAMPLES Alessio. Medians and IQR values to further describe the data before reporting the results of the.
    [Show full text]
  • T-Test, F-Test, Z-Test ,Chi Square Test. Parametric Test
    Make economics easy T-test, f-test, Z-test ,chi square test. Parametric test 1.T-test 2.Z-test 3.F-test 1.t-test T-test is a small sample test. It was developed by William Gosset in 1908 It is also called students t test(pen name) Deviation from population parameter t = Standard error of 0th sample statistics Uses of t-test/application Size of sample is small (n<30) Degree of freedom is v=n-1 T-test is used for test of significance of regression coefficient in regression model. In multiple regression with 3 individual variable the regression coefficient are to be tested by- t-test. To test the hypothesis that correlation coefficient in population is zero then we used t-test. We used t-statistics when parameter of population are normal. Population Variance are unknown. It is also called Welch t-test. 2.Z-test Z-test was given by Fisher. T-test is used when correlation coefficient of population Is zero. But if population coeff. Correlation is not zero then z-test is used. Zr-Zp Z= SEz In z-test sample size is large (n>30) Z-test is used to determine whether two population means are different ,when Population variance is known. Z-test is based on standard normal distribution. Z-test is also called large sample test.. For each significance level the z-test has single critical value ie. 1.96 for 5% two tailed. Z-test conduct various tests like.,… One sample test, Two sample test. Location test, Maximum likelihood estimate.
    [Show full text]
  • Data Analysis Methods: Average Values Evaluation of the Reliability of the Results of the Study
    DATA ANALYSIS METHODS: AVERAGE VALUES EVALUATION OF THE RELIABILITY OF THE RESULTS OF THE STUDY. CHARACTERISTICS AND ANALYSIS OF STATISTICAL ERRORS. Lecturer: Maxim V. Khorosh, PhD THE NEXT GROUPS OF METHODS ARE DISTINGUISHED: 1. Methods of calculating generalizing coefficients that characterize the different aspects of each of the features of the program: • methods of calculating of the relative values • methods of calculating of the average values • methods for assessing the reliability of relative and average values 2. Methods of comparing of the different statistical aggregates: • methods for assessing the reliability of the differences of generalizing coefficients; • methods for assessing the reliability of differences in the distribution of characteristics; • methods of standardization of generalizing coefficients. 3. Methods of differentiation, assessment of interaction and integration of factors. analysis of variance correlation analysis allow to solve the following tasks: regression analysis a) to decompose a multifactorial complex into constituent factors, highlighting factor analysis important and insignificant; principal components method b) to study the interaction of factors c) obtain an integrated assessment based on discriminant analysis a set of factors. sequential analysis 4. Methods of analysis of the dynamics of phenomena (analysis of dynamic or time series). STAGES OF DESCRIPTION OF QUANTITATIVE SIGN 1. Determining the type of distribution of the sign. 2. Estimation of the central tendency of the studied population. 3. Assessment of its diversity (spread). TO CORRECTLY CHOOSE THE PATH OF STATISTICAL ANALYSIS, IT IS NECESSARY TO KNOW THE TYPE OF DISTRIBUTION OF THE RESEARCH SIGN. Under the type of distribution of a random variable means the correspondence that is established between all possible numerical values of a random variable and the possibility of their occurrence in the aggregate.
    [Show full text]
  • Chapter 11 Outline
    Park University - EC315 - Chapter 11 Lecture Notes Page 1 Chapter 11 Lecture Notes Introduction In this chapter we continue our study of hypothesis testing. Recall that in Chapter 10 we considered hypothesis tests in which we compared the results of a single sample statistic to a population parameter. In this chapter , we expand the concept of hypothesis testing to two samples. We select random samples from two independent populations and conduct a hypothesis test to determine whether the population means are equal. We might want to test to see if there is a difference in the mean number of defects produced on the 7:00 AM to 3:00 PM shift and the 3:00 PM to 11:00 PM shift at the DaimlerChrysler Jeep Liberty plant in Toledo, Ohio. We also conduct a hypothesis tests to determine if two sample proportions come from populations which are equal. For example, we may want to determine if the proportion of Jumpin’ Java customers who purchase frozen coffee drinks is the same for New England stores versus stores in the southeast. Two-Sample Tests of Hypothesis: Independent Samples As noted above, we expand the concept of hypothesis testing to two samples. When there are two populations, we can compare two sample means to determine if they came from populations with the same or equal means. For example, a purchasing agent is considering two brands of tires for use on the company's fleet of cars. A sample of 60 Rossford tires indicates the mean useful life to be 65,000 miles. A sample of 50 Maumee tires reveals the useful life to be 68,000 miles.
    [Show full text]
  • 7 Non-Parametric Statistics 7.1 Anderson
    SEC 4 Page 1 of 6 7 NON-PARAMETRIC STATISTICS 7.1 ANDERSON - DARLING TEST: The Anderson–Darling test is a statistical test of whether a given sample of data is drawn from a given probability distribution. In its basic form, the test assumes that there are no parameters to be estimated in the distribution being tested, in which case the test and its set of critical values is distribution-free. However, the test is most often used in contexts where a family of distributions is being tested, in which case the parameters of that family need to be estimated and account must be taken of this in adjusting either the test-statistic or its critical values. When applied to testing if a normal distribution adequately describes a set of data, it is one of the most powerful statistical tools for detecting most departures from normality. K-sample Anderson–Darling tests are available for testing whether several collections of observations can be modeled as coming from a single population, where the distribution function does not have to be specified. In addition to its use as a test of fit for distributions, it can be used in parameter estimation as the basis for a form of minimum distance estimation procedure. The test is named after Theodore Wilbur Anderson (born 1918) and Donald A. Darling (born 1915), who invented it in 1952. The Anderson-Darling test for normality is one of three general normality tests designed to detect all departures from normality. While it is sometimes touted as the most powerful test, no one test is best against all alternatives and the other 2 tests are of comparable power.
    [Show full text]