Statistical Analysis of Metabolomics Data

Total Page:16

File Type:pdf, Size:1020Kb

Statistical Analysis of Metabolomics Data Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline • Introduction • Data pre-treatment 1. Normalization 2. Centering, scaling, transformation • Univariate analysis 1. Student’s t-tes 2. Volcano plot • Multivariate analysis 1. PCA 2. PLS-DA • Machine learning • Software packages 2 Results from data processing Next, analysis of the quantitative metabolite information … 3 Metabolomics data analysis • Goals – biomarker discovery by identifying significant features associated with certain conditions – Disease diagnosis via classification • Challenges – Limited sample size – Many metabolites / variables • Workflow univariate multivariate machine pre-treatment analysis analysis learning 4 Data Pre-treatment 5 Normalization (I) • Goals – to reduce systematic variation – to separate biological variation from variations introduced in the experimental process – to improve the performance of downstream statistical analysis • Sources of experimental variation – sample inhomogeneity – differences in sample preparation – ion suppression 6 Normalization (II) • Approaches – Sample-wise normalization: to make samples comparable to each other – Feature/variable-wise normalization: to make features more comparable in magnitude to each other. • Sample-wise normalization – normalize to a constant sum – normalize to a reference sample – normalize to a reference feature (an internal standard) – sample-specific normalization (dry weight or tissue volume) • Feature-wise normalization (i.e., centering, scaling, and transformation) 7 Centering, scaling, transformation 8 Centering • Converts all the concentrations to fluctuations around zero instead of around the mean of the metabolite concentrations • Focuses on the fluctuating part of the data • Is applied in combination with data scaling and transformation 9 Scaling • Divide each variable by a factor • Different variables have a different scaling factor • Aim to adjust for the differences in fold differences between the different metabolites. • Results in the inflation of small values • Two subclasses – Uses a measure of the data dispersion – Uses a size measure 10 Scaling: subclass 1 • Use data dispersion as a scaling factor – auto: use the standard deviation as the scaling factor. All the metabolites have a standard deviation of one and therefore the data is analyzed on the basis of correlations instead of covariance. – pareto: use the square root of the standard deviation as the scaling factor. Large fold changes are decreased more than small fold changes and thus large fold changes are less dominant compared to clean data. – vast: use standard deviation and the coefficient of variation as scaling factors. This results in a higher importance for metabolites with a small relative sd. – range: use (max-min) as scaling factors. Sensitive to outliers. 11 Scaling: subclass 2 • Use average as scaling factors – The resulting values are changes in percentages compared to the mean concentration. – The median can be used as a more robust alternative. 12 Transformation • Log and power transformation • Both reduce large values relatively more than the small values. • Log transformation – pros: removal of heteroscedasticity – cons: unable to deal with the value zero. • Power transformation – pros: similar to log transformation – cons: not able to make multiplicative effects additive 13 Centering, scaling, transformation A. original data B. centering C. auto D. pareto E. range F. vast G. level H. log I. power 14 Log transformation, again • Hard to do useful statistical tests with a skewed distribution. • A skewed distribution or exponentially decaying distribution can be transformed into a Gaussian distribution by applying a log transformation. 15 http://www.medcalc.org/manual/transforming_data_to_normality.php Univariate vs. multivariate analysis • Univariate analysis examines each variable separately. – t-tests – volcano plot • Multivariate analysis considers two or more variables simultaneously and takes into account relationships between variables. – PCA: Principle Component Analysis – PLS-DA: Partial Least Squares-Discriminant Analysis • Univariate analyses are often first used to obtain an overview or rough ranking of potentially important features before applying more sophisticated multivariate analyses. 16 Univariate Statistics t-test volcano plot 17 Univariate statistics • A basic way of presenting univariate data is to create a frequency distribution of the individual cases. Due to the Central Limit Theorem, many of these frequency distributions can be modeled as a normal/Gaussian distribution. 18 Gaussian distribution • The total area underneath each density curve is equal to 1. 1.0 = 0, 2= 0.2, = 0, 2= 1.0, 0.8 = 0, 2= = 2= ) 0.6 x ( 2 0.4 2 " x−µ % −$ ' 0.2 1 # σ & ϕ(x) = e 0.0 0 1 2 4 σ 2π x mean = µ variance = σ 2 0.4 0.3 standard deviation = σ 0.2 34.1% 34.1% 0.1 2.1% 2.1% 0.1% 13.6% 13.6% 0.1% 0.0 µ 19 https://en.wikipedia.org/wiki/Normal_distribution Sample statistics 1 n Sample mean: X= ∑Xi n i=1 n 2 1 2 Sample variance: S = ∑(Xi − X) n −1 i=1 Sample standard deviation: S = S2 20 https://en.wikipedia.org/wiki/Normal_distribution t-test (I) • One-sample t-test: is the sample drawn from a known population? Null hypothesis H0 : µ = µ0 Alternative hypethesis H1: µ<µ0 x − µ Test statistic: t = 0 s n n 1 2 Sample standard deviation: s = ∑(xi − x) n −1 i=1 The test statistic t follows a student’s t distribution. The distribution has n-1 degrees of freedom. 21 t-test (II): p-value When the null hypothesis is rejected, the result is said to be statistically significant. 22 t-test (III) • Two-sample t-test: are the two populations different? Null hypothesis H0 : µ1 − µ2 = 0 Alternative hypethesis H1: µ1 − µ2 ≠ 0 x − x − µ −µ Test statistic: t = ( 1 2 ) ( 1 2 ) s2 s2 1 + 2 n1 n2 • The two samples should be independent. 23 t-test (IV) Equivalent statements: • The p-value is small. • The difference between the two populations is unlikely to have occurred by chance, i.e. is statistically significant. 24 t-test (V) • The p-value is big. • The difference between the two populations are said NOT to be statistically significant. 25 t-test (VI) • Paired t-test: what is the effect of a treatment? • Measurements made on the same individuals before and after the treatment. Example: Subjects participated in a study on the effectiveness of a certain diet on serum cholesterol levels. Subject Before After Difference 1 201 200 -1 H0 : µd = 0 2 231 236 +5 Ha : µd ≠ 0 3 221 216 -5 d − µ 5 260 243 -17 Test statistic: t = d 6 228 224 -4 sd n 7 245 235 -10 26 Volcano plot (I) • Plot fold change vs. significance • y-axis: negative log of the p-value • x-axis: log of the fold change so that changes in both directions (up and down) appear equidistant from the center • Two regions of interest: those points that are found towards the top of the plot that are far to either the left- or the right- hand side. 27 Volcano plot (II) 28 Multivariate statistics PCA PLS-DA 29 PCA (I) • PCA is a statistical procedure to transform a set of correlated variables into a set of linearly uncorrelated variables. 30 PCA (II) • The uncorrelated variables are ordered in such a way that the first one accounts for as much of the variability in the data as possible and each succeeding one has the highest variance possible in the remaining variables. • These ordered uncorrelated variables are called principle components. • By discarding low-variance variables, PCA helps us reduce data dimension and visualize the data. 31 PCA (III) • The transformation matrix • The transformation st nd th • p 1 , p 2 , , p n are called the 1 , 2 , and n principle components, respectively. 32 PCA (IV) • Each original sample is represented by an n-dimensional vector: Sbefore transformation = [x1, x2,, xn ] • After the transformation – If all of the principle components are kept, then each sample is still represented by an n-dimensional vector: Safter transformation = [y1, y2,, yn ] – If only m < n principle components are kept, then each sample will be represented by an m-dimensional vector: Safter transformation = [y1, y2,, ym ] 33 PCA (V) • y are called scores. • For visualization purpose, m is usually chosen to be 2 or 3. • As a result, each sample will be represented by a 2- or 3- dimenational point in the score plot. Principal Component Analysis (Scores) 30 wt15 ● 20 wt16 ● wt22 10 ● wt18 ● PC2 wt19 wt21 ● ● ko15 ● 0 ko22 ko21 ● ● 10 − ko19 ● ko16 ● ko18 ● 34 −30 −20 −10 0 10 PC1 PCA (VI) • Loadings • [ p 1 1 , p 1 2 ] , [ p 2 1 , p 2 2 ] , , [ p n 1 , p n 2 ] are denoted as points in the loadings plot 35 Loadings plot for all variables PCA (VII) Principal Component Analysis (Loadings) 548.1/3425 324/3268 472/2954 507.2/3030316.2/3089414.2/3062 533.2/2870485.2/3041317.2/3092547.3/3019560.3/3089 0.06 594.2/3376324.1/2789567.2/3016549.3/3428341.2/3556458.2/3046596.2/3008472.2/3121567.1/3030 248.2/3240 443/2684523.1/3037595.2/3004299.2/4126551.2/3013348.2/3499420.2/2722244.9/3130361.1/2693591.3/3005344.2/3350225/3232 577.2/2861506.3/3310548.4/4232344.2/3152419.2/3807592.2/2998502.2/3032590.3/3005215.2/3478463.2/3046459.2/3047504.2/3031 363.1/3481317.2/3492444.2/2897395.2/3570560.2/3096 507.1/3017512.2/3358 466.1/3203402.1/2688554.2/3490591.2/2982219.1/2523547.2/3021573.2/3004 337.2/3438 556.2/3444 576.4/3863338.1/2789419.1/3062302.1/2818 591.5/4105310.1/2798 548.3/3015532.2/2870402.1/2672488.2/2882546.2/3018 557.1/2574562.1/3514 286.2/2855 417.1/2683502.1/4192382/2677389.2/3349573.3/3391
Recommended publications
  • UNIT 1 INTRODUCTION to STATISTICS Introduction to Statistics
    UNIT 1 INTRODUCTION TO STATISTICS Introduction to Statistics Structure 1.0 Introduction 1.1 Objectives 1.2 Meaning of Statistics 1.2.1 Statistics in Singular Sense 1.2.2 Statistics in Plural Sense 1.2.3 Definition of Statistics 1.3 Types of Statistics 1.3.1 On the Basis of Function 1.3.2 On the Basis of Distribution of Data 1.4 Scope and Use of Statistics 1.5 Limitations of Statistics 1.6 Distrust and Misuse of Statistics 1.7 Let Us Sum Up 1.8 Unit End Questions 1.9 Glossary 1.10 Suggested Readings 1.0 INTRODUCTION The word statistics has different meaning to different persons. Knowledge of statistics is applicable in day to day life in different ways. In daily life it means general calculation of items, in railway statistics means the number of trains operating, number of passenger’s freight etc. and so on. Thus statistics is used by people to take decision about the problems on the basis of different type of quantitative and qualitative information available to them. However, in behavioural sciences, the word ‘statistics’ means something different from the common concern of it. Prime function of statistic is to draw statistical inference about population on the basis of available quantitative information. Overall, statistical methods deal with reduction of data to convenient descriptive terms and drawing some inferences from them. This unit focuses on the above aspects of statistics. 1.1 OBJECTIVES After going through this unit, you will be able to: Define the term statistics; Explain the status of statistics; Describe the nature of statistics; State basic concepts used in statistics; and Analyse the uses and misuses of statistics.
    [Show full text]
  • The Landscape of R Packages for Automated Exploratory Data Analysis by Mateusz Staniak and Przemysław Biecek
    CONTRIBUTED RESEARCH ARTICLE 1 The Landscape of R Packages for Automated Exploratory Data Analysis by Mateusz Staniak and Przemysław Biecek Abstract The increasing availability of large but noisy data sets with a large number of heterogeneous variables leads to the increasing interest in the automation of common tasks for data analysis. The most time-consuming part of this process is the Exploratory Data Analysis, crucial for better domain understanding, data cleaning, data validation, and feature engineering. There is a growing number of libraries that attempt to automate some of the typical Exploratory Data Analysis tasks to make the search for new insights easier and faster. In this paper, we present a systematic review of existing tools for Automated Exploratory Data Analysis (autoEDA). We explore the features of fifteen popular R packages to identify the parts of the analysis that can be effectively automated with the current tools and to point out new directions for further autoEDA development. Introduction With the advent of tools for automated model training (autoML), building predictive models is becoming easier, more accessible and faster than ever. Tools for R such as mlrMBO (Bischl et al., 2017), parsnip (Kuhn and Vaughan, 2019); tools for python such as TPOT (Olson et al., 2016), auto-sklearn (Feurer et al., 2015), autoKeras (Jin et al., 2018) or tools for other languages such as H2O Driverless AI (H2O.ai, 2019; Cook, 2016) and autoWeka (Kotthoff et al., 2017) supports fully- or semi-automated feature engineering and selection, model tuning and training of predictive models. Yet, model building is always preceded by a phase of understanding the problem, understanding of a domain and exploration of a data set.
    [Show full text]
  • Univariate Analysis and Normality Test Using SAS, STATA, and SPSS
    © 2002-2006 The Trustees of Indiana University Univariate Analysis and Normality Test: 1 Univariate Analysis and Normality Test Using SAS, STATA, and SPSS Hun Myoung Park This document summarizes graphical and numerical methods for univariate analysis and normality test, and illustrates how to test normality using SAS 9.1, STATA 9.2 SE, and SPSS 14.0. 1. Introduction 2. Graphical Methods 3. Numerical Methods 4. Testing Normality Using SAS 5. Testing Normality Using STATA 6. Testing Normality Using SPSS 7. Conclusion 1. Introduction Descriptive statistics provide important information about variables. Mean, median, and mode measure the central tendency of a variable. Measures of dispersion include variance, standard deviation, range, and interquantile range (IQR). Researchers may draw a histogram, a stem-and-leaf plot, or a box plot to see how a variable is distributed. Statistical methods are based on various underlying assumptions. One common assumption is that a random variable is normally distributed. In many statistical analyses, normality is often conveniently assumed without any empirical evidence or test. But normality is critical in many statistical methods. When this assumption is violated, interpretation and inference may not be reliable or valid. Figure 1. Comparing the Standard Normal and a Bimodal Probability Distributions Standard Normal Distribution Bimodal Distribution .4 .4 .3 .3 .2 .2 .1 .1 0 0 -5 -3 -1 1 3 5 -5 -3 -1 1 3 5 T-test and ANOVA (Analysis of Variance) compare group means, assuming variables follow normal probability distributions. Otherwise, these methods do not make much http://www.indiana.edu/~statmath © 2002-2006 The Trustees of Indiana University Univariate Analysis and Normality Test: 2 sense.
    [Show full text]
  • Chapter Four: Univariate Statistics SPSS V11 Chapter Four: Univariate Statistics
    Chapter Four: Univariate Statistics SPSS V11 Chapter Four: Univariate Statistics Univariate analysis, looking at single variables, is typically the first procedure one does when examining first time data. There are a number of reasons why it is the first procedure, and most of the reasons we will cover at the end of this chapter, but for now let us just say we are interested in the "basic" results. If we are examining a survey, we are interested in how many people said, "Yes" or "No", or how many people "Agreed" or "Disagreed" with a statement. We aren't really testing a traditional hypothesis with an independent and dependent variable; we are just looking at the distribution of responses. The SPSS tools for looking at single variables include the following procedures: Frequencies, Descriptives and Explore all located under the Analyze menu. This chapter will use the GSS02A file used in earlier chapters, so start SPSS and bring the file into the Data Editor. ( See Chapter 1 to refresh your memory on how to start SPSS). To begin the process start SPSS, then open the data file. Under the Analyze menu, choose Descriptive Statistics and the procedure desired: Frequencies, Descriptives, Explore, Crosstabs. Frequencies Generally a frequency is used for looking at detailed information in a nominal (category) data set that describes the results. Categorical data is for variables such as gender i.e. males are coded as "1" and females are coded as "2." Frequencies options include a table showing counts and percentages, statistics including percentile values, central tendency, dispersion and distribution, and charts including bar charts and histograms.
    [Show full text]
  • Bivariate Analysis: the Statistical Analysis of the Relationship Between Two Variables
    bivariate analysis: The statistical analysis of the relationship between two variables. cell frequency: The number of cases in a cell of a cross-tabulation (contingency table). chi-square (χ2) test for independence: A test of statistical significance used to assess the likelihood that an observed association between two variables could have occurred by chance. consistency checking: A data-cleaning procedure involving checking for unreasonable patterns of responses, such as a 12-year-old who voted in the last US presidential election. correlation coefficient: A statistical measure of the strength and direction of a linear relationship between two variables; it may vary from −1 to 0 to +1. data cleaning: The detection and correction of errors in a computer datafile that may have occurred during data collection, coding, and/or data entry. data matrix: The form of a computer datafile, with rows as cases and columns as variables; each cell represents the value of a particular variable (column) for a particular case (row). data processing: The preparation of data for analysis. descriptive statistics: Procedures for organizing and summarizing data. dummy variable: A variable or set of variable categories recoded to have values of 0 and 1. Dummy coding may be applied to nominal- or ordinal-scale variables for the purpose of regression or other numerical analysis. frequency distribution: A tabulation of the number of cases falling into each category of a variable. histogram: A graphic display in which the height of a vertical bar represents the frequency or percentage of cases in each category of an interval/ratio variable. imputation: A procedure for handling missing data in which missing values are assigned based on other information, such as the sample mean or known values of other variables.
    [Show full text]
  • Instructions for Univariate Analysis
    Instructions for Univariate Analysis 1. Select a dataset from the pull‐down menu. 2. Select a table from the pull‐down menu. 3. Select a variable for which a univariate analysis is desired. 4. You may use the Add Variable feature or the Shift‐Select or Ctrl‐Select options to analyze multiple variables in a single run. 1 6/22/2012 5. Click the Run button to generate the desired descriptive statistical analysis. The results will appear at the bottom. The results are arranged in a series of tabs, providing the following information: a. Descriptive statistics (standard descriptors such as sample size, mean, median, standard deviation, coefficient of variation) b. Percentiles showing the value of the variable corresponding to every 10th percentile c. Frequency providing a detailed frequency distribution, including a cumulative distribution, for the selected variables d. Frequency distribution graph providing a graph depicting the frequency distribution, both in absolute number and in percent e. Percent distribution graph providing the frequency distribution in percent and cumulative percent 6. Use the Clear Results button to clear out all of the results and start fresh. 7. The web portal allows you to perform either unweighted or weighted data analysis. Generally, it is recommended that weighted analysis be conducted so that values reflective of population characteristics are obtained. However, if you are interested in the numbers based on the raw sample, then choose unweighted analysis. The web portal has already identified the appropriate weight variable for each data set and made it available as an option. 8. You may choose to undertake the cross tabulation analysis for any selected subsample of the survey sample.
    [Show full text]
  • Univariate Analyses Can Be Used for Which of the Following
    Chapter Coverage and Supplementary Lecture Material: Chapter 17 Please replace the hand-out in class with a print-out or on-line reading of this material. There were typos and a few errors in that hand-out. This covers (and then some) the material in pages 275-278 of Chapter 17, which is all you are responsible for at this time. There are a number of important concepts in Chapter 17 which need to be understood in order to proceed further with the exercises and to prepare for the data analysis we will do on data collected in our group research projects. I want to go over a number of them, and will be sure to cover those items that will be on the quiz. First, at the outset of the chapter, the authors explain the difference between quantitative analysis which is descriptive and that which is inferential. And then the explain the difference between univariate analysis, bivariate analysis and multivariate analysis. Descriptive data analysis, the authors point out on page 275, involves focus in on the data collected per se, which describes what Ragin and Zaret refer to as the object of the research - what has been studied (objectively one hopes). This distinction between the object of research and the subject of research is an important one, and will help us understand the distinction between descriptive and inferential data analysis. As I argue in the paper I have written: “In thinking about the topic of a research project, it is helpful to distinguish between the object and subject of research.
    [Show full text]
  • UNIT 2 DESCRIPTIVE STATISTICS Introduction to Statistics
    UNIT 2 DESCRIPTIVE STATISTICS Introduction to Statistics Structure 2.0 Introduction 2.1 Objectives 2.2 Meaning of Descriptive Statistics 2.3 Organising Data 2.3.1 Classification 2.3.2 Tabulation 2.3.3 Graphical Presentation of Data 2.3.4 Diagrammatical Presentation of Data 2.4 Summarising Data 2.4.1 Measures of Central Tendency 2.4.2 Measures of Dispersion 2.5 Use of Descriptive Statistics 2.6 Let Us Sum Up 2.7 Unit End Questions 2.8 Glossary 2.9 Suggested Readings 2.0 INTRODUCTION We have learned in the previous unit that looking at the functions of statistics point of view, statistics may be descriptive, correlational and inferential. In this unit we shall discuss the various aspects of descriptive statistics, particularly how to organise and discribe the data. Most of the observations in this universe are subject to variability, especially observations related to human behaviour. It is a well known fact that Attitude, Intelligence, Personality, etc. differ from individual to individual. In order to make a sensible definition of the group or to identify the group with reference to their observations/ scores, it is necessary to express them in a precise manner. For this purpose observations need to be expressed as a single estimate which summarises the observations. Such single estimate of the series of data which summarises the distribution are known as parameters of the distribution. These parameters define the distribution completely. In this unit we will be focusing on descriptive statistics, the characteristic features and the various statistics used in this category.
    [Show full text]
  • Normality Testing Sas Spss.Pdf
    © 2002-2008 The Trustees of Indiana University Univariate Analysis and Normality Test: 1 I n d i a n a U n i v e r s i t y U n i v e r s i t y I n f o r m a t i o n T e c h n o l o g y S e r v i c e s Univariate Analysis and Normality Test Using SAS, Stata, and SPSS* Hun Myoung Park, Ph.D. © 2002-2008 Last modified on November 2008 University Information Technology Services Center for Statistical and Mathematical Computing Indiana University 410 North Park Avenue Bloomington, IN 47408 (812) 855-4724 (317) 278-4740 http://www.indiana.edu/~statmath * The citation of this document should read: “Park, Hun Myoung. 2008. Univariate Analysis and Normality Test Using SAS, Stata, and SPSS. Working Paper. The University Information Technology Services (UITS) Center for Statistical and Mathematical Computing, Indiana University.” http://www.indiana.edu/~statmath/stat/all/normality/index.html http://www.indiana.edu/~statmath © 2002-2008 The Trustees of Indiana University Univariate Analysis and Normality Test: 2 This document summarizes graphical and numerical methods for univariate analysis and normality test, and illustrates how to do using SAS 9.1, Stata 10 special edition, and SPSS 16.0. 1. Introduction 2. Graphical Methods 3. Numerical Methods 4. Testing Normality Using SAS 5. Testing Normality Using Stata 6. Testing Normality Using SPSS 7. Conclusion 1. Introduction Descriptive statistics provide important information about variables to be analyzed. Mean, median, and mode measure central tendency of a variable.
    [Show full text]
  • Measures of Central Tendency
    © Jones & Bartlett Learning, LLC © Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION © Jones & Bartlett Learning, LLC © Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION Chapter 2 © Jones & BartlettMeasures Learning, LLC of Central© Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION Tendency © Jones & Bartlett Learning, LLC © Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION Learning Objectives ■■©Understand Jones & how Bartlett frequency Learning, tables are used LLC in statistical analysis. © Jones & Bartlett Learning, LLC ■■NOTExplain FOR the conventionsSALE OR for DISTRIBUTION building distributions. NOT FOR SALE OR DISTRIBUTION ■■ Understand the mode, median, and mean as measures of central tendency. ■■ Identify the proper measure of central tendency to use for each level of measure- ment. © Jones & Bartlett■■ Explain Learning, how to calculate LLC the mode, median, and mean.© Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION Key Terms bivariate median central tendency mode © Jones & Bartlett Learning,frequency LLC distribution © Jonesmultivariate & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTIONmean NOT univariateFOR SALE OR DISTRIBUTION Now we begin statistical analysis. Statistical analysis may be broken down into three broad categories: © Jones■■ Univariate & Bartlett analyses Learning, LLC © Jones & Bartlett Learning, LLC NOT■■ Bivariate FOR SALE analyses OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION ■■ Multivariate analyses These divisions are fairly straightforward. Univariate analyses deal with one variable at a time. Bivariate analyses compare two variables to each other to see how they dif- © Jones & Bartlett Learning, LLC © Jones & Bartlett Learning, LLC NOT FOR SALE OR DISTRIBUTION NOT FOR SALE OR DISTRIBUTION 17 © Jones & Bartlett Learning, LLC © Jones & Bartlett Learning, LLC.© NOTJones FOR SALE & OR Bartlett DISTRIBUTION.
    [Show full text]
  • A Comparative Study of Univariate and Multivariate Methodological Approaches to Educational Research Robert M
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1989 A comparative study of univariate and multivariate methodological approaches to educational research Robert M. Crawford Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Educational Assessment, Evaluation, and Research Commons, and the Teacher Education and Professional Development Commons Recommended Citation Crawford, Robert M., "A comparative study of univariate and multivariate methodological approaches to educational research " (1989). Retrospective Theses and Dissertations. 8923. https://lib.dr.iastate.edu/rtd/8923 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. INFORMATION TO USERS The most advanced technology has been used to photo­ graph and reproduce this manuscript from the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type of computer printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion.
    [Show full text]
  • Skewness to Test Normality for Mean Comparison
    International Journal of Assessment Tools in Education 2020, Vol. 7, No. 2, 255–265 https://doi.org/10.21449/ijate.656077 Published at https://dergipark.org.tr/en/pub/ijate Research Article Parametric or Non-parametric: Skewness to Test Normality for Mean Comparison Fatih Orcan 1,* 1Department of Educational Measurement and Evaluation, Trabzon University, Turkey ARTICLE HISTORY Abstract: Checking the normality assumption is necessary to decide whether a Received: Dec 06, 2019 parametric or non-parametric test needs to be used. Different ways are suggested in literature to use for checking normality. Skewness and kurtosis values are one of Revised: Apr 22, 2020 them. However, there is no consensus which values indicated a normal distribution. Accepted: May 24, 2020 Therefore, the effects of different criteria in terms of skewness values were simulated in this study. Specifically, the results of t-test and U-test are compared KEYWORDS under different skewness values. The results showed that t-test and U-test give Normality test, different results when the data showed skewness. Based on the results, using Skewness, skewness values alone to decide about normality of a dataset may not be enough. Mean comparison, Therefore, the use of non-parametric tests might be inevitable. Non-parametric, 1. INTRODUCTION Mean comparison tests, such as t-test, Analysis of Variance (ANOVA) or Mann-Whitney U test, are frequently used statistical techniques in educational sciences. The techniques used differ according to the properties of the data sets such as normality or equal variance. For example, if the data is not normally distributed Mann-Whitney U test is used instead of independent sample t-test.
    [Show full text]