Documentation of Validity for the AT-SAT Computerized Test Battery

Total Page:16

File Type:pdf, Size:1020Kb

Documentation of Validity for the AT-SAT Computerized Test Battery DOT/FAA/AM-01/6 Documentation of Validity for the AT-SAT Office of Aviation Medicine Washington, D.C. 20591 Computerized Test Battery Volume II R.A. Ramos Human Resources Research Organization Alexandria, VA 22314-1591 Michael C. Heil Carol A. Manning Civil Aeromedical Institute Federal Aviation Administration Oklahoma City, OK 73125 March 2001 Final Report This document is available to the public through the National Technical Information Service, Springfield, Virginia 22161. U.S. Department of Transportation Federal Aviation Administration NOTICE This document is disseminated under the sponsorship of the U.S. Department of Transportation in the interest of information exchange. The United States Government assumes no liability for the contents thereof. Technical Report Documentation Page 1. Report No. 2. Government Accession No. 3. Recipient's Catalog No. DOT/FAA/AM-01/6 4. Title and Subtitle 5. Report Date Documentation of Validity for the AT-SAT Computerized Test March 2001 Battery, Volume II 6. Performing Organization Code 7. Author(s) 8. Performing Organization Report No. Ramos, R.A.1, Heil, M.C.2, and Manning, C.A.2 9. Performing Organization Name and Address 10. Work Unit No. (TRAIS) 1Human Resources Research 2FAA Civil Aeromedical Institute Organization P. O. Box 25082 11. Contract or Grant No. 68 Canal Center Plaza, Suite 400 Oklahoma City, OK 73125 Alexandria, VA 22314-1591 12. Sponsoring Agency name and Address 13. Type of Report and Period Covered Office of Aviation Medicine Federal Aviation Administration 800 Independence Ave., S. W. Washington, D.C. 20591 14. Sponsoring Agency Code 15. Supplemental Notes Work was accomplished under approved subtask AM-B-99-HRR-517 16. Abstract This document is a comprehensive report on a large-scale research project to develop and validate a computerized selection battery to hire Air Traffic Control Specialists (ATCSs) for the Federal Aviation Administration (FAA). The purpose of this report is to document the validity of the Air Traffic Selection and Training (AT-SAT) battery according to legal and professional guidelines. An overview of the project is provided, followed by a history of the various job analyses efforts. Development of predictors and criterion measures are given in detail. The document concludes with the presentation of the validation of predictors and analyses of archival data. 17. Key Words 18. Distribution Statement Air Traffic Controllers, Selection, Assessment, Document is available to the public through the Job Analyses National Technical Information Service, Springfield, Virginia 22161 19. Security Classif. (of this report) 20. Security Classif. (of this page) 21. No. of Pages 22. Price Unclassified Unclassified 179 Form DOT F 1700.7 (8-72) Reproduction of completed page authorized i TABLE OF CONTENTS VOLUME II Page CHAPTER 4 - DEVELOPMENT OF CRITERION MEASURES OF AIR TRAFFIC CONTROLLER PERFORMANCE ................... 1 CBPM ..................................................................................................................................................... 1 CHAPTER 5.1 - FIELD PROCEDURES FOR CONCURRENT VALIDATION STUDY ....................................................... 13 CHAPTER 5.2 - DEVELOPMENT OF PSEUDO-APPLICANT SAMPLE ......................................................................... 17 CHAPTER 5.3 - DEVELOPMENT OF DATA BASE ................................................................................................. 21 CHAPTER 5.4 - BIOGRAPHICAL AND COMPUTER EXPERIENCE INFORMATION: DEMOGRAPHICS FOR THE VALIDATION STUDY .................................................................................................................................................. 31 Total Sample ....................................................................................................................................... 31 Controller Sample ............................................................................................................................... 31 Pseudo-Applicant Sample .................................................................................................................... 32 Computer Use and Experience Questionnaire ..................................................................................... 32 Performance Differences...................................................................................................................... 33 Relationship Between Cue-Plus and Predictor Scores .......................................................................... 33 Summary............................................................................................................................................. 35 CHAPTER 5.5 - PREDICTOR-CRITERION ANALYSES ............................................................................................. 37 CHAPTER 5.6 - ANALYSES OF GROUP DIFFERENCES AND FAIRNESS ..................................................................... 43 CHAPTER 6 - THE RELATIONSHIP OF FAA ARCHIVAL DATA TO AT-SAT PREDICTOR AND CRITERION MEASURES .. 49 Previous ATC Selection Tests .............................................................................................................. 49 Other Archival Data Obtained for ATC Candidates ........................................................................... 51 Archival Criterion Measures ................................................................................................................ 52 Historical Studies of Validity of Archival Measures .............................................................................52 Relationships Between Archival Data and AT-SAT Measures .............................................................. 54 REFERENCES ................................................................................................................................................ 61 List of Figures and Tables Figures Figure 4.1. Map of CBPM Airspace............................................................................................................. 67 Figure 4.2. Airspace Summary: Sector 05 in Hub Center ............................................................................68 Figure 4.3. Example CBPM Item ................................................................................................................ 69 Figure 4.4. Aero Center Airspace ................................................................................................................. 70 Figure 5.2.1. Sample Classified Newspaper Advertisement for Soliciting Civilian Pseudo-Applicants ............70 Figure 5.2.2. Sample flyer advertisement for soliciting civilian pseudo-applicants .......................................... 71 Figure 5.3.1. AT-SAT Data Base (*) ................................................................................................................ 72 Figure 5.3.2. CD-ROM Directory Structure of AT-SAT Data Base ................................................................73 Figure 5.5.1. Expected Performance: OPM vs. AT-SAT ..................................................................................74 Figure 5.5.2. Percentage of Selected Applicants whose Expected Performance is in the Top Third of Current Controllers: OPM vs. AT-SAT ................................................................................................... 75 Figure 5.6.1. Fairness Regression for Blacks Using AT-SAT Battery Score and Composite Criterion ..............75 iii Figure 5.6.2. Fairness Regression for Hispanics Using AT-SAT Battery Score and Composite Criterion.........76 Figure 5.6.3. Fairness Regression for Females Using AT-SAT Battery Score and Composite Criterion ............77 Figure 5.6.4. Confidence Intervals for the Slopes in the Fairness Regressions .................................................78 Figure 5.6.5. Expected Score Frequency by Applicant Group .........................................................................79 Figure 5.6.6. Percent Passing by Recruitment Strategy .................................................................................... 80 Tables Table 4.1. CBPM Development and Scaling Participants: Biographical Information .................................81 Table 4.2. CBPM Scaling Workshops: Interrater Reliability Results ...........................................................82 Table 4.3. Performance Categories for Behavior Summary Scales ...............................................................83 Table 4.4. Pilot Test Results: Computer-Based Performance Measure (CBPM) Distribution of Scores ................................................................................................................ 84 Table 4.5. Pilot Test Results: Means and Standard Deviations for Ratings on Each Dimension .................85 Table 4.6. Pilot Test Results: Interrater Reliabilities for Ratings ................................................................................................................. 85 Table 4.7. HFPM Pilot Test Results - Correlations Between Ratings for Rater Pairs (Collapsed Across Ratee) Both Across All Scenarios and Within Each Scenario .....................................................87 Table 4.8. Rater-Ratee Assignments ........................................................................................................... 88 Table 4.9. Computer-Based Performance Measure (CBPM): Distribution of Scores in Validation Sample 88 Table 4.10. Number and Percentage of Supervisor Ratings at Each
Recommended publications
  • Heterogeneous Versus Homogeneous Measures: a Meta-Analysis of Predictive Efficacy
    HETEROGENEOUS VERSUS HOMOGENEOUS MEASURES: A META-ANALYSIS OF PREDICTIVE EFFICACY A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy By SUZANNE L. DEAN M.S., Wright State University, 2008 _________________________________________ 2016 Wright State University WRIGHT STATE UNIVERSITY GRADUATE SCHOOL January 3, 2016 I HEREBY RECOMMEND THAT THE DISSERTATION PREPARED UNDER MY SUPERVISION BY Suzanne L. Dean ENTITLED Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Doctor of Philosophy. _____________________________ Corey E. Miller, Ph.D. Dissertation Director _____________________________ Scott N. J. Watamaniuk, Ph.D. Graduate Program Director _____________________________ Robert E. W. Fyffe, Ph.D. Vice President for Research and Dean of the Graduate School Committee on Final Examination _____________________________ Corey E. Miller, Ph.D. _____________________________ David LaHuis, Ph.D. _____________________________ Allen Nagy, Ph.D. _____________________________ Debra Steele-Johnson, Ph.D. COPYRIGHT BY SUZANNE L. DEAN 2016 ABSTRACT Dean, Suzanne L. Ph.D., Industrial/Organizational Psychology Ph.D. program, Wright State University, 2016. Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy. A meta-analysis was conducted to compare the predictive validity and adverse impact of homogeneous and heterogeneous predictors on objective and subjective criteria for different sales roles. Because job performance is a dynamic and complex construct, I hypothesized that equally complex, heterogeneous predictors would have stronger correlations with objective and subjective criteria than homogeneous predictors. Forty-seven independent validation studies (N = 3,378) qualified for inclusion in this study. In general, heterogeneous predictors did not demonstrate significantly stronger correlations with the performance criteria than homogeneous predictors.
    [Show full text]
  • On the Validity of Reading Assessments GOTHENBURG STUDIES in EDUCATIONAL SCIENCES 328
    On the Validity of Reading Assessments GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO © STEFAN JOHANSSON, 2013 ISBN 978-91-7346-736-0 ISSN 0436-1121 ISSN 1653-0101 Thesis in Education at the Department of Education and Special Education The thesis is also available in full text on http://hdl.handle.net/2077/32012 Photographer cover: Rebecka Karlsson Distribution: ACTA UNIVERSITATIS GOTHOBURGENSIS Box 222 SE-405 30 Göteborg, Sweden Print: Ale Tryckteam, Bohus 2013 Abstract Title: On the Validity of Reading Assessments: Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Language: English with a Swedish summary Keywords: Validity; Validation; Assessment; Teacher judgements; External tests; PIRLS 2001; Self-assessment; Multilevel models; Structural Equation Modeling; Socioeconomic status; Gender ISBN: 978-91-7346-736-0 The purpose of this thesis is to examine validity issues in different forms of assessments; teacher judgements, external tests, and pupil self-assessment in Swedish primary schools. The data used were selected from a large-scale study––PIRLS 2001––in which more than 11000 pupils and some 700 teachers from grades 3 and 4 participated. The primary method used in the secondary analyses to investigate validity issues of the assessment forms is multilevel Structural Equation Modeling (SEM) with latent variables. An argument-based approach to validity was adopted, where possible weaknesses in assessment forms were addressed.
    [Show full text]
  • Questionnaire Validity and Reliability
    Questionnaire Validity and reliability Department of Social and Preventive Medicine, Faculty of Medicine Outlines Introduction What is validity and reliability? Types of validity and reliability. How do you measure them? Types of Sampling Methods Sample size calculation G-Power ( Power Analysis) Research • The systematic investigation into and study of materials and sources in order to establish facts and reach new conclusions • In the broadest sense of the word, the research includes gathering of data in order to generate information and establish the facts for the advancement of knowledge. ● Step I: Define the research problem ● Step 2: Developing a research plan & research Design ● Step 3: Define the Variables & Instrument (validity & Reliability) ● Step 4: Sampling & Collecting data ● Step 5: Analysing data ● Step 6: Presenting the findings A questionnaire is • A technique for collecting data in which a respondent provides answers to a series of questions. • The vehicle used to pose the questions that the researcher wants respondents to answer. • The validity of the results depends on the quality of these instruments. • Good questionnaires are difficult to construct; • Bad questionnaires are difficult to analyze. •Identify the goal of your questionnaire •What kind of information do you want to gather with your questionnaire? • What is your main objective? • Is a questionnaire the best way to go about collecting this information? Ch 11 6 How To Obtain Valid Information • Ask purposeful questions • Ask concrete questions • Use time periods
    [Show full text]
  • Running Head: the STRUCTURE of PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: a Meta-An
    Running head: THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: A Meta-Analysis of the Psychopathic Personality Inventory Jared R. Ruchensky, John F. Edens, Katherine S. Corker, M. Brent Donnellan, Edward A. Witt, Daniel M. Blonigen Version: 11/17/2017 In press, Psychological Assessment © 2017, American Psychological Association. This paper is not the copy of record and may not exactly replicate the final, authoritative version of the article. The final article will be available, upon publication, via its DOI: 10.1037/pas0000520 THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 2 Abstract Which core traits exemplify psychopathic personality disorder is a hotly debated question within psychology, particularly regarding the role of ostensibly adaptive traits such as stress immunity, social potency, and fearlessness. Much of the research on the inter-relationships among putatively adaptive and more maladaptive traits of psychopathy has focused on the factor structure of the Psychopathic Personality Inventory (PPI) and its revision, the Psychopathic Personality Inventory – Revised (PPI-R). These instruments include content scales that have coalesced to form 2 higher-order factors in some (but not all) prior studies: Fearless Dominance and Self-Centered Impulsivity. Given the inconsistencies in prior research, we performed a meta- analytic factor analysis of the 8 content scales from these instruments (total N > 18,000) and found general support for these 2 dimensions in community samples. The structure among offender samples (e.g., prisoners, forensic patients) supported a three-factor model in which the Fearlessness content scale uniquely loaded onto Self-centered Impulsivity (rather than Fearless Dominance). There were also indications that the Stress Immunity content scale had different relations to the other PPI scales in offender versus community samples.
    [Show full text]
  • Validity: Theoretical Basis Importance of Validity
    Psychological Measurement mark.hurlstone @uwa.edu.au Validity Validity: Theoretical Basis Importance of Validity Classic & Contemporary Approaches PSYC3302: Psychological Measurement and Its Trinitarian View Unitary View Applications Test Content Internal Structure Response Processes Mark Hurlstone Associations With Univeristy of Western Australia Other Variables Consequences of Testing Other Week 5 Perspectives Reliability & Validity [email protected] Psychological Measurement Learning Objectives Psychological Measurement mark.hurlstone @uwa.edu.au Validity • Introduction to the concept of validity Importance of Validity • Overview of the theoretical basis of validity: Classic & Contemporary 1 Trinitarian view of validity Approaches Trinitarian View 2 Unitary view of validity Unitary View Test Content • Alternative perspectives on validity Internal Structure Response Processes • Contrasting validity and reliability Associations With Other Variables Consequences of Testing Other Perspectives Reliability & Validity [email protected] Psychological Measurement Validity in Everyday Usage Psychological Measurement mark.hurlstone @uwa.edu.au • In everyday language, we say something is valid if it is Validity sound, meaningful, or supported by evidence Importance of Validity • e.g., we may speak of a valid theory, a valid argument, Classic & or a valid reason Contemporary Approaches • In legal terminology, lawyers say something is valid if it is Trinitarian View Unitary View "executed with the proper formalities"—such as a valid
    [Show full text]
  • Criterion-Referenced Measurement for Educational Evaluation and Selection
    Criterion-Referenced Measurement for Educational Evaluation and Selection Christina Wikström Department of Educational Measurement Umeå University No. 1 Department of Educational Measurement Umeå University Thesis 2005 Printed by Umeå University March 2005 © Christina Wikström ISSN 1652-9650 ISBN 91-7305-865-3 Abstract In recent years, Sweden has adopted a criterion-referenced grading system, where the grade outcome is used for several purposes, but foremost for educational evaluation on student- and school levels as well as for selection to higher education. This thesis investigates the consequences of using criterion-referenced measurement for both educational evaluation and selection purposes. The thesis comprises an introduction and four papers that empirically investigate school grades and grading practices in Swedish upper secondary schools. The first paper investigates the effect of school competition on the school grades. The analysis focuses on how students in schools with and without competition are ranked, based on their grades and SweSAT scores. The results show that schools that are exposed to competition tend to grade their students higher than other schools. This effect is found to be related to the use of grades as quality indicators for the schools, which means that schools that compete for their students tend to be more lenient, hence inflating the grades. The second paper investigates grade averages over a six-year period, starting with the first cohort who graduated from upper secondary school with a GPA based on criterion-referenced grades. The results show that grades have increased every year since the new grading system was introduced, which cannot be explained by improved performances, selection effects or strategic course choices.
    [Show full text]
  • Validity (Content Validity) Is Measured First, and Then Validity Reliability
    the validity (content validity) is measured first, and then Validity reliability. Validity refers to a test's accuracy. “The soundness or appropriateness of a test or instrument in measuring what it is designed to measure” e.g: the exam is supposed to test the knowledge. A test is valid when it measures what it is intended to measure. The intended uses for most tests fall into one of three categories, and each category is associated with a different method for establishing validity: (these types will be discussed in detailes later on) – The test is used to obtain information about an examinee's familiarity with a particular content or behavior domain: content validity. Face validity can be considered as part of content validity. – The test is administered to determine the extent to which an examinee possesses a particular hypothetical trait: construct validity. – The test is used to estimate or predict an examinee's standing or performance on an external criterion: criterion-related validity. Types of Experimental Validity • Internal – Is the experimenter measuring the effect of the independent variable on the dependent variable? • External – Can the results be generalised to the wider population? There are 4 types of validity (Analysis done without 3) (analyzed statistically by SPSS) SPSS (manually)) 4) Criterion 1) 2) Reliability and validity are strongly related to each other. Validity of Measurement • Meaning analysis: the process of attaching meaning to concepts – Face validity – Content validity • Empirical analysis: how well our procedures (operational definition) accurately observe what we want to measure – Criterion validity. – Construct validity Done before content validity and it’s not Face Validity performed alone.
    [Show full text]
  • The Psychopathic Personality Inventory-Revised
    The Psychopathic Personality Inventory-Revised: Evaluation of Its Psychometric Properties, Incremental Validity, and Moderating Effects of Gender in a Correctional Sample by Natalia L. Nikolova M.A. (Psychology), Simon Fraser University, 2009 Dissertation Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in the Department of Psychology Faculty of Arts and Social Sciences Natalia L. Nikolova 2013 SIMON FRASER UNIVERSITY Summer 2013 Approval Name: Natalia L. Nikolova Degree: Doctor of Philosophy (Psychology) Title of Thesis: The Psychopathic Personality Inventory-Revised: Evaluation of Its Psychometric Properties, Incremental Validity, and Moderating Effects of Gender in a Correctional Sample Examining Committee: Chair: Robert Ley Associate Professor Kevin S. Douglas Senior Supervisor Associate Professor Stephen D. Hart Supervisor Professor Marlene M. Moretti Supervisor Professor William Glackman Internal Examiner Associate Professor School of Criminology Martin Sellbom External Examiner Assistant Professor Department of Psychology University of Alabama Date Defended/Approved: May 10, 2013 ii Partial Copyright Licence iii Ethics Statement The author, whose name appears on the title page of this work, has obtained, for the research described in this work, either: a. human research ethics approval from the Simon Fraser University Office of Research Ethics, or b. advance approval of the animal care protocol from the University Animal Care Committee of Simon Fraser University; or has conducted the research c. as a co-investigator, collaborator or research assistant in a research project approved in advance, or d. as a member of a course approved in advance for minimal risk human research, by the Office of Research Ethics. A copy of the approval letter has been filed at the Theses Office of the University Library at the time of submission of this thesis or project.
    [Show full text]
  • Lecture 5: Principles of Psychometrics: Validity I
    This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this site. Copyright 2006, The Johns Hopkins University and William Eaton. All rights reserved. Use of these materials permitted only in accordance with license rights granted. Materials provided “AS IS”; no representations or warranties provided. User assumes all responsibility for use, and all liability related thereto, and must independently review all materials for accuracy and efficacy. May contain materials owned by others. User is responsible for obtaining permissions for use from third parties as needed. Statistics in Psychosocial Research Lecture 5 Validity I Lecturer: William Eaton Validity The degree to which a test, scale, or assessment measures what it is supposed to measure. Outline • Face Validity • Content Validity • Construct Validity •Internal •External • Criterion Validity • Concurrent • Predictive • Validity and reliability • Validity and utility Classical Test Theory and Reliability: x = Tx + e The observed score equals the true score plus random measurement error Validity x = Tx + es + er The observed score equals the true score plus systematic error (bias) plus random measurement error Face Validity Does the item seem right? CESD: Depression: “I felt depressed” MMPI: Schizophrenia: “I believe in God” • Increases motivation among respondents • Reduces dissatisfaction among respondents • Enhances credibility of results • Improves public relations Face Validity: IQ Test Example 1. What would be the next number in this series? 1 ... 2 ... 3 ... 5 ... 8 ... 13 ... 21 ... 34 ... ?? 47 53 55 62 65 I don't know Face Validity: IQ Test Example 2.
    [Show full text]
  • Checklist to Operationalize Measurement Characteristics of Patient-Reported Outcome Measures David O
    Francis et al. Systematic Reviews (2016) 5:129 DOI 10.1186/s13643-016-0307-4 METHODOLOGY Open Access Checklist to operationalize measurement characteristics of patient-reported outcome measures David O. Francis1,2,3*, Melissa L. McPheeters3,4,5, Meaghan Noud1, David F. Penson2,4,6,7 and Irene D. Feurer2,8 Abstract Background: The purpose of this study was to advance a checklist of evaluative criteria designed to assess patient-reported outcome (PRO) measures’ developmental measurement properties and applicability, which can be used by systematic reviewers, researchers, and clinicians with a varied range of expertise in psychometric measure development methodology. Methods: A directed literature search was performed to identify original studies, textbooks, consensus guidelines, and published reports that propose criteria for assessing the quality of PRO measures. Recommendations from these sources were iteratively distilled into a checklist of key attributes. Preliminary items underwent evaluation through 24 cognitive interviews with clinicians and quantitative researchers. Six measurement theory methodological novices independently applied the final checklist to assess six PRO measures encompassing a variety of methods, applications, and clinical constructs. Agreement between novice and expert scores was assessed. Results: The distillation process yielded an 18-item checklist with six domains: (1) conceptual model, (2) content validity, (3) reliability, (4) construct validity, (5) scoring and interpretation, and (6) respondent burden and presentation. With minimal instruction, good agreement in checklist item ratings was achieved between quantitative researchers with expertise in measurement theory and less experienced clinicians (mean kappa 0.70; range 0.66–0.87). Conclusions: We present a simplified checklist that can help guide systematic reviewers, researchers, and clinicians with varied measurement theory expertise to evaluate the strengths and weakness of candidate PRO measures’ developmental properties and the appropriateness for specific applications.
    [Show full text]
  • Understanding Validity and Reliability in Classroom, School-Wide, Or District- Wide Assessments to Be Used in Teacher/Principal Evaluations
    Understanding Validity and Reliability in Classroom, School-Wide, or District- Wide Assessments to be used in Teacher/Principal Evaluations Warren Shillingburg, PhD January 2016 Introduction As expectations have risen and requirements for student growth expectations have increased across the country, more and more school districts are being asked to develop local assessments and to use them as part of their teacher evaluation process. As districts explore this process, many are concerned about how they can develop local assessments that provide information to help students learn, provide evidence of a teacher’s contribution to student growth, and create reliable and valid assessments. These terms, validity and reliability, can be very complex and difficult for many educators to understand. If you are in a district that has access to appropriate software or the luxury of hiring a statistician to work through formulas, you are in the minority. Most districts cannot afford these expenses and struggle just to find the resources and time to create any assessment, giving very little time and attention to the concepts of validity and reliability. In light of this reality, some districts are choosing to simply not go this route and instead follow the premise that everyone contributes in some way to a child’s reading and math learning, and using the state or national assessments they currently have in these areas to apply to all teachers. Although there can be some validity in this assumption, with the state expectation that all LEAs shall ensure that multiple measures of student academic progress are used (Arizona State Board of Education, 2016) in a teacher/principal evaluation, districts will need to either find other assessments or develop their own assessments to meet this requirement.
    [Show full text]
  • Understanding Validity for Classroom Teachers Slide # Title Presenter
    Presenter’s Notes for PowerPoint: Understanding Validity for Classroom Teachers Slide # Presenter Notes Title This module provides teachers with basic information they need to understand what 1. Understanding validity is and why it is so important. Both reliability and validity are necessary for a Validity for teacher to make accurate inferences about student learning based on assessment Teachers results. For this reason, you need to study both this module and the Reliability module to understand quality assessment of student learning. This commonly used illustration provides visual metaphors of the concepts of Reliability and Validity. • The first target represents a valid and reliable assessment. The bull’s eye in the center of the target is shot with accuracy again and again. The correct target is being aimed at and it is being hit consistently. So validity (or the correct target) and reliability (or consistency) are evident. • The second target shows a scenario in which the shots are consistently striking 2. Validity & in a particular location – but it isn’t on the target. So if this were an assessment it Reliability would be reliable but not valid. • The third target shows another scenario. The shots are aimed at the target, but they are striking the target in a variety of locations and some aren’t hitting the target at all. There is no consistency or reliability and because of this there is also no validity. So it is important to remember that validity and reliability interact with and influence each other. Both are necessary for quality educational tests. This module answers the essential questions: • What is test validity? • What is content-related validity? • What is criterion-related validity? 3.
    [Show full text]