Questionnaire Validity and Reliability

Total Page:16

File Type:pdf, Size:1020Kb

Questionnaire Validity and Reliability Questionnaire Validity and reliability Department of Social and Preventive Medicine, Faculty of Medicine Outlines Introduction What is validity and reliability? Types of validity and reliability. How do you measure them? Types of Sampling Methods Sample size calculation G-Power ( Power Analysis) Research • The systematic investigation into and study of materials and sources in order to establish facts and reach new conclusions • In the broadest sense of the word, the research includes gathering of data in order to generate information and establish the facts for the advancement of knowledge. ● Step I: Define the research problem ● Step 2: Developing a research plan & research Design ● Step 3: Define the Variables & Instrument (validity & Reliability) ● Step 4: Sampling & Collecting data ● Step 5: Analysing data ● Step 6: Presenting the findings A questionnaire is • A technique for collecting data in which a respondent provides answers to a series of questions. • The vehicle used to pose the questions that the researcher wants respondents to answer. • The validity of the results depends on the quality of these instruments. • Good questionnaires are difficult to construct; • Bad questionnaires are difficult to analyze. •Identify the goal of your questionnaire •What kind of information do you want to gather with your questionnaire? • What is your main objective? • Is a questionnaire the best way to go about collecting this information? Ch 11 6 How To Obtain Valid Information • Ask purposeful questions • Ask concrete questions • Use time periods based on importance of the questions • Use conventional language • Use complete sentences • Avoid abbreviations • Use shorter questions Validity and Reliability • Validity: How well does the measure or design do what it purports to do? • Reliability: How consistent or stable is the instrument? –Is the instrument dependable? Which device can measure the body temperature ? 1 37.2 37.2 2 37.2 39.3 3 37.1 34.4 4 37.2 28.2 5 37.2 43.3 6 37.2 37 7 37.2 39 Validity Logical Statistical Construct Face ContenContent Convergent Divergent/ Concurren Predictive t Discriminant t Reliability Consistency Objectivity The Major Decisions in Questionnaire Design 1. Content - What should be asked? 2. Wording - How should each question be phrased? 3. Sequence - In what order should the questions be presented? 4. Layout - What layout will best serve the research objectives? Face Validity – Face validity is the extent to which a test is subjectively viewed as covering the concept it purports to measure – As a check on face validity, test/survey items are sent to experts to obtain suggestions for modification. – Because of its vagueness and subjectivity, psychometricians have abandoned this concept for a long time. Content Validity – In psychometrics, content validity (also known as logical validity) refers to the extent to which a measure represents all facets of a given construct. – Face validity Vs Content validity: • Face validity can be established by one person • Content validity should be checked by a panel, and thus usually it goes hand in hand with inter-rater reliability (Kappa!) The Content Validity Index Content validity has been defined as follows: • (1) ‘‘...the degree to which an instrument has an appropriate sample of items for the construct being measured’’ (Polit & Beck, 2004, p. 423); • (2) ‘‘...whether or not the items sampled for inclusion on the tool adequately represent the domain of content addressed by the instrument’’ (Waltz, Strickland, & Lenz, 2005, p. 155); • (3) ‘‘...the extent to which an instrument adequately samples the research domain of interest when attempting to measure phenomena’’ (Wynd, Schmidt, & Schaefer, 2003, p. 509). Two types of CVIs. • content validity of individual items • content validity of the overall scale. • Researchers use I-CVI information to guide them in revising, deleting, or substituting items • I-CVIs tend only to be reported in methodological studies that focus on descriptions of the content validation process • Most often reported in scale development studies is the CVI S-CVI/UA Proportion of items on a scale that S-CVI achieves a relevance rating of 3 Content Validity of or 4 by all the experts CVI the overall scale Degree to which an instrument has an appropriate sample of items S-CVI/Ave for construct being I-CVI Average of the I- measured CVIs for Content Validity of individual items: Are the items representati Are the items Has each item relevance to Are the items in the ve of concepts Question clarity in term comments instruments concepts related to the of wording consistency? related to the dissertation dissertation topic? topic? Q1 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q2 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q3 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q4 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q5 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Ratings 1= not relevant 2 =somewhat relevant. 3= quite relevant 4= highly relevant. I-CVI, item-level content validity index S-CVI, content validity index for the scale Acceptable standard for the S-CVI recommended a minimum S-CVI of .80. Content validity index The I-CVI expresses the proportion of agreement on the relevancy of each item If the I-CVI is higher than 79%, the item will be appropriate. If it is between 70% and 79%, it needs revision. If it is less than 70% it is eliminated Content validity ratio 푵 풏풆− ퟐ .푪푽푹 = 푵 ퟐ In this formula: • ne = The number of specialists who have chosen the " Necessary " option • N = total number of assessors. Kappa statistic is a consensus index of inter-rater agreement that adjusts for chance agreement and is an important supplement to CVI because Kappa provides information about the degree of agreement beyond chance Evaluation criteria for Kappa is the values above 0.74= excellent between 0.60 and 0.74=good between 0.40 and 0.59= fair To calculate modified kappa statistic, the probability of chance agreement was first calculated for each item by following formula: N PC= [N! /A! (N -A)!]* . 5 . • N= number of experts in a panel and • A= number of panelists who agree that the item is relevant. After calculating I-CVI for all instrument items, finally, kappa was computed by entering the numerical values of probability of chance agreement (PC) and content validity index of each item (I-CVI) in following formula: K= (I-CVI - PC) / (1- PC). Validity Logical Statistical Construct Criterion Face ContenContent Convergent Divergent/ Concurren Predictive t Discriminant t Reliability Consistency Objectivity Criterion Validity • This type of validity is used to measure the ability of an instrument to predict future outcomes. • Validity is usually determined by comparing two instruments ability to predict a similar outcome with a single variable being measured. • There are two major types of criterion validity: – predictive – concurrent Criterion validity • A test has high criterion validity if – It correlates highly with some external benchmark (concurrent) ? – How well does the test correlated with outcome criteria (predictive)? 30 Concurrent Validity • Concurrent validity is used when the two instruments are used to measure the same event at the same time. • It correlates highly with some external benchmark (concurrent) • Example: • DASS -> measuring depression • New instrument -> measuring depression Predictive Validity • Predictive validity is used when the instrument is administered then time is allowed to pass and is measured against the another outcome. • Regression analysis can be applied to establish criterion validity. • An independent variable could be used as a predictor variable and a dependent variable, the criterion variable. • The correlation coefficient between them is called validity coefficients. How is Criterion Validity Measured? • The statistical measure or correlation coefficient tells the degree to which the instrument is valid based on the measured criteria. • What does it look like in an equation? – The symbol “r” denotes the correlation coefficient. – A higher “r” value shows a positive relationship between the instruments. – A mix of high and low “r” values shows a negative relationship. • As a rule of thumb, for absolute value of r: • 0.00-0.19: very weak • 0.20-0.39: weak • 0.40-0.59: moderate • 0.60-0.79: strong • 0.80-1.00: very strong. Predictive Validity Concurrent Validity Validity Logical Statistical Construct Criterion Face ContenContent Convergent Divergent/ Concurren Predictive t Discriminant t Reliability Consistency Objectivity Construct validity • Measuring things that are in our theory of a domain. • The construct is sometimes called a latent variable – You can’t directly observe the construct – You can only measure its surface manifestations – it is concerned with abstract and theoretical construct, construct validity is also known as theoretical validity 39 What are Latent Variables? • Most/all variables in the social world are not directly observable. • This makes them ‘latent’ or hypothetical constructs. • We measure latent variables with observable indicators, e.g. questionnaire items. • We can think of the variance of an observable indicator as being partially caused by: – The latent construct in question – Other factors (error) • I cringe when I have to go to math class. • I am uneasy about going to the board in a math Math class. • I am afraid to ask questions in math class. anxiety • I am always worried about being called on in math class. • I understand math now, but I worry that it's going to get really difficult soon. • Specifying formative versus reflective
Recommended publications
  • Internal Validity Is About Causal Interpretability
    Internal Validity is about Causal Interpretability Before we can discuss Internal Validity, we have to discuss different types of Internal Validity variables and review causal RH:s and the evidence needed to support them… Every behavior/measure used in a research study is either a ... Constant -- all the participants in the study have the same value on that behavior/measure or a ... • Measured & Manipulated Variables & Constants Variable -- when at least some of the participants in the study • Causes, Effects, Controls & Confounds have different values on that behavior/measure • Components of Internal Validity • “Creating” initial equivalence and every behavior/measure is either … • “Maintaining” ongoing equivalence Measured -- the value of that behavior/measure is obtained by • Interrelationships between Internal Validity & External Validity observation or self-report of the participant (often called “subject constant/variable”) or it is … Manipulated -- the value of that behavior/measure is controlled, delivered, determined, etc., by the researcher (often called “procedural constant/variable”) So, every behavior/measure in any study is one of four types…. constant variable measured measured (subject) measured (subject) constant variable manipulated manipulated manipulated (procedural) constant (procedural) variable Identify each of the following (as one of the four above, duh!)… • Participants reported practicing between 3 and 10 times • All participants were given the same set of words to memorize • Each participant reported they were a Psyc major • Each participant was given either the “homicide” or the “self- defense” vignette to read From before... Circle the manipulated/causal & underline measured/effect variable in each • Causal RH: -- differences in the amount or kind of one behavior cause/produce/create/change/etc.
    [Show full text]
  • Validity and Reliability of the Questionnaire for Compliance with Standard Precaution for Nurses
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Cadernos Espinosanos (E-Journal) Rev Saúde Pública 2015;49:87 Original Articles DOI:10.1590/S0034-8910.2015049005975 Marília Duarte ValimI Validity and reliability of the Maria Helena Palucci MarzialeII Miyeko HayashidaII Questionnaire for Compliance Fernanda Ludmilla Rossi RochaIII with Standard Precaution Jair Lício Ferreira SantosIV ABSTRACT OBJECTIVE: To evaluate the validity and reliability of the Questionnaire for Compliance with Standard Precaution for nurses. METHODS: This methodological study was conducted with 121 nurses from health care facilities in Sao Paulo’s countryside, who were represented by two high-complexity and by three average-complexity health care facilities. Internal consistency was calculated using Cronbach’s alpha and stability was calculated by the intraclass correlation coefficient, through test-retest. Convergent, discriminant, and known-groups construct validity techniques were conducted. RESULTS: The questionnaire was found to be reliable (Cronbach’s alpha: 0.80; intraclass correlation coefficient: (0.97) In regards to the convergent and discriminant construct validity, strong correlation was found between compliance to standard precautions, the perception of a safe environment, and the smaller perception of obstacles to follow such precautions (r = 0.614 and r = 0.537, respectively). The nurses who were trained on the standard precautions and worked on the health care facilities of higher complexity were shown to comply more (p = 0.028 and p = 0.006, respectively). CONCLUSIONS: The Brazilian version of the Questionnaire for I Departamento de Enfermagem. Centro Compliance with Standard Precaution was shown to be valid and reliable.
    [Show full text]
  • Assessment Center Structure and Construct Validity: a New Hope
    University of Central Florida STARS Electronic Theses and Dissertations, 2004-2019 2015 Assessment Center Structure and Construct Validity: A New Hope Christopher Wiese University of Central Florida Part of the Psychology Commons Find similar works at: https://stars.library.ucf.edu/etd University of Central Florida Libraries http://library.ucf.edu This Doctoral Dissertation (Open Access) is brought to you for free and open access by STARS. It has been accepted for inclusion in Electronic Theses and Dissertations, 2004-2019 by an authorized administrator of STARS. For more information, please contact [email protected]. STARS Citation Wiese, Christopher, "Assessment Center Structure and Construct Validity: A New Hope" (2015). Electronic Theses and Dissertations, 2004-2019. 733. https://stars.library.ucf.edu/etd/733 ASSESSMENT CENTER STRUCTURE AND CONSTRUCT VALIDITY: A NEW HOPE by CHRISTOPHER W. WIESE B.S., University of Central Florida, 2008 A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Psychology in the College of Sciences at the University of Central Florida Orlando, Florida Summer Term 2015 Major Professor: Kimberly Smith-Jentsch © 2015 Christopher Wiese ii ABSTRACT Assessment Centers (ACs) are a fantastic method to measure behavioral indicators of job performance in multiple diverse scenarios. Based upon a thorough job analysis, ACs have traditionally demonstrated very strong content and criterion-related validity. However, researchers have been puzzled for over three decades with the lack of evidence concerning construct validity. ACs are designed to measure critical job dimensions throughout multiple situational exercises. However, research has consistently revealed that different behavioral ratings within these scenarios are more strongly related to one another (exercise effects) than the same dimension rating across scenarios (dimension effects).
    [Show full text]
  • Construct Validity and Reliability of the Work Environment Assessment Instrument WE-10
    International Journal of Environmental Research and Public Health Article Construct Validity and Reliability of the Work Environment Assessment Instrument WE-10 Rudy de Barros Ahrens 1,*, Luciana da Silva Lirani 2 and Antonio Carlos de Francisco 3 1 Department of Business, Faculty Sagrada Família (FASF), Ponta Grossa, PR 84010-760, Brazil 2 Department of Health Sciences Center, State University Northern of Paraná (UENP), Jacarezinho, PR 86400-000, Brazil; [email protected] 3 Department of Industrial Engineering and Post-Graduation in Production Engineering, Federal University of Technology—Paraná (UTFPR), Ponta Grossa, PR 84017-220, Brazil; [email protected] * Correspondence: [email protected] Received: 1 September 2020; Accepted: 29 September 2020; Published: 9 October 2020 Abstract: The purpose of this study was to validate the construct and reliability of an instrument to assess the work environment as a single tool based on quality of life (QL), quality of work life (QWL), and organizational climate (OC). The methodology tested the construct validity through Exploratory Factor Analysis (EFA) and reliability through Cronbach’s alpha. The EFA returned a Kaiser–Meyer–Olkin (KMO) value of 0.917; which demonstrated that the data were adequate for the factor analysis; and a significant Bartlett’s test of sphericity (χ2 = 7465.349; Df = 1225; p 0.000). ≤ After the EFA; the varimax rotation method was employed for a factor through commonality analysis; reducing the 14 initial factors to 10. Only question 30 presented commonality lower than 0.5; and the other questions returned values higher than 0.5 in the commonality analysis. Regarding the reliability of the instrument; all of the questions presented reliability as the values varied between 0.953 and 0.956.
    [Show full text]
  • Statistical Analysis 8: Two-Way Analysis of Variance (ANOVA)
    Statistical Analysis 8: Two-way analysis of variance (ANOVA) Research question type: Explaining a continuous variable with 2 categorical variables What kind of variables? Continuous (scale/interval/ratio) and 2 independent categorical variables (factors) Common Applications: Comparing means of a single variable at different levels of two conditions (factors) in scientific experiments. Example: The effective life (in hours) of batteries is compared by material type (1, 2 or 3) and operating temperature: Low (-10˚C), Medium (20˚C) or High (45˚C). Twelve batteries are randomly selected from each material type and are then randomly allocated to each temperature level. The resulting life of all 36 batteries is shown below: Table 1: Life (in hours) of batteries by material type and temperature Temperature (˚C) Low (-10˚C) Medium (20˚C) High (45˚C) 1 130, 155, 74, 180 34, 40, 80, 75 20, 70, 82, 58 2 150, 188, 159, 126 136, 122, 106, 115 25, 70, 58, 45 type Material 3 138, 110, 168, 160 174, 120, 150, 139 96, 104, 82, 60 Source: Montgomery (2001) Research question: Is there difference in mean life of the batteries for differing material type and operating temperature levels? In analysis of variance we compare the variability between the groups (how far apart are the means?) to the variability within the groups (how much natural variation is there in our measurements?). This is why it is called analysis of variance, abbreviated to ANOVA. This example has two factors (material type and temperature), each with 3 levels. Hypotheses: The 'null hypothesis' might be: H0: There is no difference in mean battery life for different combinations of material type and temperature level And an 'alternative hypothesis' might be: H1: There is a difference in mean battery life for different combinations of material type and temperature level If the alternative hypothesis is accepted, further analysis is performed to explore where the individual differences are.
    [Show full text]
  • Heterogeneous Versus Homogeneous Measures: a Meta-Analysis of Predictive Efficacy
    HETEROGENEOUS VERSUS HOMOGENEOUS MEASURES: A META-ANALYSIS OF PREDICTIVE EFFICACY A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy By SUZANNE L. DEAN M.S., Wright State University, 2008 _________________________________________ 2016 Wright State University WRIGHT STATE UNIVERSITY GRADUATE SCHOOL January 3, 2016 I HEREBY RECOMMEND THAT THE DISSERTATION PREPARED UNDER MY SUPERVISION BY Suzanne L. Dean ENTITLED Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Doctor of Philosophy. _____________________________ Corey E. Miller, Ph.D. Dissertation Director _____________________________ Scott N. J. Watamaniuk, Ph.D. Graduate Program Director _____________________________ Robert E. W. Fyffe, Ph.D. Vice President for Research and Dean of the Graduate School Committee on Final Examination _____________________________ Corey E. Miller, Ph.D. _____________________________ David LaHuis, Ph.D. _____________________________ Allen Nagy, Ph.D. _____________________________ Debra Steele-Johnson, Ph.D. COPYRIGHT BY SUZANNE L. DEAN 2016 ABSTRACT Dean, Suzanne L. Ph.D., Industrial/Organizational Psychology Ph.D. program, Wright State University, 2016. Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy. A meta-analysis was conducted to compare the predictive validity and adverse impact of homogeneous and heterogeneous predictors on objective and subjective criteria for different sales roles. Because job performance is a dynamic and complex construct, I hypothesized that equally complex, heterogeneous predictors would have stronger correlations with objective and subjective criteria than homogeneous predictors. Forty-seven independent validation studies (N = 3,378) qualified for inclusion in this study. In general, heterogeneous predictors did not demonstrate significantly stronger correlations with the performance criteria than homogeneous predictors.
    [Show full text]
  • Validity and Reliability in Quantitative Studies Evid Based Nurs: First Published As 10.1136/Eb-2015-102129 on 15 May 2015
    Research made simple Validity and reliability in quantitative studies Evid Based Nurs: first published as 10.1136/eb-2015-102129 on 15 May 2015. Downloaded from Roberta Heale,1 Alison Twycross2 10.1136/eb-2015-102129 Evidence-based practice includes, in part, implementa- have a high degree of anxiety? In another example, a tion of the findings of well-conducted quality research test of knowledge of medications that requires dosage studies. So being able to critique quantitative research is calculations may instead be testing maths knowledge. 1 School of Nursing, Laurentian an important skill for nurses. Consideration must be There are three types of evidence that can be used to University, Sudbury, Ontario, given not only to the results of the study but also the demonstrate a research instrument has construct Canada 2Faculty of Health and Social rigour of the research. Rigour refers to the extent to validity: which the researchers worked to enhance the quality of Care, London South Bank 1 Homogeneity—meaning that the instrument mea- the studies. In quantitative research, this is achieved University, London, UK sures one construct. through measurement of the validity and reliability.1 2 Convergence—this occurs when the instrument mea- Correspondence to: Validity sures concepts similar to that of other instruments. Dr Roberta Heale, Validity is defined as the extent to which a concept is Although if there are no similar instruments avail- School of Nursing, Laurentian accurately measured in a quantitative study. For able this will not be possible to do. University, Ramsey Lake Road, example, a survey designed to explore depression but Sudbury, Ontario, Canada 3 Theory evidence—this is evident when behaviour is which actually measures anxiety would not be consid- P3E2C6; similar to theoretical propositions of the construct ered valid.
    [Show full text]
  • On the Validity of Reading Assessments GOTHENBURG STUDIES in EDUCATIONAL SCIENCES 328
    On the Validity of Reading Assessments GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO © STEFAN JOHANSSON, 2013 ISBN 978-91-7346-736-0 ISSN 0436-1121 ISSN 1653-0101 Thesis in Education at the Department of Education and Special Education The thesis is also available in full text on http://hdl.handle.net/2077/32012 Photographer cover: Rebecka Karlsson Distribution: ACTA UNIVERSITATIS GOTHOBURGENSIS Box 222 SE-405 30 Göteborg, Sweden Print: Ale Tryckteam, Bohus 2013 Abstract Title: On the Validity of Reading Assessments: Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Language: English with a Swedish summary Keywords: Validity; Validation; Assessment; Teacher judgements; External tests; PIRLS 2001; Self-assessment; Multilevel models; Structural Equation Modeling; Socioeconomic status; Gender ISBN: 978-91-7346-736-0 The purpose of this thesis is to examine validity issues in different forms of assessments; teacher judgements, external tests, and pupil self-assessment in Swedish primary schools. The data used were selected from a large-scale study––PIRLS 2001––in which more than 11000 pupils and some 700 teachers from grades 3 and 4 participated. The primary method used in the secondary analyses to investigate validity issues of the assessment forms is multilevel Structural Equation Modeling (SEM) with latent variables. An argument-based approach to validity was adopted, where possible weaknesses in assessment forms were addressed.
    [Show full text]
  • Relations Between Inductive Reasoning and Deductive Reasoning
    Journal of Experimental Psychology: © 2010 American Psychological Association Learning, Memory, and Cognition 0278-7393/10/$12.00 DOI: 10.1037/a0018784 2010, Vol. 36, No. 3, 805–812 Relations Between Inductive Reasoning and Deductive Reasoning Evan Heit Caren M. Rotello University of California, Merced University of Massachusetts Amherst One of the most important open questions in reasoning research is how inductive reasoning and deductive reasoning are related. In an effort to address this question, we applied methods and concepts from memory research. We used 2 experiments to examine the effects of logical validity and premise– conclusion similarity on evaluation of arguments. Experiment 1 showed 2 dissociations: For a common set of arguments, deduction judgments were more affected by validity, and induction judgments were more affected by similarity. Moreover, Experiment 2 showed that fast deduction judgments were like induction judgments—in terms of being more influenced by similarity and less influenced by validity, compared with slow deduction judgments. These novel results pose challenges for a 1-process account of reasoning and are interpreted in terms of a 2-process account of reasoning, which was implemented as a multidimensional signal detection model and applied to receiver operating characteristic data. Keywords: reasoning, similarity, mathematical modeling An important open question in reasoning research concerns the arguments (Rips, 2001). This technique can highlight similarities relation between induction and deduction. Typically, individual or differences between induction and deduction that are not con- studies of reasoning have focused on only one task, rather than founded by the use of different materials (Heit, 2007). It also examining how the two are connected (Heit, 2007).
    [Show full text]
  • Construct Validity in Psychological Tests
    CONSTRUCT VALIDITY IN PSYCHOLOGICAL TF.STS - ----L. J. CRONBACH and P. E. MEEID..----- validity seems only to have stirred the muddy waters. Portions of the distinctions we shall discuss are implicit in Jenkins' paper, "Validity for 'What?" { 33), Gulliksen's "Intrinsic Validity" (27), Goo<lenough's dis­ tinction between tests as "signs" and "samples" (22), Cronbach's sepa· Construct Validity in Psychological Tests ration of "logical" and "empirical" validity ( 11 ), Guilford's "factorial validity" (25), and Mosier's papers on "face validity" and "validity gen­ eralization" ( 49, 50). Helen Peak ( 52) comes close to an explicit state­ ment of construct validity as we shall present it. VALIDATION of psychological tests has not yet been adequately concep· Four Types of Validation tua1ized, as the APA Committee on Psychological Tests learned when TI1e categories into which the Recommendations divide validity it undertook (1950-54) to specify what qualities should be investigated studies are: predictive validity, concurrent validity, content validity, and before a test is published. In order to make coherent recommendations constrnct validity. The first two of these may be considered together the Committee found it necessary to distinguish four types of validity, as criterion-oriented validation procedures. established by different types of research and requiring different interpre· TI1e pattern of a criterion-oriented study is familiar. The investigator tation. The chief innovation in the Committee's report was the term is primarily interested in some criterion which he wishes to predict. lie constmct validity.* This idea was first formulated hy a subcommittee administers the test, obtains an independent criterion measure on the {Meehl and R.
    [Show full text]
  • Running Head: the STRUCTURE of PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: a Meta-An
    Running head: THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: A Meta-Analysis of the Psychopathic Personality Inventory Jared R. Ruchensky, John F. Edens, Katherine S. Corker, M. Brent Donnellan, Edward A. Witt, Daniel M. Blonigen Version: 11/17/2017 In press, Psychological Assessment © 2017, American Psychological Association. This paper is not the copy of record and may not exactly replicate the final, authoritative version of the article. The final article will be available, upon publication, via its DOI: 10.1037/pas0000520 THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 2 Abstract Which core traits exemplify psychopathic personality disorder is a hotly debated question within psychology, particularly regarding the role of ostensibly adaptive traits such as stress immunity, social potency, and fearlessness. Much of the research on the inter-relationships among putatively adaptive and more maladaptive traits of psychopathy has focused on the factor structure of the Psychopathic Personality Inventory (PPI) and its revision, the Psychopathic Personality Inventory – Revised (PPI-R). These instruments include content scales that have coalesced to form 2 higher-order factors in some (but not all) prior studies: Fearless Dominance and Self-Centered Impulsivity. Given the inconsistencies in prior research, we performed a meta- analytic factor analysis of the 8 content scales from these instruments (total N > 18,000) and found general support for these 2 dimensions in community samples. The structure among offender samples (e.g., prisoners, forensic patients) supported a three-factor model in which the Fearlessness content scale uniquely loaded onto Self-centered Impulsivity (rather than Fearless Dominance). There were also indications that the Stress Immunity content scale had different relations to the other PPI scales in offender versus community samples.
    [Show full text]
  • Reliability, Validity, and Trustworthiness
    © Jones & Bartlett Learning, LLC. NOT FOR SALE OR DISTRIBUTION Chapter 12 Reliability, Validity, and Trustworthiness James Eldridge © VLADGRIN/iStock/Thinkstock Chapter Objectives At the conclusion of this chapter, the learner will be able to: 1. Identify the need for reliability and validity of instruments used in evidence-based practice. 2. Define reliability and validity. 3. Discuss how reliability and validity affect outcome measures and conclusions of evidence-based research. 4. Develop reliability and validity coefficients for appropriate data. 5. Interpret reliability and validity coefficients of instruments used in evidence-based practice. 6. Describe sensitivity and specificity as related to data analysis. 7. Interpret receiver operand characteristics (ROC) to describe validity. Key Terms Accuracy Content-related validity Concurrent validity Correlation coefficient Consistency Criterion-related validity Construct validity Cross-validation 9781284108958_CH12_Pass03.indd 339 10/20/15 5:59 PM © Jones & Bartlett Learning, LLC. NOT FOR SALE OR DISTRIBUTION 340 | Chapter 12 Reliability, Validity, and Trustworthiness Equivalency reliability Reliability Interclass reliability Sensitivity Intraclass reliability Specificity Objectivity Stability Observed score Standard error of Predictive validity measurement (SEM) Receiver operand Trustworthiness characteristics (ROC) Validity n Introduction The foundation of good research and of good decision making in evidence- based practice (EBP) is the trustworthiness of the data used to make decisions. When data cannot be trusted, an informed decision cannot be made. Trustworthiness of the data can only be as good as the instruments or tests used to collect the data. Regardless of the specialization of the healthcare provider, nurses make daily decisions on the diagnosis and treatment of a patient based on the results from different tests to which the patient is subjected.
    [Show full text]