Questionnaire Validity and Reliability
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Internal Validity Is About Causal Interpretability
Internal Validity is about Causal Interpretability Before we can discuss Internal Validity, we have to discuss different types of Internal Validity variables and review causal RH:s and the evidence needed to support them… Every behavior/measure used in a research study is either a ... Constant -- all the participants in the study have the same value on that behavior/measure or a ... • Measured & Manipulated Variables & Constants Variable -- when at least some of the participants in the study • Causes, Effects, Controls & Confounds have different values on that behavior/measure • Components of Internal Validity • “Creating” initial equivalence and every behavior/measure is either … • “Maintaining” ongoing equivalence Measured -- the value of that behavior/measure is obtained by • Interrelationships between Internal Validity & External Validity observation or self-report of the participant (often called “subject constant/variable”) or it is … Manipulated -- the value of that behavior/measure is controlled, delivered, determined, etc., by the researcher (often called “procedural constant/variable”) So, every behavior/measure in any study is one of four types…. constant variable measured measured (subject) measured (subject) constant variable manipulated manipulated manipulated (procedural) constant (procedural) variable Identify each of the following (as one of the four above, duh!)… • Participants reported practicing between 3 and 10 times • All participants were given the same set of words to memorize • Each participant reported they were a Psyc major • Each participant was given either the “homicide” or the “self- defense” vignette to read From before... Circle the manipulated/causal & underline measured/effect variable in each • Causal RH: -- differences in the amount or kind of one behavior cause/produce/create/change/etc. -
Validity and Reliability of the Questionnaire for Compliance with Standard Precaution for Nurses
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Cadernos Espinosanos (E-Journal) Rev Saúde Pública 2015;49:87 Original Articles DOI:10.1590/S0034-8910.2015049005975 Marília Duarte ValimI Validity and reliability of the Maria Helena Palucci MarzialeII Miyeko HayashidaII Questionnaire for Compliance Fernanda Ludmilla Rossi RochaIII with Standard Precaution Jair Lício Ferreira SantosIV ABSTRACT OBJECTIVE: To evaluate the validity and reliability of the Questionnaire for Compliance with Standard Precaution for nurses. METHODS: This methodological study was conducted with 121 nurses from health care facilities in Sao Paulo’s countryside, who were represented by two high-complexity and by three average-complexity health care facilities. Internal consistency was calculated using Cronbach’s alpha and stability was calculated by the intraclass correlation coefficient, through test-retest. Convergent, discriminant, and known-groups construct validity techniques were conducted. RESULTS: The questionnaire was found to be reliable (Cronbach’s alpha: 0.80; intraclass correlation coefficient: (0.97) In regards to the convergent and discriminant construct validity, strong correlation was found between compliance to standard precautions, the perception of a safe environment, and the smaller perception of obstacles to follow such precautions (r = 0.614 and r = 0.537, respectively). The nurses who were trained on the standard precautions and worked on the health care facilities of higher complexity were shown to comply more (p = 0.028 and p = 0.006, respectively). CONCLUSIONS: The Brazilian version of the Questionnaire for I Departamento de Enfermagem. Centro Compliance with Standard Precaution was shown to be valid and reliable. -
Assessment Center Structure and Construct Validity: a New Hope
University of Central Florida STARS Electronic Theses and Dissertations, 2004-2019 2015 Assessment Center Structure and Construct Validity: A New Hope Christopher Wiese University of Central Florida Part of the Psychology Commons Find similar works at: https://stars.library.ucf.edu/etd University of Central Florida Libraries http://library.ucf.edu This Doctoral Dissertation (Open Access) is brought to you for free and open access by STARS. It has been accepted for inclusion in Electronic Theses and Dissertations, 2004-2019 by an authorized administrator of STARS. For more information, please contact [email protected]. STARS Citation Wiese, Christopher, "Assessment Center Structure and Construct Validity: A New Hope" (2015). Electronic Theses and Dissertations, 2004-2019. 733. https://stars.library.ucf.edu/etd/733 ASSESSMENT CENTER STRUCTURE AND CONSTRUCT VALIDITY: A NEW HOPE by CHRISTOPHER W. WIESE B.S., University of Central Florida, 2008 A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Psychology in the College of Sciences at the University of Central Florida Orlando, Florida Summer Term 2015 Major Professor: Kimberly Smith-Jentsch © 2015 Christopher Wiese ii ABSTRACT Assessment Centers (ACs) are a fantastic method to measure behavioral indicators of job performance in multiple diverse scenarios. Based upon a thorough job analysis, ACs have traditionally demonstrated very strong content and criterion-related validity. However, researchers have been puzzled for over three decades with the lack of evidence concerning construct validity. ACs are designed to measure critical job dimensions throughout multiple situational exercises. However, research has consistently revealed that different behavioral ratings within these scenarios are more strongly related to one another (exercise effects) than the same dimension rating across scenarios (dimension effects). -
Construct Validity and Reliability of the Work Environment Assessment Instrument WE-10
International Journal of Environmental Research and Public Health Article Construct Validity and Reliability of the Work Environment Assessment Instrument WE-10 Rudy de Barros Ahrens 1,*, Luciana da Silva Lirani 2 and Antonio Carlos de Francisco 3 1 Department of Business, Faculty Sagrada Família (FASF), Ponta Grossa, PR 84010-760, Brazil 2 Department of Health Sciences Center, State University Northern of Paraná (UENP), Jacarezinho, PR 86400-000, Brazil; [email protected] 3 Department of Industrial Engineering and Post-Graduation in Production Engineering, Federal University of Technology—Paraná (UTFPR), Ponta Grossa, PR 84017-220, Brazil; [email protected] * Correspondence: [email protected] Received: 1 September 2020; Accepted: 29 September 2020; Published: 9 October 2020 Abstract: The purpose of this study was to validate the construct and reliability of an instrument to assess the work environment as a single tool based on quality of life (QL), quality of work life (QWL), and organizational climate (OC). The methodology tested the construct validity through Exploratory Factor Analysis (EFA) and reliability through Cronbach’s alpha. The EFA returned a Kaiser–Meyer–Olkin (KMO) value of 0.917; which demonstrated that the data were adequate for the factor analysis; and a significant Bartlett’s test of sphericity (χ2 = 7465.349; Df = 1225; p 0.000). ≤ After the EFA; the varimax rotation method was employed for a factor through commonality analysis; reducing the 14 initial factors to 10. Only question 30 presented commonality lower than 0.5; and the other questions returned values higher than 0.5 in the commonality analysis. Regarding the reliability of the instrument; all of the questions presented reliability as the values varied between 0.953 and 0.956. -
Statistical Analysis 8: Two-Way Analysis of Variance (ANOVA)
Statistical Analysis 8: Two-way analysis of variance (ANOVA) Research question type: Explaining a continuous variable with 2 categorical variables What kind of variables? Continuous (scale/interval/ratio) and 2 independent categorical variables (factors) Common Applications: Comparing means of a single variable at different levels of two conditions (factors) in scientific experiments. Example: The effective life (in hours) of batteries is compared by material type (1, 2 or 3) and operating temperature: Low (-10˚C), Medium (20˚C) or High (45˚C). Twelve batteries are randomly selected from each material type and are then randomly allocated to each temperature level. The resulting life of all 36 batteries is shown below: Table 1: Life (in hours) of batteries by material type and temperature Temperature (˚C) Low (-10˚C) Medium (20˚C) High (45˚C) 1 130, 155, 74, 180 34, 40, 80, 75 20, 70, 82, 58 2 150, 188, 159, 126 136, 122, 106, 115 25, 70, 58, 45 type Material 3 138, 110, 168, 160 174, 120, 150, 139 96, 104, 82, 60 Source: Montgomery (2001) Research question: Is there difference in mean life of the batteries for differing material type and operating temperature levels? In analysis of variance we compare the variability between the groups (how far apart are the means?) to the variability within the groups (how much natural variation is there in our measurements?). This is why it is called analysis of variance, abbreviated to ANOVA. This example has two factors (material type and temperature), each with 3 levels. Hypotheses: The 'null hypothesis' might be: H0: There is no difference in mean battery life for different combinations of material type and temperature level And an 'alternative hypothesis' might be: H1: There is a difference in mean battery life for different combinations of material type and temperature level If the alternative hypothesis is accepted, further analysis is performed to explore where the individual differences are. -
Heterogeneous Versus Homogeneous Measures: a Meta-Analysis of Predictive Efficacy
HETEROGENEOUS VERSUS HOMOGENEOUS MEASURES: A META-ANALYSIS OF PREDICTIVE EFFICACY A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy By SUZANNE L. DEAN M.S., Wright State University, 2008 _________________________________________ 2016 Wright State University WRIGHT STATE UNIVERSITY GRADUATE SCHOOL January 3, 2016 I HEREBY RECOMMEND THAT THE DISSERTATION PREPARED UNDER MY SUPERVISION BY Suzanne L. Dean ENTITLED Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF Doctor of Philosophy. _____________________________ Corey E. Miller, Ph.D. Dissertation Director _____________________________ Scott N. J. Watamaniuk, Ph.D. Graduate Program Director _____________________________ Robert E. W. Fyffe, Ph.D. Vice President for Research and Dean of the Graduate School Committee on Final Examination _____________________________ Corey E. Miller, Ph.D. _____________________________ David LaHuis, Ph.D. _____________________________ Allen Nagy, Ph.D. _____________________________ Debra Steele-Johnson, Ph.D. COPYRIGHT BY SUZANNE L. DEAN 2016 ABSTRACT Dean, Suzanne L. Ph.D., Industrial/Organizational Psychology Ph.D. program, Wright State University, 2016. Heterogeneous versus Homogeneous Measures: A Meta-Analysis of Predictive Efficacy. A meta-analysis was conducted to compare the predictive validity and adverse impact of homogeneous and heterogeneous predictors on objective and subjective criteria for different sales roles. Because job performance is a dynamic and complex construct, I hypothesized that equally complex, heterogeneous predictors would have stronger correlations with objective and subjective criteria than homogeneous predictors. Forty-seven independent validation studies (N = 3,378) qualified for inclusion in this study. In general, heterogeneous predictors did not demonstrate significantly stronger correlations with the performance criteria than homogeneous predictors. -
Validity and Reliability in Quantitative Studies Evid Based Nurs: First Published As 10.1136/Eb-2015-102129 on 15 May 2015
Research made simple Validity and reliability in quantitative studies Evid Based Nurs: first published as 10.1136/eb-2015-102129 on 15 May 2015. Downloaded from Roberta Heale,1 Alison Twycross2 10.1136/eb-2015-102129 Evidence-based practice includes, in part, implementa- have a high degree of anxiety? In another example, a tion of the findings of well-conducted quality research test of knowledge of medications that requires dosage studies. So being able to critique quantitative research is calculations may instead be testing maths knowledge. 1 School of Nursing, Laurentian an important skill for nurses. Consideration must be There are three types of evidence that can be used to University, Sudbury, Ontario, given not only to the results of the study but also the demonstrate a research instrument has construct Canada 2Faculty of Health and Social rigour of the research. Rigour refers to the extent to validity: which the researchers worked to enhance the quality of Care, London South Bank 1 Homogeneity—meaning that the instrument mea- the studies. In quantitative research, this is achieved University, London, UK sures one construct. through measurement of the validity and reliability.1 2 Convergence—this occurs when the instrument mea- Correspondence to: Validity sures concepts similar to that of other instruments. Dr Roberta Heale, Validity is defined as the extent to which a concept is Although if there are no similar instruments avail- School of Nursing, Laurentian accurately measured in a quantitative study. For able this will not be possible to do. University, Ramsey Lake Road, example, a survey designed to explore depression but Sudbury, Ontario, Canada 3 Theory evidence—this is evident when behaviour is which actually measures anxiety would not be consid- P3E2C6; similar to theoretical propositions of the construct ered valid. -
On the Validity of Reading Assessments GOTHENBURG STUDIES in EDUCATIONAL SCIENCES 328
On the Validity of Reading Assessments GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO GOTHENBURG STUDIES IN EDUCATIONAL SCIENCES 328 On the Validity of Reading Assessments Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Stefan Johansson ACTA UNIVERSITATIS GOTHOBURGENSIS LOGO © STEFAN JOHANSSON, 2013 ISBN 978-91-7346-736-0 ISSN 0436-1121 ISSN 1653-0101 Thesis in Education at the Department of Education and Special Education The thesis is also available in full text on http://hdl.handle.net/2077/32012 Photographer cover: Rebecka Karlsson Distribution: ACTA UNIVERSITATIS GOTHOBURGENSIS Box 222 SE-405 30 Göteborg, Sweden Print: Ale Tryckteam, Bohus 2013 Abstract Title: On the Validity of Reading Assessments: Relationships Between Teacher Judgements, External Tests and Pupil Self-assessments Language: English with a Swedish summary Keywords: Validity; Validation; Assessment; Teacher judgements; External tests; PIRLS 2001; Self-assessment; Multilevel models; Structural Equation Modeling; Socioeconomic status; Gender ISBN: 978-91-7346-736-0 The purpose of this thesis is to examine validity issues in different forms of assessments; teacher judgements, external tests, and pupil self-assessment in Swedish primary schools. The data used were selected from a large-scale study––PIRLS 2001––in which more than 11000 pupils and some 700 teachers from grades 3 and 4 participated. The primary method used in the secondary analyses to investigate validity issues of the assessment forms is multilevel Structural Equation Modeling (SEM) with latent variables. An argument-based approach to validity was adopted, where possible weaknesses in assessment forms were addressed. -
Relations Between Inductive Reasoning and Deductive Reasoning
Journal of Experimental Psychology: © 2010 American Psychological Association Learning, Memory, and Cognition 0278-7393/10/$12.00 DOI: 10.1037/a0018784 2010, Vol. 36, No. 3, 805–812 Relations Between Inductive Reasoning and Deductive Reasoning Evan Heit Caren M. Rotello University of California, Merced University of Massachusetts Amherst One of the most important open questions in reasoning research is how inductive reasoning and deductive reasoning are related. In an effort to address this question, we applied methods and concepts from memory research. We used 2 experiments to examine the effects of logical validity and premise– conclusion similarity on evaluation of arguments. Experiment 1 showed 2 dissociations: For a common set of arguments, deduction judgments were more affected by validity, and induction judgments were more affected by similarity. Moreover, Experiment 2 showed that fast deduction judgments were like induction judgments—in terms of being more influenced by similarity and less influenced by validity, compared with slow deduction judgments. These novel results pose challenges for a 1-process account of reasoning and are interpreted in terms of a 2-process account of reasoning, which was implemented as a multidimensional signal detection model and applied to receiver operating characteristic data. Keywords: reasoning, similarity, mathematical modeling An important open question in reasoning research concerns the arguments (Rips, 2001). This technique can highlight similarities relation between induction and deduction. Typically, individual or differences between induction and deduction that are not con- studies of reasoning have focused on only one task, rather than founded by the use of different materials (Heit, 2007). It also examining how the two are connected (Heit, 2007). -
Construct Validity in Psychological Tests
CONSTRUCT VALIDITY IN PSYCHOLOGICAL TF.STS - ----L. J. CRONBACH and P. E. MEEID..----- validity seems only to have stirred the muddy waters. Portions of the distinctions we shall discuss are implicit in Jenkins' paper, "Validity for 'What?" { 33), Gulliksen's "Intrinsic Validity" (27), Goo<lenough's dis tinction between tests as "signs" and "samples" (22), Cronbach's sepa· Construct Validity in Psychological Tests ration of "logical" and "empirical" validity ( 11 ), Guilford's "factorial validity" (25), and Mosier's papers on "face validity" and "validity gen eralization" ( 49, 50). Helen Peak ( 52) comes close to an explicit state ment of construct validity as we shall present it. VALIDATION of psychological tests has not yet been adequately concep· Four Types of Validation tua1ized, as the APA Committee on Psychological Tests learned when TI1e categories into which the Recommendations divide validity it undertook (1950-54) to specify what qualities should be investigated studies are: predictive validity, concurrent validity, content validity, and before a test is published. In order to make coherent recommendations constrnct validity. The first two of these may be considered together the Committee found it necessary to distinguish four types of validity, as criterion-oriented validation procedures. established by different types of research and requiring different interpre· TI1e pattern of a criterion-oriented study is familiar. The investigator tation. The chief innovation in the Committee's report was the term is primarily interested in some criterion which he wishes to predict. lie constmct validity.* This idea was first formulated hy a subcommittee administers the test, obtains an independent criterion measure on the {Meehl and R. -
Running Head: the STRUCTURE of PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: a Meta-An
Running head: THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 1 Evaluating the Structure of Psychopathic Personality Traits: A Meta-Analysis of the Psychopathic Personality Inventory Jared R. Ruchensky, John F. Edens, Katherine S. Corker, M. Brent Donnellan, Edward A. Witt, Daniel M. Blonigen Version: 11/17/2017 In press, Psychological Assessment © 2017, American Psychological Association. This paper is not the copy of record and may not exactly replicate the final, authoritative version of the article. The final article will be available, upon publication, via its DOI: 10.1037/pas0000520 THE STRUCTURE OF PSYCHOPATHIC PERSONALITY 2 Abstract Which core traits exemplify psychopathic personality disorder is a hotly debated question within psychology, particularly regarding the role of ostensibly adaptive traits such as stress immunity, social potency, and fearlessness. Much of the research on the inter-relationships among putatively adaptive and more maladaptive traits of psychopathy has focused on the factor structure of the Psychopathic Personality Inventory (PPI) and its revision, the Psychopathic Personality Inventory – Revised (PPI-R). These instruments include content scales that have coalesced to form 2 higher-order factors in some (but not all) prior studies: Fearless Dominance and Self-Centered Impulsivity. Given the inconsistencies in prior research, we performed a meta- analytic factor analysis of the 8 content scales from these instruments (total N > 18,000) and found general support for these 2 dimensions in community samples. The structure among offender samples (e.g., prisoners, forensic patients) supported a three-factor model in which the Fearlessness content scale uniquely loaded onto Self-centered Impulsivity (rather than Fearless Dominance). There were also indications that the Stress Immunity content scale had different relations to the other PPI scales in offender versus community samples. -
Reliability, Validity, and Trustworthiness
© Jones & Bartlett Learning, LLC. NOT FOR SALE OR DISTRIBUTION Chapter 12 Reliability, Validity, and Trustworthiness James Eldridge © VLADGRIN/iStock/Thinkstock Chapter Objectives At the conclusion of this chapter, the learner will be able to: 1. Identify the need for reliability and validity of instruments used in evidence-based practice. 2. Define reliability and validity. 3. Discuss how reliability and validity affect outcome measures and conclusions of evidence-based research. 4. Develop reliability and validity coefficients for appropriate data. 5. Interpret reliability and validity coefficients of instruments used in evidence-based practice. 6. Describe sensitivity and specificity as related to data analysis. 7. Interpret receiver operand characteristics (ROC) to describe validity. Key Terms Accuracy Content-related validity Concurrent validity Correlation coefficient Consistency Criterion-related validity Construct validity Cross-validation 9781284108958_CH12_Pass03.indd 339 10/20/15 5:59 PM © Jones & Bartlett Learning, LLC. NOT FOR SALE OR DISTRIBUTION 340 | Chapter 12 Reliability, Validity, and Trustworthiness Equivalency reliability Reliability Interclass reliability Sensitivity Intraclass reliability Specificity Objectivity Stability Observed score Standard error of Predictive validity measurement (SEM) Receiver operand Trustworthiness characteristics (ROC) Validity n Introduction The foundation of good research and of good decision making in evidence- based practice (EBP) is the trustworthiness of the data used to make decisions. When data cannot be trusted, an informed decision cannot be made. Trustworthiness of the data can only be as good as the instruments or tests used to collect the data. Regardless of the specialization of the healthcare provider, nurses make daily decisions on the diagnosis and treatment of a patient based on the results from different tests to which the patient is subjected.