Construct Validity

Total Page:16

File Type:pdf, Size:1020Kb

Construct Validity Associate Prof. Dr Anne Yee Dr Mahmoud Danaee 1 What does this resemble? 2 Rorschach test • At the end of the test, the tester says … – you need therapy – or you can't work for this company 3 Psychological Testing • Occurs widely … – in personnel selection – in clinical settings – in education What constitutes a good test? 4 Validity and Reliability • Validity: How well does the measure or design do what it purports to do? • Reliability: How consistent or stable is the instrument? –Is the instrument dependable? Validity AKA Criterion Logical Statistical Construc t Face ContenConten Convergent Divergent/ Concurren Predictive tt Discriminant t Reliability Consistency Objectivity Face Validity – Infers that a test is valid by face value – It is clear that the test measures what it is supposed to – As a check on face validity, test/survey items are sent to experts to obtain suggestions for modification. – Because of its vagueness and subjectivity, psychometricians have abandoned this concept for a long time. Content Validity – Infers that the test measures all aspects contributing to the variable of interest – Face validity Vs Content validity: • Face validity can be established by one person • Content validity should be checked by a panel, and thus usually it goes hand in hand with inter- rater reliability (Kappa!) Example: • Computer literacy includes skills in operating system, word processing, spreadsheet, database, graphics, internet, and many others. • It is difficult to administer a test covering all aspects of computing. Therefore, only several tasks are sampled from the universe of computer skills. • A test of computer literacy should be written or reviewed by computer science professors or senior programmers in the IT industry because it is assumed that computer scientists should know what are important in his own discipline. Overall: A logically valid test simply appears to measure the right variable in its entirety? Subjective!!! The Content Validity Index Content validity has been defined as follows: • (1) ‘‘...the degree to which an instrument has an appropriate sample of items for the construct being measured’’ (Polit & Beck, 2004, p. 423); • (2) ‘‘...whether or not the items sampled for inclusion on the tool adequately represent the domain of content addressed by the instrument’’ (Waltz, Strickland, & Lenz, 2005, p. 155); • (3) ‘‘...the extent to which an instrument adequately samples the research domain of interest when attempting to measure phenomena’’ (Wynd, Schmidt, & Schaefer, 2003, p. 509). Two types of CVIs. • content validity of individual items • content validity of the overall scale. • Researchers use I-CVI information to guide them in revising, deleting, or substituting items • I-CVIs tend only to be reported in methodological studies that focus on descriptions of the content validation process • Most often reported in scale development studies is the CVI S-CVI/UA Proportion of items on a scale that achieves a S-CVI relevance rating of 3 or 4 by all the Content Validity of experts the overall scale CVI Degree to which an instrument has an appropriate sample of items for construct S-CVI/Ave being measured Average of the I- I-CVI CVIs for Content Validity of individual items: Are the items representati Are the items Has each item relevance to Are the items in the ve of concepts Question clarity in term comments instruments concepts related to the of wording consistency? related to the dissertation dissertation topic? topic? Q1 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q2 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q3 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q4 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Q5 ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ ① ② ③ ④ Ratings 1= not relevant 2 =somewhat relevant. 3= quite relevant 4= highly relevant. I-CVI, item-level content validity index S-CVI, content validity index for the scale Acceptable standard for the S-CVI recommended a minimum S-CVI of .80. If the I-CVI is higher than 79%, the item will be appropriate. If it is between 70% and 79%, it needs revision. If it is less than 70% it is eliminated Kappa statistic is a consensus index of inter-rater agreement that adjusts for chance agreement and is an important supplement to CVI because Kappa provides information about the degree of agreement beyond chance Evaluation criteria for Kappa is the values above 0.74= excellent between 0.60 and 0.74=good between 0.40 and 0.59= fair Validity Criterion Logical Statistical Construc t Face ContenConten Convergent Divergent/ Concurren Predictive tt Discriminant t Reliability Consistency Objectivity Criterion Validity • This type of validity is used to measure the ability of an instrument to predict future outcomes. • Validity is usually determined by comparing two instruments ability to predict a similar outcome with a single variable being measured. • There are two major types of criterion validity predictive or concurrent forms of validity. Criterion validity ‘Warwick spider phobia questionnaire’ positive correlation with SPQ • A test has high criterion validity if – It correlates highly with some external benchmark (concurrent) ? – How well does the test correlated with outcome criteria (predictive)? • Eg – You have lost 30 pounds if your scale reported that you lost 30 pounds, you would expect that your clothes would also feel looser 24 Concurrent Criterion Validity • Concurrent criterion validity is used when the two instruments are used to measure the same event at the same time. • Example: Predictive Criterion Validity • Predictive validity is used when the instrument is administered then time is allowed to pass and is measured against the another outcome. • Example: Criterion validity • When the focus of the test is on criterion validity, we draw an inference from test scores to performance. A high score of a valid test indicates that the test taker has met the performance criteria. • Regression analysis can be applied to establish criterion validity. An independent variable could be used as a predictor variable and a dependent variable, the criterion variable. The correlation coefficient between them is called validity coefficients. How is Criterion Validity Measured? • The statistical measure or correlation coefficient tells the degree to which the instrument is valid based on the measured criteria. • What does it look like in an equation? – The symbol “r” denotes the correlation coefficient. – A higher “r” value shows a positive relationship between the instruments. – A mix of high and low “r” values shows a negative relationship. Predictive Validity Concurrent Validity • As a rule of thumb, for absolute value of r: • 0.00-0.19: very weak • 0.20-0.39: weak • 0.40-0.59: moderate • 0.60-0.79: strong • 0.80-1.00: very strong. Validity AKA Criterion Logical Statistical Construct Face ContenConten Convergent Divergent/ Concurren Predictive tt Discriminant t Reliability Consistency Objectivity Construct validity • Measuring things that are in our theory of a domain. • The construct is sometimes called a latent variable – You can’t directly observe the construct – You can only measure its surface manifestations – it is concerned with abstract and theoretical construct, construct validity is also known as theoretical validity 32 What are Latent Variables? • Most/all variables in the social world are not directly observable. • This makes them ‘latent’ or hypothetical constructs. • We measure latent variables with observable indicators, e.g. questionnaire items. • We can think of the variance of an observable indicator as being partially caused by: – The latent construct in question – Other factors (error) • I cringe when I have to go to math class. • I am uneasy about going to the board in a math class. Math • I am afraid to ask questions in math anxiety class. • I am always worried about being called on in math class. • I understand math now, but I worry that it's going to get really difficult soon. • Specifying formative versus reflective constructs is a critical preliminary step prior to further statistical analysis. Specification follows these guidelines: • Formative – Direction of causality is from measure to construct – No reason to expect the measures are correlated – Indicators are not interchangeable • Reflective – Direction of causality is from construct to measure – Measures expected to be correlated – Indicators are interchangeable – An example of formative versus reflective constructs is given in the figure below. Factor model • A factor model identifies the relationship between observed items and latent factors. For example, when a psychologist wants to study the causal relationships between Math anxiety and job performance, first he/she has to define the constructs “Math anxiety” and “job performance.” To accomplish this step, about the psychologist needs to develop items that measure the defined construct. Construct, dimension, subscale, factor, component • This construct has eight dimensions (e.g. Intelligence has eight aspects) • This scale has eight subscales (e.g. the survey measures different but weakly related things) • The factor structure has eight factors/components (e.g. in factor analysis/PCA) Exploratory Factor Analysis • (EFA) is a statistical approach to determining the correlation among the variables in a dataset. • This type of analysis provides a factor structure (a grouping of variables based on strong correlations). • EFA is good for detecting "misfit" variables. In general, an EFA prepares the variables to be used for cleaner structural equation modeling. An EFA should always be conducted
Recommended publications
  • “Is the Test Valid?” Criterion-Related Validity– 3 Classic Types Criterion-Related Validity– 3 Classic Types
    “Is the test valid?” Criterion-related Validity – 3 classic types • a “criterion” is the actual value you’d like to have but can’t… Jum Nunnally (one of the founders of modern psychometrics) • it hasn’t happened yet and you can’t wait until it does claimed this was “silly question”! The point wasn’t that tests shouldn’t be “valid” but that a test’s validity must be assessed • it’s happening now, but you can’t measure it “like you want” relative to… • It happened in the past and you didn’t measure it then • does test correlate with “criterion”? -- has three major types • the specific construct(s) it is intended to measure Measure the Predictive Administer test Criterion • the population for which it is intended (e.g., age, level) Validity • the application for which it is intended (e.g., for classifying Concurrent Measure the Criterion folks into categories vs. assigning them Validity Administer test quantitative values) Postdictive Measure the So, the real question is, “Is this test a valid measure of this Administer test Validity Criterion construct for this population for this application?” That question can be answered! Past Present Future The advantage of criterion-related validity is that it is a relatively Criterion-related Validity – 3 classic types simple statistically based type of validity! • does test correlate with “criterion”? -- has three major types • If the test has the desired correlation with the criterion, then • predictive -- test taken now predicts criterion assessed later you have sufficient evidence for criterion-related
    [Show full text]
  • Validity and Reliability in Social Science Research
    Education Research and Perspectives, Vol.38, No.1 Validity and Reliability in Social Science Research Ellen A. Drost California State University, Los Angeles Concepts of reliability and validity in social science research are introduced and major methods to assess reliability and validity reviewed with examples from the literature. The thrust of the paper is to provide novice researchers with an understanding of the general problem of validity in social science research and to acquaint them with approaches to developing strong support for the validity of their research. Introduction An important part of social science research is the quantification of human behaviour — that is, using measurement instruments to observe human behaviour. The measurement of human behaviour belongs to the widely accepted positivist view, or empirical- analytic approach, to discern reality (Smallbone & Quinton, 2004). Because most behavioural research takes place within this paradigm, measurement instruments must be valid and reliable. The objective of this paper is to provide insight into these two important concepts, and to introduce the major methods to assess validity and reliability as they relate to behavioural research. The paper has been written for the novice researcher in the social sciences. It presents a broad overview taken from traditional literature, not a critical account of the general problem of validity of research information. The paper is organised as follows. The first section presents what reliability of measurement means and the techniques most frequently used to estimate reliability. Three important questions researchers frequently ask about reliability are discussed: (1) what Address for correspondence: Dr. Ellen Drost, Department of Management, College of Business and Economics, California State University, Los Angeles, 5151 State University Drive, Los Angeles, CA 90032.
    [Show full text]
  • Validity of the Patient Health Questionnaire-9 in Assessing Depression Following Traumatic Brain Injury
    J Head Trauma Rehabil Vol. 20, No. 6, pp. 501–511 c 2005 Lippincott Williams & Wilkins, Inc. Validity of the Patient Health Questionnaire-9 in Assessing Depression Following Traumatic Brain Injury Jesse R. Fann, MD, MPH; Charles H. Bombardier, PhD; Sureyya Dikmen, PhD; Peter Esselman, MD; Catherine A. Warms, PhD; Erika Pelzer, BS; Holly Rau, BS; Nancy Temkin, PhD Objective: To test the validity and reliability of the Patient Health Questionnaire-9 (PHQ-9) for diag- nosing major depressive disorder (MDD) among persons with traumatic brain injury (TBI). Design: Prospective cohort study. Setting: Level I trauma center. Participants: 135 adults within 1 year of complicated mild, moderate, or severe TBI. Main Outcome Measures: PHQ-9 Depression Scale, Structured Clinical Interview for Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (SCID). Results: Using a screening criterion of at least 5 PHQ-9 symptoms present at least several days over the last 2 weeks (with one being depressed mood or anhedonia) maximizes sensitivity (0.93) and specificity (0.89) while providing a positive predictive value of 0.63 and a negative predictive value of 0.99 when compared to SCID diagnosis of MDD. Pearson’s correlation between the PHQ-9 scores and other depression measures was 0.90 with the Hopkins Symptom Checklist depression subscale and 0.78 with the Hamilton Rating Scale for Depression. Test-retest reliability of the PHQ-9 was r = 0.76 and κ = 0.46 when using the optimal screening method. Conclusions: The PHQ-9 is a valid and reliable screening tool for detecting MDD in persons with TBI.
    [Show full text]
  • Reliability and Validity of the Measurement--Two Concepts Which We Will Discuss Further Later in This Lab
    INTRODUCTION Sociologist James A. Quinn states that the tasks of scientific method are related directly or indirectly to the study of similarities of various kinds of objects or events. One of the tasks of scientific method is that of classifying objects or events into categories and of describing the similar characteristics of members of each type. A second task is that of comparing variations in two or more characteristics of the members of a category. Indeed, it is the discovery, formulation, and testing of generalizations about the relations among selected variables that constitute the central task of scientific method. Fundamental to the performance of these tasks is a system of measurement. S.S. Stevens defines measurement as "the assignment of numerals to objects or events according to rules." This definition incorporates a number of important distinctions. It implies that if rules can be set up, it is theoretically possible to measure anything. Further, measurement is only as good as the rules that direct its application. The "goodness" of the rules reflects on the reliability and validity of the measurement--two concepts which we will discuss further later in this lab. Another aspect of definition given by Stevens is the use of the term numeral rather than number. A numeral is a symbol and has no quantitative meaning unless the researcher supplies it through the use of rules. The researcher sets up the criteria by which objects or events are distinguished from one another and also the weights, if any, which are to be assigned to these distinctions. This results in a scale.
    [Show full text]
  • Convergent and Discriminant Validity with Formative Measurement: a Mediator Perspective Xuequn Wang Murdoch University, Perth, AUS, [email protected]
    Journal of Modern Applied Statistical Methods Volume 14 | Issue 1 Article 11 5-1-2015 Convergent and Discriminant Validity with Formative Measurement: A Mediator Perspective Xuequn Wang Murdoch University, Perth, AUS, [email protected] Brian F. French Washington State University, [email protected] Paul F. Clay Fort Lewis College, [email protected] Follow this and additional works at: http://digitalcommons.wayne.edu/jmasm Recommended Citation Wang, Xuequn; French, Brian F.; and Clay, Paul F. (2015) "Convergent and Discriminant Validity with Formative Measurement: A Mediator Perspective," Journal of Modern Applied Statistical Methods: Vol. 14 : Iss. 1 , Article 11. DOI: 10.22237/jmasm/1430453400 Available at: http://digitalcommons.wayne.edu/jmasm/vol14/iss1/11 This Regular Article is brought to you for free and open access by the Open Access Journals at DigitalCommons@WayneState. It has been accepted for inclusion in Journal of Modern Applied Statistical Methods by an authorized editor of DigitalCommons@WayneState. Journal of Modern Applied Statistical Methods Copyright © 2015 JMASM, Inc. May 2015, Vol. 14, No. 1, 83-106. ISSN 1538 − 9472 Convergent and Discriminant Validity with Formative Measurement: A Mediator Perspective Xuequn Wang Brian F. French Paul F. Clay Murdoch University Washington State University Fort Lewis College Perth, Australia Pullman, WA Durango, CO The ability to validate formative measurement has increased in importance as it is used to develop and test theoretical models. A method is proposed to gather convergent and discriminant validity evidence of formative measurement. Survey data is used to test the proposed method. Keywords: Causal indicators, formative measurement, construct validity, convergent validity, discriminant validity, mediator Introduction There has been a vigorous debate and discussion about the issues surrounding the application of formative measurement (Bollen, 2007; Howell et al., 2007a, 2007b; Petter et al., 2007) and how to validate this specific kind of measurement model (Hardin et al.
    [Show full text]
  • Assessing Convergent and Discriminant Validity of the Motivation Construct for the Technology Integration Education (TIE) Model
    Assessing Convergent and Discriminant Validity of the Motivation Construct for the Technology Integration Education (TIE) Model Kevin A. Engellant University of Montana Western Denise D. Holland University of Montana Western Randy T. Piper Motivation researchers have given high priority to the nomological validity of the motivation construct by creating diverse models such as the technology integration education (TIE) model as part of the technological, pedagogical, and content knowledge (TPACK) research agenda. However, researchers have not given equally high priority to convergent validity and discriminant validity. This study incorporated the multitrait-multimethod (MTMM) matrix to discover evidence of convergent validity and discriminant validity for self-determination theory motivation and social cognitive theory motivation. The researchers found limited evidence of convergent validity and discriminant validity for the motivation construct. INTRODUCTION Motivation is perennially important because it has diverse implications for academic research in education, psychology, business, and public administration (Greene, 2005; Perry & Hondeghem, 2008; Rainey, 2014) and real-world applications for teachers, parents, coaches, and others (Ryan & Deci, 2000a, 2000b). Motivation is among the most studied constructs in the social sciences. In industrial and organizational psychology, along with job satisfaction (Cranny, Smith, & Stone, 1992; Locke, 1969; Spector, 1992, 1987, 1997), motivation may be the most studied construct (Latham, 2012; Pinder, 2008). In educational psychology, along with self-efficacy (Bandura, 1977a, 1977b, 1997, 2009; Bandura & Locke, 2003; Gist & Mitchell, 1992), motivation may be the most studied construct (Brophy, 2010; Schunk, Pintrich, & Meece, 2008; Stipek, 2002; Weiner, 1972, 1980, 1986, 1990). Education motivation textbooks have identified expectancy value, attribution, social cognitive, goal- setting, and self-determination theory as leading theories (Brophy, 2010; Schunk et al., 2008; Stipek, 2002).
    [Show full text]
  • Construct Validation 1 Construct Validation in Organizational
    Construct Validation 1 Construct Validation in Organizational Behavior Research Jeffrey R. Edwards University of North Carolina The author thanks Richard P. DeShon, Fritz Drasgow, and Larry J. Williams for their helpful comments on an earlier version of this chapter. In J. Greenberg (Ed.), Organizational behavior: The state of the science (2nd ed., pp. 327-371). Mahwah, NJ: Erlbaum, 2003. Construct Validation 2 A theory comprises two sets of relationships, one that links constructs that constitute the substance of the theory, and another that maps constructs onto phenomena that can be directly observed and measured (Bagozzi & Phillips, 1982; Costner, 1969). In organizational behavior (OB) research, theory development emphasizes relationships among constructs but devotes relatively little attention to relationships between constructs and measures (Schwab, 1980). These latter relationships are crucial to theory development, because they provide the means by which constructs become accessible to empirical research and theories are rendered testable. Moreover, because the relationships between constructs and measures are integral to a theory, theory testing is incomplete unless these relationships are scrutinized. Thus, the relationships between constructs and measures constitute an auxiliary theory that itself is subject to empirical testing and falsification (Costner, 1969; Cronbach & Meehl, 1955; Schwab, 1980). Relationships between constructs and measures are the essence of construct validity. At its most fundamental level, construct validity concerns the degree to which a measure captures its intended theoretical construct (Cronbach & Meehl, 1955). Although the notion of construct validity is straightforward, procedures used to assess construct validity are complex and have evolved considerably during the past several decades. These procedures present a potentially bewildering array of choices for OB researchers confronted with crucial task of establishing the correspondence between theoretical constructs and their measures.
    [Show full text]
  • Quantifying Construct Validity: Two Simple Measures
    Quantifying Construct Validity: Two Simple Measures The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Westen, Drew, and Robert Rosenthal. 2003. Quantifying construct validity: Two simple measures. Journal of Personality and Social Psychology 84, no. 3: 608-618. Published Version http://dx.doi.org/10.1037/0022-3514.84.3.608 Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:3708469 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA Journal of Personality and Social Psychology Copyright 2003 by the American Psychological Association, Inc. 2003, Vol. 84, No. 3, 608–618 0022-3514/03/$12.00 DOI: 10.1037/0022-3514.84.3.608 Quantifying Construct Validity: Two Simple Measures Drew Westen Robert Rosenthal Emory University University of California, Riverside and Harvard University Construct validity is one of the most central concepts in psychology. Researchers generally establish the construct validity of a measure by correlating it with a number of other measures and arguing from the pattern of correlations that the measure is associated with these variables in theoretically predictable ways. This article presents 2 simple metrics for quantifying construct validity that provide effect size estimates indicating the extent to which the observed pattern of correlations in a convergent-discriminant validity matrix matches the theoretically predicted pattern of correlations. Both measures, based on contrast analysis, provide simple estimates of validity that can be compared across studies, constructs, and measures meta-analytically, and can be implemented without the use of complex statistical proce- dures that may limit their accessibility.
    [Show full text]
  • Constructing Indices and Scales
    Constructing Indices and Scales Hsueh-Sheng Wu CFDR Workshop Series June 8, 2015 1 Outline • What are scales and indices? • Graphical presentation of relations between items and constructs for scales and indices • Why do sociologists need scales and indices? • Similarities and differences between scales and indices • Constructions of scales and indices • Criteria for evaluating a composite measure • Evaluation of scales and indices • How to obtain the sum score of a scale or an index? • Conclusion 2 What Are Scales and Indices? • Scales and indices are composite measures that use multiple items to collect information about a construct. These items are then used to rank individuals. • Examples of scales: – Depression scales – Anxiety scale – Mastery scale • Examples of indices: – Socio-Economic Status (SES) index – Consumer price index – Stock market index – Body mass Index 3 Graphical Presentation of Relations between Items and Constructs for Scales and Indices • Scale: Feeling sad e Depression Sleepless e Suicidal Ideation e • Index: Education Social Income Economic e Status Occupation 4 Why Do Sociologists Need Scales and Indices? • Most social phenomenon of interest are multi- dimensional constructs and cannot be measured by a single question, for example: – Well-being – Violence • When a single question is used, the information may not be very reliable because people may have different responses to a particular word or idea in the question. • The variation of one question may not be enough to differentiate individuals. • Scales and indices allow researchers to focus on large theoretical constructs rather than individual empirical indicator. 5 Similarities and Differences between Scales and Indices Similarities: • Both try to measure a composite construct or constructs.
    [Show full text]
  • Convergent and Discriminant Validity of the Test of Sensitivity
    University of Montana ScholarWorks at University of Montana Graduate Student Theses, Dissertations, & Professional Papers Graduate School 1975 Convergent and discriminant validity of the test of sensitivity Marianne Thomas Spitzform The University of Montana Follow this and additional works at: https://scholarworks.umt.edu/etd Let us know how access to this document benefits ou.y Recommended Citation Spitzform, Marianne Thomas, "Convergent and discriminant validity of the test of sensitivity" (1975). Graduate Student Theses, Dissertations, & Professional Papers. 5607. https://scholarworks.umt.edu/etd/5607 This Thesis is brought to you for free and open access by the Graduate School at ScholarWorks at University of Montana. It has been accepted for inclusion in Graduate Student Theses, Dissertations, & Professional Papers by an authorized administrator of ScholarWorks at University of Montana. For more information, please contact [email protected]. CONVERGENT AND DISCRIMINANT VALIDITY OF THE TEST OF SENSITIVITY By Marianne Spitzform B.S. Allegheny College, 1968 M.T.S. Harvard Divinity School, 1970 Presented in partial fulfillment of the requirements for the degree of Master of Arts UNIVERSITY OF MONTANA 1975 Approved by; , Board of Examiners Deaia'L/Gradu UMI Number: EP41071 All rights reserved INFORMATION TO ALL USERS The quality of this reproduction is dependent upon the quality of the copy submitted. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if material had to be removed, a note will indicate the deletion. Dissertation Publishing UMI EP41071 Published by ProQuest LLC (2014). Copyright in the Dissertation held by the Author.
    [Show full text]
  • The Validity of Measurement Methods, Text Version
    University of York Department of Health Sciences Measuring Health and Disease The validity of measurement methods Validity In this lecture I shall discuss some of the statistical procedures used in the validation of measurement techniques. There is no universally accepted definition of validity, but we shall regard a measurement technique as valid if it measures what we want it to measure. Because of the great variety of measurement techniques, there is no strategy of validation which can be used in all cases. Validation is very much an ad hoc process. For example, if we want to validate a new sphygmomanometer we can compare readings directly with those made with a random zero instrument, which we regard as a valid method. If there is good agreement between the two instruments, we can conclude that the new instrument is valid. However, if we want to validate a set of respiratory symptom questions used in the study of possible effects on children of air pollution or passive smoking, we cannot compare the answers to these questions with any objective measurement of respiratory distress. We must rely on more indirect methods to assess validity, such as the relationship between answers to similar questions asked to children and their parents, or between answers and measured lung function. Because of the great variety of measures for which we seek to investigate the validity, there are many terms used. Some of these do not have consistent interpretations and may overlap. They include concurrent validity, construct validity, content validity, convergent validity, criterion validity, discriminant validity, divergent validity, face validity, and predictive validity.
    [Show full text]
  • Convergent and Divergent Validity of the Learning Transfer Questionnaire. Annette Irving Bookter Louisiana State University and Agricultural & Mechanical College
    Louisiana State University LSU Digital Commons LSU Historical Dissertations and Theses Graduate School 1999 Convergent and Divergent Validity of the Learning Transfer Questionnaire. Annette Irving Bookter Louisiana State University and Agricultural & Mechanical College Follow this and additional works at: https://digitalcommons.lsu.edu/gradschool_disstheses Recommended Citation Bookter, Annette Irving, "Convergent and Divergent Validity of the Learning Transfer Questionnaire." (1999). LSU Historical Dissertations and Theses. 7068. https://digitalcommons.lsu.edu/gradschool_disstheses/7068 This Dissertation is brought to you for free and open access by the Graduate School at LSU Digital Commons. It has been accepted for inclusion in LSU Historical Dissertations and Theses by an authorized administrator of LSU Digital Commons. For more information, please contact [email protected]. INFORMATION TO USERS This manuscript has been reproduced from the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type of com puter printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note wiH indicate the deletion. Oversize materials (e.g., maps, drawings, charts) are reproduced by sectioning the original, beginning at the upper left-hand comer and continuing from left to right in equal sections with small overlaps.
    [Show full text]