1 Estimating the Prevalence of Transparency and Reproducibility
Total Page:16
File Type:pdf, Size:1020Kb
Estimating the prevalence of transparency and reproducibility-related research practices in psychology (2014-2017) Tom E. Hardwicke1,2, Robert T. Thibault3,4, Jessica E. Kosie5,6, Joshua D. Wallach7,8, Mallory C. Kidwell9, & John P. A. Ioannidis2,10,11 1 Department of Psychology, University of Amsterdam 2 Meta-Research Innovation Center Berlin (METRIC-B), QUEST Center for Transforming Biomedical Research, Berlin Institute of Health, Charité – Universitätsmedizin Berlin 3 School of Psychological Science, University of Bristol 4 MRC Integrative Epidemiology Unit at the University of Bristol 5 Department of Psychology, University of Oregon 6 Department of Psychology, Princeton University 7 Department of Environmental Health Sciences, Yale School of Public Health 8 Collaboration for Research Integrity and Transparency, Yale School of Medicine 9 Department of Psychology, University of Utah 10 Departments of Medicine, of Health Research and Policy, of Biomedical Data Science, and of Statistics, Stanford University 11 Meta-Research Innovation Center at Stanford (METRICS), Stanford University Author note Tom Hardwicke was based at Stanford University when the study began and is now based at the University of Amsterdam. Jessica Kosie was based at University of Oregon when the study began and is now based at Princeton University. Correspondence concerning this article should be addressed to Tom Hardwicke, Nieuwe Achtergracht 129B, Department of Psychology, University of Amsterdam, 1018 WT Amsterdam, The Netherlands. E-mail: [email protected] 1 Abstract Psychologists are navigating an unprecedented period of introspection about the credibility and utility of their discipline. Reform initiatives have emphasized the benefits of several transparency and reproducibility-related research practices; however, their adoption across the psychology literature is unknown. To estimate their prevalence, we manually examined a random sample of 250 psychology articles published between 2014-2017. Over half of the articles were publicly available (154/237, 65% [95% confidence interval, 59%-71%]); however, sharing of research materials (26/183, 14% [10%-19%]), study protocols (0/188, 0% [0%-1%]), raw data (4/188, 2% [1%-4%]), and analysis scripts (1/188, 1% [0%-1%]) was rare. Pre-registration was also uncommon (5/188, 3% [1%-5%]). Many articles included a funding disclosure statement (142/228, 62% [56%-69%]), but conflict of interest statements were less common (88/228, 39% [32%-45%]). Replication studies were rare (10/188, 5% [3%-8%]) and few studies were included in systematic reviews (21/183, 11% [8%-16%]) or meta-analyses (12/183, 7% [4%-10%]). Overall, the results suggest that transparency and reproducibility-related research practices were far from routine. These findings establish a baseline which can be used to assess future progress towards increasing the credibility and utility of psychology research. Keywords: transparency, reproducibility, meta-research, psychology, open science 2 Introduction Serious concerns about the credibility and utility of some scientific literature (Ioannidis, 2005; Ioannidis et al., 2014) have prompted calls for increased adoption of research practices that enhance reproducibility and transparency (Miguel et al., 2014; Munafò et al., 2017; Nosek et al., 2015; Wallach et al., 2018). Close scrutiny of psychology in particular has suggested that standard research and publication practices have rendered the discipline highly exposed to bias, potentially resulting in a large volume of exaggerated and misleading results (Ioannidis et al., 2014; John et al., 2012; Open Science Collaboration, 2015; Pashler & Wagenmakers, 2012; Simmons et al., 2011; Szucs & Ioannidis, 2017). This realization within the field has also led to a number of reform efforts (Hardwicke & Ioannidis, 2018a; Hardwicke et al., 2019; Nelson et al., 2018; Vazire, 2018), which have the potential to improve efficiency (Chalmers & Glasziou, 2009), facilitate self-correction (Ioannidis, 2012), and enhance credibility (Vazire, 2017). A central focus of reform initiatives has been to encourage scientists to share more information about the studies they perform. Journal articles are only the most visible facade of deeper layers of scholarship that may include protocols, original research materials, raw data, and analysis scripts – resources that are not necessarily shared with other scientists (Buckheit & Donoho, 1995; Klein et al., 2018). Even journal articles themselves may only be accessible to those with institutional access or the ability to pay a fee (Piwowar et al., 2018). Additionally, potential sources of bias, such as conflicts of interest and funding sources, may not be disclosed (Bekelman et al., 2003; Cristea & Ioannidis, 2018). However, there is a growing (or re-emerging, David, 2008) appreciation that the scientific community needs to be able to access all of this information in order to comprehensively evaluate, interpret, and independently verify scientific claims (Vazire, 2017; Munafò et al., 2017). Furthermore, access to this information enables 3 replication, evidence synthesis, and discovery activities that may ultimately accelerate scientific progress (Ioannidis, 2012; Klein et al., 2018). The burgeoning discipline of meta-research (‘research on research’) has already begun to evaluate the impact of various reform initiatives (Hardwicke et al., 2020). For example, journal data sharing policies have been associated with moderate to substantial increases in data sharing (Hardwicke et al., 2018; Kidwell et al., 2016; Nuijten et al., 2018; Rowhani-Farid & Barnett, 2016; Naudet et al., 2018). However, the prevalence of transparency and reproducibility-related research practices across the psychology literature is largely unknown. Based on previous investigations in biomedicine (Iqbal et al., 2016; Wallach et al., 2018) and the social sciences (Hardwicke et al., 2019), we manually examined a random sample of 250 articles to estimate the prevalence of several transparency and reproducibility-related indicators in psychology articles published between 2014-2017. The indicators were open access to published articles; availability of study materials, study protocols, raw data, and analysis scripts; pre-registration; disclosure of funding sources and conflicts of interest; conduct of replication studies; and cumulative synthesis of evidence in meta-analyses and systematic reviews. Methods Design This was a retrospective observational study with a cross-sectional design. Sampling units were individual articles. Measured variables are shown in Table 1. 4 Table 1. Measured variables. The variables measured for an individual article depended on the study design classification. For articles that were not available (the full text could not be retrieved) or non-English language, only article characteristics were obtained. The exact operational definitions and procedures for data extraction/coding are available in the structured form here: https://osf.io/x9rmy/ Applicable study designs Article characteristics Subject area, year of publication, study design, All country of origin (based on corresponding author's affiliation), human/animal subjects, 2017 journal impact factor (according to Thomson Reuters Journal Citation Reports) Articles Accessibility and Retrieval Method (can the All article be accessed, is there a public version or is paywall access required?) Protocols Availability statement (is availability, or lack of, Study designs involving primary data#, study explicitly declared?) designs involving secondary data (commentaries with analysis and meta-analyses). Content (what aspects of the study are included in the protocol?) Materials Availability statement (is availability, or lack of, Study designs involving primary data# explicitly declared?) Retrieval Method (e.g., upon request or via online repository) Accessibility (can the materials be accessed?) Raw data Availability statement (is availability, or lack of, explicitly declared?) 5 Retrieval Method (e.g., upon request or via online Study designs involving primary data#, study repository) designs involving secondary data (commentaries with analysis and meta-analyses). Accessibility (can the data be accessed?) Content (has all relevant data been shared?) Documentation (are the data understandable?) Analysis scripts Availability statement (is availability, or lack of, Study designs involving primary data#, study explicitly declared?) designs involving secondary data (commentaries with analysis and meta-analyses). Retrieval Method (e.g., upon request or via online repository) Accessibility (can the scripts be accessed?) Pre-registration Availability statement (is availability, or lack of, Study designs involving primary data#, study explicitly declared?) designs involving secondary data (commentaries with analysis and meta-analyses). Retrieval Method (which registry was used?) Accessibility (can the pre-registration be accessed?) Content (what was pre-registered?) Funding Disclosure statement (are funding sources, or lack All of, explicitly declared?) Conflicts of interest Disclosure statement (are conflicts of interest, or All lack of, explicitly declared?) Replication Statement (does the article claim to report a All replication?) 6 Citation history (has the article been cited by a Study designs involving primary data# study that claims to be a replication?) Evidence synthesis Meta-analysis citation history* (has the article Study designs involving primary data# been cited by, and included