Data Sharing and Reanalysis of Randomized Controlled Trials In

Total Page:16

File Type:pdf, Size:1020Kb

Data Sharing and Reanalysis of Randomized Controlled Trials In RESEARCH Data sharing and reanalysis of randomized controlled trials in leading biomedical journals with a full data sharing policy: survey of studies published in The BMJ and PLOS Medicine Florian Naudet,1 Charlotte Sakarovitch,2 Perrine Janiaud,1 Ioana Cristea,1,3 Daniele Fanelli,1,4 David Moher,1,5 John P A Ioannidis1,6 1Meta-Research Innovation ABSTRACT in contacting corresponding authors and lack of Center at Stanford (METRICS), OBJECTIVES resources on their behalf in preparing the datasets. Stanford University, Stanford, To explore the effectiveness of data sharing by In addition, there was a range of different data California, USA randomized controlled trials (RCTs) in journals with sharing practices across study groups. 2Quantitative Sciences Unit, Division of Biomedical a full data sharing policy and to describe potential CONCLUSIONS Informatics Research, difficulties encountered in the process of performing Data availability was not optimal in two journals Department of Medicine, reanalyses of the primary outcomes. Stanford University, Stanford, with a strong policy for data sharing. When CA, USA DESIGN investigators shared data, most reanalyses largely 3Department of Clinical Survey of published RCTs. reproduced the original results. Data sharing Psychology and Psychotherapy, practices need to become more widespread and Babes-Bolyai University, SETTING Romania PubMed/Medline. streamlined to allow meaningful reanalyses and 4 reuse of data. Department of Methodology, ELIGIBILITY CRITERIA London School of Economics RCTs that had been submitted and published by The TRIAL REGISTRATION and Political Science, UK Open Science Framework osf.io/c4zke. 5Centre for Journalology, Clinical BMJ and PLOS Medicine subsequent to the adoption Epidemiology Program, Ottawa of data sharing policies by these journals. Hospital Research Institute, MAIN OUTCOME MEASURE Introduction Ottawa, ON, Canada The primary outcome was data availability, defined 6Departments of Medicine, of Patients, medical practitioners, and health policy Health Research and Policy, of as the eventual receipt of complete data with clear analysts are more confident when the results and Biomedical Data Science, and labelling. Primary outcomes were reanalyzed to assess conclusions of scientific studies can be verified. For of Statistics, Stanford University, to what extent studies were reproduced. Difficulties Stanford, California, USA a long time, however, verifying the results of clinical encountered were described. Correspondence to: trials was not possible, because of the unavailability J P A Ioannidis RESULTS of the data on which the conclusions were based. [email protected] 37 RCTs (21 from The BMJ and 16 from PLOS Medicine) Data sharing practices are expected to overcome this Additional material is published published between 2013 and 2016 met the eligibility problem and to allow for optimal use of data collected online only. To view please visit the journal online. criteria. 17/37 (46%, 95% confidence interval 30% to in trials: the value of medical research that can inform Cite this as: BMJ 2018;360:k400 62%) satisfied the definition of data availability and clinical practice increases with greater transparency http://dx.doi.org/10.1136/bmj.k400 14 of the 17 (82%, 59% to 94%) were fully reproduced and the opportunity for external researchers to Accepted: 8 January 2018 on all their primary outcomes. Of the remaining RCTs, reanalyze, synthesize, or build on previous data. errors were identified in two but reached similar In 2016 the International Committee of Medical conclusions and one paper did not provide enough Journal Editors (ICMJE) published an editorial1 stating information in the Methods section to reproduce the that “it is an ethical obligation to responsibly share analyses. Difficulties identified included problems data generated by interventional clinical trials because participants have put themselves at risk.” The ICMJE WHat IS ALREADY KNOWN ON THIS TOPIC proposed to require that deidentified individual patient The International Committee of Medical Journal Editors requires that a data data (IPD) are made publicly available no later than sharing plan be included in each paper (and prespecified in study registration) six months after publication of the trial results. This 2-7 Two leading general medical journals, The BMJ and PLOS Medicine, already have proposal triggered debate. In June 2017, the ICMJE a stronger policy, expressly requiring data sharing as a condition for publication stepped back from its proposal. The new requirements of randomized controlled trials (RCTs) do not mandate data sharing but only a data sharing Only a small number of reanalyses of RCTs has been published to date; of these, plan to be included in each paper (and prespecified in 8 a minority was conducted by entirely independent authors study registration). Because of this trend toward a new norm where WHat THIS stUDY ADDS data sharing for randomized controlled trials (RCTs) Data availability was not optimal in two journals with a strong policy for data becomes a standard, it seems important to assess sharing, but the 46% data sharing rate observed was higher than elsewhere in how accessible the data are in journals with existing the biomedical literature data sharing policies. Two leading general medical When reanalyses are possible, these mostly yield similar results to the original journals, The BMJ9 10 and PLOS Medicine,11 already analysis; however, these reanalyses used data at a mature analytical stage have a policy expressly requiring data sharing as a Problems in contacting corresponding authors, lack of resources in preparing condition for publication of clinical trials: data sharing the datasets, and important heterogeneity in data sharing practices are barriers became a requirement after January 2013 for RCTs on to overcome drugs and devices9 and July 2015 for all therapeutics10 the bmj | BMJ 2018;360:k400 | doi: 10.1136/bmj.k400 1 RESEARCH at The BMJ, and after March 2014 for all types of we sent a standardized email (https://osf.io/h9cas/). interventions at PLOS Medicine. Initial emails were sent from a professional email We explored the effectiveness of RCT data sharing in address ([email protected]), and three additional both journals in terms of data availability, feasibility, reminders were sent to each author two or three weeks and accuracy of reanalyses and describe potential apart, in case of non-response. difficulties encountered in the process of performing reanalyses of the primary outcomes. We focused on Data availability RCTs because they are considered to represent high Our primary outcome was data availability, defined as quality evidence and because availability of the data the eventual receipt of data presented with sufficient is crucial in the evaluation of health interventions. information to reproduce the analysis of the primary RCTs represent the most firmly codified methodology, outcomes of the included RCTs (ie, complete data with which also allows data to be most easily analyzed. clear labelling). Additional information was collected Moreover, RCTs have been the focus of transparency on type of data sharing (request by email, request and data sharing initiatives,12 owing to the importance using a specific website, request using a specific of primary data availability in the evaluation of register, available on a public register, other), time for therapeutics (eg, for IPD meta-analyses). collecting the data (in days, time between first attempt to success of getting a database), deidentification Methods of data (concerning name, birthdate, and address), The methods were specified in advance. They were type of data shared (from case report forms to directly documented in a protocol submitted for review on 12 analyzable datasets),13 sharing of analysis code, and November 2016 and subsequently registered with the reasons for non-availability in case data were not Open Science Framework on 15 November 2016 (https:// shared. osf.io/u6hcv/register/565fb3678c5e4a66b5582f67). Reproducibility Eligibility criteria When data were available, a single researcher (FN) We surveyed publications of RCTs, including cluster carried out a reanalysis of the trial. For each study, trials and crossover studies, non-inferiority designs, analyses were repeated exactly as described in the and superiority designs, that had been submitted and published report of the study. Whenever insufficient published by The BMJ and PLOS Medicine subsequent to details about the analysis was provided in the the adoption of data sharing policies by these journals. study report, we sought clarifications from the trial investigators. We considered only analyses concerning Search strategy and study selection the primary outcome (or outcomes, if multiple primary We identified eligible studies from PubMed/Medline. For outcomes existed) of each trial. Any discrepancy The BMJ we used the search strategy: “BMJ”[jour] AND between results obtained in the reanalysis and (“2013/01/01”[PDAT]: “2017/01/01”[PDAT]) AND those reported in the publication was examined in Randomized Controlled Trial[ptyp]. For PLOS Medicine consultation with a statistician (CS). This examination we used: “PLoS Med”[jour]) AND (“2014/03/01”[PDAT]: aimed to determine if, based on both quantitative “2017/01/01”[PDAT]) AND Randomized Controlled (effect size, P values) and qualitative (clinical Trial[ptyp]. judgment) consideration, the discrepant results of the Two reviewers (FN and PJ) performed the eligibility
Recommended publications
  • Data Management and Data Sharing in Science and Technology Studies
    Article Science, Technology, & Human Values 1-18 ª The Author(s) 2018 Data Management Article reuse guidelines: sagepub.com/journals-permissions DOI: 10.1177/0162243918798906 and Data Sharing journals.sagepub.com/home/sth in Science and Technology Studies Katie Jane Maienschein1, John N. Parker2, Manfred Laubichler1 and Edward J. Hackett3 Abstract This paper presents reports on discussions among an international group of science and technology studies (STS) scholars who convened at the US National Science Foundation (January 2015) to think about data sharing and open STS. The first report, which reflects discussions among members of the Society for Social Studies of Science (4S), relates the potential benefits of data sharing and open science for STS. The second report, which reflects dis- cussions among scholars from many professional STS societies (i.e., European Association for the Study of Science and Technology [EASST], 4S, Society for the History of Technology [SHOT], History of Science Society [HSS], and Philosophy of Science Association [PSA]), focuses on practical and conceptual issues related to managing, storing, and curating STS data. As is the case for all reports of such open discussions, a scholar’s presence at the meeting does not necessarily mean that they agree with all aspects of the text to follow. 1Center for Biology and Society, School of Life Sciences, Arizona State University, Tempe, AZ, USA 2Barrett, The Honors College, Arizona State University, Tempe, AZ, USA 3Heller School for Social Policy and Management, Brandeis University, Waltham, MA, USA Corresponding Author: John N. Parker, Barrett, The Honors College, Arizona State University, PO Box 871612, Tempe, AZ 85287, USA.
    [Show full text]
  • Sharing Is Caring—Data Sharing Initiatives in Healthcare
    International Journal of Environmental Research and Public Health Review Sharing Is Caring—Data Sharing Initiatives in Healthcare Tim Hulsen Department of Professional Health Solutions & Services, Philips Research, 5656AE Eindhoven, The Netherlands; [email protected] Received: 28 February 2020; Accepted: 24 April 2020; Published: 27 April 2020 Abstract: In recent years, more and more health data are being generated. These data come not only from professional health systems, but also from wearable devices. All these ‘big data’ put together can be utilized to optimize treatments for each unique patient (‘precision medicine’). For this to be possible, it is necessary that hospitals, academia and industry work together to bridge the ‘valley of death’ of translational medicine. However, hospitals and academia often are reluctant to share their data with other parties, even though the patient is actually the owner of his/her own health data. Academic hospitals usually invest a lot of time in setting up clinical trials and collecting data, and want to be the first ones to publish papers on this data. There are some publicly available datasets, but these are usually only shared after study (and publication) completion, which means a severe delay of months or even years before others can analyse the data. One solution is to incentivize the hospitals to share their data with (other) academic institutes and the industry. Here, we show an analysis of the current literature around data sharing, and we discuss five aspects of data sharing in the medical domain: publisher requirements, data ownership, growing support for data sharing, data sharing initiatives and how the use of federated data might be a solution.
    [Show full text]
  • ORION OPEN SCIENCE FACTSHEETS Brief, Informative, and Easy-To-Understand One-Page Factsheets on Open Science Topics
    ORION OPEN SCIENCE FACTSHEETS Brief, informative, and easy-to-understand one-page factsheets on Open Science topics Created by the Max-Delbrück-Centre for Molecular Medicine in the Helmholtz Association Design by Vetenskap & Allmänhet This project has received funding from the European Union’s Horizon 2020 research and innovation programme under Grant Agreement No. 741527 and runs from May 2017 to April 2021. CONTENTS 1. Biohacking and DIYBio Research .......................................................................................... 3 2. Career Benefits of Open Science ......................................................................................... 4 3. Citizen Science ....................................................................................................................... 5 4. Commercialisation of Research ............................................................................................ 6 5. Communicating Animal Research ......................................................................................... 7 6. Crowd Science ....................................................................................................................... 8 7. Research Data Management ................................................................................................. 9 8. Open Access ......................................................................................................................... 10 9. Open Research Data ...........................................................................................................
    [Show full text]
  • Standards for Design and Measurement Would Make Clinical
    PAPER Standards for design and measurement would make COLLOQUIUM clinical research reproducible and usable Kay Dickersina,1 and Evan Mayo-Wilsona aDepartment of Epidemiology, Johns Hopkins Bloomberg School of Public Health, Baltimore, MD 21205 Edited by David B. Allison, Indiana University Bloomington, Bloomington, IN, and accepted by Editorial Board Member Susan T. Fiske September 10, 2017 (received for review June 20, 2017) We find standards useful in everyday life and in science, although In research, standards facilitate cooperation and better overall we do not always follow them. Adopting new standards can be results—namely, good science. Across disciplines, scientists tacitly expensive, so there may be a strong incentive to maintain the status and formally agree to common standards for both the conduct and quo rather than adopt new standards. The scientific community has reporting of our work, ranging from units of measurement to many standards encompassing both doing clinical research and principles about research integrity (3). Standards may have im- reporting it, including standards for design and measurement. Al- portant benefits without incurring substantial costs or limiting though existing research standards have improved both research scientific creativity. For example, the American Heart Association and its reporting, we need to unify existing standards and to fill developed standards for measuring blood pressure (4); their use the gaps between steps throughout the research process. Existing may improve the consistency of data collection across times and gaps include implementation of standards and links between stan- places, thus improving the comparability of clinical trials. Failing dards for study registration (to know about all studies undertaken), to apply minimum standards can lead to calamitous errors.
    [Show full text]
  • Biodiversa Citizen Science Toolkit P
    The ERA-NET promoting European research on biodiversity, ecosystem services and Nature-based solutions BiodivERsA CITIZEN SCIENCE TOOLKIT for biodiversity scientists 1 To cite this report Photography credits Goudeseune, L., Eggermont, H., Groom, Q., Le Cover picture: Glacier NPS Roux, X., Paleco, C., Roy, H.E., van Noordwijk, p.2: © Edwin Brossens C.G.E. (2020). BiodivERsA Citizen Science Toolkit p. 4/5: © Märt Kose For Biodiversity Scientists. BiodivERsA report, 44 pp. p.7: © NPS Photo by Ivie Metzen DOI: https://doi.org/10.5281/zenodo.3979343 p.9: Pexels p.13: Unsplash p.14/15: © Park Ranger - Cascades Butterfly Main contact for this toolkit ProjectTeam p.16: Picjumbo [email protected] p.17: © PNR Millevaches en Limousin p.18: © Antonio Dinis Fereira Layout p.19: Pixabay p.20: Rawpixel.com Thibaut Lochu p.23: Pexels p.24: © Michael Krikorev p.25: © Johan Enqvist p.26: © Vlad Karpinskiy p.27: Unsplash p.28/29: Pixabay p.32: Pexels p.33: © Albert de Jong. p.34/35: © PNRP - F. Picandet - Perché dans le Perche p.42/43: Pexels TO CONTACT BIODIVERSA BiodivERsA Chair & Coordinator Xavier Le Roux [email protected] Tel: +33 (0)6 31 80 38 20 BiodivERsA Secretariat [email protected] Fondation pour la Recherche sur la Biodiversité 195 rue St Jacques, 75005 Paris, France www.biodiversa.org FOLLOW US: @BiodivERsA3 /BiodivERsA3 2 TABLE OF CONTENTS PART I: INTRODUCTION 4 A About this toolkit 6 B What is Citizen Science? 7 C The rise of Citizen Science 8 D Different types of Citizen Science projects 10 E Common Citizen
    [Show full text]
  • Data Sharing and the Future of Science Who Benefits from Sharing Data? the Scientists of Future Do, As Data Sharing Today Enables New Science Tomorrow
    EDITORIAL OPEN Data sharing and the future of science Who benefits from sharing data? The scientists of future do, as data sharing today enables new science tomorrow. Far from being mere rehashes of old datasets, evidence shows that studies based on analyses of previously published data can achieve just as much impact as original projects. ata sharing has a long history makes new types of research possible. in many areas of research. Consider, for instance, research using the Although the push to encou- Human Connectome Project (HCP) data- Drage social and biological sci- set, one of the data sharing initiatives entists to share and pool their included in the Milham et al. study. The results is a recent one1, in other fields the HCP currently contains extensive fMRI, use of shared data has been the norm for structural MRI and behavioural data from some time. For over a century, much of 1200 healthy young adult volunteers 1234567890():,; economics and meteorology have been (https://www.humanconnectome.org/ based on publicly shared data, for example. study/hcp-young-adult), and is expanding However, trepidation in relation to data- to encompass child, adolescent and older sharing is still prevalent in the scientific adult brains. These data are made available community, particularly in certain dis- to any interested researcher. ciplines. The issues that While data sharing had a somewhat “There is a strong argument to be make some researchers rocky start in the world of cognitive neu- reluctant to share their roscience4, the success of the HCP and the made that leaving data unshared is own data have been many influential studies based on it shows much discussed2,but that its time has come.
    [Show full text]
  • A Mixed-Methods Investigation Into Barriers for Sharing Geospatial and Resilience Flood Data in the UK
    water Article A Mixed-Methods Investigation into Barriers for Sharing Geospatial and Resilience Flood Data in the UK Luke Waterman 1,Mónica Rivas Casado 1,* , Emma Bergin 2 and Gary McInally 2 1 School of Water, Energy and Environment, Cranfield University, College Road, Cranfield, Bedfordshire MK43 0AL, UK; Luke.J.Waterman@cranfield.ac.uk 2 Flood Re, 75 King William Street, London EC4N 7BE, UK; emma.bergin@floodre.co.uk (E.B.); gary.mcinally@floodre.co.uk (G.M.) * Correspondence: m.rivas-casado@cranfield.ac.uk Abstract: With increases in average temperature and rainfall predicted, more households are expected to be at risk of flooding in the UK by 2050. Data and technologies are increasingly playing a critical role across public-, private- and third-sector organisations. However, barriers and constraints exist across organisations and industries that limit the sharing of data. We examine the international context for data sharing and variations between data-rich and data-sparse countries. We find that local politics and organisational structures influence data sharing. We focus on the case study of the UK, and on geospatial and flood resilience data in particular. We use a series of semi-structured interviews to evaluate data sharing limitations, with particular reference to geospatial and flood resilience data. We identify barriers and constraints when sharing data between organisations. We find technological, security, privacy, cultural and commercial barriers across different use cases and data points. Finally, we provide three long-term recommendations to improve the overall accessibility to flood data and enhance outcomes for organisations and communities. Citation: Waterman, L.; Rivas Casado, M.; Bergin, E.; McInally, G.
    [Show full text]
  • CODATA Task Group on Data Citation Standards
    1 CODATA Task Group on data citation standards and best practices: Research and analysis results As part of the activities of the CODATA Data Citation Standards and Practices Task Group, since its inception in 2010, we conducted an inventory of existing literature on data citation best practices and attribution activities. This document is the result of the collection of bibliographic sources, subsequent research, and corresponding analysis. Prior versions of this bibliography have been published on the Task Group’s website. [http://www.codata.org/taskgroups/TGdatacitation/Bilbliography_Links.html]. This version integrates all additional references gathered after the publication of the Summary Report of our August 2011 Workshop (sponsored jointly with the Board on Research Data and Information of the National Academy of Sciences), National Research Council. For Attribution -- Developing Data Attribution and Citation Practices and Standards: Summary of an International Workshop. Washington, DC: The National Academies Press, 2012. The corpus of this bibliography was created by Task Group members and consultants who performed online database searches and captured information sources that are directly or peripherally focused on data citation practices and attribution. These contributions were then shared via email or via Zotero, a reference management software tool. Many new entries were contributed over time by authors and reviewers of the Task Group’s reports, interviewees from various stakeholder communities, and other interested persons. As of this version, we have found 441 resources and organized them into 16 different formats that cover the many facets of citation, such as policies, infrastructure, research practices, and best practices development. We concentrated our efforts on sources that were published during the past five years, with the occasional older, seminal item included because of additional context and background.
    [Show full text]
  • Digital Research Data Sharing and Management
    NSB-11-79 December 14, 2011 Prepublication Copy Digital Research Data Sharing and Management December 2011 Task Force on Data Policies Committee on Strategy and Budget National Science Board National Science Board Ray M. Bowen, Chairman, President Emeritus, Texas A&M University, College Station, Texas, and Visiting Distinguished Professor, Rice University, Houston, Texas Esin Gulari, Vice Chairman, Dean of Engineering and Science, Clemson University, Clemson, South Carolina Mark R. Abbott, Dean and Professor, College of Earth, Ocean, and Atmospheric Sciences, Oregon State University, Corvallis, Oregon Dan E. Arvizu, Director and Chief Executive, National Renewable Energy Laboratory, Golden, Colorado Bonnie Bassler*, Howard Hughes Medical Institute Investigator, Squibb Professor of Molecular Biology Princeton University, Princeton, New Jersey Camilla P. Benbow, Patricia and Rodes Hart Dean of Education and Human Development, Peabody College of Education and Human Development, Vanderbilt University, Nashville, Tennessee John T. Bruer, President, The James S. McDonnell Foundation, Saint Louis, Missouri France A. Córdova, President, Purdue University, West Lafayette, Indiana Kelvin K. Droegemeier, Vice President for Research, Regents’ Professor of Meteorology and Weathernews Chair Emeritus, University of Oklahoma, Norman, Oklahoma Patricia D. Galloway, Chief Executive Officer, Pegasus Global Holdings, Inc., Cle Elum, Washington José-Marie Griffiths, Vice President for Academic Affairs and University Professor, Bryant University, Smithfield, Rhode Island Louis J. Lanzerotti*, Distinguished Research Professor of Physics, Center for Solar Terrestrial Research, Department of Physics, New Jersey Institute of Technology, Newark, New Jersey Alan I. Leshner, Chief Executive Officer, Executive Publisher, Science, American Association for the Advancement of Science, Washington, DC W. Carl Lineberger, Fellow of JILA, E. U. Condon Distinguished Professor of Chemistry, University of Colorado, Boulder, Colorado G.P.
    [Show full text]
  • Data Papers As a New Form of Knowledge Organization in the Field of Research Data Joachim Schöpfel, Dominic Farace, Hélène Prost, Antonella Zane
    Data papers as a new form of knowledge organization in the field of research data Joachim Schöpfel, Dominic Farace, Hélène Prost, Antonella Zane To cite this version: Joachim Schöpfel, Dominic Farace, Hélène Prost, Antonella Zane. Data papers as a new form of knowledge organization in the field of research data. 12ème Colloque international d’ISKO-France : Données et mégadonnées ouvertes en SHS : de nouveaux enjeux pour l’état et l’organisation des con- naissances ?, ISKO France, Oct 2019, Montpellier, France. halshs-02284548 HAL Id: halshs-02284548 https://halshs.archives-ouvertes.fr/halshs-02284548 Submitted on 7 Oct 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Data Papers as a New Form of Knowledge Organization in the Field of Research Data Joachim Schöpfel University of Lille, GERiiCO laboratory Associate professor in information sciences, ORCID 0000-0002-4000-807X [email protected] Dominic Farace GreyNet International, Amsterdam Director of the Grey Literature Network Service, ORCID 0000-0003-2561-3631 [email protected] Hélène Prost CNRS, GERiiCO laboratory Information engineer, ORCID 0000-0002-7982-2765 [email protected] Antonella Zane University of Padova, Library System Head of the Antonio Vallisneri Bio-medical Library, ORCID 0000-0001-7218-6068 [email protected] Abstract Data papers have been defined as scholarly journal publications whose primary purpose is to describe research data.
    [Show full text]
  • A Content Analysis on the Rhetorical Moves Used in Data Paper Abstracts K
    The data paper as a socio-linguistic epistemic object: A content analysis on the rhetorical moves used in data paper abstracts Kai Li School of Information Resource Management, Renmin University of China, Beijing, China 59 Zhongguancun St, Haidian District, Beijing, China, 100872 +86 13161285125; [email protected] Chenyue Jiao School of Information Sciences, University of Illinois Urbana-Champaign, Illinois, USA 501 E. Daniel St, Champaign, IL, USA, 61820 1-217-305-2995; [email protected] Abstract The data paper is an emerging academic genre that focuses on the description of research data objects. However, there is a lack of empirical knowledge about this rising genre in quantitative science studies, particularly from the perspective of its linguistic features. To fill this gap, this research aims to offer a first quantitative examination of which rhetorical moves—rhetorical units performing a coherent narrative function—are used in data paper abstracts, as well as how these moves are used. To this end, we developed a new classification scheme for rhetorical moves in data paper abstracts by expanding a well-received system that focuses on English- language research article abstracts. We used this expanded scheme to classify and analyze rhetorical moves used in two flagship data journals, Scientific Data and Data in Brief. We found that data papers exhibit a combination of IMRaD- and data-oriented moves and that the usage differences between the journals can be largely explained by journal policies concerning abstract and paper structure. This research offers a novel examination of how the data paper, a novel data-oriented knowledge representation, is composed, which greatly contributes to a deeper understanding of data and data publication in the scholarly communication system.
    [Show full text]
  • Toward Open Data Policies in Phonetics
    Toward open data policies in phonetics: What we can gain and how we can avoid pitfalls Marc Garellek, Matthew Gordon, James Kirby, Wai-Sum Lee, Alexis Michaud, Christine Mooshammer, Oliver Niebuhr, Daniel Recasens, Timo Roettger, Adrian Simpson, et al. To cite this version: Marc Garellek, Matthew Gordon, James Kirby, Wai-Sum Lee, Alexis Michaud, et al.. Toward open data policies in phonetics: What we can gain and how we can avoid pitfalls. Journal of Speech Science, 2020, 9 (1), pp.3 - 16. halshs-02894375v2 HAL Id: halshs-02894375 https://halshs.archives-ouvertes.fr/halshs-02894375v2 Submitted on 10 Sep 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution - NonCommercial - ShareAlike| 4.0 International License Journal of Speech Sciences (9): 03-16. 2020. Available at: http://revistas.iel.unicamp.br/joss LETTER TO THE EDITOR: TOWARD OPEN DATA POLICIES IN PHONETICS: WHAT WE CAN GAIN AND HOW WE CAN AVOID PITFALLS GARELLEK, Marc1 GORDON, Matthew2 KIRBY, James3 LEE, Wai-Sum4 MICHAUD, Alexis5* MOOSHAMMER,
    [Show full text]