A Study of Current Policies, Practices and Infrastructure Supporting the Sharing of Data to Prevent and Respond to Epidemic and Pandemic Threats

Total Page:16

File Type:pdf, Size:1020Kb

A Study of Current Policies, Practices and Infrastructure Supporting the Sharing of Data to Prevent and Respond to Epidemic and Pandemic Threats January 2018 A study of current policies, practices and infrastructure supporting the sharing of data to prevent and respond to epidemic and pandemic threats Elizabeth Pisani, Amrita Ghataure, Laura Merson Summary: Data sharing in public health emergencies: Wellcome/GloPID-R ii Executive summary Introduction Discussions around sharing public health research data have been running for close to a decade, and yet when the Ebola epidemic hit West Africa in 2014, data sharing remained the exception, not the norm. In response, the GloPID-R consortium of research funders commissioned a study to take stock of current data sharing practices involving research on pathogens with pandemic potential. The study catalogued different data sharing practices, and investigated the governance and curation standards of the most frequently used platforms and mechanisms. In addition, it sought to identify specific areas of support or investment which could lead to more effective sharing of data to prevent or limit future epidemics. Methods The study proceeded in three phases: a search for shared data, interviews with investigators, and a review of policies and statements about data sharing. We began by searching the academic literature, trials registries and research repositories for papers and data related to 12 pathogens named by the WHO as of priority concern because of their epidemic or pandemic potential. Chikungunya, Crimean Congo Haemorrhagic Fever (CCHF), Ebola, Lassa Fever, Marburg, Middle East Respiratory Syndrome Coronavirus (MERS-CoV), Severe Acute Respiratory Syndrome (SARS), Nipah, Rift Valley Fever (RVF), Severe Fever with Thrombocytopenia Syndrome (SFTS), Zika. For biomedical research involving these pathogens in humans, animals, in vitro and in silico, we sought to access the data underlying each paper, including through a short, on-line survey sent to 110 corresponding authors. A total of 28 people conducting or supporting research on the pathogens of interest were interviewed for 30-90 minutes about their data sharing practices and policies. In addition, we reviewed 105 institutional policies, discussion documents and academic commentaries about standards and norms in data sharing. Data sharing practice Extent of sharing We identified a total of 787 research studies related to the priority pathogens published since 2003. The majority were case studies involving a single patient. Of the remaining 319, 98 provided the underlying data in an openly accessible form, and a further 15 authors said they would make data available on request. Excluding the case reports, two thirds of all papers were based on undiscoverable data. Some 161 clinical trials relating to the priority pathogens were registered in either clinicaltrials.gov or the Pan-African registry of trials; 58 of them have registered completion, but 41 of those remain unpublished. Two trialists posted links to aggregate data several years after completion. None linked to individual patient data. Just over half of all published papers gave dates for data collection. Authors who shared data published faster than those who didn't (18 vs. 30 months median time lag to publication). Summary: Data sharing in public health emergencies: Wellcome/GloPID-R iii Open sharing mechanisms The most commonly used mechanism for the formal sharing of data was embedding supplementary material in journal articles, followed by deposit in general-purpose academic repositories. Very few of the datasets shared were in structured, machine- readable formats. Over half of the files provided as supplementary information included data in pdf format. Datasets in academic repositories were more likely to be structured, though many used proprietary software formats. Data shared on genetic sequences or protein structures databases are shared more commonly than clinical data for the priority pathogens, generally through extremely well-resourced community-designated databases such as GenBank or Protein Data Bank. However even these data do not appear to be shared rapidly; often, the public release of sequence data is timed to coincide with paper publication, even for pathogens with pandemic potential. Though registries do list some data repositories with a special interest in priority pathogens, uptake was limited amongst the researchers interviewed. Other formal sharing mechanisms included journal preprints and institutional repositories, together with "Big Data" approaches which collate information from public sources. Closed sharing mechanisms Our quantitative search strategy, based largely on publication, was biased towards data sharing mechanisms that were formal, findable and open. However, in interviews with the research community, we identified a second set of mechanisms for sharing data: closed consortia, regional collaborations and informal professional networks. We were unable to quantify the sharing of data through these mechanisms, but it is clear that for most of the priority pathogens informal or consortium sharing is currently more extensive, and faster, than through formal open mechanisms. In short: Most research data related to priority pathogens are not being shared for reuse through formal, discoverable means. Data that are shared through publication or open repositories are most often shared in formats that are not inter-operable or easily reusable. Closed consortia and informal, trust-based networks are favoured by researchers, particularly those working in resource-constrained settings. Table A summarises the strengths and weaknesses of some of the different sharing mechanisms. Broadly speaking, those that invest most in curation are most likely to provide long-term access to data that can most readily be reused for additional knowledge generation. However, they also tend to be the most costly. Sharing through informal networks of trusted colleagues is popular and can happen rapidly, but it lacks transparency, greatly restricts potential users, and essentially excludes machine learning. Summary: Data sharing in public health emergencies: Wellcome/GloPID-R iv Table A: Platforms and mechanisms used for the sharing of data about priority pathogens Data sharing mechanism Advantages Disadvantages Well-established Genomic/ structural Discoverable Expensive databases Standardised metadata Slow Supplementary material in Findable by humans, with some effort Not machine readable, limiting discoverability journals Public investment not vital Depends on electronic access to papers; Often paywalled Diverse, largely unusable data formats Not discoverable even by humans Journal preprints Faster than traditional publishing models Diverse, largely unusable data formats Standardised metadata Extensive set-up costs Disease-specific curated Community support Sustainability questionable repositories Greatly facilitates reuse and interoperability Duplication of effort General purpose academic Standardised metadata rare Potentially cost-effective repositories Sustainability uncertain Duplication of effort Strong support for data management and curation Institutional repositories Not easily discoverable Easy entry point for academics reluctant to share Standardised metadata rare Standardised metadata Excludes reuse by non-members Facilitates reuse and interoperability. Set-up costs Closed consortia Protects equity and research interests, so strong Duplication of effort community support Sustainability questionable Informal professional Trust-based, so strong community support Zero transparency networks Big Data approaches Cost-effective Neither sensitive nor specific Summary: Data sharing in public health emergencies: Wellcome/GloPID-R v Table B: FREE-FAIRER Principles and their inclusion in institutional data sharing policies Notes Findable Most policies address "findability" and require data to be linked to a DOI. In practice DOIs Data should be described with rich, machine-readable metadata, persistently are linked to papers, rarely to data. No policy requires standardised ontologies for metadata. and uniquely identified, and indexed in a searchable resource No policies apply to surveillance data. Few institutional policies exist specifically for research relevant to potential emergencies. Rapidly available Most are tied to publication models and other quality control norms that are inimical to speed. Data should be shared rapidly to maximise their potential utility in outbreak Preprints increase speed but reduce findability. Structures for rapid release of surveillance situations. data are dysfunctional. Ethical Discussions relating to ethics prioritise the protection of individuals. Few address the ethics of Data are collected with appropriate consent procedures. Data sharing protects withholding data or public health importance. individual privacy, without undermining the rapid protection of public health. Some policies include aspirational statements about equity, but few provide guidance on how Equitable to operationalize them. For interviewees, especially in research-poor settings, this is the Data sharing delivers fair benefit to those who collected the data, and to the biggest obstacle to sharing. In those policies that mention it, 'equity' is often reduced to norms communities from which data were collected for citation and authorship. Forever Many policies address persistent identifiers; few address long-term preservation or Data are shared through a mechanism that uses persistent identifiers and has sustainability of sharing platforms. clear, securely funding strategies to ensure the long-term
Recommended publications
  • Get Laid, Get Paid, Get High, Survive
    Get Laid, Get Paid, Get High, Survive An analytical and personal response to Elizabeth Pisani's deconstruction of HIV and AIDS Simone LeClaire SOC407: Contemporary Theory Dr. Lee Garth Vigilant March 7, 2012 LeClaire 1 The Wisdom of Whores: Bureaucrats, Brothels, and the Business of AIDS explores the phenomenon of HIV and AIDS through the enduringly personal account of epidemiologist Elizabeth Pisani. Pisani's voice weaves itself energetically through what could easily become an overwhelming amount of meaningless statistics and numbers—but instead remains a compelling and eye-opening illustration of her own experience and knowledge of the global problem of HIV and AIDS, including all kinds of the personal, social, and political complications linked to the issue. Thus, here is an powerful example of standpoint epistemology, which is an approach that, according to Dorothy Smith, "first, places sociologists where we are actually situated...and second, makes our direct embodied experience of the everyday world the primary ground of our knowledge". (1974) In other words, this first-person account is based on Pisani's own experience, and she is careful to record the context and shortcomings of that experience throughout the entirety of her account. Pisani sets the tone of her book by immediately identifying her career as "sex and drugs", and then gives a brief introduction of how she moved through a degree in Chinese, a bout of investigative journalism, and a public health education, to eventually become involved with epidemiology and specifically HIV research. The main issue that she immediately saw within the field was a gap between knowledge, and communication or action.
    [Show full text]
  • Recursive Definitions Across the Footprint That a Straight- Line Program Manipulated, E.G
    natural proofs for verification of dynamic heaps P. MADHUSUDAN UNIV. OF ILLINOIS AT URBANA-CHAMPAIGN (VISITING MSR, BANGALORE) WORKSHOP ON MAKING FORMAL VERIFICATION SCALABLE AND USEABLE CMI, CHENNAI, INDIA JOINT WORK WITH… Xiaokang Qiu .. graduating! Gennaro Andrei Pranav Garg Parlato Stefanescu GOAL: BUILD RELIABLE SOFTWARE Build systems with proven reliability and security guarantees. (Not the same as finding bugs!) Deductive verification with automatic theorem proving • Floyd/Hoare style verification • User supplied modular annotations (pre/post cond, class invariants) and loop invariants • Resulting verification conditions are derived automatically for each linear program segment (using weakest-pre/strongest-post) Verification conditions are then proved valid using mostly automatic theorem proving (like SMT solvers) TARGET: RELIABLE SYS TEMS SOFTWARE Operating Systems Angry birds Browsers Excel VMs Mail clients … App platforms Systems Software Applications Several success stories: • Microsoft Hypervisor verification using VCC [MSR, EMIC] • A secure foundation for mobile apps [@Illinois, ASPLOS’13] A SECURE FOUNDATION FOR MOBILE APPS In collaboration with King’s systems group at Illinois First OS architecture that provides verifiable, high-level abstractions for building mobile applications. Application state is private by default. • Meta-data on files controls access by apps; only app with appropriate permissions can access files • Every page is encrypted before sending to the storage service. • Memory isolation between apps • A page fault handler serves a file- backed page for a process, the fille has to be opened by the same process. • Only the current app can write to the screen buffer. • IPC channel correctness • … VERIFICATION TOOLS ARE MATURE And works! ENOUGH TO BUILD RELIABLE S/W.
    [Show full text]
  • Type 2 Diabetes Prediction Using Machine Learning Algorithms
    Type 2 Diabetes Prediction Using Machine Learning Algorithms Parisa Karimi Darabi*, Mohammad Jafar Tarokh Department of Industrial Engineering, K. N. Toosi University of Technology, Tehran, Iran Abstract Article Type: Background and Objective: Currently, diabetes is one of the leading causes of Original article death in the world. According to several factors diagnosis of this disease is Article History: complex and prone to human error. This study aimed to analyze the risk of Received: 15 Jul 2020 having diabetes based on laboratory information, life style and, family history Revised: 1 Aug 2020 with the help of machine learning algorithms. When the model is trained Accepted: 5 Aug 2020 properly, people can examine their risk of having diabetes. *Correspondence: Material and Methods: To classify patients, by using Python, eight different machine learning algorithms (Logistic Regression, Nearest Neighbor, Decision Parisa Karimi Darabi, Tree, Random Forest, Support Vector Machine, Naive Bayesian, Neural Department of Industrial Network and Gradient Boosting) were analyzed. Were evaluated by accuracy, Engineering, K. N. Toosi University sensitivity, specificity and ROC curve parameters. of Technology, Tehran, Iran Results: The model based on the gradient boosting algorithm showed the best [email protected] performance with a prediction accuracy of %95.50. Conclusion: In the future, this model can be used for diagnosis diabete. The basis of this study is to do more research and develop models such as other learning machine algorithms. Keywords: Prediction, diabetes, machine learning, gradient boosting, ROC curve DOI : 10.29252/jorjanibiomedj.8.3.4 people die every year from diabetes disease. Introduction By 2045, it will reach 629 million (1).
    [Show full text]
  • Scalable Fault-Tolerant Elastic Data Ingestion in Asterixdb
    UC Irvine UC Irvine Electronic Theses and Dissertations Title Scalable Fault-Tolerant Elastic Data Ingestion in AsterixDB Permalink https://escholarship.org/uc/item/9xv3435x Author Grover, Raman Publication Date 2015 License https://creativecommons.org/licenses/by-nd/4.0/ 4.0 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California UNIVERSITY OF CALIFORNIA, IRVINE Scalable Fault-Tolerant Elastic Data Ingestion in AsterixDB DISSERTATION submitted in partial satisfaction of the requirements for the degree of DOCTOR OF PHILOSOPHY in Information and Computer Science by Raman Grover Dissertation Committee: Professor Michael J. Carey, Chair Professor Chen Li Professor Sharad Mehrotra 2015 © 2015 Raman Grover DEDICATION To my wonderful parents... ii TABLE OF CONTENTS Page LIST OF FIGURES vi LIST OF TABLES viii ACKNOWLEDGMENTS ix CURRICULUM VITAE x ABSTRACT OF THE DISSERTATION xi 1 Introduction 1 1.1 Challenges in Data Feed Management . .3 1.1.1 Genericity and Extensibility . .3 1.1.2 Fetch-Once Compute-Many Model . .5 1.1.3 Scalability and Elasticity . .5 1.1.4 Fault Tolerance . .7 1.2 Contributions . .8 1.3 Organization . .9 2 Related Work 10 2.1 Stream Processing Engines . 10 2.2 Data Routing Engines . 11 2.3 Extract Transform Load (ETL) Systems . 12 2.4 Other Systems . 13 2.4.1 Flume . 14 2.4.2 Chukwa . 15 2.4.3 Kafka . 16 2.4.4 Sqoop . 17 2.5 Summary . 17 3 Background and Preliminaries 19 3.1 AsterixDB . 19 3.1.1 AsterixDB Architecture . 20 3.1.2 AsterixDB Data Model . 21 3.1.3 Querying Data .
    [Show full text]
  • Anomaly Detection and Identification Scheme for VM Live Migration in Cloud Infrastructure✩
    Future Generation Computer Systems 56 (2016) 736–745 Contents lists available at ScienceDirect Future Generation Computer Systems journal homepage: www.elsevier.com/locate/fgcs Anomaly detection and identification scheme for VM live migration in cloud infrastructureI Tian Huang a, Yongxin Zhu a,∗, Yafei Wu a, Stéphane Bressan b, Gillian Dobbie c a School of Microelectronics, Shanghai Jiao Tong University, China b School of Computing, National University of Singapore, Singapore c Department of Computer Science, University of Auckland, New Zealand h i g h l i g h t s • We extend Local Outlier Factors to detect anomalies in time series and adapt to smooth environmental changes. • We propose Dimension Reasoning LOF (DR-LOF) that can point out the ``most anomalous'' dimension of the performance profile. DR-LOF provides clues for administrators to pinpoint and clear the anomalies. • We incorporate DR-LOF and Symbolic Aggregate ApproXimation (SAX) measurement as a scheme to detect and identify the anomalies in the progress of VM live migration. • In experiments, we deploy a typical computing application on small scale clusters and evaluate our anomaly detection and identification scheme by imposing two kinds of common anomalies onto VMs. article info a b s t r a c t Article history: Virtual machines (VM) offer simple and practical mechanisms to address many of the manageability Received 3 March 2015 problems of leveraging heterogeneous computing resources. VM live migration is an important feature Received in revised form of virtualization in cloud computing: it allows administrators to transparently tune the performance 11 May 2015 of the computing infrastructure. However, VM live migration may open the door to security threats.
    [Show full text]
  • A Decidable Logic for Tree Data-Structures with Measurements
    A Decidable Logic for Tree Data-Structures with Measurements Xiaokang Qiu and Yanjun Wang Purdue University fxkqiu,[email protected] Abstract. We present Dryaddec, a decidable logic that allows reason- ing about tree data-structures with measurements. This logic supports user-defined recursive measure functions based on Max or Sum, and re- cursive predicates based on these measure functions, such as AVL trees or red-black trees. We prove that the logic's satisfiability is decidable. The crux of the decidability proof is a small model property which al- lows us to reduce the satisfiability of Dryaddec to quantifier-free lin- ear arithmetic theory which can be solved efficiently using SMT solvers. We also show that Dryaddec can encode a variety of verification and synthesis problems, including natural proof verification conditions for functional correctness of recursive tree-manipulating programs, legal- ity conditions for fusing tree traversals, synthesis conditions for con- ditional linear-integer arithmetic functions. We developed the decision procedure and successfully solved 220+ Dryaddec formulae raised from these application scenarios, including verifying functional correctness of programs manipulating AVL trees, red-black trees and treaps, check- ing the fusibility of height-based mutually recursive tree traversals, and counterexample-guided synthesis from linear integer arithmetic specifi- cations. To our knowledge, Dryaddec is the first decidable logic that can solve such a wide variety of problems requiring flexible combination of measure-related, data-related and shape-related properties for trees. 1 Introduction Logical reasoning about tree data-structures has been needed in various applica- tion scenarios such as program verification [32,24,26,42,49,4,14], compiler opti- mization [17,18,9,44] and webpage layout engines [30,31].
    [Show full text]
  • Data Intensive Computing with Linq to Hpc Technical
    DATA INTENSIVE COMPUTING WITH LINQ TO HPC TECHNICAL OVERVIEW Ade Miller, Principal Program Manager, LINQ to HPC Agenda Introduction to LINQ to HPC Using LINQ to HPC Systems Management Integrating with Other Data Technologies The Data Spectrum One extreme is analyzing raw, unstructured data. The analyst does not know exactly what the data contains, nor LINQ to HPC what cube would be justified. The analyst needs to do ad-hoc analyses that may never be run again. Another extreme is analytics targeting a traditional data warehouse. The analyst Parallel Data knows the cube he or she wants to build, Warehouse and the analyst knows the data sources. What kind of Data? Large Data Volume . 100s of TBs to 10s of PBs New Questions & Non-Traditional data Types New Technologies New Insights . Unstructured & Semi structured . How popular is my product? . Weak relational schema . Distributed Parallel Processing . What is the best ad to serve? . Text, Images, Videos, Logs Frameworks . Is this a fraudulent transaction? . Easy to Scale on commodity hardware . MapReduce-style programming models New Data Sources . Sensors & Devices . Traditional applications . Web Servers . Public data Overview MOVING THE DATA TO THE COMPUTE So how does it work? FIRST, STORE THE DATA Data Intensive Computing with HPC Server Windows HPC Server 2008 R2 Service Pack 2 So how does it work? SECOND, TAKE THE PROCESSING TO THE DATA Data Intensive Computing with HPC Server var logentries = from line in logs where !line.StartsWith("#") select new LogEntry(line); var user = from access
    [Show full text]
  • Cloud Computing Paradigms for Pleasingly Parallel Biomedical
    Cloud Computing Paradigms for Pleasingly Parallel Biomedical Applications Thilina Gunarathne, Tak-Lon Wu, Jong Youl Choi, Seung-Hee Bae, Judy Qiu School of Informatics and Computing / Pervasive Technology Institute Indiana University, Bloomington. {tgunarat, taklwu, jychoi, sebae, xqiu}@indiana.edu Abstract Cloud computing offers exciting new approaches for scientific computing that leverages the hardware and software investments on large scale data centers by major commercial players. Loosely coupled problems are very important in many scientific fields and are on the rise with the ongoing move towards data intensive computing. There exist several approaches to leverage clouds & cloud oriented data processing frameworks to perform pleasingly parallel computations. In this paper we present three pleasingly parallel biomedical applications, 1) assembly of genome fragments 2) sequence alignment and similarity search 3) dimension reduction in the analysis of chemical structures, implemented utilizing cloud infrastructure service based utility computing models of Amazon Web Services and Microsoft Windows Azure as well as utilizing MapReduce based data processing frameworks, Apache Hadoop and Microsoft DryadLINQ. We review and compare each of the frameworks and perform a comparative study among them based on performance, efficiency, cost and the usability. Cloud service based utility computing model and the managed parallelism (MapReduce) exhibited comparable performance and efficiencies for the applications we considered. We analyze the variations in cost between the different platform choices (eg: EC2 instance types), highlighting the need to select the appropriate platform based on the nature of the computation. 1. Introduction Scientists are overwhelmed with the increasing amount of data processing needs arising from the storm of data that is flowing through virtually every field of science.
    [Show full text]
  • ERC Document on Open Research Data and Data Management Plans
    Open Research Data and Data Management Plans Information for ERC grantees by the ERC Scientific Council Version 4.0 11 August 2021 This document is regularly updated in order to take into account new developments in this rapidly evolving field. Comments, corrections and suggestions should be sent to the Secretariat of the ERC Scientific Council Working Group on Open Science via the address [email protected]. The table below summarizes the main changes that this document has undergone. HISTORY OF CHANGES Version Publication Main changes Page (in the date relevant version) 1.0 23.02.2018 Initial version 2.0 24.04.2018 Part ‘Open research data and data deposition in the 15-17 Physical Sciences and Engineering domain’ added Minor editorial changes; faulty link corrected 6, 10 Contact address added 2 3.0 23.04.2019 Name of WG updated 2 Added text to the section on ‘Data deposition’ 5 Reference to FAIRsharing moved to the general part 7 from the Life Sciences part and extended Added example of the Austrian Science Fund in the 8 section on ‘Policies of other funding organisations’; updated links related to the German Research Foundation and the Arts and Humanities Research Council; added reference to the Science Europe guide Small changes to the text on ‘Image data’ 9 Added reference to the Ocean Biogeographic 10 Information (OBIS) Reformulation of the text related to Biostudies 11 New text in the section on ‘Metadata’ in the Life 11 Sciences part Added reference to openICPSR 13 Added references to ioChem-BD and ChemSpider
    [Show full text]
  • Universal Health Coverage UNIVERSAL HEALTH COVERAGE
    September 2019 A collection published by The Harvard Global Health Institute and The BMJ Universal health coverage UNIVERSAL HEALTH COVERAGE EDITORIAL 1 Delivering on the promise of universal health coverage Ashish Jha, Fiona Godlee, Kamran Abbasi ANALYSIS 2 Rethinking assumptions about delivery of healthcare: implications for universal health coverage September 2019 EDITORIAL OFFICES Jishnu Das, Liana Woskie, Ruma Rajbhandari, Kamran Abbasi, Ashish Jha The Editor, The BMJ BMA House, Tavistock Square 7 Climate change threatens the achievement of effective universal healthcare London, UK, WC1H 9JR Email: [email protected] Renee N Salas, Ashish K Jha Tel: + 44 (0) 20 7387 4410 Fax: + 44 (0) 20 7383 6418 BMJ - Beijing 13 Achieving universal health coverage for mental disorders A1203 Tian Yuan Gang Center Vikram Patel, Shekhar Saxena East 3rd Ring North Road Chaoyang District 16 Motivating provision of high quality care: it is not all about the money Beijing 100027 China Mylène Lagarde, Luis Huicho, Irene Papanicolas Telephone: +86 (10) 5722 7209 BMJ - Hoboken BMJ Publishing Inc 21 Overcoming distrust to deliver universal health coverage: lessons from Ebola Two Hudson Place Liana R Woskie, Mosoka P Fallah Hoboken, NJ 07030 Tel: 1- 855-458-0579 email [email protected] 26 Global health security and universal health coverage: from a marriage of convenience BMJ - Mumbai 102, Navkar Chamber, A Wing to a strategic, effective partnership Marol, Andheri - Kurla Road Andheri (East) Mumbai 400059 Clare Wenham, Rebecca Katz, Charles Birungi, Lisa Boden,
    [Show full text]
  • In the Eye of the Beholder: to Make Global Health Estimates Useful, Make Them More Socially Robust
    Global Health Action æ CURRENT DEBATE In the eye of the beholder: to make global health estimates useful, make them more socially robust Elizabeth Pisani1* and Maarten Kok2 1The Policy Institute, King’s College London, London, UK; 2Institute for Health Policy and Management, Erasmus University, Rotterdam, Netherlands A plethora of new development goals and funding institutions have greatly increased the demand for internationally comparable health estimates in recent years and have brought important new players into the field of health estimate production. These changes have rekindled debates about the validity and legitimacy of global health estimates. This paper draws on country case studies and personal experience to support our opinion that the production and use of estimates are deeply embedded in specific social, economic, political, and ideational contexts, which vary at different levels of the global health architecture. Broadly, most global health estimates tend to be made far from the localities where the data upon which they are based are collected and where the results of estimation processes must ultimately be used if they are to make a difference to the health of individuals. Internationally standardised indicators are necessary, but they are no substitute for data that meet local needs and that fit with local ideas of what is credible and useful Á in other words, data that are both technically and socially robust for those who make key decisions about health. We suggest that greater engagement of local actors (and local data) in the formulation, communication, and interpretation of health estimates would increase the likelihood that these data will be used by those most able to translate them into health gains for the longer term.
    [Show full text]
  • Open Science Toolbox
    Open Science Toolbox A Collection of Resources to Facilitate Open Science Practices Lutz Heil [email protected] LMU Open Science Center April 14, 2018 [last edit: July 10, 2019] Contents Preface 2 1 Getting Started 3 2 Resources for Researchers 4 2.1 Plan your Study ................................. 4 2.2 Conduct your Study ............................... 6 2.3 Analyse your Data ................................ 7 2.4 Publish your Data, Material, and Paper .................... 10 3 Resources for Teaching 13 4 Key Papers 16 5 Community 18 1 Preface The motivation to engage in a more transparent research and the actual implementation of Open Science practices into the research workflow can be two very distinct things. Work on the motivational side has been successfully done in various contexts (see section Key Papers). Here the focus rather lays on closing the intention-action gap. Providing an overview of handy Open Science tools and resources might facilitate the implementation of these practices and thereby foster Open Science on a practical level rather than on a theoretical. So far, there is a vast body of helpful tools that can be used in order to foster Open Science practices. Without doubt, all of these tools add value to a more transparent research. But for reasons of clarity and to save time which would be consumed by trawling through the web, this toolbox aims at providing only a selection of links to these resources and tools. Our goal is to give a short overview on possibilities of how to enhance your Open Science practices without consuming too much of your time.
    [Show full text]