Generalist Repository Comparison Chart

Total Page:16

File Type:pdf, Size:1020Kb

Generalist Repository Comparison Chart This chart is designed to assist researchers in finding a generalist repository should no domain Generalist Repository repository be available to preserve their research data. Generalist repositories accept data regardless of data type, format, content, or disciplinary focus. For this chart, we included a repository available to all Comparison Chart researchers specific to clinical trials (Vivli) to bring awareness to those in this field. doi: 10.5281/zenodo.3946720 https://fairsharing.org/collection/GeneralRepositoryComparison HARVARD TOPIC DRYAD FIGSHARE MENDELEY DATA OSF VIVLI ZENODO DATAVERSE Brief Description Harvard Dataverse is Open-source, A free, open access, data Mendeley Data is a free OSF is a free and Vivli is an independent, Powering Open Science, a free data repository community-led data repository where users repository specialized open source project non-profit organization built on Open Source. open to all researchers curation, publishing, and can make all outputs of for research data. management tool that that has developed a Built by reserachers for from any discipline, both preservation platform for their research available in Search more than 20+ supports researchers global data-sharing researchers. Run from inside and outside of CC0 publicly available a discoverable, reusable, million datasets indexed throughout their entire and analytics platform. the CERN data centre, the Harvard community, research data and citable manner. from 1000s of data project lifecycle in open Our focus is on sharing whose purpose is long where you can share, Dryad is an independent Users can upload files repositories and collect science best practices. individual participant- term preservation archive, cite, access, and non-profit that works of any type and are and share datasets with level data from for the High Energy explore research data. directly with: able to share diverse the research community completed clinical trials Physics discipline, one research products following the FAIR data to serve the international of the largest scientific • researchers to publish including datasets, principles. research community. datasets in the world datasets utlizing best code, multimedia files, practices for discovery workflows, posters, and reuse presentations, and • publishers to support more. With discoverable the integration of data metadata supporting availability statements FAIR principles, file and data citations into visualizations, and their workflows integrations, researchers • institutions to enable can make their work scalable campus more impactful and support for research move research further data managment best faster. practices at low cost Size limits No byte size limit 300GB/dataset Soft limit of 20GB/file for 10GB per dataset Projects currently have If more than 10GB per 50GB per dataset, per dataset. Harvard free accounts. System not storage limit. There study data, reach out contact us via https:// Dataverse currently sets limit of 5000GB/file. is a 5GB/file upload limit to us zenodo.org/support for a file size limit of 2.5GB. Unlimited storage of for native OSF Storage. higher limits public data but 20GB There is no limit imposed storage for private data by OSF for the amount for free accounts. Email of storage used across [email protected] to add-ons connected to a have upload and storage given project. limits raised. Storage space per 1 TB per researcher No limit No limit No limit No limit No limit No limit researcher Persistent, Unique DOI, Handle DOI DOI DOI DOI DOI DOI Identifier Support HARVARD TOPIC DRYAD FIGSHARE MENDELEY DATA OSF VIVLI ZENODO DATAVERSE Licensing Options By default, datasets CC0 Default licenses Default licenses The following 14 licenses https://vivli.org/ Content is available are published in the supported: supported: are available: resources/vivli-data-use- publicly under any one of public domain (CC0). CC0 1.0 CC0 1.0 No License - is a agreement/ 400 open licences (from Depositors can change opendefinition.org and CC BY 4.0 CC BY 4.0 copyright license for this to apply their own spdx.org). Restricted and MIT CC BY NC 3.0 the project authors and licenses. Closed content is also Apache 2.0 MIT contributors supported. GPL v3 Apache 2.0 CC0 1.0 GPL v2 BSD 3-Clause CC-By 4.0 BSD 2-Clause MIT GPL v3 Apache 2.0 GPL v2 BSD 2-Clause LGPL BSD 3-Clause MPL-2.0 GPL 3.0 CeCILL GPL 2.0 CeCILL-B Artistic 2.0 CERN OHL Eclipse 1.0 TAPR OHL LGPL 3.0 LGPL 2.1 Mozilla 2.0 Other- user defines a license in a .txt file and uploads to the project (not available on registrations or collections) Costs to the Harvard Dataverse is Costs covered by Free Free Free https://vivli.org/about/ Free up to 50GB per researcher free for all researchers institutional, publisher, share-data/ All fees dataset. Larger datasets worldwide (up to 1 TB) and funder members, to share, store, and possible by arrangement otherwise a one-time access COVID-19 clinical fee of $120 for authors research are waived. to cover cost of curation and preservation doi: 10.5281/zenodo.3946720 2 https://fairsharing.org/collection/GeneralRepositoryComparison HARVARD TOPIC DRYAD FIGSHARE MENDELEY DATA OSF VIVLI ZENODO DATAVERSE Equitable free and Use of CC0 and free CC0, public available, Metadata is CC0. All Metadata is licensed Open, public API to Managed Access as Metadata is licensed ongoing access to data access highly broadly indexed files and metadata can CC0. support broad indexing, Human Subject Data. CC0. Content is both data encouraged. Support research data with long be accessed from docs. Datasets are and will partnership with Internet No charge period for online on disk and offline for broad indexing and term preservation in a figshare.com. Figshare continue to be free Archive for long-term data that is accessible on tape as part of long- long-term preservation CoreTrustSeal-certified has also partnered access preservation, $250k only with a research term preservation policy strategies repository with DuraSpace and preservation fund. environment. Costs after Long-term access in Chronopolis to offer no charge time period the event of cease of further assurances ends. https://vivli.org/ operations granted that public data will resources/vivli-secure- by DANS. Access to be archived under research-environment-2/ archived datasets will the stewardship of be provided for free in Chronopolis. In the perpetuity highly unlikely event of multiple AWS S3 failures, Figshare can restore public user content from Chronopolis. Version Support Yes, including version Yes, versioning is Yes Yes, including version Yes for OSF Storage, Coming soon Yes, with "Concept" comparison and W3C supported comparison, to easily see when supported for DOI to represent "all provenance support what has changed from integrated storage versions" version to version providers Characteristics - Standards Supported Metadata Dublin Core, Data DataCite, schema.org Dublin Core (oai_dc), Dublin Core, DataCite, Datacite, Crossref DataCite DataCite Schemas Documentation Initiative Datacite (oai_datacite), Schema.org preprint (DDI Codebook), RDF (rdf), CERIF DataCite, OpenAIRE, XML (cerif), Qualified Schema.org, OAI_ORE Dublin Core (qdc) (hasPart support), Metadata Encoding and Transmission Standard (mets) and UKETD_DC (uketd_dc) Formats Supported JSON, XML Fully documented API Dublin Core (oai_dc), JSON, XML JSON, API, JSONAPI DataCite, Dublin for Export available for direct Datacite (oai_datacite), Core, DCAT-AP, JSON, integration. Exports RDF (rdf), CERIF JSON-LD, GeoJSON, available in JSON, XML, XML (cerif), Qualified MARCXML, Citation schema.org Dublin Core (qdc) Style Language JSON (hasPart support), + support for custom Metadata Encoding and metadata formats Transmission Standard (mets) and UKETD_DC (uketd_dc) doi: 10.5281/zenodo.3946720 3 https://fairsharing.org/collection/GeneralRepositoryComparison HARVARD TOPIC DRYAD FIGSHARE MENDELEY DATA OSF VIVLI ZENODO DATAVERSE Supported ISO 3166-1 CV for ORCID - Authors ORCID - Authors, Omniscience taxonomy BePress 3 tier taxonomy Cochrane Linked Data ORCID - Authors, Community geospatial metadata, ISO ROR - Institutions Dimensions - Funding, for subject categories for preprints and custom Vocabulary FundRef + OpenAIRE Vocabulary or 639-1 CV for languages, Field of Research codes, and keywords, and taxonomy mapped to Projects Database - Open Funder Registry - Taxonomy DataCite’s dataset GRID ids, Implementing additional custom Bepress on preprints Funding Funding contributor vocab, and ROR and Make Data vocabularies can be subsets of the OBI PLOS Thesaurus - Count loaded on institutional Ontology and NCBI Keywords version Taxonomy for Organisms Characteristics Useful for Linking data to other relevant digital information Support for creators/ ORCID, ISNI, LCNA, VIAF, ORCID required for ORCID ORCID - Mendeley ID - ORCID, ResearcherID ORCID ORCID authors identifiers GND, DAI, ResearcherID corresponding author, Scopus Author ID and ScopusID co-author ORCID supported Support linking to Yes DataCite relation types DataCite relation types Yes, DataCite and Scholix Preprints link to Yes DataCite relation types related publications relation types related Peer-reviewed publications using Crossef 'isPrerint of' Linking of derived Yes DataCite relation types DataCite relation types Yes, DataCite and Scholix No DataCite relation types products from relation types another Grant ID(s) Yes Yes Yes, hard-linked to Yes, customizable on No Yes Dimensions.ai database institutional version Grant ID Yes Open Funder Registry Yes, hard-linked to Yes,
Recommended publications
  • Recursive Definitions Across the Footprint That a Straight- Line Program Manipulated, E.G
    natural proofs for verification of dynamic heaps P. MADHUSUDAN UNIV. OF ILLINOIS AT URBANA-CHAMPAIGN (VISITING MSR, BANGALORE) WORKSHOP ON MAKING FORMAL VERIFICATION SCALABLE AND USEABLE CMI, CHENNAI, INDIA JOINT WORK WITH… Xiaokang Qiu .. graduating! Gennaro Andrei Pranav Garg Parlato Stefanescu GOAL: BUILD RELIABLE SOFTWARE Build systems with proven reliability and security guarantees. (Not the same as finding bugs!) Deductive verification with automatic theorem proving • Floyd/Hoare style verification • User supplied modular annotations (pre/post cond, class invariants) and loop invariants • Resulting verification conditions are derived automatically for each linear program segment (using weakest-pre/strongest-post) Verification conditions are then proved valid using mostly automatic theorem proving (like SMT solvers) TARGET: RELIABLE SYS TEMS SOFTWARE Operating Systems Angry birds Browsers Excel VMs Mail clients … App platforms Systems Software Applications Several success stories: • Microsoft Hypervisor verification using VCC [MSR, EMIC] • A secure foundation for mobile apps [@Illinois, ASPLOS’13] A SECURE FOUNDATION FOR MOBILE APPS In collaboration with King’s systems group at Illinois First OS architecture that provides verifiable, high-level abstractions for building mobile applications. Application state is private by default. • Meta-data on files controls access by apps; only app with appropriate permissions can access files • Every page is encrypted before sending to the storage service. • Memory isolation between apps • A page fault handler serves a file- backed page for a process, the fille has to be opened by the same process. • Only the current app can write to the screen buffer. • IPC channel correctness • … VERIFICATION TOOLS ARE MATURE And works! ENOUGH TO BUILD RELIABLE S/W.
    [Show full text]
  • Type 2 Diabetes Prediction Using Machine Learning Algorithms
    Type 2 Diabetes Prediction Using Machine Learning Algorithms Parisa Karimi Darabi*, Mohammad Jafar Tarokh Department of Industrial Engineering, K. N. Toosi University of Technology, Tehran, Iran Abstract Article Type: Background and Objective: Currently, diabetes is one of the leading causes of Original article death in the world. According to several factors diagnosis of this disease is Article History: complex and prone to human error. This study aimed to analyze the risk of Received: 15 Jul 2020 having diabetes based on laboratory information, life style and, family history Revised: 1 Aug 2020 with the help of machine learning algorithms. When the model is trained Accepted: 5 Aug 2020 properly, people can examine their risk of having diabetes. *Correspondence: Material and Methods: To classify patients, by using Python, eight different machine learning algorithms (Logistic Regression, Nearest Neighbor, Decision Parisa Karimi Darabi, Tree, Random Forest, Support Vector Machine, Naive Bayesian, Neural Department of Industrial Network and Gradient Boosting) were analyzed. Were evaluated by accuracy, Engineering, K. N. Toosi University sensitivity, specificity and ROC curve parameters. of Technology, Tehran, Iran Results: The model based on the gradient boosting algorithm showed the best [email protected] performance with a prediction accuracy of %95.50. Conclusion: In the future, this model can be used for diagnosis diabete. The basis of this study is to do more research and develop models such as other learning machine algorithms. Keywords: Prediction, diabetes, machine learning, gradient boosting, ROC curve DOI : 10.29252/jorjanibiomedj.8.3.4 people die every year from diabetes disease. Introduction By 2045, it will reach 629 million (1).
    [Show full text]
  • Scientometrics1
    Scientometrics1 Loet Leydesdorff a and Staša Milojević b a Amsterdam School of Communication Research (ASCoR), University of Amsterdam, Kloveniersburgwal 48, 1012 CX Amsterdam, The Netherlands; [email protected] b School of Informatics and Computing, Indiana University, Bloomington 47405-1901, United States; [email protected]. Abstract The paper provides an overview of the field of scientometrics, that is: the study of science, technology, and innovation from a quantitative perspective. We cover major historical milestones in the development of this specialism from the 1960s to today and discuss its relationship with the sociology of scientific knowledge, the library and information sciences, and science policy issues such as indicator development. The disciplinary organization of scientometrics is analyzed both conceptually and empirically. A state-of-the-art review of five major research threads is provided. Keywords: scientometrics, bibliometrics, citation, indicator, impact, library, science policy, research management, sociology of science, science studies, mapping, visualization Cross References: Communication: Electronic Networks and Publications; History of Science; Libraries; Networks, Social; Merton, Robert K.; Peer Review and Quality Control; Science and Technology, Social Study of: Computers and Information Technology; Science and Technology Studies: Experts and Expertise; Social network algorithms and software; Statistical Models for Social Networks, Overview; 1 Forthcoming in: Micheal Lynch (Editor), International
    [Show full text]
  • Making Institutional Repositories Work “Making Institutional Repositories Work Sums It up Very Well
    Making Institutional Repositories Work “Making Institutional Repositories Work sums it up very well. This book, the first of its kind, explains how IRs work and how to get the greatest re- sults from them. As many of us know, numerous IRs launched with high hopes have in fact languished with lackluster results. Faculty have little in- terest, and administrators see little promise. But the many chapter authors of this very well edited book have made their IRs successful, and here they share their techniques and successes. This is a necessary book for anyone contemplating starting an IR or looking to resurrect a moribund one.” — Richard W. Clement Dean, College of University Libraries & Learning Sciences University of New Mexico “This volume presents an interesting cross-section of approaches to in- stitutional repositories in the United States. Just about every view and its opposite makes an appearance. Readers will be able to draw their own con- clusions, depending on what they see as the primary purpose of IRs.” — Stevan Harnad Professor, University of Québec at Montréal & University of Southampton “Approaching this volume as one of ‘those of us who have been furiously working to cultivate thriving repositories,’ I am very excited about what this text represents. It is a broad compilation featuring the best and brightest writing on all the topics I’ve struggled to understand around re- positories, and it also marks a point when repository management and de- velopment is looking more and more like a core piece of research library work. Callicott, Scherer, and Wesolek have pulled together all the things I wished I’d been able to read in my first year as a scholarly communication librarian.
    [Show full text]
  • Will Sci-Hub Kill the Open Access Citation Advantage and (At Least for Now) Save Toll Access Journals?
    Will Sci-Hub Kill the Open Access Citation Advantage and (at least for now) Save Toll Access Journals? David W. Lewis October 2016 © 2016 David W. Lewis. This work is licensed under a Creative Commons Attribution 4.0 International license. Introduction It is a generally accepted fact that open access journal articles enjoy a citation advantage.1 This citation advantage results from the fact that open access journal articles are available to everyone in the word with an Internet collection. Thus, anyone with an interest in the work can find it and use it easily with no out-of-pocket cost. This use leads to citations. Articles in toll access journals on the other hand, are locked behind paywalls and are only available to those associated with institutions who can afford the subscription costs, or who are willing and able to purchase individual articles for $30 or more. There has always been some slippage in the toll access journal system because of informal sharing of articles. Authors will usually send copies of their work to those who ask and sometime post them on their websites even when this is not allowable under publisher’s agreements. Stevan Harnad and his colleagues proposed making this type of author sharing a standard semi-automated feature for closed articles in institutional repositories.2 The hashtag #ICanHazPDF can be used to broadcast a request for an article that an individual does not have access to.3 Increasingly, toll access articles are required by funder mandates to be made publically available, though usually after an embargo period.
    [Show full text]
  • Mapping Scientometrics (1981-2001)
    Mapping Scientometrics (1981 -2001) Chaomei Chen College of Information Science and Technology, Drexel University, Philadelphia, PA 19104, USA. Email: [email protected] Katherine McCain College of Information Science and Technology, Drexel University, Philadelphia, PA 19104, USA. Email: kate .mccai n@cis .drexel . ed u Howard White College of Information Science and Technology, Drexel University, Philadelphia, PA 19104, USA. Email: [email protected] Xia Lin College of Information Science and Technology, Drexel University, Philadelphia, PA 19104, USA. Email: xia .I in@cis . d rexel. edu We investigate an integrated approach to co-citation analysis could lead to a clearer picture of the scientometric studies with emphasis to the use of cognitive content of publications (Braam, Moed, & Raan, information visualization and animation 1991a, 1991b). techniques. This study draws upon citation and In Little Science, Big Science, Derek de Solla Price co-citation patterns derived from articles (1963) raised some of the most fundamental questions that published in the journal Scientornetrics (1981- have led to the scientometric study today: Why should we 2001). The modeling and visualization takes an not turn the tools of science on science itself? Why not evolutionary and historical perspective. The measure and generalize, make hypotheses, and derive design of the visualization model adapts a virtual conclusions? He used the metaphor of studying the landscape metaphor with document cocitation behavior of gas in thermodynamics as an analogue of the networks as the base map and annual citation science of science. Thermodynamics studies the behavior of rates as the thematic overlay. The growth of gas under various conditions of temperature and pressure, citation rates is presented through an animation but the focus is not on the trajectory of a specific molecule.
    [Show full text]
  • Scalable Fault-Tolerant Elastic Data Ingestion in Asterixdb
    UC Irvine UC Irvine Electronic Theses and Dissertations Title Scalable Fault-Tolerant Elastic Data Ingestion in AsterixDB Permalink https://escholarship.org/uc/item/9xv3435x Author Grover, Raman Publication Date 2015 License https://creativecommons.org/licenses/by-nd/4.0/ 4.0 Peer reviewed|Thesis/dissertation eScholarship.org Powered by the California Digital Library University of California UNIVERSITY OF CALIFORNIA, IRVINE Scalable Fault-Tolerant Elastic Data Ingestion in AsterixDB DISSERTATION submitted in partial satisfaction of the requirements for the degree of DOCTOR OF PHILOSOPHY in Information and Computer Science by Raman Grover Dissertation Committee: Professor Michael J. Carey, Chair Professor Chen Li Professor Sharad Mehrotra 2015 © 2015 Raman Grover DEDICATION To my wonderful parents... ii TABLE OF CONTENTS Page LIST OF FIGURES vi LIST OF TABLES viii ACKNOWLEDGMENTS ix CURRICULUM VITAE x ABSTRACT OF THE DISSERTATION xi 1 Introduction 1 1.1 Challenges in Data Feed Management . .3 1.1.1 Genericity and Extensibility . .3 1.1.2 Fetch-Once Compute-Many Model . .5 1.1.3 Scalability and Elasticity . .5 1.1.4 Fault Tolerance . .7 1.2 Contributions . .8 1.3 Organization . .9 2 Related Work 10 2.1 Stream Processing Engines . 10 2.2 Data Routing Engines . 11 2.3 Extract Transform Load (ETL) Systems . 12 2.4 Other Systems . 13 2.4.1 Flume . 14 2.4.2 Chukwa . 15 2.4.3 Kafka . 16 2.4.4 Sqoop . 17 2.5 Summary . 17 3 Background and Preliminaries 19 3.1 AsterixDB . 19 3.1.1 AsterixDB Architecture . 20 3.1.2 AsterixDB Data Model . 21 3.1.3 Querying Data .
    [Show full text]
  • Anomaly Detection and Identification Scheme for VM Live Migration in Cloud Infrastructure✩
    Future Generation Computer Systems 56 (2016) 736–745 Contents lists available at ScienceDirect Future Generation Computer Systems journal homepage: www.elsevier.com/locate/fgcs Anomaly detection and identification scheme for VM live migration in cloud infrastructureI Tian Huang a, Yongxin Zhu a,∗, Yafei Wu a, Stéphane Bressan b, Gillian Dobbie c a School of Microelectronics, Shanghai Jiao Tong University, China b School of Computing, National University of Singapore, Singapore c Department of Computer Science, University of Auckland, New Zealand h i g h l i g h t s • We extend Local Outlier Factors to detect anomalies in time series and adapt to smooth environmental changes. • We propose Dimension Reasoning LOF (DR-LOF) that can point out the ``most anomalous'' dimension of the performance profile. DR-LOF provides clues for administrators to pinpoint and clear the anomalies. • We incorporate DR-LOF and Symbolic Aggregate ApproXimation (SAX) measurement as a scheme to detect and identify the anomalies in the progress of VM live migration. • In experiments, we deploy a typical computing application on small scale clusters and evaluate our anomaly detection and identification scheme by imposing two kinds of common anomalies onto VMs. article info a b s t r a c t Article history: Virtual machines (VM) offer simple and practical mechanisms to address many of the manageability Received 3 March 2015 problems of leveraging heterogeneous computing resources. VM live migration is an important feature Received in revised form of virtualization in cloud computing: it allows administrators to transparently tune the performance 11 May 2015 of the computing infrastructure. However, VM live migration may open the door to security threats.
    [Show full text]
  • Piracy of Scientific Papers in Latin America: an Analysis of Sci-Hub Usage Data
    Developing Latin America Piracy of scientific papers in Latin America: An analysis of Sci-Hub usage data Juan D. Machin-Mastromatteo Alejandro Uribe-Tirado Maria E. Romero-Ortiz This article was originally published as: Machin-Mastromatteo, J.D., Uribe-Tirado, A., and Romero-Ortiz, M. E. (2016). Piracy of scientific papers in Latin America: An analysis of Sci-Hub usage data. Information Development, 32(5), 1806–1814. http://dx.doi.org/10.1177/0266666916671080 Abstract Sci-Hub hosts pirated copies of 51 million scientific papers from commercial publishers. This article presents the site’s characteristics, it criticizes that it might be perceived as a de-facto component of the Open Access movement, it replicates an analysis published in Science using its available usage data, but limiting it to Latin America, and presents implications caused by this site for information professionals, universities and libraries. Keywords: Sci-Hub, piracy, open access, scientific articles, academic databases, serials crisis Scientific articles are vital for students, professors and researchers in universities, research centers and other knowledge institutions worldwide. When academic publishing started, academies, institutions and professional associations gathered articles, assessed their quality, collected them in journals, printed and distributed its copies; with the added difficulty of not having digital technologies. Producing journals became unsustainable for some professional societies, so commercial scientific publishers started appearing and assumed printing, sales and distribution on their behalf, while academics retained the intellectual tasks. Elsevier, among the first publishers, emerged to cover operations costs and profit from sales, now it is part of an industry that grew from the process of scientific communication; a 10 billion US dollar business (Murphy, 2016).
    [Show full text]
  • Open Access Initiatives and Networking in the Global South Iryna Kuchma
    Open Access Initiatives and Networking in the Global South Iryna Kuchma This short study highlights the impact of open access in the Global South. Featuring collaborative open access initiatives in Algeria, Kenya, Myanmar, Nigeria, Nepal, Palestine, Tanzania, Uganda and Latin American countries, it showcases success and describes the challenges that we still face. It also questions a notion of a journal article – perhaps already becoming obsolete – and discusses the growing preprints initiatives to speed up the availability of research results. The value of regional journal and repository networks enhancing open access content in Europe and Latin America is also discussed as well as the impact human networks make in the Global South. Health research dissemination Dr. Bessie Mukami is a general physician at Embu General Provincial Hospital, a relatively large teaching hospital in Embu, a town located approximately 120 kilometres northeast of Nairobi towards Mount Kenya. Embu serves as the provincial headquarters of Eastern Province in Kenya and is also the county headquarters of Embu County. “You have, maybe, one doctor to ten thousand people”, says Dr. Mukami. And as she speaks, her fingers click through pages of open access medical journals on a laptop. Subscribing to medical journals is very expensive, and it can be difficult for doctors to consult each other because of the long distances between hospitals. Open access is solving one of the biggest problems Dr. Mukami has: “Instead of calling other doctors for information, the information is open and available, and you search for what you really want”, she says. Dr. Gerald Nderitu, the medical superintendent at the same hospital and an experienced surgeon of 16 years, also relies on open access research output to help his patients.
    [Show full text]
  • A Decidable Logic for Tree Data-Structures with Measurements
    A Decidable Logic for Tree Data-Structures with Measurements Xiaokang Qiu and Yanjun Wang Purdue University fxkqiu,[email protected] Abstract. We present Dryaddec, a decidable logic that allows reason- ing about tree data-structures with measurements. This logic supports user-defined recursive measure functions based on Max or Sum, and re- cursive predicates based on these measure functions, such as AVL trees or red-black trees. We prove that the logic's satisfiability is decidable. The crux of the decidability proof is a small model property which al- lows us to reduce the satisfiability of Dryaddec to quantifier-free lin- ear arithmetic theory which can be solved efficiently using SMT solvers. We also show that Dryaddec can encode a variety of verification and synthesis problems, including natural proof verification conditions for functional correctness of recursive tree-manipulating programs, legal- ity conditions for fusing tree traversals, synthesis conditions for con- ditional linear-integer arithmetic functions. We developed the decision procedure and successfully solved 220+ Dryaddec formulae raised from these application scenarios, including verifying functional correctness of programs manipulating AVL trees, red-black trees and treaps, check- ing the fusibility of height-based mutually recursive tree traversals, and counterexample-guided synthesis from linear integer arithmetic specifi- cations. To our knowledge, Dryaddec is the first decidable logic that can solve such a wide variety of problems requiring flexible combination of measure-related, data-related and shape-related properties for trees. 1 Introduction Logical reasoning about tree data-structures has been needed in various applica- tion scenarios such as program verification [32,24,26,42,49,4,14], compiler opti- mization [17,18,9,44] and webpage layout engines [30,31].
    [Show full text]
  • Open Research Data: SNSF Monitoring Report 2017-2018
    Open Research Data: SNSF monitoring report 2017-2018 Katrin Milzow, Martin von Arx, Corn´eliaSommer, Julia Cahenzli, and Lionel Periniy February 2020 Abstract The Swiss National Science Foundation (SNSF) introduced its policy on Open Research Data (ORD) in 2017. This monitoring report gives an overview of the first experiences and provides an analysis of the adherence of researchers to the ORD policy. Data collected in this report come from the applications funded in the project funding scheme, some career funding schemes, and the Sin- ergia programme between October 2017 and December 2018. The report provides an analysis of the plausibility of the data management plans (DMPs), the requested costs for granting access to research data and the characteristics of the mentioned data repositories. Keywords: Open Research Data, Data Management Plan, Data sharing, Data repositories, Science Policy. y Open Research Data Group, Swiss National Science Foundation (SNSF). E-mail address: [email protected] Contents 1 Introduction 2 2 ORD Policy and Processes 2 2.1 ORD Policy . .2 2.1.1 Core Principles . .2 2.1.2 FAIR Data Principles . .3 2.2 Processes . .3 2.2.1 Costs for Granting Access to Research Data . .3 2.2.2 Assessment of the DMP . .4 3 Data 4 4 Descriptive Analysis 4 4.1 Costs for Granting Access to Research Data . .4 4.2 Overall Assessment of the DMPs . .5 4.3 Detailed Assessment of the DMPs . .5 4.4 Data Repositories . .7 5 Conclusion 10 6 Challenges and Perspectives 11 7 References 12 1 SNSF monitoring report Open Research Data 1 Introduction The Swiss National Science Foundation (SNSF) introduced its policy on Open Research Data (ORD) in 2017.
    [Show full text]