Policy-Making for Research Data in Repositories: a Guide
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Data and Services Data and Services
Data and services Data and services The UK Data Service provides researchers with flexible research support, training and access to the UK’s largest collection of social, economic and population data. We are a critical part of the UK’s research infrastructure, established to facilitate high quality, impactful research and education across all sectors. We enable researchers, teachers and policy makers to get an in-depth understanding of social and economic issues – and greater insights into tackling social challenges, now and in the future. A partnership of five universities and Jisc, we are funded by the Economic and Social Research Council (ESRC) and the collections of national data resources we provide access to have been built up over 50 years. Our holdings include major UK government-sponsored surveys, longitudinal studies, UK census data, international macrodata, business microdata, cross-national surveys, qualitative data and administrative data. Services that meet Data Assessment your needs We offer significant expertise across ukdataservice.ac.uk the data publishing life cycle, from data appraisal, quality and disclosure risk Teaching assessment, through to preparation The use of real data in coursework and documentation. adds authenticity and relevance to teaching resources and gives students We work closely with data owners to the data analysis skills necessary help triage and share data with the to make significant contributions appropriate governance measures to society. in place. Research We offer data cleaning, rescue and We provide a Trusted Digital Repository data 'archaeology' services, which we for social and economic researchers’ undertake alongside our core data data through our lead partner, the UK preparation activities. -
A Data Management Life-Cycle
science for a changing world Central Energy Resources Team Data Management A Data Management Life-Cycle Introduction Documented, reliable, and accessible data and information are essential building blocks supporting scientific research and applications that enhance society's knowledge base (fig. 1). The U.S. Geological Survey (USGS), a leading provider of science data, information, and knowledge, is uniquely positioned to integrate science and natural resource information to address societal needs. The USGS Central Energy Resources Team (USGS-CERT) provides critical information and knowledge on tnu iuu the quantity, quality, and distribution of the Nation's and the 10101010110 world's oil, gas, and coal resources. 1C001111001 By using a life-cycle model, the USGS-CERT Data Manage 1001001"*: ment Project is developing an integrated data management system 10^19*1- to (1) promote access to energy data and information, (2) increase data documentation, and (3) streamline product delivery to the public, scientists, and decision makers. The project incorporates web-based technology, data cataloging systems, data processing Data Information routines, and metadata documentation tools to improve data Figure 1. Access to organized and documented data and access, enhance data consistency, and increase office efficiency. information increases knowledge. Implementation Summary and Future Directions In a life-cycle model, data and information are transformed into tangible products and knowledge by a continuous flow in which The Data Management Project addresses USGS science the output of one process becomes the input of others (fig. 2). \ goals and affects how the Central Energy Resources Team manages data and carries out its programs. Our task-oriented "Find, Get, Use, Deliver, and Maintain" The tools and processes developed are based on strategy incorporates life-cycle concepts and directs USGS-CERT the life-cycle approach and our "Find, Get, Use, data management tasks and implementation. -
Collections Management Plan for the U.S. Geological Survey Woods Hole Coastal and Marine Science Center Data Library
7 Collections Management Plan for the U.S. Geological Survey Woods Hole Coastal and Marine Science Center Data Library Open-File Report 2015–1141 U.S. Department of the Interior U.S. Geological Survey Collections Management Plan for the U.S. Geological Survey Woods Hole Coastal and Marine Science Center Data Library By Kelleen M. List, Brian J. Buczkowski, Linda P. McCarthy, and Alice M. Orton Open-File Report 2015–1141 U.S. Department of the Interior U.S. Geological Survey U.S. Department of the Interior SALLY JEWELL, Secretary U.S. Geological Survey Suzette M. Kimball, Acting Director U.S. Geological Survey, Reston, Virginia: 2015 For more information on the USGS—the Federal source for science about the Earth, its natural and living resources, natural hazards, and the environment—visit http://www.usgs.gov/ or call 1–888–ASK–USGS (1–888–275–8747). For an overview of USGS information products, including maps, imagery, and publications, visit http://www.usgs.gov/pubprod/. Any use of trade, firm, or product names is for descriptive purposes only and does not imply endorsement by the U.S. Government. Although this information product, for the most part, is in the public domain, it also may contain copyrighted materials as noted in the text. Permission to reproduce copyrighted items must be secured from the copyright owner. Suggested citation: List, K.M., Buczkowski, B.J., McCarthy, L.P., and Orton, A.M., 2015, Collections management plan for the U.S. Geological Survey Woods Hole Coastal and Marine Science Center Data Library: U.S. Geological Survey Open- File Report 2015–1141, 16 p., http://dx.doi.org/10.3133/ofr20151141. -
Relevance of Data Mining in Digital Library
International Journal of Future Computer and Communication, Vol. 2, No. 1, February 2013 Relevance of Data Mining in Digital Library R. N. Mishra and Aishwarya Mishra Abstract—Data mining involves significant process of Data Mining concept has been delimitated in multiple identifying the extraction of hidden predictive information ways by different organizations, scientists etc. Wikipedia, from vast array of databases and it is an authoritative new has visualized data mining a non-trivial extraction of technology with potentiality to facilitate the Libraries and implicit and potentially useful information from data and Information Centers to focus on the most important the science of extracting useful information from large data information in their data warehouses. It is a viable tool to predict future trends and behaviors in the field of library and sets or databases (http://en.wikipedia. org/wiki/Data information service for deducing proactive, knowledge-driven mining). Marketing Dictionary defines data mining as a decisions. Mechanized, prospective analyses of data mining process for extraction of customer information from a vast move beyond the analyses of past events provided by gamut of databases with the help of the feasible software to retrospective tools typical of decision support systems. Data isolate and identify previously unknown patterns or trends. Mining, Process of Data Mining, Knowledge Discovery in Multiple techniques with the help of technologies are Databases, DBMS, Data Mining Techniques etc. etc. have been discussed in this paper. employed in data mining for extraction of data from the heave of databases. Intelligence Encyclopedia has, however, Index Terms—Data mining, KDD, artificial neural Networks, defined data mining as a statistical analysis technique to sequential pattern, modeling, DMT. -
What Is an Institutional Repository to Do? Implementing Open Access Harvesting Workflows
publications Article What Is an Institutional Repository to Do? Implementing Open Access Harvesting Workflows Rachel Smart Office of Digital Research and Scholarship, Florida State University, Tallahassee, FL 32306, USA; [email protected] Received: 16 March 2019; Accepted: 23 May 2019; Published: 27 May 2019 Abstract: In 2016, Florida State University adopted an institutional Open Access policy, and the library staff were tasked with implementing an outreach plan to contact authors and collect publication post-prints. In 2018, I presented at Open Repositories in Bozeman to share our workflow, methods, and results with the repository community. This workflow utilizes both restricted and open source methods of obtaining and creating research metadata and reaching out to authors to make their work more easily accessible and citable. Currently, post-print deposits added using this workflow are still in the double digits for each year since 2016. Like many institutions before us, participation rates of article deposit in the institutional repository are low and it may be too early in the implementation of this workflow to expect a real change in faculty participation. Keywords: open access policy; metadata; author outreach; web of science; open source; repositories 1. Introduction This content recruitment workflow is a component of a larger Open Access (OA) implementation plan proposed to the Florida State University (FSU) Faculty Senate by University Libraries staff following the adoption of an institutional OA policy in 2016. The challenge and goal of this plan is scaling repository operations by gathering author content and encouraging faculty to submit their manuscripts without the robust technical services of a commercial solution, such as Symplectic Elements. -
Making Institutional Repositories Work “Making Institutional Repositories Work Sums It up Very Well
Making Institutional Repositories Work “Making Institutional Repositories Work sums it up very well. This book, the first of its kind, explains how IRs work and how to get the greatest re- sults from them. As many of us know, numerous IRs launched with high hopes have in fact languished with lackluster results. Faculty have little in- terest, and administrators see little promise. But the many chapter authors of this very well edited book have made their IRs successful, and here they share their techniques and successes. This is a necessary book for anyone contemplating starting an IR or looking to resurrect a moribund one.” — Richard W. Clement Dean, College of University Libraries & Learning Sciences University of New Mexico “This volume presents an interesting cross-section of approaches to in- stitutional repositories in the United States. Just about every view and its opposite makes an appearance. Readers will be able to draw their own con- clusions, depending on what they see as the primary purpose of IRs.” — Stevan Harnad Professor, University of Québec at Montréal & University of Southampton “Approaching this volume as one of ‘those of us who have been furiously working to cultivate thriving repositories,’ I am very excited about what this text represents. It is a broad compilation featuring the best and brightest writing on all the topics I’ve struggled to understand around re- positories, and it also marks a point when repository management and de- velopment is looking more and more like a core piece of research library work. Callicott, Scherer, and Wesolek have pulled together all the things I wished I’d been able to read in my first year as a scholarly communication librarian. -
The Application of File Identification, Validation, and Characterization Tools in Digital Curation
THE APPLICATION OF FILE IDENTIFICATION, VALIDATION, AND CHARACTERIZATION TOOLS IN DIGITAL CURATION BY KEVIN MICHAEL FORD THESIS Submitted in partial fulfillment of the requirements for the degree of Master of Science in Library and Information Science in the Graduate College of the University of Illinois at Urbana-Champaign, 2011 Urbana, Illinois Advisers: Research Assistant Professor Melissa Cragin Assistant Professor Jerome McDonough ABSTRACT File format identification, characterization, and validation are considered essential processes for digital preservation and, by extension, long-term data curation. These actions are performed on data objects by humans or computers, in an attempt to identify the type of a given file, derive characterizing information that is specific to the file, and validate that the given file conforms to its type specification. The present research reviews the literature surrounding these digital preservation activities, including their theoretical basis and the publications that accompanied the formal release of tools and services designed in response to their theoretical foundation. It also reports the results from extensive tests designed to evaluate the coverage of some of the software tools developed to perform file format identification, characterization, and validation actions. Tests of these tools demonstrate that more work is needed – particularly in terms of scalable solutions – to address the expanse of digital data to be preserved and curated. The breadth of file types these tools are anticipated to handle is so great as to call into question whether a scalable solution is feasible, and, more broadly, whether such efforts will offer a meaningful return on investment. Also, these tools, which serve to provide a type of baseline reading of a file in a repository, can be easily tricked. -
IRUS-UK: Standardised Institutional Repository Usage Statistics
IRUS-UK: standardised institutional repository usage statistics Paul Needham, IRUS-UK and COUNTER Two Approaches for Accurately Counting IR Usage: RAMP and IRUS-UK CARL and COAR Webinar Tuesday, October 3, 1:00-2:00 p.m. ET (19:00-20:00 CEST) Who is responsible for IRUS-UK? • Funded by Jisc • Team Members: – Jisc – Service Management & Host – Cranfield University – Service Development & Maintenance – Evidence Base, Birmingham City University – User Engagement & Evaluation Bringing together key repository services to deliver a connected national infrastructure to support OA What is IRUS-UK? A national aggregation service for UK Institutional Repository Usage Statistics: • Collects raw download data from UK IRs for *all item types* within repositories • Processes raw data into COUNTER-conformant statistics - so produced on the same basis as ‘traditional’ scholarly publishers About COUNTER • Counting Online Usage of Networked Electronic Resources – an international initiative serving librarians, publishers and intermediaries – setting standards that facilitate the recording and reporting of online usage statistics – consistent, credible and comparable • COUNTER Codes of Practice: – Release 4 of the COUNTER Code of Practice for e-Resources • Books, databases , journals and multimedia content – Release 1 of the COUNTER Code of Practice for Articles • Journal articles • The Codes specify: – How raw data should be processed into statistics – How statistics reports should be formatted and delivered The Tracker Protocol • To enable repositories -
Eprints Institutional Repository Software: a Review
Partnership: the Canadian Journal of Library and Information Practice and Research, vol. 5, no. 2 (2010) Eprints Institutional Repository Software: A Review Mike Beazley Academic librarian Vaughan Memorial Library, Acadia University [email protected] Keywords Institutional Repository, Eprints, Review, Software, Open-Source Abstract Setting up an institutional repository (IR) can be a daunting task. There are many software packages out there, some commercial, some open source, all of which offer different features and functionality. This article will provide some thoughts about one of these software packages: Eprints. Eprints is open-source, and the software is easy to modify. This presents clear advantages for institutions will smaller budgets and that have programmers on staff. Installation and initial configuration are straightforward and once the IR is up and running, users can easily upload documents by filling out a simple web form. Eprints is an excellent choice for any institution looking to get an IR up and running quickly and easily, although it is less clear that an institution with an existing IR based on another software package should migrate to Eprints. Introduction Setting up an institutional repository (IR) can be a daunting task. There are many software packages out there, some commercial, some open source, all of which offer different features and functionality. This article will provide some thoughts about one of these software packages: Eprints. Eprints was one of the first IR software packages to appear and has been available for 10 years. It is under continual development by its creators at the University of Southampton and the current version is v3.2.3. -
Using Numeric Datasets in Learning and Teaching Final Report, 2002
An enquiry into the use of numeric data in learning & teaching S e p t e m b e r, 2001 Robin Rice (Project Manager) Edinburgh UniversityData Library Peter Burnhill (Project Director) Edinburgh UniversityData Library, EDINA Melanie Wright (Project Te a m ) The UK Data Archive Sean Townsend (Project Te a m ) British Library of Political and Economic Science Part 1: Report and recommendations C o n t e n t s Executive summary 3 Introduction and participants 4 Aims and objectives of the project 5 Nature of the enquiry (or why it matters) 5 Methods used in the enquiry 6 Discussion of results 7 Recommendations and conclusions 17 References 20 Appendix A: Related developments 21 Appendix B: Case studies of current practice 23 Part 2: Teachers’ survey results 4 1 An enquiry into the use of numeric data in learning & teaching page 3 Executive summary Within UK higher education the renewed attention to learning and teaching is an impetus for change. Advances in information technology create new space for learning beyond the traditional classroom lecture format. New initiatives are creating networked teaching materials for shared use across institutions. But little is known about the readiness of teachers and students to take advantage of these resources for teaching and study. Are universities providing the support needed for using these networked resources in classrooms, computer labs, and independent study? An academic Task Force on the use of numeric data in learning and teaching has issued a report on the barriers faced by teachers and students to using national data services across a number of disciplines, including but not limited to the social sciences. -
Skills, Not Just Diplomas
Restoring and sustaining growth in Eastern Europe and Central Asia requires reforms to Managing Education for Results in Eastern EuropeSkills, and Central Not Asia Just Diplomas boost competitiveness and increase labor productivity. Among the required changes are reforms to education. In surveys conducted immediately before the economic crisis, companies in the region reported shortage of skills as one of the most significant bottle- necks in their operations, suggesting that education systems in Eastern Europe and Central Asia— with a reputation for high enrollment rates and well-trained teachers—still need to improve their performance. In fact, international test results show that many students—outside of a handful of coun- tries in the region—are failing to acquire more than the most basic literacy and numeracy skills. Anecdotal evidence also indicates that the rapid expansion in higher education has led to a decline in the quality and relevance of education provided. At the same time, there are few opportunities for adults to retrain, upgrade, or acquire new skills—the life-long learning needed for employability. As Skills, Not Just Diplomas suggests, the shortage of skills is a wake-up call to reform education and training systems to provide higher quality education with the flexibility for students and training institutions to better respond to market signals. Such deep reform will have to center on the following: DIRECTIONSINDEVELOPMENT • Focusing more on measuring whether students learn and graduates find jobs, and Human Development using this information to actively improve teaching and learning. • Using incentives across the education system, including granting greater autonomy to institutions on curriculum, teaching methods, resource use and institutional mission, and increasing accountability for learning. -
The Statistical Data and Metadata Exchange Standard (SDMX)
UNITED NATIONS ECONOMIC AND SOCIAL COMMISSION FOR ASIA AND THE PACIFIC Expert Group Meeting: Opportunities and advantages of enhanced collaboration on statistical information management in Asia and the Pacific 20-22 June 2011 Swissotel Nailert Park, Bangkok Tuesday, 21 June 2011 Session 3: Emerging generic business process models and common frameworks and terminology – A basis for practical cooperation? Integrating statistical information systems: The Statistical Data and Metadata eXchange Standard (SDMX) Brian Studman Australian Bureau of Statistics [email protected] Session Overview •SDMX – Background –History – What can it represent – SDMX & Other Standards (particularly DDI) • ABS & SDMX: Integration via IMTP (ABS major business programme) • Using SDMX in the context of GSBPM • Examples of Use: ABS main projects that utilise SDMX • A quick preview of GSIM v 0.1 (if we have time) 1 SDMX • Origins – Bank for International Settlements, European Central Bank, Eurostat, IMF, UN, OECD, and the World Bank (SDMX Consortium) • Format for statistical returns for aggregated data • Version 2.1 (April 2011 public comment) • http://sdmx.org/ (best starting place) • SDMX comes from the international agencies (OECD, IMF, Eurostat, UNSD, World Bank, ECB, BIS) – they get aggregate statistical tables from many countries regularly over time – they wanted to automate and manage the process • they need standard agreed definitions and classifications, standard agreed table structures, standard agreed formats for both data and metadata – They commissioned