POSITIONING AND POWER IN ACADEMIC PUBLISHING: PLAYERS, AGENTS AND AGENDAS This page intentionally left blank
Positioning and Power in Academic Publishing: Players, Agents and Agendas Proceedings of the 20th International Conference on Electronic Publishing
Edited by Fernando Loizides Emerging Interactive Technologies Lab, University of Wolverhampton, UK and Birgit Schmidt University of Göttingen, State and University Library, Germany
Amsterdam • Berlin • Washington, DC
© 2016 The authors and IOS Press.
This book is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License.
ISBN 978-1-61499-648-4 (print) ISBN 978-1-61499-649-1 (online) Library of Congress Control Number: 2016940262
Publisher IOS Press BV Nieuwe Hemweg 6B 1013 BG Amsterdam Netherlands fax: +31 20 687 0019 e-mail: [email protected]
Distributor in the USA and Canada IOS Press, Inc. 4502 Rachael Manor Drive Fairfax, VA 22032 USA fax: +1 703 323 3668 e-mail: [email protected]
Cover design: IOS Press Cover image: Cultural Studies Divisional Library, Göttingen State and University Library; photo by Claudia Hake.
LEGAL NOTICE The publisher is not responsible for the use which might be made of the following information.
PRINTED IN THE NETHERLANDS v
Preface
Running a technology-informed conference such as the International Conference on Electronic Publishing (Elpub) for the twentieth time could be taken as a sign of saturation and maturity. However, if we consider technology as only one cultural aspect of our current scholarly communication ecosystem, we have to note that we continue to be in the middle of a digital transformation. Technologies come and go due to social reasons; the positioning of stakeholders and the distribution of economic, technological and discursive power continues to be negotiated. And at times a seemingly given fact of publishing like the transfer of intellectual property rights to third parties gets heavily questioned – as the recent discussion around the shadow library Sci-Hub indicates. To provide room for discussion beyond technology and embed technology in its social and cultural framework, Elpub 2016 will open the floor for emerging alternatives in how scholars and citizens interact with scholarly content and the role dissemination and publishing plays in these interactions. What is the core of publishing today? How does agenda-setting in emerging frameworks like Open Science work and what is the nature of power of the surrounding scholarly discourses? How does this relate to the European and world-wide Open Science and Open Innovation agenda of funders and institutions, and what does this look like in publishing practice? Asking such questions promises to widen our horizons. Elpub reaches a great milestone with its 20th anniversary this year, to be held in Göttingen, Germany. Since its beginnings twenty years ago, the conference has been a leading forum for electronic publishing topics, attracting people from around the world and facilitating active collaboration and knowledge exchange. Twenty years on, the conference brings together leading stakeholders such as academics, practitioners, policy makers, students and entrepreneurs from a wide variety of fields and countries. The conference once again has an exciting programme in store for attendees and readers of the proceedings alike. The conference opens with a reflection and celebration of the last twenty years. This year, 17 research papers and 9 posters will be presented. The programme covers a wide variety of topics, including how to maintain the quality of electronic publications, modelling processes, and implementation issues regarding open access. These subjects, and especially the latter, become even more prevalent with reforms such as Britain’s Research Excellence Framework rule which allows only open access articles to be eligible for submission – and on this basis, deposit in repositories (as a core element of institutional research information systems) becomes the norm. At Elpub, there will be several new publishing systems and repositories presented and tested, as well as datasets for the delegates to examine. In addition, four workshops will offer delegates the opportunity to explore “Open Peer Review: Models, Benefits and Limitations” (co-organized by OpenAIRE), “Opening up the collection – reuse and publishing” (LIBER), “Entering the publishing system – Junior Scientist Day (FOSTER)” and “OJS 3.0 and OMP 1.2: The latest in open source software for academia-controlled publishing (PKP)”. We are delighted to have three captivating keynote talks this year: Jean-Claude Guédon from the University of Montréal, Canada, will speak on the topic of “Whither Open Access? Four scenarios and four choices”, and our second keynote speaker, Tara vi
Andrews from the University of Bern in Switzerland talks on the topic of “After the Spring: digital forms of scholarship and the publication ecosystem”. The third speaker Prateek Mahalwar shares with us his views on opportunities and challenges for early career researchers in the context of Open Science. Finally, a panel of experts will investigate the conference topics in an open forum with short stakeholder perspectives and the opportunity for the audience to engage with the discussion. We would like to express our sincere gratitude to all members of the Elpub Executive Committee who, together with the Programme Committee, helped us to bring together such a diverse and exciting programme. We would also like our sponsors – Altmetric, MDPI and Copernicus Publications (at the time of writing) – for their support as well as their openness to discussion and cooperation in bringing forward the Open Science agenda. We wish everyone an inspiring conference and a happy 20th Anniversary with many more to come. We look forward to continuing the discussion and seeing you again at the 21st edition of the conference in Cyprus!
Fernando Loizides and Birgit Schmidt 1 June 2016 vii
Organisation
Conference Host University of Göttingen
General Chair Birgit Schmidt, University of Göttingen, Germany
Programme Chair Fernando Loizides, University of Wolverhampton, UK
Programme Committee Members Ana Alice Baptista University of Minho Portugal Margo Bargheer University of Göttingen Germany George Buchanan City University London UK Leslie Chan University of Toronto Canada Jan Engelen Katholieke Universiteit Leuven Belgium Paola Gargiulo CINECA Italy Stamatios Giannoulakis Cyprus University of Technology Cyprus Herbert Gruttemeier CNRS France Arunas Gudinavicius Vilnius University Lithuania Puneet Kishor Independent USA Alexia Kounoudes Cyprus University of Technology Cyprus Peter Linde Blekinge Institute of Technology Sweden Natalia Manola University of Athens Greece Eva Méndez Rodríguez Universidad Carlos III de Madrid Spain Pierre Mounier Open Edition France Panayiota Polydoratou Alexander Technological Education Greece Institute of Thessaloniki Andreas Rauber TU Wien Austria Laurent Romary DARIAH Germany Anthony Ross-Hellauer University of Göttingen Germany Andrea Scharnhorst DANS-KNAW Netherlands viii
John Smith University of Kent UK Josef Steinberger University of West Bohemia Czech Republic Niels Stern The Nordic Council of Ministers Denmark Xenia van Edig Copernicus Publications Germany Jens Vigen CERN Switzerland Marios Zervas Cyprus University of Technology Cyprus
Executive Committee Ana Alice Baptista University of Minho Portugal Leslie Chan University of Toronto Canada Sely Costa University of Brasilia Brazil Jan Engelen Katholieke Universiteit Leuven Belgium Turid Hedlund Hanken School of Economics Finland Peter Linde Blekinge Institute of Technology Sweden Mícheál Mac an Airchinnigh University of Dublin Ireland Bob Martens Vienna University of Technology Austria Panayiota Polydoratou Alexander Technological Education Greece Institute of Thessaloniki Yasar Tonta Hacettepe University Turkey
ix
Sponsors
This page intentionally left blank xi
Contents
Preface v Fernando Loizides and Birgit Schmidt Organisation vii Sponsors ix
Time to Adopt: Librarians’ New Skills and Competency Profiles 1 Birgit Schmidt, Pascal Calarco, Iryna Kuchma and Kathleen Shearer SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics 9 Angelika Kutz COAR Case Study Controlled Vocabularies and PHAIDRA International 16 Imma Subirat, Iryna Solodovnik, Paolo Budroni, Raman Ganguly and Rastislav Hudak “If There Are Documents You Really Care About: Print Them Out!” (Vint Cerv, 2015) 23 Bernd Kulawik Interactive Public Peer ReviewTM: An Innovative Approach to Scientific Quality Assurance 28 Xenia Van Edig Renegotiating Open-Access-Licences for Scientific Films 34 Elke Brehm ROAD: The Directory of Open Access Scholarly Resources to Promote Open Access Worldwide 37 Nathalie Cornic Sustaining the Growth of Library Scholarly Publishing 42 Graham Stone Genome Sharing Projects Around the World: How You Find Data for Your Research 48 Fiona Nielsen and Nadezda Kovalevskaya Stakeholders in Academic Publishing: Text and Data Mining Perspective and Potential 51 Maria Eskevich Referencing of Complex Software Environments as Representations of Research Data 58 Sven Bingert, Stefan Buddenbohm and Daniel Kurzawe Towards New Metrics for Bioresource Use 69 Laurence Mabile, Paola De Castro, Elena Bravo, Barbara Parodi, Mogens Thomsen, Samuel Moore and Anne Cambon-Thomsen xii
Open Access, Open Science, Open Society 75 Thomas Margoni, Roberto Caso, Rossana Ducato, Paolo Guarda and Valentina Moscon Jupyter Notebooks – A Publishing Format for Reproducible Computational Workflows 87 Thomas Kluyver, Benjamin Ragan-Kelley, Fernando Pérez, Brian Granger, Matthias Bussonnier, Jonathan Frederic, Kyle Kelley, Jessica Hamrick, Jason Grout, Sylvain Corlay, Paul Ivanov, Damián Avila, Safia Abdalla, Carol Willing and Jupyter Development Team Peer Review on the Move from Closed to Open 91 Birgit Schmidt, Arvid Deppe, Julien Bordier and Tony Ross-Hellauer UCL Press: A New Model for Open Access University Presses 99 Lara Speicher Identifying and Improving Dataset References in Social Sciences Full Texts 105 Behnam Ghavimi, Philipp Mayr, Sahar Vahdati and Christoph Lange Battling for ‘Openness’. Applying Situational Analysis to Negotiations Between Dutch Universities and Elsevier 115 Elena Šimukovič Insights from Over a Decade of Electronic Publishing Research 119 Fernando Loizides and Sam A.M. Jones Article Processing Charges: A New Route to Open Access? 125 Gerald Beasley A Glance on Presence of Shahid Beheshti University Scholars in Research Gate 131 Amir Reza Asnafi FOSTER’s Open Science Training Tools and Best Practices 135 Astrid Orth, Nancy Pontika and David Ball Stepping Up Towards Greater Alignment of Repository Networks 142 Katharina Müller, Arvid Deppe, Maxie Gottschling, Eloy Rodrigues and Kathleen Shearer
Subject Index 147 Author Index 149
Positioning and Power in Academic Publishing: Players, Agents and Agendas 1 F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-1 Time to Adopt: Librarians’ New Skills and Competency Profiles
1 Birgit SCHMIDTa, , Pascal CALARCOb, Iryna KUCHMAc, and Kathleen SHEARERd
a University of Göttingen, State and University Library b University of Waterloo c Electronic Information for Libraries (EIFL) d Confederation of Open Access Repositories (COAR)
Abstract. On the one hand, libraries are at the forefront of the digital transformation and digital information infrastructures, on the other, they manage and curate cultural heritage collections. This brings about new ways of engagement with information and knowledge and the need to rethink skills and competency profiles – which enable librarians to support e-research all along the research cycle. This paper presents findings of the joint Task Force on Librarians’ Competencies in Support of E-Research and Scholarly Communication.
Keywords. e-research, competencies, job profiles, libraries.
1. Introduction
Rapid changes in technology and associated shifts in research and scholarly communications are profoundly changing the role of libraries in the 21st century. The emergence of e-research, for example, is bringing about new ways of doing science across the globe, compelling libraries to adopt new services, such as assisting with the development of research data management plans, hosting collaborative virtual research environments, managing institutional repositories, and disseminating research outputs through open access mechanisms. These novel services require a range of new skills and expertise within the library community as well as a shift in organizational models for libraries. In August 2013, the Association of Research Libraries (ARL), the Canadian Association of Research Libraries (CARL), the Association of European Research Libraries (LIBER), and the Confederation of Open Access Repositories (COAR) launched the joint Task Force on Librarians’ Competencies in Support of E-Research and Scholarly Communication.2 Since then, the Task Force has been working on identifying emerging specialty roles, through performing literature reviews and collaboratively preparing a series of service areas and competencies documents for research data management, scholarly communication and Open Access, digital curation and preservation and support for digital scholarship.
1 Corresponding Author. E-mail: [email protected]. 2 https://www.coar-repositories.org/activities/support-and-training/task-force-competencies. 2 B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles
2. Competencies and Skills under Review
The growing abundance of digital information and data affects the whole research workflow, including methods and tools as well as enabling infrastructures. Accordingly, there is an emerging need for a new type of workforce, and existing and emerging staff competencies and skills are under scrutiny. In particular, a lot of attention has been attracted by “data science”, in particular in relation to “big data”; with Hal Varian, chief economist at Google declaring “the sexy job of the next 10 years will be statisticians” (Lohr, 2009). With a view on customers Affelt (2015) pointed out how librarians and information professionals – which have always been well-versed in working with data – can leverage their skills and training for big data applications that 3 resonate with stakeholders. 3 Librarians manage different types of published information and data, and also curate a wealth of information that awaits further exploration and exploitation based on digital methods and tools. To step up skills and competencies of librarians and to some degree research staff, several initiatives have looked into specific areas, e.g. open science, research data management, digital curation, digital humanities, eResearch, data 4 science, etc.4 Some of these initiatives focus on professional training, others target the development of higher education curricula and explore how librarians can contribute.
2.1. Defining Competencies and Skills
According to the European e-Competence Framework (e-CF) competence is the “demonstrated ability to apply knowledge, skills and attitudes to achieve observable results”. Hence, a competence is not a skill; on the contrary, a competence embeds skills. Whilst competencies are holistic concepts, skills are precise and definite abilities, either hard technical, e.g. make a cost / benefit analysis, develop user interfaces; or soft, e.g. deploy empathy to customer needs, negotiate contract terms and conditions (e-CF, 2014). Job profiles typically combine several competencies, and one single competence may be assigned to a number of different job profiles. A core idea in this context is that competencies can be grouped by areas (plan, build, run, enable, manage) and can be categorized by proficiency levels, ranging from the ability to apply knowledge and skills to solve straightforward problems to the overall accountability and responsibility, and to solve critical problems in an innovative way (the e-CF’s levels e-1 to e-5 are related to the European Qualifications Framework’s levels 3 to 8) (e-CF, 2014). In North America, ARL members have been engaged in identifying the issues around evolving competencies needed for the work of research librarians, and how these roles intersect with new functional specialists, documented in a series of
3 See also: Florida Library Webinars, 31 August 2015, http://floridalibrarywebinars.org/the-accidental- data-scientist-a-new-role-for-librarians-and-info-pros-ondemand 4 E.g. Facilitate Open Science Training for European Research (FOSTER), https://www.fosteropenscience.eu/; MANTRA Research Data Management Training, http://datalib.edina.ac.uk/mantra/; Essentials for Data Support, http://datasupport.researchdata.nl/en/; Curriculum Framework for Digital Curation, Digital Curator Vocational Education Europe (DigCurV), 2013, http://www.digcur-education.org; EDISON, http://www.edison-project.eu.
B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles 3 publications entitled “New Roles for New Times” (Janguszewski and Williams, 2013; Covert-Vail and Collard, 2012) and by Rockenbach et al. (2015).
3. Mapping E-Research and Service Areas
A range of descriptions of services areas and related competencies have been developed and shared with the community for comments by the joint Task Force on Librarians’ Competencies in Support of E-Research and Scholarly Communication. Consolidated versions will be published in spring 2016.
3.1. Managing Research Data
Research data management (RDM) involves services and infrastructures in order to support the handling of research data across the data lifecycle (i.e. creating/collecting, processing, analyzing, publishing, archiving/preserving, re-using data). The various aspects of RDM are often distributed across different support services (research office, IT services, library) and academic departments. Interviews with researchers demonstrate that, while researchers need support in numerous areas across the entire research lifecycle: planning, organizing, security, documenting and sharing, preparing datasets for deposit and long-term preservation, as well as issues related to copyright, licensing and intellectual property more generally (e.g. Wilson, 2014; Parsons et al., 2013). Research data management encompasses a large group of activities that may differ significantly across the research data lifecycle. Generally it requires a high level of interaction with researchers and also working with other support services including technical services and research officers. There are various strategies for service development and operation, some concentrate on discipline-specific services developed in the context of projects, others highlight multidisciplinary perspectives (e.g. Molloy and Snow, 2012; Carlson, 2015). For RDM training close work with disciplinary experts is recommended to ensure that the terminology used is accurate and clear; discipline-specific examples and good practices are also highly valuable for engaging the audience and for putting basic principles in context (Molloy and Snow, 2012). Based on funder requirements, the need to support researchers in creating and implementing data management plans has substantially grown over the last years. Several libraries have set up a service to support such needs, often in collaboration with other service units (e.g. research office, IT services, legal advisor, ethics committee). The development of such a service can even serve as a training ground for librarians and other institutional stakeholders (Davis and Cross, 2015). Libraries’ activities in research data management can be usefully conceptualized as falling into three broad categories: providing access to data; supporting researchers and students in managing their data; and managing a data collection. There are overlaps, but each of these areas has some distinctive roles for librarians. Providing access to data mainly involves consultation and reference services, e.g. to identify datasets, provide advice on discovery and analytic tools as well as how to cite/reference data. Advocacy and support for managing data is a wide area of activities ranging from promoting the institutional data policy, providing support and training, e.g. on how to 4 B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles write a data management plan or how to identify and use data repositories, to develop data curation profiles, and to manage software related to data. The management of collections targets activities such as the preparation of data, its preservation, sharing and publishing. Further details on the core competencies needed to cover these areas can be found in the RDM skills and competency profile (cf. note 1).
3.2. Scholarly Communication and Open Access
Scholarly communication and Open Access (OA) involves changing modes of communication of research made possible in the digital environment. For example, the evolution from the traditional commercial publishing model where the author signs away their copyright to a work of original research scholarship to subscription-based journals to one of several OA models that have been emerging over the past two decades. Scholarly communication can be defined as “the system through which research and other scholarly writings are created, evaluated for quality, disseminated to the scholarly community, and preserved for future use. The system includes both formal means of communication, such as publication in peer-reviewed journals, and informal channels, such as electronic listservs” (ARL). Other informal scholarly communication channels include posts in social media, e.g. blogs, tweets, etc. One of the most widely used definitions of OA is that from the Budapest Open Access Initiative, from a conference of that name held in 2001: “By ‘open access’ to this literature, we mean its free availability on the public internet, permitting any users to read, download, copy, distribute, print, search, or link to the full texts of these articles, crawl them for indexing, pass them as data to software, or use them for any other lawful purpose, without financial, legal, or technical barriers other than those inseparable from gaining access to the internet itself. The only constraint on reproduction and distribution, and the only role for copyright in this domain, should be to give authors control over the integrity of their work and the right to be properly acknowledged and cited” (BOAI, 2002). Some examples of how libraries have been involved in this process are: Providing consultation and training that encourages and enables faculty to manage their own copyright and improve the economics of, and access to, published research. Developing and contributing to scientific information infrastructures for the dissemination and linking of research outputs, e.g. digital repositories and their networks – on institutional, national and international levels (e.g. SHARE, OpenAIRE, LA Referencia, COAR).5 Offering OA journal and/or book publishing services and other value-added services to scholars on their campus (work with the university press if there is any). Working with the acquisition department, library consortia and research funders to develop and maintain a publication fund, OA memberships and OA clauses in licenses. Providing access to services and resources that help measure quality and impact of scholarship, from traditional bibliometrics to emerging altmetrics.
5 SHARE, http://www.share-research.org/; OpenAIRE, https://www.openaire.eu/; LA Referencia, http://lareferencia.redclara.net/; COAR, https://www.coar-repositories.org.
B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles 5
Libraries’ activities in scholarly communication and open access typically fall into one of these categories: scholarly publishing services; copyright and open access advocacy and outreach; scholarly resource assessment. Some level of subject knowledge is required in most of these roles. In particular, librarians will need to have a broad perspective and understanding of the traditional (commercial, society) and open access models of publishing, intellectual property issues, and economics of scholarly publishing. For example, librarians in this role may support graduate students and faculty members who wish to start new Open Access journals that the library may host, coordinate education and advocacy events such as Open Access Week, and serve on campus copyright committees to assist with campus policy development and interpretation.
3.3. Digital Humanities
Digital Humanities (DH) describes a multidisciplinary research community involving the application of computational methods to humanistic topics of inquiry, or, more broadly, the intersection between the arts and humanities and information technology and/or digital expression (for a discussion of research trends and views on DH see e.g (Burdick et al., 2012; Holm et al., 2015). Digital humanists utilize computational methods and/or digital tools to advance research and pedagogy. Methods and tools include, but are not limited to: 3-D representation, digital mapping, electronic textual analysis, digitization of materials, data visualization techniques, and interactive digital media including game-based systems. Over the last decade, several universities have stepped up their support of 6 humanities research by creating research centres and workspaces,6 many of them in close collaboration and physically located with their respective university libraries. Engagement with Digital Humanities is an evolving specialization in librarianship, one that requires a combination of a strong academic background in the arts and humanities (domain expertise), technical grounding in technologies and tools to support computational models of research and teaching in Humanities, and project management. The role has other important components such as advisor, advocate, and partner for special collections curators. The digital humanities librarian’s role is also directly related to scholarly communication and data management. The library’s organizational structure will affect the services offered and roles played by digital humanities librarians, but across various models, the digital humanities librarian will likely work in a team environment, making collaboration and the ability to perform in a changing and dynamic role core competencies for this specialization. When it comes to services and responsibilities digital humanities librarians engage in scholarly communication and publishing, technical services (in particular related to interaction with digital resources and collections), partner with faculty and student for
6 Cf. centerNet, an international network of digital humanities centres, http://www.dhcenternet.org/, and European Association for Digital Humanities, Digital Humanities Centres, http://eadh.org/education/digital- humanities-centres. From a much broader perspective an increasing number of universities and colleges are also establishing digital scholarship centres to support researchers in their work with digital tools and large datasets, such as data visualization in the environmental sciences, data mining of large corpora of texts in the humanities, and developing GIS or other geolocation data representations in the social sciences (Lippincott, 2014).
6 B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles digital humanities research and consulting, provide teaching and training activities, and develop and manage spaces (labs, collaboratories) for digital humanities work. Accordingly, digital humanities librarians bring together a wide range of compentencies and skills. A ba s e la ye r is typically advanced academic subject expertise and professional training in library and information science, particularly in scholarly communication and data management. When it comes to technical skills, many job descriptions note the emerging and evolving state of technology by requiring general competencies such as “demonstrated ability and interest in exploring and evaluating emerging technologies in support of digital humanities,” and a “willingness to remain current with changing technology and its applications” (cited from job descriptions, cf. note 1). Technical skills and competencies include e.g. data visualization, text mining, metadata standards and schema, text markup and encoding, semantic web technologies. Essential for direct involvement and/or consultation with research activities are also project and program development and collaboration skills, e.g. grant writing and the development of technology-rich work spaces.
4. Conclusions
Not surprisingly, a number of other new areas could benefit from librarians’ support – and this again comes with a need for developing/expanding skills to fulfill these new roles. One area which is evolving very fast is text and data mining (TDM), and libraries might already have a range of subscriptions and collections which come with appropriate licenses but have not yet stepped up to provide practical support for researchers to exploit these riches (Okerson, 2013). As already mentioned above support for TDM plays a key role in digital humanities research, e.g. allowing new views on texts, but other research areas such as economical and social sciences will benefit as well (Liber, 2014; ASIS&T, 2015). In our discussion of competency profiles we have only briefly touched how librarians acquire the skills and competencies needed for these evolving and sometimes already well-established service areas. Strategies will vary depending on institutional and personnel resources and range from attending workshops or conferences and/or joining working groups, the development of institutional training programs for individuals and/or groups, participation in online learning course (e.g. MOOCs), etc. Most beneficial might be to combine newly hired experts and long-term staff in new teams which dedicate their efforts to developing and delivering new types of services. Such teams will often combine staff with different backgrounds, and bring new skills sets to the institution, e.g. from the publishing industry to build up / enhance a publishing unit, or from research disciplines or technology experts to develop and promote specialized data infrastructures and digital work environments. Not surprisingly, involvement in collaborative projects, national and internationally, are a good instrument to contribute own expertise and learn from others to build up prototypical services. However, additional effort will be needed to assess the results of these efforts and for sustaining both personnel and infrastructures. It should be noted that job descriptions can be excessively demanding in terms of experiences and skills, as if the search is for the “Unicorn Librarian – that magical creature who can be all things to all people” (Johnson, 2014). Therefore, individuals and employers should consult the task force’s competency profiles with some caution. B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles 7
Typically it will be a group of individuals that bring together these competencies and skills, a collaborative work force which strengthens the library’s capacities and which may also be an element of new organizational structures.
Acknowledgement.The authors would like to thank Rob Grim, Alicia López Medina, Susan Reilly, Judy Ruttenberg, and Dominic Tate for their contributions to the work of the Task Force and their comments to this paper.
References
Affelt A. (2015) The Accidental Data Scientist: Big Data Applications and Opportunities for Librarians and Information Professionals. Medford, New Jersey: Information Today, Inc. ARL – Association of Research Libraries, Focus areas: Scholarly communication. Available at: http://www.arl.org/focus-areas/scholarly-communication (Accessed: 11 March 2016).
ASIS&T (2015) ‘Text Mining in Libraries, How Librarians Develop Skills Required to Support this Evolving Form of Research’, ASIS&T webinar, blog post, 2 November 2015. Available at: https://www.asist.org/events/webinars/text-mining-in-libraries (Accessed: 11 March 2016).
BOAI – Budapest Open Access Initiative (2002): Declaration, Budapest, Hungary, 14 February 2002. Available at: http://www.budapestopenaccessinitiative.org/read (Accessed: 11 March 2016).
Burdick, A., Drucker, J., Lunenfeld, P., Presner, T. and Schnapp, J. (2012) ‘Digital Humanities’, Cambridge, MA: MIT Press. Available at: https://mitpress.mit.edu/sites/default/files/titles/content/ 9780262018470_Open_Access_Edition.pdf (Accessed: 11 March 2016).
Carlson, J., Sapp Nelson, M. Johnston, L. R. and Koshoffer, A. (2015) ‘Developing Data Literacy Programs: Working with Faculty, Graduate Students and Undergraduates’, Bulletin of the American Society for Information Science and Technology 41(6), pp 14–17. doi:10.1002/bult.2015.1720410608.
Covert-Vail, L. and Collard, S. (2012) ‘New Roles for New Times: Research Library Services for Graduate Students’, Association of Research Libraries (ARL), http://www.arl.org/storage/documents/ publications/nrnt-grad-roles-20dec12.pdf (Accessed: 11 March 2016).
Davis, H. M. and Cross, W. M. (2015) ‘Using a Data Management Plan Review Service as a Training Ground for Librarians’, Journal of Librarianship and Scholarly Communication 3(2), eP1243, doi: 10.7710/2162-3309.1243. eCF – European e-Competence Framework 3.0 (2014) A common European Framework for ICT Professionals in all industry sectors, CWA 16234:2014, Part 1. Available at: http://www.ecompetences.eu/e-cf-3-0-download (Accessed: 11 March 2016).
Holm, P., Jarrick, A. and Scott, D. (2015) ‘The Digital Humanities’, Humanities World Report 2015, pp 64– 83. Palgrave Macmillan UK. doi: 10.1007/978-1-137-50028-1.
Janguszewski J. M. and Williams, K. (2013) ‘New Roles for New Times: Transforming Liaison Roles in Research Libraries’, report prepared for the Association of Research Libraries (ARL), August 2013. Available at: http://www.arl.org/storage/documents/publications/nrnt-liaison-roles-revised.pdf (Accessed: 11 March 2016).
Johnson A. (pseudonym) (2014) ‘Hiring Data Librarians: Notes on hiring and being hired as a data librarian’. Available at: http://www.scribd.com/doc/265015825/Hiring-Data-Librarians (Accessed: 11 March 2016).
LIBER – Association of European Research Libraries (2014), ‘Text and Data Mining: The need for change in Europe’. Available at: http://libereurope.eu/text-data-mining (Accessed: 11 March 2016).
8 B. Schmidt et al. / Time to Adopt: Librarians’ New Skills and Competency Profiles
Lippincott, J., Hemmasi, H. and Lewis, V. M. (2014) ‘Trends in Digital Scholarship Centers’, Educause Review, 16 June 2014, http://er.educause.edu/articles/2014/6/trends-in-digital-scholarship-centers (Accessed: 11 March 2016).
Lohr S. (2009) ‘For Today’s Graduate, Just One Word: Statistics’, New York Times, 5 August 2009. Available at: http://www.nytimes.com/2009/08/06/technology/06stats.html (Accessed: 11 March 2016). [16] Molloy, L. and K. Snow, K. (2012) ‘The Data Management Skills Support Initiative: Synthesising Postgraduate Training in Research Data Management’, The International Journal of Digital Curation 7(2), pp 101–109. doi: 10.2218/ijdc.v7i2.233.
Okerson, A. (2013) ‘Text & Data Mining – A Librarian Overview’, paper presented at IFLA World Library and Information Congress, 17-23 August 2013, Singapore. Available at: http://library.ifla.org/252/1/165-okerson-en.pdf (Accessed: 11 March 2016).
Parsons, T., Grimshaw, S. and Williamson, L. (2013) ‘Research Data Management Survey’, University of Nottingham, February 2013. Available at: http://admire.jiscinvolve.org/wp/files/2013/02/ADMIRe- Survey-Results-and-Analysis-2013.pdf (Accessed: 11 March 2016).
Rockenbach, B., Ruttenberg, J., Tancheva, K. and Vine, R. (2015) ‘Pilot Library Liaison Institute’, Association of Research Libraries/Columbia University/Cornell University/University of Toronto, Final Report, June 2015. Available at: http://www.arl.org/storage/documents/publications/library- liaison-institute-final-report-dec2015.pdf (Accessed: 11 March 2016).
Wilson, J. A. J. (2014) ‘University of Oxford Research Data Management Infrastructure’, LIBER Research Data Management Case Study, June 2014. Available at: http://libereurope.eu/wp- content/uploads/2014/06/LIBER-Case-Study-UOX.pdf (Accessed: 11 March 2016). Positioning and Power in Academic Publishing: Players, Agents and Agendas 9 F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-9 SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics
Angelika KUTZ LL.M.1
TIB Hannover (German National Library of Science and Technology) Germany
Abstract. SCOAP3 is a global partnership which converts high-quality subscription journals in the field of High-Energy Physics to Open Access through redirection of existing subscription funds. Since January 1, 2014 the SCOAP3 Gold Open Access Repository is providing free access to scientific articles in high quality journals in the field of High Energy Physics. This article describes this international pilot which flips the current subscription-based financing of scientific publication to an output-based financing model (fair share). This includes a description of the unique mechanisms of SCOAP3 as well as its governance structure and short view on the national German contributing partners, especially the German universities (SCOAP3-DH).
Keywords. SCOAP3, SCOAP3-DH, Gold Open Access, High Energy Physics, Flipping Model, CC-BY, No Costs for Authors, Text- and Datamining, CERN, TIB Hannover
1. Introduction
SCOAP3 stands for Sponsoring Consortium for Open Access Publishing in Particle Physics. The goal of this worldwide open access project initiated by CERN (European Organization for Nuclear Research) is to convert scientific articles in the high energy physics field published in high quality journals into Gold Open Access by redirecting subscription fees for the flipping of the financing system. Its first phase from 2014-2016 was made possible due to the common effort of about 3000 libraries and consortia worldwide, research organisations, cooperative publishers and funding agencies in many countries.
SCOAP3-DH stands for Sponsoring Consortium for Open Access Publishing in Particle Physics – German Universities. Organised by TIB (German National Library of Science and Technology) in Hannover in Germany this national project SCOAP3-DH is taking care of the participation of the German universities in the worldwide project led by CERN.
1 Corresponding Author.: [email protected] 10 A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics
2. SCOAP3: a worldwide Gold Open Access Flipping Model
This worldwide pilot originally initiated by CERN includes over three thousand libraries and consortia and is supported by funding agencies in various countries. The thereby implemented Gold Open Access articles can be read and re-used by anyone as long as the author is named as all articles are published under a CC-BY licence (https://creativecommons.org/licenses/by/4.0/). During its first phase SCOAP3 redirects the subscription payments into the SCOAP3 fund out of which the publishers are centrally paid by CERN for their services. These concrete services were defined during the tendering process and laid down in a technical specification which is available online under: http://scoap3.org/files/Technical_Specification.pdf. In order to be able to provide funds for the SCOAP3 fund the participating publishers reduced the proportional amount of the converted journals in the bills sent to the libraries thereby enabling them to redirect these released amounts to the SCOAP3 fund. All final versions of peer-reviewed articles published in the SCOAP3 journals are immediately available for everyone on the internet and free of charge on both the publishers’ websites and the SCOAP³ Repository (http://repo.scoap3.org).
2.1. Start on January 1, 2014
Parameters of SCOAP3: Gold Open Access (worldwide free accessibility over the internet). Key Journals in High Energy Physics have – fully or partially – been converted into Gold Open Access. No costs for authors. The copyright remains with the author. No administrative burden for the author. No administrative burden for participating institutions due to National Contact Points dealing with SCOAP³/SCOAP³-DH centrally. CC-BY Licences (the copyright remains with the author). CC0 for metadata. Text- and datamining allowance. Constantly growing number of articles available upon publication in the SCOAP³ Repositorium (end of 2015: more than 9.000 Gold Open Access articles) and the publishers’ websites
These parameters have been implemented due to an international tendering process led by CERN for the first phase of SCOAP3.
A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics 11
2.2. Participating Countries
Currently there are over 30 countries worldwide participating, more are expected to join soon. The current international SCOAP3 Partners are listed on the CERN website: http://scoap3.org/participating-countries.
2.3. The capping mechanism of SCOAP3leading to reasonable average APCs
The SCOAP³ tendering process included a maximum amount paid to the individual publisher for a certain (“capped”) amount of articles of the specific journal (“capping mechanism”). Any article above this “cap” will be published under the same conditions and services (see parameters described in 2.1) but the publisher will not receive any additional amounts of money for any article exceeding this cap. This capping mechanism thereby steadily leads to sinking average article processing charges (APC) as the publisher receives only a maximum amount for each journal contract. Due to this capping mechanism the average APC for SCOAP³ Articles (articles published in SCOAP³ Journals) is currently about 1.100€.
2.4. Value added services of SCOAP3
SCOAP3 reached a very high compliance corresponding with the prerequisites (like CC-BY, XML and CC0 for metadata) implemented by the tendering process due to the central control function of a single well trained team at CERN. SCOAP3 can offer a near to 100% compliance rate (99.98%) compared to other similar initiatives (e.g. Welcome Trust; 61%).
2.5. Advantages of SCOAP3
SCOAP³/SCOAP³-DH offers the following advantages for SCOAP³-Partners:
Advantages for the libraries: No administration effort as after one single payment to SCOAP3 everything regarding the publication is taken care of by SCOAP3/SCOAP³-DH. There is no need for a single library to deal with individual APC-payments for each single article published and which relieves them from any administrative burden. At the same time the necessary involvement of time, personnel and costs are eased.
Advantages for Authors: Authors can publish their articles in high quality journals Gold Open Access upon publication and at no cost. They do not have to deal with any of the administrative burden connected with APC-payment in other publication models.
2.6. First Phase of SCOAP3 2014-2016
During the first phase of SCOAP3 a model of redirection of subscription cost was implemented. This included that publishers reduced their bills according to the SCOAP3 contribution of each institution. The bills of the respective institution were 12 A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics reduced according to their participation portion. By using the money already contained in the system in form of subscription costs SCOAP3 enabled a real flip from closed access to Gold Open Access in one single step.
The Technical Specification during the international tendering process led by CERN set inter alia the following preconditions: CC-BY Capping mechanism API OAI-PMH
There is the important note for articles to be published in journals which have been partially converted: authors have to upload their articles to arXiv before submitting it to the respective publisher together with the arXiv number received by uploading it to arXiv.
2.7. Second Phase of SCOAP3 (2017-2019)
The second phase of SCOAP3 is stepping even further. The next phase will change the current model to a mere output-orientated model in which institutions pay a lump sum (flat-rate) according to the share of publications made by their institutions (fair share) in order to enable their authors to publish freely and without any organizational or administrative burden with regards to any Article Processing Charge (APC) handling. By doing this SCOAP3 enables a kind of an “all-you-can-publish-model” for authors of High Energy Physics Gold Open Access scientific articles in high quality journals.
3. Publishers, Journals and APCs (Article Processing Charges)
The following publishers and journals are participating in SCOAP3 during its first phase of from 2014-2016. The table reflects the results of the international tendering process held by CERN during 2012 and 2013 in order to reach competitive quality parameters as well as reasonable APCs (Article Processing Charges).
Table 3. Publishers. Journals. Article Processing Charges (APCs) Publisher Journal APCs Elsevier Nuclear Physics B 2.000 USD Elsevier Physics Letters B 1.800 USD Hindawi Advances in High Energy 1.000 USD Physics Institute of Physics Publishing/ Chinese Physics C 1.000 GBP Chinese Academy of Science Institute of Physics Publishing/ New Journal of Physics 1.200 GBP Deutsche Physikalische Gesellschaft Institute of Physics Publishing/ Journal of Cosmology and 1.400 GPB SISSA Astroparticle Physics Jagiellonian University Acta Physica Polonica B 500 EUR Oxford University Press/ Progress of Theoretical and A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics 13
Physical Society of Japan experimental Physics Springer/SISSA Journal of High Energy Physics 1.200 EUR Springer/Società Italiana di European Physical Journal C 1.500 EUR Fisica
Due to the SCOAP3 inherent capping mechanism which pays only for a certain amount of articles determined upfront with the publishers the current number of over 9.000 articles published via SCOAP3 lead to an average APC of ca. 1.100 EUR.
4. SCOAP3 Repository
The technical advantages of the SCOAP3 Repository (http://repo.scoap3.org) are: OAI-PMH RSS feeds API
5. SCOAP3 Panels
The SCOAP3 Governance comprises three panels taking care of the decisions concerning future steps of SCOAP3 as well as the day-to-day administration. SCOAP3 Executive Committee SCOAP3 Governing Council SCOAP3 Forum The SCOAP3 Executive Committee comprises five to seven members and oversees the SCOAP3 operations. The SCOAP3 Governing Council has up to forty-five representatives from contributing countries. They take their decisions about the direction of SCOAP3 during regular meetings at CERN in Geneva at least twice a year. The SCOAP3 Forum is a panel open to anyone interested in the latest developments in SCOAP3.
6. Why arXiv and SCOAP3?
Both arXiv and SCOAP3 are necessary publication tools for scientists. Both are Open Access tools enhancing the visibility and quick dissemination of scientific information which is advantageous for all scientists. There are good reasons to maintain these two important instruments in order to support the worldwide swift towards more and more Open Access in scientific publication.
6.1. Good reasons for arXiv arXiv constitutes the daily tool for scientist. Preprints uploaded by the scientists in arXiv are visible very quickly that is why their content is known sometimes long before 14 A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics being published in a journal. This influences the velocity of spreading scientific information in a positive way.
6.2. Good reasons for SCOAP3
Quality Journals as they are participating in the Open Access project SCOAP3 are highly necessary and important as the quality of the citation index depends on the quality of the journal an article is published in. Quality Journals provide for the necessary evaluation by the reputation of the journal itself which means quality which is important for both the scientists’ careers and their ability to get funding for their research.
7. National Project SCOAP3–DH supporting SCOAP3
7.1. German Partners for SCOAP3
On a national level there are three German partners supporting SCOAP3. SCOAP3-DH – participation of the German universities organised by TIB (German National Library of Science and Technology) HGF (Helmholtz Association)/DESY (German Electron Synchrotron) MPG (Max Planck Society)/MPDL (Max Planck Digital Library) All three organisations are National Contact Points (NCPs) and participate in the SCOAP³ panels for the institutions they represent.
7.2. Advantages of SCOAP3 for the German universities
Compared to an individual publication fund at each single university SCOAP3-DH offers the advantage of no administrative burden after having paid the lump-sum and “flat-rate” for as much publications in the SCOAP3 Journals as ever wanted.
Further advantages in short are: SCOAP3-DH is an all-inclusive solution for the HEP-Community (lump-sum). All-you-can-publish-flat-rate for HEP-Publications in SCOAP3-Journals. Reasonably-priced administration – due to centralization at CERN/TIB. No administrative burden regarding individual APC-handling for each university through publication funds with their financial restrictions.
All in all SCOAP3 as well as SCOAP3-DH are quite ahead of their time.
7.3. German Universities supporting SCOAP3
During this current first phase of SCOAP3 thirty university libraries plus one consortium are SCOAP3 partners contributing financially to SCOAP3-DH by redirecting their former subscription costs to the SCOAP3 fund. These current participants are listed on the national website under the following link: http://www.scoap3.de/scoap3-partner/nationale-scoap3-partner. A. Kutz / SCOAP3/SCOAP3-DH – Gold Open Access in High Energy Physics 15
Once the change to the output-based financing mechanism will have taken place the structure of SCOAP3-DH for the German universities might change slightly due to the fact that each publishing community of a university will have to support SCOAP3 financially and politically.
8. The future of SCOAP3/SCOAP3-DH
Currently all international partners are preparing a second phase of SCOAP3 which is planned to consist of prolonged contracts with the current publishers and journals in order to reach Gold Open Access for the respective high energy articles and journals for another three-year-period (2017-2019).
The decision whether APS (American Physical Society) which comprises further important High Energy Physics publications will join SCOAP3 for its second phase is still pending.
With or without APS the continuation of SCOAP3 is highly depending on the commitment of the scientific community to keep up both Open Access publication tools, arXiv as well as SCOAP3, and to enable its further financing in order to support and uphold quick as well as reputation-driven quality articles and their publication.
9. Repository, Websites and Information
Direct link to the SCOAP3 Repository: http://repo.scoap3.org
Further information about the international project SCOAP3 can be found under: www.scoap3.org
Information about the national project SCOAP3-DH is provided under: www.scoap3.de
SCOAP3 Newsletters provided by CERN can be found under: http://scoap3.org/news/scoap3-newsletter
Further Articles about SCOAP3: - http://cds.cern.ch/record/1735210/files/SCOAP3-APC.pdf - http://www.scoap3.de/fileadmin/dateien/Dokumente/Prof._Heuer_PJ02_2 012_31_PDF.pdf 16 Positioning and Power in Academic Publishing: Players, Agents and Agendas F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-16 COAR Case Study Controlled Vocabularies and PHAIDRA International
Imma SUBIRATa,1, Iryna SOLODOVNIKa, Paolo BUDRONIb, Raman GANGULYc, c Rastislav HUDAK
a Food and Agriculture Organization of the United Nations, Italy; b Department PHAIDRA, Vienna University Library and Archive Services, Austria;c Vienna University Computer Center, Austria
Abstract. Following the COAR-SPARC conference in Porto, the COAR Controlled Vocabularies Interest Group met on the 16th of April 2015 and had a detailed discussion about the new set of COAR controlled vocabularies, while proposing detailed actions to solve the remaining issues (mainly in the Resource Type vocabulary). Echoing aspects discussed about sustainability and organization (long-term technical support), as well as dissemination and implementation of COAR controlled vocabularies, the immediate projection of these issues has seemed to be quite feasible on PHAIDRA International long-term ecosystem. Nowadays this OAI-PMH environment consists of fourteen Open Access repositories disseminating their contents to EUROPEANA, OpenAIRE, OAPENLibrary, e-infrastructures Austria, and national CRIS etc. PHAIDRA International is taking all necessary steps to be constantly aligned with Trusted Digital Repositories criteria and harmonized (technically and semantically) in line with COAR Roadmap: Future Directions for Repository Interoperability. The COAR Roadmap stresses that still many challenges remain with improving interoperability. These involve standardization of controlled vocabularies in use, as well as metadata and indicators, in connection with state-of-the-art interoperability approaches supporting Linked Open Data. There is an urgent need for PHAIDRA International to harmonize the encoding description (metadata properties) according to LOD-enabling strategies, and to adhere to multilingual COAR controlled vocabularies” shared registry (Knowledge Base) services. COAR Controlled Vocabularies can be easily implemented in PHAIDRA in fixed formats and mapped to related persistent RDF/SKOS versions.
Keywords. COAR, Controlled Vocabularies, Linked Open Data, LODE-BD, PHAIDRA
1. PHAIDRA International
PHAIDRA (Permanent Hosting, Archiving and Indexing of Digital Resources and 2 Assets) International consists of fourteen Open Archives Initiative (OAI) persistent
1 Corresponding Author; E-mail: [email protected] 2 Available at: http://phaidra.org/; http://bibliothek.univie.ac.at/english/phaidra.html (Accessed: 19 February I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International 17
digital repositories in use internationally at universities in Austria, Serbia, Montenegro and Italy. PHAIDRA is involved in content aggregation, reliable storing/archiving, linking for creation and archiving of multi-resource content types with a view to long- term data preservation (LIBER, 2014). Documentation of digital contents - uploaded and aggregated in PHAIDRA repositories is accomplished by means of metadata assignment on several levels: single file (individual item), collection, container (multiple content datastreams) and paper (single file with relations to other objects). The metadata structure of PHAIDRA permits clear assignment of data covering contextual and provenance information (both analog and digital). The descriptive metadata structure of PHAIDRA consists of three widely-endorsed descriptive metadata standard schemes such as Metadata Object Description Schema (MODS) crosswalked to Dublin Core (DC), and Learning Object Metadata (LOM). PHAIDRA metadata includes also locally developed metadata (e.g., UWmetadata, crosswalked to DC and some other metadata solutions) implemented as minimum information needed to fulfill requirements of different designated communities (Digital Humanities Centers, OAPEN Foundation, OpenAIRE, e- Infrastructures Austria, EUROPEANA Libraries, local CRIS, institutional repositories). The overall interoperability between PHAIDRA and designated communities is in accordance with policies (based on technical, content, organization agreements) about the perpetual use of data, Open Access (OA) and specific community recommendations such as, for example, COAR Roadmap: Future Directions for Repository Interoperability (COAR, 2015) providing common meaning to the requested and exchanged services and data in order to lower possible (technical, syntactic, functional, semantic) conflicts.
2. Open Repositories towards Linked Open Data
As stated by Zeng & Chan (2015, p.5), “within the spectrum of different perspectives on interoperability, semantic interoperability lies at the heart of all matters”, and goes “beyond those implementing the canonical search paradigm f o r seeking relevant 3 information” . Metadata and Knowledge Organization Systems (KOS; names and subject heading authorities, ontologies, classifications, thesauri and other controlled vocabularies) are two related areas of most interoperability efforts. Without semantic interoperability, the meaning of the language, terminology, and metadata values in use cannot be negotiated or correctly understood. Different KOS have been already: shared through different registries (e.g., VEST, BARTOC.org, CKAN DataHub, European Union Open Data Portal, CIARD RING; COAR controlled vocabularies); aligned with each other (e.g., according to ISO 25964); linked on the frontline of the Semantic Web (SW, Web of Data), with the help of web-based tools (VocBench, SILK, PoolParty, Amalgame, Protégé) and other concept and (meta)data harmonizing approaches; published as Linked Open Data. Linked Data (LD) offers great potential to connect open repository 4 users to a
3 2nd Linked Open Data-enabled Recommender Systems Challenge (2015). Available at: http://sisinflab.poliba.it/events/lod-recsys-challenge-2015 (Accessed: 19 February 2016). 4 OR 2015 10th International Conference on Open Repositories (2015). Retrieved from: http://www.or2015.net/ (Accessed: 19 February 2016). 18 I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International vast array of interconnected heterogeneous sources found inside and outside of repository virtual walls. With LD every repository can provide access to linked specific datasets, without converting full metadata records. Metadata statements - rather than whole records - can be aligned and mashed up from a variety of datasets. Where the repository is based on highly specialised formats for the exchange of its metadata and authority files, open web standards (LD included) and programming can be used to process them (LOD-LAM Project, 2015). The COAR Roadmap stresses that Linked Open Data (LOD) and their provision have the most important interoperability function in the SW environment and, consequently, to be considered the best candidates for interoperable web services for repositories. By engaging both LOD-enabled metadata and KOS data values published as LOD trustworthy (i.e. maintained in long term by authoritative organizations) sources, repositories can augment access to the best (most relevant and most reliable) sources of the required information. The “supporting Linked (Open) Data” aspect in the COAR Roadmap is cross- linked with other (cross-independent) aspects, such as: improving metadata quality; extending/replacing metadata exposition protocols; supporting long-term preservation and archiving; extending usage of visualization tools (along with implementing best practices for search engine optimization); handling of complex/compound/nested repository objects (e.g., Enhanced Publications); monitoring OA mandate compliance; exposing versioning information; extending the bi-directional connectivity of repositories and their services (technical issues and strategic benefit). Information, data and knowledge modeled and processed according to LD techniques “practice at different levels of semantic precision surpass the current syntax-based possibilities in a qualitative fashion and thus can be processed, exchanged, referred and linked to different statement levels” (Isaac, Baker, 2015, p.35). To experience how LOD-ready data (metadata properties and controlled vocabulary value data pairs) represent and support navigation of (external) related contents, one can access AGRIS (International Information System for the Agricultural Sciences and Technology) platform5 and search for some contents. AGRIS leverages the power of descriptive LOD-enabled metadata (encoded according to LODE-BD Recommendations (Subirats, Zeng, 2015)) as well as AGROVOC LOD Thesaurus mapped (through s k o s : c l o s e M a t c h , s k o s : e x a c t M a t c h R D F / S K O S s y n t a x e s ) to other sixteen LOD-ready controlled vocabularies on its backbone (Subirats, Zeng, 2014), and openly shared via CKAN Data Hub6.
2.1. PHAIDRA towards Linked Open Data
SKOS, a modern well established SW standard, can potentially support: formal alignments and hierarchical grouping of concepts using different SKOS relations (e.g. skos:exactMatch, skos:closeMatch, skos:narrower, skos:broader, skos:related); translation of concept labels; URI-based mapping to similar concepts in other KOS. (Open Metadata Handbook, 2012).
5 Available at: http://agris.fao.org/agris-search/index.do (Accessed: 19 February 2016). 6 Available at: http://aims.fao.org/vest-registry/vocabularies/agrovoc-multilingual-agricultural-thesaurus; http://datahub.io/dataset/agrovoc-skos (Accessed: 19 February 2016).
I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International 19
Controlled vocabularies implemented in PHAIDRA International can be processed and aligned with existing trustworthy (based on persistent URIs) controlled vocabularies – such as EUROVOC Thesaurus and Dewey classification encoded in RDF using SKOS, Association for Computing Machinery (ACM) classification and Art & Architecture Thesaurus in LOD version – with help of a web-based, multilingual, editing and workflow tool VocBench v.2.3 (2015). All concepts controlled vocabularies in use should be rendered/identified with persistent URIs and stored in triple store Terminology server, in order to be reused.
2.2. Linked Open Data-enabled bibliographic data
Without syntactic interoperability - underpinned primary by metadata - data and information cannot be handled properly with regard to formats, encodings, properties, values, and data types; and therefore, they can neither be merged nor exchanged. Despite the importance of syntactic interoperability, still different repository data can be difficult to find on the Web. This means, that a number of repositories are losing public influence and impact rather than providing much needed leadership to the information age. To address this dilemma, the following questions may arise: (i) is there any consistent well-structured metadata modeling and encoding methodology (in spite of concurrent proliferation of metadata standards) that can ensure maximum interoperability among digital objects on the frontline of LOD?; (ii) are there significant metadata properties related to the reuse of digital data/information/knowledge and how can these be identified and expressed?; (iii) could repository management and the user experience really benefit from bibliographic data published as LOD? The already mentioned LODE-BD Recommendations were born to provide clear practical solutions for these cutting edge issues. LODE-BD is a reference tool providing practical support on significant metadata modeling decisions (in both depth and detail), metadata encoding (enabling data re-use) and implementation (satisfying local and/or specific needs), while insuring sharing of meaningful (with clear purport) and comprehensive (both to humans and web engines) bibliographic data. LODE-BD also clarifies what kind of KOS values should be used to produce high-quality LOD- enabled bibliographic data, easily exchangeable through open repositories and sharable on the SW. The present use-case proposes to align the encoding of metadata properties in all PHAIDRA RDF-aware repositories according to LOD-BD strategies encouraging the reuse of different LD-enabled KOS data values in property–value statement pairs.
3. COAR Vocabularies. Some alignment and implementation issues
Focusing on open repository interoperability issues, COAR vocabularies (2015) - published in traditional/fixed formats and anchored to URI-RDF/SKOS with the help of the VocBench tool - represent a high recommendable set of standardized multilingual controlled vocabularies. Concept labels of these latter can be easily used to encode metadata properties, thus providing the cornerstone of effective representation of bibliographical records for different communities of users. 20 I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International
The COAR Resource type controlled vocabulary has largely been compared to other vocabularies (used in repositories), to address a number of alignment issues, as follows: resource type: comparison performed with DataCite metadata kernel, Elsevier CrossMark, CASRAI Dictionary for publications, CERIF Semantic Vocabulary, dcmi_terms, e-LIS, Gateway to ResearchData (GtR), PubMed, PURE, RIOXX, SWAP schemes; access rights and versions: comparison performed with RIOXX, NISO- ALPS-JAV schemes, EPrints access rights vocabulary; grant agreement identifiers: comparison performed with RIOXX and FundRef schemes; resource identifier schemes: comparison performed with LOC identifiers vocabulary; date types and value:comparison performed with DataCite, dcmi_terms, SWAP, Bibo-Ontology; classification concepts: LOC classification scheme vocabulary.
COAR vocabularies formats are replicating good practices of other open communities (e.g. CLARI77) publishing controlled vocabularies. COAR vocabularies adhere to the following principles: open vocabularies and open standards (under open licenses) are preferred over proprietary standards; formats and protocols are well- documented, verifiable, proven (being used in practice). COAR vocabularies can be implemented in repositories through plug in to import an RDF into the systems and integration through XML, with the name space for the vocabularies
Figure 1. Data flow components for the COAR Controlled Vocabularies.
7 Available at: https://www.clarin.eu/content/standards-and-formats; https://openskos.meertens.knaw.nl/clavas/web-report/documentation/clavas-overview.html (Accessed: 19 February 2016).
I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International 21
COAR vocabularies will be exported in PHAIDRA in RDF and HTML formats. Afterwards a list of all required (from COAR community) export functionalities will be investigated and a survey to identify the needs in terms of data (vocabulary) ingestion and data publication will be prepared.
Conclusions
In respect to the use-case PHAIDRA International - on which COAR controlled vocabulary community, is working on - there are some final considerations/proposals that can become actions/tasks to be undertaken: (1) all concepts of KOS implemented in PHAIDRA fixed formats should be turned to (persistent) URI-RDF/SKOS notations. Existing URI-RDF/SKOS versions of KOS should be re-used; (2) all concepts of KOS turned into URI-RDF/SKOS should be mapped/switched-off/crosswalked/cross- referenced on COAR vocabulary backbone based on LOD-ready concepts. The implementation of COAR controlled vocabularies in PHAIDRA repositories will not only expand semantic expressivity of metadata in use, but also provide normalization and validation of metadata properties, according to standard COAR community requirements included as default for research contents, firstly at the European level. Normalization and validation of PHAIDRA (meta)data properties through COAR vocabularies will certificate PHAIDRA as a trustworthy member (Solodovnik, Budroni, 2015) of European research community, highly and reciprocally tuned with OpenAIRE/Zenodo, DANS, Dataverse Network and other national and international research infrastructures. Last but not least, COAR vocabularies in PHAIDRA International will serve as input for language tools to manipulate metadata records at international level, as well as to improve usability of PHAIDRA multilingual user interfaces according to hierarchical levels of the COAR vocabularies. By focusing on PHAIDRA International use case, COAR controlled vocabulary group (in cooperation with other COAR working groups) is planning to document serialization of specific formats, and fetching strategies to retrieve COAR vocabularies from web services and to import them into repository platforms.
References
COAR Vocabularies. Available at: https://www.coar-repositories.org/activities/repository- interoperability/ig-controlled-vocabularies-for-repository-assets/coar-vocabularies/. (Accessed: 19 February 2016).
COAR (2015) COAR Roadmap: Future Directions for Repository Interoperability. Available at: https://www.coar-repositories.org/activities/repository-interoperability/ (Accessed: 19 February 2016). Isaac, A., Baker, T. (2015). Linked data practice at different levels of semantic precision: The perspective of libraries, archives and museums, Bulletin of the Association for Information Science and Technology, 41(4), 34-39. Retrieved from: https://www.asist.org/publications/bulletin/apr-15/ (Accessed: 19 February 2016).
Kent State University (updated 2015). LOD-LAM Project. Retrieved from http://lod- lam.slis.kent.edu/about.html (Accessed: 19 February 2016).
22 I. Subirat et al. / COAR Case Study Controlled Vocabularies and PHAIDRA International
LIBER (2014). LIBER Case Study: Factors for Enabling Sharing and Reuse of Research Data – Library and Archive Services at the University of Vienna. Retrieved from: http://libereurope.eu/wp- content/uploads/2014/06/LIBER-Case-Study-UNVIE1.pdf (Accessed: 19 February 2016). Open Metadata Handbook (2012), in Wikibooks. Retrieved from: http://en.wikibooks.org/wiki/Open_Metadata_Handbook/Recommendations#Switching-accross (Accessed: 19 February 2016).
Solodovnik I., Budroni P. (2015) “Preserving digital heritage: At the crossroads of Trust and Linked Open Data ‘, IFLA Journal, 41 (3). Available at: http://www.ifla.org/files/assets/hq/publications/ifla- journal/ifla-journal-41-3_2015.pdf (Accessed: 19 February 2016).
Subirats, I., Zeng, M. L. (2015) “LODE-BD Recommendations 2.0: How to select appropriate encoding strategies for producing Linked Open Data (LOD)-enabled bibliographic data”. Food and Agriculture Organization of United Nations. Retrieved from: http://aims.fao.org/lode/bd; http://eprints.rclis.org/22454/1/LODE-BD-2.pdf; http://aims.fao.org/lode/bd-2/step-forward#why (Accessed: 19 February 2016).
Subirats, I., Zeng, M.L. (2014). ‚Using KOS as the Connectors of Linked Datasets”. The 77th Annual Meeting of the Association for Information Science and Technology, Seattle, WA, USA. Retrieved from http://www.asis.org/asist2014/ (Accessed: 19 February 2016).
VocBench 2.3 (2015). Available at: http://vocbench.uniroma2.it/downloads/; http://vocbench.uniroma2.it/; http://vocbench.uniroma2.it/support/; http://lists.w3.org/Archives/Public/public-esw- thes/2015Apr/0004.html; VocBench User Manual: http://vocbench.uniroma2.it/documentation/VocBench_v2.1_user_manual.pdf (Accessed: 19 February 2016).
Zeng, M. L., Chan, L.M. (2015) ‘Semantic Interoperability”, in Encyclopedia of Library and Information Sciences 4th ed
Positioning and Power in Academic Publishing: Players, Agents and Agendas 23 F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-23 “If there are documents you really care about: Print them out!” (Vint Cerv, 2015)
Bernd KULAWIK1 Stiftung Bibliothek Werner Oechslin, Einsiedeln, Switzerland
Abstract. With theses words (only “documents” substituted for the original “photos”) Vint Cerf, one of the ‘fathers of the internet’ and now Google Vice President, warned in 2015 that all our photos – and obviously, documents and research data, too – might disappear soon and that our century may become the “Digital Dark Age”. To avoid this, Cerf is working on a solution named “digital vellum”: It shall provide a platform that can preserve any documents, the software used to create and work with them, the operating system needed for this software and even an emulation of the appropriate hardware. But it may take quite some time before this platform will be available. In the meantime, the good old paper is the only medium that surely can and will survive more than 50 years – the maximum now expected for simple formats like .txt and .pdf files. Even Microfilms (also not usable without technical means) may not survive more than 200 years. But how do we print out digital documents created for and by research: short miscellanea, articles and papers, collections of them and monographs, and in recent years even facebook postings or twitter messages? We write these documents in a dedicated (text) program, sent them to the publisher, who may forward them after several transmissions forth and back with the author(s) to a layouter, again followed by some corrections requiring exchange of the file(s) … and finally they may appear in print and / or online repositories. Taking into account that all participants in the process today are (or should be) well familiar with web-based Content Management Systems and – hopefully – the concept of markup languages, it is simply astonishing that there is no system yet combining the advantages of both. Such a combination could not only serve to shorten the publishing process but also provide the ecosystem for online repositories and web- based collaboration while the results – printable documents – could be updated regularly and made available via book-on-demand and as ePublications. There may be some solutions providing such a system used by publishers “in-house”, but if so, they are not available for free. The paper will propose such a system based on Free and Open Source Software with a simple proof-of-concept.
Keywords. LaTeX, Content Management Systems, Free Software,
Electronic publication in the humanities and other fields is still a process following the centuries-old model developed for paper:
1. Authors write papers about research results, including images and tables, using a ‘word processor’ producing a digital document in a proprietary format that hardly can be opened with other software without formatting or information loss. 2. The digital document is sent to the publisher.
1 Corresponding Author: [email protected] 24 B. Kulawik / “If There Are Documents You Really Care About: Print Them Out!”
3. The publisher (a person working at the publishing house and responsible for this publication) sends the electronic file to someone checking it for errors and guidelines. 4. Before the paper is regarded as ‘finished’, it is sent at least once back to the author for some sort of ‘polishing’. 5. Steps 2 – 4 usually are repeated several times … 6. Finally, the ‘final’ version of the paper is sent to the layouter who transforms the file from the word processor’s format into another usually also commercial and proprietary format, and reworks the entire text according to the layout guidelines. 7. The result may be sent back to the author again for final corrections in a format that he can not change. 8. After the last final ‘final’ reworkings, the file is transformed again into a format suitable for printing and sent to the printer. 9. The printing machine produces the (e.g.) book with binding etc. in the number of the first edition. 10. The printed books are sent to the bookstores or kept in storage until they are ordered. 11. For a few years now, this process is split up after step 7 into two: the second way is the publication in an e-book format.
Except for the usage of digital files sent around several times in several versions via e-mail or some cloud storage, all of this is still identical with the ‘good ol’ paper process’ – and often authors and publishers repeat some of these steps by using prints on paper. So, could this really be called “electronic publishing”? My answer would be: No! In 1991 Tim Berners Lee developed the World Wide Web mostly based on already existing techniques and HTML, to shorten this process. Since the year 2000 many steps of this process could be shortened and united with the help of web-based, Content Management Systems. But 25 and 15 years later, respectively, these systems with databases and versioning are still not used for (printable, well-formatted) publication(s). The title of my paper is a slight derivation of a quotation from Vint Cerf. Cerf is the (co-) developer of the basic protocol used for the internet, the TCP/IP, and, therefore, one of the ‘fathers of the internet’. Since the early 1970s he took part in crucial developments. Today he is one of Google’s Vice Presidents. Therefore, we should take his warnings regarding the looming ‘Digital Dark Age’ of our data as well as the lack of any solution for the long-time preservation of digital at least seriously. He said these words in February 2015 at the annual meeting of the AAAS. The suggested solution he is working on is the “digital vellum”: It shall provide a soft- and hardware environment able to preserve not only digital documents but also the software used for their creation and the operating systems as well as the (emulation of) the hardware needed. Let’s put aside the (crucial) questions of (commercial) licences which usually would not allow to run the software in everchanging virtual environments. And let’s also put aside other questions about those digital data that are not documents in a broad sense: Most of the databases today may still be able to print out the entries in one set of data, but its form surely will not be sufficient to be regard as a (printed) scientific publication. And, of course, the pure amount of data (sets) will make it impossible to print them out after every change. Let’s in addition put aside questions regarding online B. Kulawik / “If There Are Documents You Really Care About: Print Them Out!” 25 platforms where data, layout styles and their definitions as well as software for specific functions etc. are distributed over servers all over the world and combined ad-hoc. (This should cause a fundamental problem for the “digital vellum”.) So, after letting aside all of these fundamental questions and problems: should we follow Vint Cerf’s advice anyway and print out our photographs and other documents as long as his “digital vellum” is not available yet? — Of course! Because no-one can guarantee the preservation (let alone: usability) of any digital data format for more than 20, let alone 50 years. Archives plan to make sure that simple and open formats like TXT and PDF and image formats like TIFF or JPEG will be available for up to 50 years, but even this is not sure. File formats like Microsoft Word’s .doc are even definitely excluded by archives. Based on experiences with digital data formats from the past 30 years, I surely would doubt any possibility for such a long timespan of preservation: When I started programming in the early 1980s, the answer to the question “How should we store your data for a longest time possible?” would have been: “I put the paper punched tape into a dry and cold armoire.” — So, at the moment there is no way to make sure that the digital data we create and work with will be usable even during our own lifetime! Should the creation of such data not be regarded as waste, even willful destruction of life and working time as well as resources? So: Let’s print them out, at least those worth surviving the next 50 years! But with this decision, another problem arises: The layout of our digital documents as they appear on the screen is usually not very satisfactory, often not even acceptable for scientific publishing. This is even more astonishing when we take into account that the basic tools for scientific publication, mostly word processors and type setting programs, have by now been available for more than 25 years. In addition, the tools for web-based content management able to replicate the publishing processes described above also have been available as Content Management Systems by now for at least 15 years. And both, publishing tools as well as web-based CMS are using Markup Languages. So why is there no unification of both, one widely spread and used as a standard tool and based on Free Software? As far as I know, at the moment there is no such software system available. Some publishing houses use web-based editing tools that their authors may use, but — according to their warnings — even those tools do not generate a document that looks exactly like the one that finally will be printed. And in some of the cases versioning or commenting seems to be difficult. But these tools are ‘private’, closed source, ‘in-house’ applications, not available to others. The by far larger group of professional publishing houses does not even offer such tools: They require their authors to follow their specific guidelines, written in MS Word or PDF documents with more than one hundred pages that authors are expected to read first and fully keep in mind! Some publishers offer MS Word templates that can be filled by the authors with their own texts. Only a small number of publishers also offers LaTeX templates. But these are mostly directed to the natural sciences; to humanists and historians, LaTeX and its free interface tools usually are unknown. So, especially these authors are bound (or bind themselves) to non-free word processing software and operating systems and regularly encounter problems should they, for instance, try to reuse or re-work their own documents written some 10–15 years ago with earlier versions of this very same office software. The solution, from my point of view, would be a combination of a free and open source Content Management System like Plone (based on the free Web Application Server ZOPE and its object-oriented database ZODB, both written in Python) and 26 B. Kulawik / “If There Are Documents You Really Care About: Print Them Out!”
LaTeX: Both, Plone and LaTeX, can be run on any operating system, and a working proof of concept solution already exists: It is called ftw.book, provided as a Plone module by the Swiss software company 4teamwork. Because it is free, it could be extended by anyone with some experience in Python and LaTeX into a more general tool, e.g. offering different LaTeX-based designs and document classes or new layouts preferred by the publisher or institution.
What are the advantages of such an extended version of ftw.book? Because of the customizable user rights and role management, the entire process described above can be applied to the CMS and adapted to almost any special need of institutions or publishers. The CMS versioning allows to go forth and back in the editing process and keep control over the versions at any time. No document versions have to be sent multiple times between the participants of the publishing process, because they remain in one place, available to any authorised person. The available formatting functions offered in the web interface can be limited to avoid authors breaking them — a big problem in all word processing programs, causing lots of additional work. The final print layout is always available for controls. This process is protected by the CMS against unauthorised access. The final document can be made (un) available over the internet with a few clicks, even if a print version is not yet on or intended for the market. But any authorised reader may print a PDF copy. Changes can be easily done while the original is still available. Different editions are available at any time, so that links set to an old edition will not break because a new one has been published. This would allow, e.g., to update a book on an almost daily basis: When a change has been made to its content, it could not only immediately be online, but also appear in the next printed copy.
While these and other advantages regard the publishing process of scientific documents, there are even more important advantages:
Not only publishing houses could use such a system, but any institution, group or private person. The software could be used to build up scientific repositories, e. g. for Open Access strategies. Because all components of the software scale very well from laptops to (groups of) servers, it would be possible to have a copy of the system run on the personal computers of members of an institution or students, e.g.: They could work on their texts even when they are offline, syncing all preserved versions later while observing the layout required by their institution or publisher. A simple syncing tool available for ZOPE guarantees the identity and integrity of the documents on the ‘official’ servers with those on the local computers or laptops. With the rapidly growing technical possibilities of handhelds, these should be able to run the entire software system very soon and serve the data to the B. Kulawik / “If There Are Documents You Really Care About: Print Them Out!” 27
internet or synchronise them with the server(s). (Any Ubuntu-based tablet already could do this today.) The freedom of all components allows for constant development and adaptation of the entire system: from the underlying operating systems to the hardware. Of course, it would be useful to have large research institutions provide central repositories of the freely available parts. Those institutions could even provide hosting services for projects and, vice versa, require these projects to make their results available online via their repositories in any Open Access strategy suitable. Of course: not every paper, article, book etc. would have to be printed: But every one could be printed and, therefore, according to Vint Cerf’s suggestion, be preserved even for a distant future.
So, everything seems to be wonderful with this suggested solution — or are there disadvantages? Of course, there are some: For instance, if the usability and standard conformity of the system should be preserved, this would radically restrict the many ‘bells & whistles’ often used in the research projects: Everything that does not fit on a (large) page would have to be excluded. Well, not completely: It would be possible, e.g., to have large images with very high resolutions, annotations, links etc. connected to the reduced images or the data in the printable version. But, of course, such high resolution images, documents or data sets surely will not survive as long as the printed counterpart or ‘mother document’. Another problem could arise from projects where data and information are intrinsically very closely linked to each other. This would make it almost impossible to represent them in a printable form. In these cases a solution could lie in the generation of reduced ‘abstracts’ or reduced data sets that would be imported automatically from the original database(s) into the suggested system and then be formatted for printing. Again, one would lose some data — but, depending on the decisions made regarding the exported data sets, at least part of the work and resources put into these projects could be preserved for ‘eternity’. The proposed system would not only establish a real environment for electronic publishing for the first time, but also provide a solution for the looming dangers of the ‘Digital Dark Age’ that Vint Cerf and others are warning about and archivists and librarians are or should be aware of. One could even imagine that such system could develop into a general standard for publishing and digital preservation. Commercial software then would have to offer plugins to allow its users to publish their texts without having to leave their ‘familiar’ word processor. For scientific database projects it could offer a solution in the form of repositories that would help to avoid masses of data compiled over years being lost after a short time — just because, e.g., the financial support has been turned off. In cases where server systems spread all over the world are used, e.g. som ‘facebook’ of science, there should at least be plug-ins to the suggested solution to create printable documents at any time. For such already or soon also very common cases, I do not even see a future in Vint Cerf’s “digital vellum”.
28 Positioning and Power in Academic Publishing: Players, Agents and Agendas F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-28 Interactive Public Peer ReviewTM: an innovative approach to scientific quality assurance
a, 1 Xenia VAN EDIG aCopernicus Publications, Bahnhofsallee 1e, 37075 Göttingen, Germany [email protected]
Abstract. Besides providing open access to the article, Copernicus Publications provides open access to the peer review via its Interactive Public Peer ReviewTM. In this process, a public discussion among the author, two independent referees, and interested members of the scientific community builds the core of the peer- review process. Keywords. Peer review, open access, transparency
1. Introduction The discussions surrounding peer review are ongoing. Several authors are claiming a crisis of peer review with regard to its length (Nguyen et al. 2015; Powell 2016) and effectiveness (Lee et al. 2013; Walker R. and Rocha da Silva, 2015), and researchers are calling for more openness in the process (Aleksic et al. 2015). Copernicus Publications already developed a new form of peer review in 2001 (Pöschl 2012). Since then, the process has been implemented in different scientific disciplines and enhanced continuously. Today, 18 open-access journals published by Copernicus Publications apply this form of peer review. In addition, an economy journal also applies this kind of peer review. In the following, the initial idea and the development of the process of Interactive Public Peer ReviewTM are described.
2. Interactive Public Peer ReviewTM When the concept of interactive open-access publishing and Interactive Public Peer ReviewTM was developed by Ulrich Pöschl and his fellow scientists in 2000, they faced the problem that the traditional journal publication and peer-review process were not sufficient for thorough quality assurance, constructive discussion, and integration of scientific knowledge: the majority of studies did not build on related earlier publications, and some studies were not even self-consistent even though they had been published in reputable journals with high impact factors. After long discussion, Pöschl and his colleagues were convinced that public review on the Internet would provide the 1 Corresponding author: [email protected] X. Van Edig / Interactive Public Peer ReviewTM 29 opportunity to resolve or at least improve many of these issues. With the Nobel Prize winner Paul Crutzen, the new concept found a prominent supporter (Pöschl 2011). Through the rapid publication after a swift access review, scientists receive a fast record of their research as a discussion paper. The process enhances transparency as referee comments, author comments, and the comments of the scientific community are published in the interactive public discussion (online and open access). However, the process meets the criteria of traditional quality insurance as papers undergo revisions and are only published as final revised papers in the journal after final acceptance by the editor. In summary, the process is designed to • foster scientific discussion; • maximize the effectiveness and transparency of scientific quality assurance; • enable rapid publication of new scientific results; • make scientific publications freely accessible. Thus, the new process was intended to provide both rapid scientific exchange and thorough quality assurance (Pöschl 2012). In contrast to post-publication peer review, the process of scientific quality assurance takes place prior to the formal journal publication. The discussion paper is just the manuscript submitted by the authors and therefore the starting point of the peer-review process. In addition, reviewers can disclose their names, but they do not have to do so as in open peer review. In 2001, the first journal to apply this new peer-review process, Atmospheric Chemistry and Physics (ACP), was launched by Copernicus Publications with the support of the European Geophysical Society (EGS), which has been part of the European Geosciences Union (EGU) since 2002 (Pöschl 2011). Since 2001, 17 other journals (14 sister journals of ACP and 3 journals not affiliated to EGU) have adopted this innovative review process. In addition to the journals published by Copernicus Publications, the Economic E-Journal has also adopted this form of peer review. But how does it work?
TM Figure 1. Example workflow of Interactive Public Peer Review 30 X. Van Edig / Interactive Public Peer ReviewTM
2.1 Access review After submission, the manuscript is swiftly reviewed by the topical editor who agreed to handle the review process. In this first assessment, the topical editor decides whether to start the discussion or not. Reasons for not starting the discussion might be a lack of basic scientific or language quality or the manuscript is not within the journal’s scope. Some journals provide the possibility to request the feedback of independent referees already at this stage. However, experience shows that consulting referees at this point unnecessarily prolongs the process. In addition, referees who are not used to the process sometimes already provide full referee reports, which are not needed prior to the discussion. During this stage, only technical corrections or minor revisions can be requested. 2.2 Interactive public discussion After a positive outcome of the access review, the author’s manuscript is published as a discussion paper. At least two independent referees – who are nominated by the topical editor – review the manuscript and post their referee reports as referee comments (RCs) on an online discussion forum. This forum is openly accessible on the Internet. While the reports are open access, the referees can decide whether they want to disclose their names during the discussion or not. Research shows that about 80% of the referees decide to stay anonymous during the Interactive Public Peer ReviewTM, while ca. 20% of them decide to disclose their name. In addition to the referees, the scientific community is invited to participate in the discussion and to post short comments (SCs). The authors of short comments have to register, and their names and contact details are shown in the discussion (Pöschl 2012). Usually, the interactive public discussion lasts 6–8 weeks depending on the journal. Before a discussion can be closed, at least the two RCs have to be published alongside the discussion paper. The author can decide to answer each comment individually or to address all comments collectively. Figure 2. Example of an interactive public discussion X. Van Edig / Interactive Public Peer ReviewTM 31
To guarantee the author’s publication precedence and to provide a lasting record of the review process, every discussion paper and its comments remain online and are individually citable (Pöschl 2012). This occurs regardless of whether or not a manuscript is accepted for publication as a final revised paper in the journal. 2.3 Final response and peer-review completion After the discussion has ended, the author should address all comments in a final response, if he or she did not do so during the open discussion. During this stage also the editor has the opportunity to post comments and suggestions (Pöschl 2012). Formal editorial recommendations and decisions shall be made only after the authors have had an opportunity to respond to all comments, or if they request editorial advice before responding.
Depending on the journal, the next step is one of the following: • The authors submit their revised manuscript. In this case, the topical editor – in view of the access peer review and interactive public discussion – either directly accepts/rejects the revised manuscript for publication in the journal or consults referees in the same way as during the completion of the traditional peer- review process. If necessary, additional revisions may be requested during peer-review completion until a final decision about acceptance/rejection for the journal is reached (Atmospheric Chemistry and Physics, website, 2016). • The topical editor makes a post-discussion decision in which he or she, based on the responses, either invites the authors to submit a revised manuscript or directly rejects the manuscript. If necessary, he or she may also consult referees in the same way as during the completion of the traditional peer- review process (Biogeosciences, website, 2016). 2.4 Publication of final revised paper In the case of acceptance, the final revised paper is typeset and proofread. Then it is published on the journal’s website, and the preceding discussion paper and the interactive discussion are displayed in a “peer-review tab” alongside the article. In addition, many journals display all referee and associate editor reports, the authors' response, as well as the different manuscript versions of the peer-review completion. All publications (original paper, interactive comments, and final revised paper) are permanently archived and remain accessible to the public via the Internet, and final revised papers are also available as print copies. The articles are also distributed via various abstracting and indexing services as well as other databases worldwide. 2.5 Interactive Public Peer ReviewTM in various disciplines This model is mainly utilized in the geosciences. However, it is also applied to other disciplines such as drinking water engineering and wind energy science. In the table below all journals published by Copernicus Publications that apply the Interactive Public Peer ReviewTM are listed. One can see that it is applied in various subdisciplines within the geosciences ranging from geophysics to atmospheric sciences: 32 X. Van Edig / Interactive Public Peer ReviewTM
Table 1. Journals applying the Interactive Public Peer ReviewTM published by Copernicus Publications
Title Access review with Post-discussion referee quick reports editor decision Atmospheric Chemistry and Physics (ACP) yes no Atmospheric Measurement Techniques (AMT) yes no Biogeosciences (BG) yes yes Climate of the Past (CP) no yes Drinking Water Engineering and Science (DWES) no yes Earth Surface Dynamics (ESurf) no no Earth System Dynamics (ESD) no no Earth System Science Data (ESSD) no no Geoscientific Instrumentation, Methods and Data Systems (GI) yes no Geoscientific Model Development (GMD) no no Hydrology and Earth System Sciences (HESS) no yes Natural Hazards and Earth System Sciences (NHESS) no yes Nonlinear Processes in Geophysics (NPG) no no Ocean Science (OS) yes no SOIL (SOIL) no yes Solid Earth (SE) no no The Cryosphere (TC) no no Wind Energy Science (WES) no yes
In the past years, a range of journals have switched from the traditional peer- review model to public peer review. In many cases, the interactive public discussion only consists of two referee comments and the author’s reply. However, providing the scientific community withthe opportunity to contribute to the discussion is a crucial aspect. Other papers on “hot topics” such as climate change or radioactivity sometimes receive 30–100 comments. There is an overview of these “most commented papers” in each journal's online library. Prior to the discussion (i.e. during the access the review) rejection rates vary among journals and are found to be 8–37%. After the discussion, which aims to improve the quality of the manuscripts, the rejection rate is only 8% on average. 3. Recent developments After the implementation of the accelerated access review (i.e. the access review without the possibility of consulting referees), the launch of the post-discussion editor decision (more guidance for authors after the discussion), and the adoption of the post- discussion report publication by most journals (i.e. disclosure of all reports from peer- review completion after final acceptance), a major adjustment to the concept was the decision to no longer typeset discussion papers from 2016 onwards and to merge the libraries of discussion papers and final revised papers. Thus, discussion papers now look less like a publication and more like pre-print papers. The discussion paper is the PDF uploaded by the author, with an added header indicating the journal to which the manuscript was submitted for review. The manuscript is still citable, but the citation will indicate that the paper is under review (Copernicus Publications 2015). The discussion papers do not receive a subsequent pagination anymore, but a DOI is still registered for them. In order to emphasize that the discussion paper is only the first step to the final X. Van Edig / Interactive Public Peer ReviewTM 33 paper, discussion papers are no longer archived in volumes and issues in a separate online library. With the new concept, a final revised paper and its corresponding discussion paper are archived together. There is a main page that includes all the information relating to the paper in separate tabs, such as metrics, related articles, and the list of peer-review comments and the discussion paper (Copernicus Publications, 2015) These actions should address two main obstacles which occurred in the past: on the one hand, it should prevent authors from citing the discussion paper instead of the final paper; on the other hand, it should help authors whose discussion papers were rejected by indicating more clearly that discussion papers are not to be regarded as formal publications and thus can be submitted to other journals. With the new concept for our interactive journals, we also introduced a new payment concept. Before 2016, authors were obliged to pay solely for the publication of their discussion paper if the respective journal had APCs. This concept is now obsolete since we no longer provide formatting services for discussion papers. Furthermore, funders welcome the new payment concept since now they are paying for the final revised paper and hence the version of record.
References Aleksic J, Alexa A, Attwood TK et al. (2015) An Open Science Peer Review Oath [version 2; referees: 4 approved, 1 approved with reservations] F1000Research, 3:271, doi: 10.12688/f1000research.5686.2
Atmospheric Chemistry and Physics (2016) http://www.atmospheric-chemistry-and-physics.net/peer_review/ interactive_review_process.html (accessed 07.01.2016)
Biogeosciences (2016) http://www.biogeosciences.net/peer_review/interactive_review_process.html (accessed 07.01.2016)
Copernicus Publications (2015) New concept for interactive journals, news item of 14 October 2015, http://www.copernicus.org/news_and_press/2015-10-14_new-concept-for-interactive-journals.html (accessed 07.01.2016)
Lee, C. J., Sugimoto, C. R., Zhang, G. and Cronin, B. (2013) Bias in peer review. J. Am. Soc. Inf. Sci., 64, pp. 2–17, doi: 10.1002/asi.22784
Nguyen V.M., Haddaway N.R., Gutowsky L.F.G., Wilson A.D.M., Gallagher A.J., Donaldson M.R., et al. (2015) How Long Is Too Long in Contemporary Peer Review? Perspectives from Authors Publishing in Conservation Biology Journals. PLoS ONE 10(8), e0132557, doi:10.1371/journal.pone.0132557
Powell, K (2016) The Waiting Game, Nature, 530, pp. 148–151, doi:10.1038/530148a
Pöschl U. (2012) Multi-stage open peer review: scientific evaluation integrating the strengths of traditional peer review with the virtues of transparency and self-regulation, Front. Comput. Neurosci, 6:33, doi: 10.3389/fncom.2012
Pöschl U. (2011) On the origin and development of Interactive Open Access Publishing, A short History of Interactive Open Access Publishing, Copernicus Publications, Göttingen
Walker R. and Rocha da Silva P. (2015) Emerging trends in peer review—a survey. Front. Neurosci. 9:169, doi: 10.3389/fnins.2015.00169, 34 Positioning and Power in Academic Publishing: Players, Agents and Agendas F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-34 Renegotiating Open-Access-Licences for Scientific Films
1 Elke BREHM Technische Informationsbibliothek (TIB) // German National Library of Science and Technology
Abstract. Scientific publishing is not limited to text any more, but more and more extends also to digital audio-visual media. Thus services for publishing these media in portals designed for scientific content, oriented towards the demands of scientists and which comply with the requirements of Open Access Licenses must be provided. Among others, it is the goal of the Competence Centre for Non-textual-materials of TIB to collect, archive and provide access to scientific audio-visual media in the TIB AV-Portal under the best possible (open) conditions. This applies to older films, as for example the film collection of the former IWF Knowledge and Media gGmbH i. L. (IWF) and to new films. However, even if the acquisition of the necessary rights for audio-visual media is complex, the renegotiation of Open-Access- Licenses for older films is very successful. This paper focuses on the role of Open Access in the licensing strategy of TIB regarding scientific films, the respective experience of TIB and the presentation in the AV-Portal, but also touches upon prerequisites and procedures for the use of Orphan Works.
Keywords. Open Access, Scientific Films, IWF Knowledge and Media, Licensing
1. Introduction
Scientific publishing is not limited to text anymore but more and more comprises other objects such as research data, 3D objects and digital audiovisual material among others. Thus opportunities are needed to publish different types of scientific content in an environment specifically designed for the object type and adapted to the needs of scientists. Among others, TIB collects, archives and provides access to scientific audiovisual media in its TIB AV-Portal2. TIB puts a special emphasis on Open Access and strives to offer its content under free and open conditions as close to the definition of Open Access contained in the Berlin Declaration of 20033 as possible. This is valid for older films such as the film collection of the former IWF Knowledge and Media as well as for the new scientific films which TIB acquires continuously.
When libraries acquire analogue film copies, the legal situation is relatively clear because the possibilities of utilisation are already defined in the German Copyright Act and copying constitutes a practical hurdle for analogue film formats. The legal situation for digital films and films posted on the internet, however, is more complex: The permits for analogue material regulated in the German Copyright Act cannot simply be
1 Corresponding Author: [email protected] . 2 TIB AV-Portal. Available at https://av.tib.eu (Accessed: 4 March 2016). 3 Berlin Declaration on Open Access to Knowledge in Sciences and Humanities. E. Brehm / Renegotiating Open-Access-Licences for Scientific Films 35 transferred to digital material. In order to be able to make a film available online, all of the permits required must be agreed upon in license agreements. In contrast, creating copies of digital copyrighted material is child’s play. There is a great danger that films will be passed on as digital copies (against the creator’s will) because the process is such a simple one. In particular with regard to the publication of films on the internet, not only copyright has to be observed, but also publicity and personality rights of the filmed or vocally recorded individuals. This factor can usually be neglected in the case of texts. At the end of 2012, TIB was entrusted with taking on the scientific film collection of the former IWF Knowledge and Media and, ideally, making it available to the public via its online portals. The collection comprises around 11,500 analogue and digital scientific films related to various subjects. The collection revolves mainly around technical and scientific subjects, as well as biology and ethnology. Although most of the publications were created between the 1950s and 1980s, the collection also contains a number of earlier cinematographic works. Unfortunately not all films are available in digital form. In addition to a lack of funding for digitizing such materials, TIB faces the problem of having to clarify how the films may be used. The rights in the IWF Knowledge and Media Film collection are very heterogeneous. After all, particularly when films are rather old, the agreements with the creators involved in the film production were also concluded at a time before digital use options on the internet became known and customary. In addition, many necessary changes were made in copyright law to adapt it to modern needs in the digital age. These options are therefore not included in the agreements concluded at the time with the creators. Moreover, usually many different people are involved in many different ways in the creation of a film – people who may potentially have rights to the film. Scriptwriter, director, cameraman, cutter, producer, performing artists, narrator … Often it is not easy, or no longer possible in retrospect, to establish who was actually involved in the production of the film. When attempts are made to renegotiate rights, many film authors cannot be traced. Thus, before being able to make the films available to the public via the TIB AV- Portal or through other services of TIB, it is obligatory that the legal situation regarding each film is examined and each author or rightsholder is contacted individually. As far as necessary and feasible, TIB renegotiates the necessary rights to be able to offer an up-to-date service to the customers. The renegotiation of rights among other things focuses on the rights needed to offer access to the films via the TIB AV-Portal, to create derivates in alternative film formats and perform the video analysis in the TIB AV-Portal. It has become customary to make bibliographic metadata freely available under the Creative-Commons-License CC0 1.0 Universal Public Domain to spread information and knowledge about the collection. When contacting each author individually, TIB offers the authors and rightsholders the possibility to release the films under an open access-license of the Non-Profit-Organisation Creative Commons4 in the TIB AV-Portal and allow for a greater and facilitated distribution of the films in the public.
4 Creative Commons. Available at: http://creativecommons.org/ (Accessed: 4 March 2016). 36 E. Brehm / Renegotiating Open-Access-Licences for Scientific Films
Challenges for the renegotiation of the rights are manifold:
The effort undertaken needs to be in balance with the results that can be achieved.
In many cases it is impossible to trace the original film authors.
Generally, the Creative Commons Licenses are unknown to this particular group of scientists.
Many films contain material, to which the rights are not held by the original film authors (copyright of third parties, personality and publicity rights, etc.).
Film authors expect a financial compensation for the use of the films.
A lot of films are only available in analogue formats.
In spite of the challenges, the renegotiation of Creative Commons Licenses for many of the films is very successful: 59 % of all film authors whose films can be made available online, decide to release the films under the Creative Commons License. In the process, films for which the copyright protection has expired and orphan works are identified and made available online and through other services of TIB. In the poster, more detailed information is given about the background, process and success rates.
Positioning and Power in Academic Publishing: Players, Agents and Agendas 37 F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-37 ROAD: the Directory of Open Access Scholarly Resources to Promote Open Access Worldwide
Nathalie CORNICa,1
a Data, Network & Standards Department, ISSN International Centre
Abstract. ROAD, the Directory of Open Access scholarly Resources, is a new service implemented by the ISSN International Centre. ROAD provides a free access to a selection of worldwide, multidisciplinary scholarly resources in open access that have been identified by the ISSN Network. This paper will present ROAD background, its innovative concept and how it is positioned in the open access ecosystem.
Keywords. open access, scholarly resources, serial publications, ISSN
1. Introduction
ROAD 2, the Directory of Open Access scholarly Resources, gathers all serial publications that can be identified by an ISSN such as journals, conference proceedings, monographic series, academic repositories and blogs. Launched in December 2013 by the ISSN International Centre3 and supported by the Communication and Information Sector of UNESCO4, ROAD provides a free access to a subset of the ISSN Register5. This subset gathers nearly 14,000 bibliographic records describing and pointing to worldwide, multidisciplinary scholarly resources in open access that have been identified by the ISSN Network6. ROAD innovative concept is that ISSN bibliographic records are enhanced with external information aggregated from data sources like indexing and abstracting services, metrics and registries. These OA resources are thus selected for their compliance with the core components of the open access spectrum guide7.
1 Corresponding Author, Data, Network & Standards Department, ISSN International Centre, 45 rue de Turbigo, 75003 Paris, France; E-mail: [email protected] 2 http://road.issn.org/en 3 http://www.issn.org/ 4 http://www.unesco.org/new/en/communication-and-information/ 5 The ISSN international Register gathers 1,9 million bibliographic records available on subscription: http://www.issn.org/understanding-the-issn/the-issn-international-register/ 6 The ISSN Network comprises 89 national centres worldwide: http://www.issn.org/the-centre-and- the- network/our-organization/le-reseau-issn-en/ 7 In terms of readership, reuse, copyright, author and automatic posting, and machine readability, according to the guide How Open Is It? https://www.plos.org/open-access/howopenisit/ 38 N. Cornic / ROAD: The Directory of Open Access Scholarly Resources to Promote OA Worldwide
2. Project background
A significant growth of open access scholarly resources have been observed over the previous years.
2.1 The needs of the scholarly community
Scholars and librarians have increasing difficulties to select reliable, open access resources for their research in this ocean of publications. The ISSN standard (ISO 3297 : 2007) mainly used to identify serial resources in an unambiguous way and like ISBN8, it does not guarantee any scientific and editorial quality. However, a key role remains to be played for the ISSN Network. Facing the emergence of questionable publishing practices, ISSN national centres receive recurring questions about the reliability of OA journals, and there is some confusion around ISSN being mistakenly associated to a quality label by some young researchers (Pelegrin 2014). There is also a need for scholars: to find out how many journals are indexed or ranked in their country, which OA scholarly resources are available in their discipline in a specific language. And of course, scholars need to be guided to select the trusted journals they should submit their research to.
2.2 The gaps to be filled
In 2013, despite the existence of many complementary (Stenson 2012) directories like the DOAJ9, the DOAB10, OpenDOAR11, there was no global system referencing both, on the one hand, the resources published through the green and gold open access models, and on the other hand, the digitized versions of dead print scholarly journals made available online for free by institutions. As a matter of fact, the scope of these existing directories is focused on a certain type of resources, and consequently, of disciplines, STM being predominant in journal publishing, whereas humanities and social sciences are monograph-centred. Some other directories like AJOL12 and Latindex13 have a specific specific geographical coverage. Referring to scholarly communication evaluation, in 2013, there was no comprehensive, global and multidisciplinary directory enabling to know in which indexes or databases a given publication appears in. Conventional indicators of reputation are traditionally very narrowly defined, and built mainly around one scholar activity. And the numerous emerging reputation platforms like ResearchGate or Kudos are still in their infancy, and none cover the whole gamut of activities (Nickolas 2015).
8 International Standard Book Number: https://www.isbn-international.org/ 9 The Directory of Open Access Journals: https://doaj.org/ 10 The Directory of Open Access Books: http://www.doabooks.org/ 11 The Directory of Open Access Repositories: http://www.opendoar.org/ 12 African Journals Online: http://www.ajol.info/ 13 Latindex: http://www.latindex.org/ N. Cornic / ROAD: The Directory of Open Access Scholarly Resources to Promote OA Worldwide 39
Existing tools (directories, abstracting and indexing services, metrics) may lack accuracy in the bibliographic description, relying on the metadata provided by the publishers. Finally, UNESCO needed some global, worldwide statistics about open access scholarly communication. These are the reasons why ROAD was conceived.
3. ROAD's main innovative features
3.1 Identifying scholarly resources is a constant priority for the ISSN system
The ISSN Network, along its 40 years of expertise in identifying serials and maintaining the authoritative database for serials14, has gathered 1,9 million records, and among them, 160,000 online publications. Identifying scholarly resources has been a constant priority, as the 89 ISSN centres are hosted by National libraries and National Centres for Scientific Information and Technology. Thus, a strong relationship exists between the ISSN Network and the scientific community, through a widespread use of the ISSN. As seen on Table 1 and Figure 1 below, the cooperation of the ISSN national centres is essential to make ROAD a unique service based on a continuous identification of 5 types of multidisciplinary resources. Not only content criteria for inclusion in ROAD are strict15, but also, in a second phase of acceptance, ISSN IC performs some quality control on the metadata in terms of bibliographic accuracy, updates and relevancy to meet the inclusion criteria.
Table 1. Top 10 participating countries in ROAD
Countries No of resources India 1400 Brazil 1126 United Kingdom 817 United States 793 France Poland 551 Germany 478 Russian Federation 466 Iran 448 Spain 448 423 Source: ISSN Register
14 The ISSN International Register: http://www.issn.org/understanding-the-issn/the-issn-international- register/ 15 Criteria applied: open access to the whole content of the resource, no moving wall, the resources comprise research papers addressed to the scholarly community. 40 N. Cornic / ROAD: The Directory of Open Access Scholarly Resources to Promote OA Worldwide
Figure 1. Number of resources in ROAD by thematic area and type as of December 2015.
3.2 Partnering databases
Another strength of ROAD is that ISSN bibliographic records are enriched by external data sources16 through long-standing ISSN partnerships like with Edina17 and Latindex, or through new collaborations with DOAJ and Scopus for instance. Indexing and abstracting services, evaluation services and registries complete the bibliographic metadata provided by ISSN centres, adding publishing information not traditionally collected by libraries, like business models, article processing charges, re-use licences and metrics about the visibility and the quality. All these external databases aggregate peer reviewed content harvested by the ROAD system. Partnering databases are chosen for their specialized scopes, the learned societies they are supported by, and for their positive, white list approach.
3.3 Statistics
On UNESCO's request, ROAD provides statistics computed through automated searches against the database. They cover all types of OA resources, showing the evolution in the number of publications per type, geographic and thematic area, and countries, as well as the coverage of the resources by indexing and abstracting services.
16 data sources so far have been associated to ROAD: see http://road.issn.org/ Data Sources 17 Edina maintains The Keepers Registry: http://thekeepers.org/registry.asp 18 The Global Open Access Portal (GOAP), http://www.unesco.org/new/en/communication-and- information/portals-and-platforms/goap/
N. Cornic / ROAD: The Directory of Open Access Scholarly Resources to Promote OA Worldwide 41
4. Current status and future steps
The ISSN International Centre is well positioned to create a global, comprehensive, multidisciplinary portal of open access continuing resources. The support of UNESCO was crucial, ROAD being complementary with, and accessible through the Global 18 Open Access Portal. ROAD development strategy is oriented on diversifying the thematic content and balancing the non-journals resources through ISSN assignment campaigns among the network and through implementing a data quality plan, to strive to become the authoritative, global database about open access resources.
References
Pelegrin, F.X., ROAD: A New Free Service for Identifying and Selecting OA Scholarly Resources, 2014, Charleston Conference
Stenson, L., (2012). Why all these directories? An introduction to DOAJ and DOAB. Insights. 25(3), pp.251– 256. DOI: http://doi.org/10.1629/2048-7754.25.3.251
Nicholas, D., Herman, E., Jamali, H., Rodríguez-bravo, b., Boukacem-Zeghmouri, c., dobrowolski, t. And pouchot, s. (2015), New ways of building, showcasing, and measuring scholarly reputation. Learned Publishing, 28: 327. http://ciber- research.eu/Dave_Nicholas_Publications.html
42 Positioning and Power in Academic Publishing: Players, Agents and Agendas F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-42 Sustaining the growth of library scholarly publishing
a,1 Graham STONE a University of Huddersfield
Abstract. In 2012, the University of Huddersfield Press presented a paper at the 16th International Conference on Electronic Publishing on its new open access journals platform. At the time, the Press was one of the only New University Presses (NUP) in the UK and one of the first to publish open access journals, open access monographs and sound recordings. This paper will develop Hahn’s programme and publication level business plan and relate this to the sustainability of the Press. It will demonstrate how the Press has been able to show value to the University in order to secure funding. The paper will conclude with a discussion around the need for collaboration between library led NUPs.
Keywords. Library, publishing, university press, open access, business models
1. Introduction The University of Huddersfield Press was re-launched in 2010 as a library led publishing initiative with decisions taken by an academic led Editorial Board. In 2012, the Press presented a paper at the 16th International Conference on Electronic Publishing on its new open access journals platform (Stone, 2011). At the time, the Press was one of the only New University Presses (NUP) in the UK and one of the first to publish open access journals, open access monographs and sound recordings. Since then it has published seven journals, ten scholarly monographs and nine music recordings. The library as publisher or library scholarly publishing is now a growing worldwide movement (Simser, Stockham & Turtle, 2015) and Huddersfield has followed the lead from NUPs in the United States and Australia (Lynch, 2010). This paper develops Hahn’s (2008) programme and publication level business plan and relates this to the sustainability of the Press. It demonstrates how the Press has been able to show value to the University in order to secure future funding. It concludes with a discussion about the need for collaboration between library led NUPs.
2. Business models Business model development for NUPs is an area that needs significant work (Hahn, 2008; Withey et al., 2011). Issues with open access business models have also been discussed an refined for much of the last decade (Thatcher, 2007). However, they are still based on the principles of rigorous peer review and close engagement with faculty
1 Corresponding Author: [email protected] G. Stone / Sustaining the Growth of Library Scholarly Publishing 43 and strategic leadership through an advisory board with representatives from all faculties (Missingham & Kanellopoulos, 2014). Like many NUPs, the University of Huddersfield Press developed without a clear business model in its early years. This has issues for sustainability and scalability. 3. Sustainability In 2012, a report to SPARC found that only 15% of libraries surveyed had a documented sustainability plan (Mullins et al., 2012). Hahn (2008) found that very few library publishers were able to ‘…support even 10 journal titles or more than a handful of monographic works’ (p.25). Thus, library presses hesitate in more aggressive marketing due to fears that this could generate more demand than could be satisfied. This leads to the question of scalability. If a library publisher wishes to expand, it has to identify the resources needed and this is a long-term commitment. This could result in resources being diverted from other areas (Xia, 2009). A more successful press will create a need to reallocate greater staffing resources unless new resources are identified. 4. Programme level planning Regarding the development of the business model, Hahn (2008) suggests two levels of business plans for library publishers: • Programme level planning • Publication level planning A NUP operating without a business model at the programme level is effectively operating at a publication level. Moving from one publication to the next without a clear plan. Staffing and funding challenges need to be resolved at a programme level for the library as publisher to be sustainable. In addition, planning is needed at both programme level and publication level in order for the initiative to become a success. 4.1. Scalability of library publishing services NUPs offer a truncated list of services when compared to traditional publishers (Hahn, 2008). However, this represents a leaner version of traditional ‘legacy’ publishers. Once presses begin to grow there is a question of scalability and sustainability and this is what programme level planning provides. This is the case for the University of Huddersfield Press, and is echoed by comments made by other library presses (Mullins et al., 2012). There is a fear that greater demand could lead to the press becoming a victim of its own success. 4.2. Staffing The issue of staffing and the resulting effect of increased success verses a limited staff base have been the focus of discussion for many successful presses as over time this inhibits growth (Perry et al., 2011). The SPARC study found the number of staff 44 G. Stone / Sustaining the Growth of Library Scholarly Publishing allocated to publishing activities ranged between 0.9-2.4 FTE, with staff dedicated to library publishing programmes described as relatively rare (Mullins et al., 2012). 4.3. Business models and funding As part of the UK Crossick report (London Economics, 2015), theoretical tests established that each open access business model has its own strengths, weaknesses, opportunities and threats. This was further developed as part of the OAPEN-UK project (Beech & Milloy, 2015). The predominant business model for the University of Huddersfield Press is the institutional subsidy model where the Press receives subsidies from the University, either centrally, from faculty or the library, or from a funder. 5. Publication level planning In order for publication level planning to work, programme level planning needs to be in place. For example, planning at the programme level leads towards a business plan. This plan can outline the case for growth of the press. The plan at Huddersfield suggests a more robust funding allocation and modest increase in staffing. This in turn supports a greater number of publications and improved publication level planning. An annual plan, which includes a budget, key dates and an evaluation process, could then be produced. An example of this at Huddersfield is Fields: journal of Huddersfield student research (Stone, Jensen & Beech, 2016). The Press worked with the University’s Teaching and Learning Institute to ring-fence funding for the publication. Publication level planning helps to address issues that have arisen in the process. The journal is now entering its third year of publication and lessons learned from volume 1 have led to a revision in the notes for contributors, a writing retreat for authors, conference attendance for student authors and marketing around campus. 6. Cash flow and profit and loss forecast At Huddersfield, a paper on staffing was taken to the Press Board in 2015. As a result annual staffing costs for the Press of around £40K have been absorbed by Computing and Library Services (CLS) as part of the staffing budget. Institutional repository costs (the publication platform for the Press) are also covered by CLS. As part of the Press business plan, the following costs were identified in order to grow the Press at a sustainable level. • DOI costs for seven existing journals, with a growth rate of an extra two journals per year • Set-up costs for the additional journals • Two monographs to be published in 2016, three in 2017, four in 2018 and five in 2019 • Recurrent costs including appropriate memberships and marketing Sales forecasts were also included for print copies of monographs, although these are not guaranteed. Income from print sales would enable the Press to publish additional titles to those highlighted above. This model also allows the Press to run a G. Stone / Sustaining the Growth of Library Scholarly Publishing 45 fee waiver model for peer reviewed monographs and journals from Huddersfield authors as this would be underwritten by programme level funding. This model is also being adopted by other NUPs in the UK such as UCL (2015) and the recently launched White Rose University Press (2016). 7. Value and impact NUPs are not for-profit enterprises, they are an exercise in scholarly communication. In order to attract programme level funding and to justify a local subsidy the press must demonstrate its value to the university rather than monetarize the work of the press. An example of how to do this is to show that financial returns, which do not come back to the Press directly, have the potential to earn research income for the university. At Huddersfield this has been done by showing how the Press can contribute to ‘quality- related research funding’ (QR funding) from the 2014 Research Excellence Framework (REF) (HEFCE, 2015). As part of the 2014 REF, the University submitted 100 research outputs from its staff to the music Unit of Assessment (REF, 2014). University Press publications were included in eleven of these outputs (some of these were as part of portfolio outputs). While REF scores cannot be associated with individual outputs, 85% of music research was judged to be internationally excellent (3* and 4*), which attracts QR funding. The assumption here is that at least some of the Press output was ranked in these categories. In addition, Press output also contributed to the wider impact and environment statements, which were also ranked highly. If all 100 outputs were treated equally, then six outputs from the Press (three books and three CDs) have contributed to 11% of the University’s QR funding for music in 2016. This is a not inconsiderable sum, indeed far more than the overheads of the Press for all publications forecast in the Press’s four year plan. In February 2016 a discussion paper was tabled at the University of Huddersfield Press Editorial Board. It invited the Board to discuss the four-year plan, which outlined the funding required at programme level in order for the Press to become sustainable and scalable. This included a detailed cash flow and profit and loss forecast and evidence of the value and impact of the Press on QR funding in the University. It was suggested that there was potential for this to have impact on other disciplines such as history, politics and English, which rely on monograph publishing as the gold standard. The Board approved the proposal for a programme level funding model in principle. As a result the Press has now had funding confirmed for the 2016/17 and 2017/18 academic years. This is in addition to staffing costs and will allow the Press to finance additional monographs and journals described above as part of a programme level plan. The Press will also be able to offer a fee waiver to researchers at Huddersfield who submit proposals for new titles subject to satisfying the Press’s peer review process. 8. Collaboration In addition to programme and publication level planning, NUPs also need to collaborate to achieve scalability. The 2012 report to SPARC recommends that collaborations should be used to, “…leverage resources within campuses, across institutions, and between university presses, scholarly societies, and other partners” 46 G. Stone / Sustaining the Growth of Library Scholarly Publishing
(Mullins et al., 2008, p.19). This paper suggests that the follow areas of collaboration are required.
• Landscape survey. In the UK there is uncertainty as to how many library led open access university presses are operating. Huddersfield, White Rose and UCL presses have all been mentioned in this paper. However, there are others emerging in both the UK and the rest of Europe. A data gathering exercise is required in order to assess the current state of play regarding NUPs and library publishing ventures in Europe
• A Library Publishing Coalition for Europe. This paper suggests that NUPs in Europe establish a European Library Publishing Coalition (LPC). This would be based upon the LPC in the United States (Educopia Institute, 2013) and could become a hub for best practice and innovative approaches
• Best practice/efficiencies in the workflows. The Landscape study will give intelligence on where the new and proposed library presses are, a LPC would help to establish a community. It is hoped that this will lead to further collaboration and therefore sustainability for NUPs. It is suggested that a series of best practice guidelines could be developed providing useful tools for NUPs. For example, licences, workflows, business models and recommendations for appropriate membership, e.g. COPE, OASPA, DOAJ and DOAB. Best practice around establishing value and impact would also allow these NUPs to flourish in the future. 9. Conclusion This paper has shown how the University of Huddersfield Press has used evidence of value and impact based on REF output to secure funding for the next two years. An understanding of Hahn’s programme and publication level business plan has allowed the Press to achieve sustainability going forward. This will allow the scaling up of publications with a view to the next REF in the UK. The next steps for the Press are to produce a plan for the next two years in order to secure further funding going forward. In addition, the Press needs to work alongside other NUPs in order to establish best practice for library-led open access publishing. References Beech, D. and Milloy, C. (2015) Business models for open access monographs: a summary document. Available at http://oapen-uk.jiscebooks.org/files/2015/04/Annex-A-OAPEN-UK-Business- Models-SWOT-Workshop-preparatory-document.docx (Accessed: 7 March 2016).
Educopia Institute (2013) Library Publishing Coalition. Available at http://librarypublishing.org/ (Accessed: 7 March 2016).
Hahn, K.L. (2008) Research library publishing services: new options for university publishing, Washington DC: Association of Research Libraries. Available at http://www.arl.org/storage/documents/ publications/research-library-publishing- services-mar08.pdf (Accessed: 7 March 2016). G. Stone / Sustaining the Growth of Library Scholarly Publishing 47
HEFCE (2015) How we fund research. Available at http://www.hefce.ac.uk/rsrch/funding/mainstream/ (Accessed: 7 March 2016). London Economics (2015) Economic analysis of business models for open access monographs: Annex 4 to the Report of the HEFCE Monographs and Open Access Project. London: HEFCE. Availableat http://www.hefce.ac.uk/media/hefce/content/pubs/indirreports/ 2015/Monographs,and,o pen,access/2014_monographs4.pdf (Accessed: 7 March 2016).
Lynch, C. (2010) ‘Imagining a university press system to support scholarship in the digital age’, Journal of electronic publishing, 13(2). http://dx.doi.org/10.3998/3336451.0013.207
Missingham, R. and Kanellopoulos, L. (2014) ‘University presses in libraries: Potential for successful marriages’, OCLC Systems & Services: International Digital Library Perspectives, 30(3), 158-166. http://dx.doi.org/10.1108/OCLC-01-2014-0001
Mullins, J.L. et al. (2012) Library publishing services: Strategies for success. Final research report. Washington, DC: SPARC. Available at http://docs.lib.purdue.edu/purduepress_ebooks/24/ (Accessed: 7 March 2016).
Perry, A.M., Borchert, C.A., Deliyannides, T.S., Kosavic, A., Kennison, R. and Dyas- Correia, S. (2011) ‘Libraries as journal publishers’, Serials Review, 37(3), 196-204. http://dx.doi.org/10.1016/ j.serrev.2011.06.006
Research Excellence Framework (2014) UoA 35 Music, Drama, Dance and Performing Arts, University of Huddersfield. Available at http://results.ref.ac.uk/Results/BySubmission/2521 (Accessed: 7 March 2016).
Simser, C.N., Stockham, M.G. and Turtle, E. (2015) ‘Libraries as publishers: a winning combination’, OCLC Systems & Services: International Digital Library Perspectives, 31(2), 69-75. http://dx.doi.org/ 10.1108/OCLC-01-2014-0006
Stone, G. (2011) ‘Huddersfield Open Access Publishing’, Information Services and Use, 31(3/4), 215- 223. http://dx.doi.org/10.3233/ISU-2012-0651 Stone, G., Jensen, K. and Beech, M. (2016) ‘Publishing undergraduate research: linking teaching and research through a dedicated peer reviewed open access journal’, Journal of scholarly publishing, 47(2), 147-170. http://dx.doi.org/10.3138/ jsp.47.2.147
Thatcher, S.G. (2007) ‘The challenge of open access for university presses’, Learned Publishing, 20(3), 165-172. http://dx.doi.org/10.1087/095315107X205084 UCL (2015) UCL launches UK’s first fully Open Access university press. Available at https://www.ucl.ac.uk/news/news-articles/0515/270515-ucl- press (Accessed: 7 March 2016).
White Rose University Press (2016) White Rose University Press, 2016. Available at http://universitypress.whiterose.ac.uk/ (Accessed: 7 March 2016).
Withey, L. et al. (2011) ‘Sustaining scholarly publishing: New business models for university presses: a report of the AAUP task force on economic models for scholarly publishing’, Journal of Scholarly Publishing, 42(4), 397-441. http://doi.org/10.1353/scp.2011.0035
Xia, J. (2009) ‘Library publishing as a new model of scholarly communication’, Journal of Scholarly Publishing, 40(4), 370-383. http://dx.doi.org/10.3138/jsp.40.4.370 48 Positioning and Power in Academic Publishing: Players, Agents and Agendas F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-48 Genome sharing projects around the world: how you find data for your research
1 a,b Fiona NIELSENa,b, and Nadezda KOVALEVSKAYA a Repositive Ltd, Future Business Centre, Kings Hedges Road, Cambridge CB4 2HY, United Kingdom b DNAdigest, Future Business Centre, Kings Hedges Road, Cambridge CB4 2HY, United Kingdom
Abstract. Access to raw experimental research data and data reuse is a common hurdle in scientific research. Despite the mounting requirements from funding agencies that the raw data is deposited as soon as (or even before) the paper is published, multiple factors often prevent data from being accessed and reused by other researchers. The situation with the human genomic data is even more dramatic, since on the one hand human genomic data is probably the most important data to share - it lies at the heart of efforts to combat major health issues such as cancer, genetic diseases, and genetic predispositions for complex diseases like heart disease and diabetes. On the other hand, since it is sensitive and personal information, it is often exempt from data sharing requirements. DNAdigest investigates the barriers for ethical and efficient genomic data sharing and engages with all stakeholder groups, including researchers, librarians, data managers, software developers, policy makers, and the general public interested in genomics. Repositive offers services and tools that reduce the barriers for data access and reuse for the research community in academia, industry, and clinics. To address the most pressing problem for public genomic data: that of data discoverability, Repositive has built an online platform (repositive.io) providing a single point of entry to find and access available genomic research data. Keywords. Genomic data, data access, data sharing, genomic data repositories, tools
1. Introduction: data access and reuse is a common hurdle in scientific research
Research organisations, both public and private, are producing ever increasing volumes of data. Irrespective of whether the research is funded publicly or privately, there is increasing pressure to provide evidence that the maximum benefit is obtained from generated data. Recent years have seen a concerted effort by providers of public funds for research to require that the results of that research be publicly available (Collection of UK funders’ policies 2015). While the benefits of data sharing are becoming more widely accepted (Toronto International Data Release Workshop Authors 2009), human genomic data (i.e.,
1 Corresponding Author: [email protected] F. Nielsen and N. Kovalevskaya / Genome Sharing Projects Around the World 49 information about the composition of our DNA and RNA) is often exempt from data sharing requirements from major funders that all experimental data must be placed in publicly accessible repositories. This is because of concerns that making human genomic data public exposes potentially sensitive personal information to the world (Richards 2015).
2. The special case of human genomic data: it is there but it is mainly inaccessible
It is estimated that, in 2015, the world human genome sequencing capacity will exceed 80 petabyte of sequence a year. However, as of 2014, the largest public repository for human genomics data (the NIH database of genotypes and phenotypes dbGaP) holds only about 0.5 petabytes of clinical genomics data. This gap between the availability of genomic information and the production of it can be at least partially attributed to the absence of tangible benefits for the individuals who make data available and, at the same time, to the existence of sanctions for improper handling of personal information. However, when data donors give consent for their data to be used for research, they set their expectation that the data will actually be used for this purpose. To not utilise their data in the best possible way within the consent given goes against the data donor’s interests and expectations. Ironically, human genomic data is probably the most important data to share, since it lies at the heart of efforts to combat major health issues such as cancer, genetic diseases, and genetic predispositions for complex diseases like heart disease and diabetes. In particular, the promise of personalised medicine (where treatment is tailored to the individual) is unlikely to be realised without widespread access to large amounts of genomic data. Existing data sharing initiatives generally take the form of some kind of repository for storing data or some kind of service to help find collaborators or data. Examples of repositories include publicly funded repositories (e.g. SRA, ENA, dbGaP, EGA, ArrayExpress etc), biobanks, and data repositories set up by individual institutions or projects (e.g. LOVD). Examples of services to help find data include, for example, GenomeConnect, PhenomeCentral and the Beacon project. Public data repositories do an excellent job of storing data, a crucial task to enable data availability. The mentioned services do great jobs at servicing specific needs, e.g. connecting clinicians who have found similar phenotypes for their patients with genetic diseases. Currently no public initiatives are have successfully addressed the problem of discovering the existence of datasets (data discoverability) for specific diseases and specific data types across locations and repositories. In addition, all of the mentioned initiatives face challenges of funding and sustainability of their initiatives, due to their reliance on research grants.
3. A solution to increase access to human genomic data: the community platform
To address the most pressing problem for public genomic data: that of data discoverability (van Schaik 2014), Repositive has built an online platform 50 F. Nielsen and N. Kovalevskaya / Genome Sharing Projects Around the World
(repositive.io) providing a single point entry to search and access public genomic data sources. The Repositive platform enables users to search through all its indexed data sources in a single click via an easy-to-use interface free of charge. To address the problem of varying quality and type of metadata associated with data across data sources and public repositories, the Repositive platform allows users to comment on the content and quality of datasets and add descriptions to the listed metadata. If a research scientist has data that he/she would like to share but cannot for any reason, he/she can announce the existence of the data on the Repositive platform. In this case, other scientists that have similar or complementary data can contact the author to start a collaboration or to discuss for instance the conditions under which they can exchange their data. Similarly, a user can post a request for data and another user, who has the data stored but not used, can respond and find an application for their otherwise unused data. By listening to our users and concentrating on a specific use case for the genetics researcher – the problem of finding and accessing human genomic research data - and supporting best practices for data annotation, accessibility and reuse, we offer the Repositive platform and services as a contribution to ease the workflow for research in human genomics for health and disease. The Repositive business model is built around the Repositive freemium features of the online platform for data discovery. The online platform is open for all to sign up and search for free (see above), but at the same time Repositive offers premium products and services to both data providers and data consumer organisations. Our premium services include: customised data scouting; data access applications; automating data access workflows; setting up and maintaining public data catalogues; setting up data collaborations between different organisations, e.g. across industry and academia, etc. With this business model, Repositive can deliver a free service on the online platform which supports researchers across academia and industry, while our revenue comes from related professional products and custom services.
References
Collection of UK funders’ policies. In: Research Data Management Blog. (2015) Available at: http://www.data.cam.ac.uk/funders. (Accessed: 15 February 2016).
Richards, M., Anderson, R., Hinde, S., Kaye, J., Lucassen, A. et al. (2015) ‘The collection, linking and use of data in biomedical research and health care: ethical issues’. Nuffield Council on Bioethics. Report. Available at: http://nuffieldbioethics.org/wp- content/uploads/Biological_and_health_data_web.pdf. (Accessed: 15 February 2016). van Schaik, T.A., Kovalevskaya, N.V., Protopapas, E., Wahid, H. and Nielsen, FGG (2014) ‘The need to redefine genomic data sharing: A focus on data accessibility’. Applied&Translational Genomics, 3(4), pp.100-104 (doi:10.1016/j.atg.2014.09.013)
Toronto International Data Release Workshop Authors (2009) ‘Prepublication data sharing’. Nature, 461, pp. 168-170. (doi:10.1038/461168a) Positioning and Power in Academic Publishing: Players, Agents and Agendas 51 F. Loizides and B. Schmidt (Eds.) © 2016 The authors and IOS Press. This article is published online with Open Access by IOS Press and distributed under the terms of the Creative Commons Attribution License. doi:10.3233/978-1-61499-649-1-51 Stakeholders in academic publishing: text and data mining perspective and potential
Maria ESKEVICH1
Radboud University, Nijmegen, The Netherlands
Abstract. In this paper we discuss the concept of open access in academic publishing with the focus on the right to mine the data once the right to read is granted. Thus we envisage the roles and types of the stakeholders in academic publishing from the perspective of the potential text and data mining (TDM) applications. Further on, we briefly introduce FutureTDM project that aims to improve TDM uptake in Europe.
Keywords. digital libraries, text and data mining, FutureTDM project
1. Introduction
The main incentive for academic publishing is to share the knowledge acquired through experimental and/or empirical observations with an overall aim to promote further scientific development and knowledge distribution. Hence, initially the publishing empowered more researchers and thinkers to improve their expertise and to expand the overall knowledge ontology. The dialogue was set up between the content creators through the medium of written and printed text providers. However, the societal and technological development in combination with population growth over the recent centuries lead to a more complicated framework of agents in the field of scientific knowledge sharing, as the same agents can play different roles. Th The publication process, while having a target to broaden the access to the knowledge across communities, is a service that is being provided, and thus over the years it converted into a business model which raised a pay wall between the ultimate content consumers, i.e. researchers, general audience, and the content itself. As a vast amount of research is being carried out on public funding, the new frameworks for efficient scientific knowledge transfer are discussed and promoted, with the Open Access (OA) strategy being the main focus (De Grandis, Lomazzi, Rettberg). Following the OA principles defined in Budapest and Berlin Declarations, the European Commission (EC) defines Open Access as ’the practice of providing online access to scientific information that is free of charge to the end user and that is reusable’, where scientific information can refer to (i) peer-reviewed scientific research articles (published in scholarly journals) or (ii) research data (data