The Impact of Preprint Servers in the Formation of Novel Ideas
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Medical Journals Drastically Speed up Their Publication Process for Covid-19
bioRxiv preprint doi: https://doi.org/10.1101/2020.04.18.045963; this version posted April 18, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. Pandemic Publishing: Medical journals drastically speed up their publication process for Covid-19 Author Serge P.J.M. Horbach*,1,2 [email protected] +31243652730 ORCID: 0000-0003-0406-6261 *Corresponding author 1 Radboud University Nijmegen, Faculty of Science, Institute for Science in Society, P.O. box 9010, 6500 GL Nijmegen, The Netherlands 2 Centre for Science and Technology Studies (CWTS), Faculty of Social Sciences, Leiden University, Wassenaarseweg 62A, 2333 AL Leiden, The Netherlands Abstract In times of public crises, including the current Covid-19 pandemic, rapid dissemination of relevant scientific knowledge is of paramount importance. The duration of scholarly journals’ publication process is one of the main factors hindering quick delivery of new information. While proper editorial assessment and peer review obviously require some time, turnaround times for medical journals can be up to several months, which is undesirable in the era of a crisis. Following initiatives of medical journals and scholarly publishers to accelerate their publication process, this study assesses whether medical journals have indeed managed to speed up their publication process for Covid-19 related articles. It studies the duration of 14 medical journals’ publication process both during and prior to the current pandemic. Assessing a total of 669 articles, the study concludes that medical journals have indeed drastically accelerated the publication process for Covid-19 related articles since the outbreak of the pandemic. -
Is Sci-Hub Increasing Visibility of Indian Research Papers? an Analytical Evaluation Vivek Kumar Singh1,*, Satya Swarup Srichandan1, Sujit Bhattacharya2
Journal of Scientometric Res. 2021; 10(1):130-134 http://www.jscires.org Perspective Paper Is Sci-Hub Increasing Visibility of Indian Research Papers? An Analytical Evaluation Vivek Kumar Singh1,*, Satya Swarup Srichandan1, Sujit Bhattacharya2 1Department of Computer Science, Banaras Hindu University, Varanasi, Uttar Pradesh, INDIA. 2CSIR-National Institute of Science Technology and Development Studies, New Delhi, INDIA. ABSTRACT Sci-Hub, founded by Alexandra Elbakyan in 2011 in Kazakhstan has, over the years, Correspondence emerged as a very popular source for researchers to download scientific papers. It is Vivek Kumar Singh believed that Sci-Hub contains more than 76 million academic articles. However, recently Department of Computer Science, three foreign academic publishers (Elsevier, Wiley and American Chemical Society) have Banaras Hindu University, filed a lawsuit against Sci-Hub and LibGen before the Delhi High Court and prayed for Varanasi-221005, INDIA. complete blocking these websites in India. It is in this context, that this paper attempts to Email id: [email protected] find out how many Indian research papers are available in Sci-Hub and who downloads them. The citation advantage of Indian research papers available on Sci-Hub is analysed, Received: 16-03-2021 with results confirming that such an advantage do exist. Revised: 29-03-2021 Accepted: 25-04-2021 Keywords: Indian Research, Indian Science, Black Open Access, Open Access, Sci-Hub. DOI: 10.5530/jscires.10.1.16 INTRODUCTION access publishing of their research output, and at the same time encouraging their researchers to publish in openly Responsible Research and Innovation (RRI) has become one accessible forms. -
Sci-Hub Provides Access to Nearly All Scholarly Literature
Sci-Hub provides access to nearly all scholarly literature A DOI-citable version of this manuscript is available at https://doi.org/10.7287/peerj.preprints.3100. This manuscript was automatically generated from greenelab/scihub-manuscript@51678a7 on October 12, 2017. Submit feedback on the manuscript at git.io/v7feh or on the analyses at git.io/v7fvJ. Authors • Daniel S. Himmelstein 0000-0002-3012-7446 · dhimmel · dhimmel Department of Systems Pharmacology and Translational Therapeutics, University of Pennsylvania · Funded by GBMF4552 • Ariel Rodriguez Romero 0000-0003-2290-4927 · arielsvn · arielswn Bidwise, Inc • Stephen Reid McLaughlin 0000-0002-9888-3168 · stevemclaugh · SteveMcLaugh School of Information, University of Texas at Austin • Bastian Greshake Tzovaras 0000-0002-9925-9623 · gedankenstuecke · gedankenstuecke Department of Applied Bioinformatics, Institute of Cell Biology and Neuroscience, Goethe University Frankfurt • Casey S. Greene 0000-0001-8713-9213 · cgreene · GreeneScientist Department of Systems Pharmacology and Translational Therapeutics, University of Pennsylvania · Funded by GBMF4552 PeerJ Preprints | https://doi.org/10.7287/peerj.preprints.3100v2 | CC BY 4.0 Open Access | rec: 12 Oct 2017, publ: 12 Oct 2017 Abstract The website Sci-Hub provides access to scholarly literature via full text PDF downloads. The site enables users to access articles that would otherwise be paywalled. Since its creation in 2011, Sci- Hub has grown rapidly in popularity. However, until now, the extent of Sci-Hub’s coverage was unclear. As of March 2017, we find that Sci-Hub’s database contains 68.9% of all 81.6 million scholarly articles, which rises to 85.2% for those published in toll access journals. -
Open Access Availability of Scientific Publications
Analytical Support for Bibliometrics Indicators Open access availability of scientific publications Analytical Support for Bibliometrics Indicators Open access availability of scientific publications* Final Report January 2018 By: Science-Metrix Inc. 1335 Mont-Royal E. ▪ Montréal ▪ Québec ▪ Canada ▪ H2J 1Y6 1.514.495.6505 ▪ 1.800.994.4761 [email protected] ▪ www.science-metrix.com *This work was funded by the National Science Foundation’s (NSF) National Center for Science and Engineering Statistics (NCSES). Any opinions, findings, conclusions or recommendations expressed in this report do not necessarily reflect the views of NCSES or the NSF. The analysis for this research was conducted by SRI International on behalf of NSF’s NCSES under contract number NSFDACS1063289. Analytical Support for Bibliometrics Indicators Open access availability of scientific publications Contents Contents .............................................................................................................................................................. i Tables ................................................................................................................................................................. ii Figures ................................................................................................................................................................ ii Abstract ............................................................................................................................................................ -
On the Relation Between the Wos Impact Factor, the Eigenfactor, the Scimago Journal Rank, the Article Influence Score and the Journal H-Index
1 On the relation between the WoS impact factor, the Eigenfactor, the SCImago Journal Rank, the Article Influence Score and the journal h-index Ronald ROUSSEAU 1 and the STIMULATE 8 GROUP 2 1 KHBO, Dept. Industrial Sciences and Technology, Oostende, Belgium [email protected] 2Vrije Universiteit Brussel (VUB), Brussels, Belgium The STIMULATE 8 Group consists of: Anne Sylvia ACHOM (Uganda), Helen Hagos BERHE (Ethiopia), Sangeeta Namdev DHAMDHERE (India), Alicia ESGUERRA (The Philippines), Nguyen Thi Ngoc HOAN (Vietnam), John KIYAGA (Uganda), Sheldon Miti MAPONGA (Zimbabwe), Yohannis MARTÍ- LAHERA (Cuba), Kelefa Tende MWANTIMWA (Tanzania), Marlon G. OMPOC (The Philippines), A.I.M. Jakaria RAHMAN (Bangladesh), Bahiru Shifaw YIMER (Ethiopia). Abstract Four alternatives to the journal Impact Factor (IF) indicator are compared to find out their similarities. Together with the IF, the SCImago Journal Rank indicator (SJR), the EigenfactorTM score, the Article InfluenceTM score and the journal h- index of 77 journals from more than ten fields were collected. Results show that although those indicators are calculated with different methods and even use different databases, they are strongly correlated with the WoS IF and among each other. These findings corroborate results published by several colleagues and show the feasibility of using free alternatives to the Web of Science for evaluating scientific journals. Keywords: WoS impact factor, Eigenfactor, SCImago Journal Rank (SJR), Article Influence Score, journal h-index, correlations Introduction STIMULATE stands for Scientific and Technological Information Management in Universities and Libraries: an Active Training Environment. It is an international training programme in information management, supported by the Flemish Interuniversity Council (VLIR), aiming at young scientists and professionals from 2 developing countries. -
Do You Speak Open Science? Resources and Tips to Learn the Language
Do You Speak Open Science? Resources and Tips to Learn the Language. Paola Masuzzo1, 2 - ORCID: 0000-0003-3699-1195, Lennart Martens1,2 - ORCID: 0000- 0003-4277-658X Author Affiliation 1 Medical Biotechnology Center, VIB, Ghent, Belgium 2 Department of Biochemistry, Ghent University, Ghent, Belgium Abstract The internet era, large-scale computing and storage resources, mobile devices, social media, and their high uptake among different groups of people, have all deeply changed the way knowledge is created, communicated, and further deployed. These advances have enabled a radical transformation of the practice of science, which is now more open, more global and collaborative, and closer to society than ever. Open science has therefore become an increasingly important topic. Moreover, as open science is actively pursued by several high-profile funders and institutions, it has fast become a crucial matter to all researchers. However, because this widespread interest in open science has emerged relatively recently, its definition and implementation are constantly shifting and evolving, sometimes leaving researchers in doubt about how to adopt open science, and which are the best practices to follow. This article therefore aims to be a field guide for scientists who want to perform science in the open, offering resources and tips to make open science happen in the four key areas of data, code, publications and peer-review. The Rationale for Open Science: Standing on the Shoulders of Giants One of the most widely used definitions of open science originates from Michael Nielsen [1]: “Open science is the idea that scientific knowledge of all kinds should be openly shared as early as is practical in the discovery process”. -
Downloads Presented on the Abstract Page
bioRxiv preprint doi: https://doi.org/10.1101/2020.04.27.063578; this version posted April 28, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under aCC-BY 4.0 International license. A systematic examination of preprint platforms for use in the medical and biomedical sciences setting Jamie J Kirkham1*, Naomi Penfold2, Fiona Murphy3, Isabelle Boutron4, John PA Ioannidis5, Jessica K Polka2, David Moher6,7 1Centre for Biostatistics, Manchester Academic Health Science Centre, University of Manchester, Manchester, United Kingdom. 2ASAPbio, San Francisco, CA, USA. 3Murphy Mitchell Consulting Ltd. 4Université de Paris, Centre of Research in Epidemiology and Statistics (CRESS), Inserm, Paris, F-75004 France. 5Meta-Research Innovation Center at Stanford (METRICS) and Departments of Medicine, of Epidemiology and Population Health, of Biomedical Data Science, and of Statistics, Stanford University, Stanford, CA, USA. 6Centre for Journalology, Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Canada. 7School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Canada. *Corresponding Author: Professor Jamie Kirkham Centre for Biostatistics Faculty of Biology, Medicine and Health The University of Manchester Jean McFarlane Building Oxford Road Manchester, M13 9PL, UK Email: [email protected] Tel: +44 (0)161 275 1135 bioRxiv preprint doi: https://doi.org/10.1101/2020.04.27.063578; this version posted April 28, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. -
“Publish Or Perish” As Citation Metrics Used to Analyze Scientific Output In
ai-6.qxd 11/4/08 2:52 PM Page 1 VARIA – SCIENTOMETRICS Arch. Immunol. Ther. Exp., 2008, 56, 1–9 DOI 10.1007/s00005-008-0043-0 PL ISSN 0004-069X “Publish or Perish” as citation metrics used to analyze scientific output in the humanities: International case studies in economics, geography, social sciences, philosophy, and history Audrey Baneyx Institut Francilien “Recherche, Innovation et Société” (IFRIS)1, Paris, France Received: 2008.10.16, Accepted: 2008.10.27 Abstract Traditionally, the most commonly used source of bibliometric data is the Thomson ISI Web of Knowledge, in particular the (Social) Science Citation Index and the Journal Citation Reports, which provide the yearly Journal Impact Factors. This database used for the evaluation of researchers is not advantageous in the humanities, mainly because books, conference papers, and non-English journals, which are an important part of scientific activity, are not (well) covered. This paper pre- sents the use of an alternative source of data, Google Scholar, and its benefits in calculating citation metrics in the humani- ties. Because of its broader range of data sources, the use of Google Scholar generally results in more comprehensive cita- tion coverage in the humanities. This presentation compares and analyzes some international case studies with ISI Web of Knowledge and Google Scholar. The fields of economics, geography, social sciences, philosophy, and history are focused on to illustrate the differences of results between these two databases. To search for relevant publications in the Google Scholar database, the use of “Publish or Perish” and of CleanPoP, which the author developed to clean the results, are compared. -
Ten Good Reasons to Use the Eigenfactor Metrics in Our Opinion, There Are Enough Good Reasons to Use the Eigenfactor Method to Evaluate Journal Influence: 1
Ten good reasons to use the EigenfactorTMmetrics✩ Massimo Franceschet Department of Mathematics and Computer Science, University of Udine Via delle Scienze 206 – 33100 Udine, Italy [email protected] Abstract The Eigenfactor score is a journal influence metric developed at the De- partment of Biology of the University of Washington and recently intro- duced in the Science and Social Science Journal Citation Reports maintained by Thomson Reuters. It provides a compelling measure of journal status with solid mathematical background, sound axiomatic foundation, intriguing stochastic interpretation, and many interesting relationships to other ranking measures. In this short contribution, we give ten reasons to motivate the use of the Eigenfactor method. Key words: Journal influence measures; Eigenfactor metrics; Impact Factor. 1. Introduction The Eigenfactor metric is a measure of journal influence (Bergstrom, 2007; Bergstrom et al., 2008; West et al., 2010). Unlike traditional met- rics, like the popular Impact Factor, the Eigenfactor method weights journal citations by the influence of the citing journals. As a result, a journal is influential if it is cited by other influential journals. The definition is clearly recursive in terms of influence and the computation of the Eigenfactor scores involves the search of a stationary distribution, which corresponds to the leading eigenvector of a perturbed citation matrix. The Eigenfactor method was initially developed by Jevin West, Ben Alt- house, Martin Rosvall, and Carl Bergstrom at the University of Washington ✩Accepted for publication in Information Processing & Management. DOI: 10.1016/j.ipm.2010.01.001 Preprint submitted to Elsevier January 18, 2010 and Ted Bergstrom at the University of California Santa Barbara. -
COVID-19 Preprints and Their Publishing Rate: an Improved Method
medRxiv preprint doi: https://doi.org/10.1101/2020.09.04.20188771; this version posted September 7, 2020. The copyright holder for this preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY-ND 4.0 International license . COVID-19 Preprints and Their Publishing Rate: An Improved Method Francois Lachapelle, PhD Candidate Department of Sociology, University of British Columbia Abstract Context: As the COVID-19 pandemic persists around the world, the scientific community continues to produce and circulate knowledge on the deadly disease at an unprecedented rate. During the early stage of the pandemic, preprints represented nearly 40% of all English-language COVID-19 scientific corpus (6, 000+ preprints | 16, 000+ articles). As of mid-August 2020, that proportion dropped to around 28% (13, 000+ preprints | 49, 000+ articles). Nevertheless, preprint servers remain a key engine in the efficient dissemination of scientific work on this infectious disease. But, giving the ‘uncertified’ nature of the scientific manuscripts curated on preprint repositories, their integration to the global ecosystem of scientific communication is not without creating serious tensions. This is especially the case for biomedical knowledge since the dissemination of bad science can have widespread societal consequences. Scope: In this paper, I propose a robust method that will allow the repeated monitoring and measuring of COVID-19 preprint’s publication rate. I also introduce a new API called Upload-or-Perish. It is a micro- API service that enables a client to query a specific preprint manuscript’s publication status and associated meta-data using a unique ID. -
Overview of the Altmetrics Landscape
Purdue University Purdue e-Pubs Charleston Library Conference Overview of the Altmetrics Landscape Richard Cave Public Library of Science, [email protected] Follow this and additional works at: https://docs.lib.purdue.edu/charleston Part of the Library and Information Science Commons An indexed, print copy of the Proceedings is also available for purchase at: http://www.thepress.purdue.edu/series/charleston. You may also be interested in the new series, Charleston Insights in Library, Archival, and Information Sciences. Find out more at: http://www.thepress.purdue.edu/series/charleston-insights-library-archival- and-information-sciences. Richard Cave, "Overview of the Altmetrics Landscape" (2012). Proceedings of the Charleston Library Conference. http://dx.doi.org/10.5703/1288284315124 This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] for additional information. Overview of the Altmetrics Landscape Richard Cave, Director of IT and Computer Operations, Public Library of Science Abstract While the impact of article citations has been examined for decades, the “altmetrics” movement has exploded in the past year. Altmetrics tracks the activity on the Social Web and looks at research outputs besides research articles. Publishers of scientific research have enabled altmetrics on their articles, open source applications are available for platforms to display altmetrics on scientific research, and subscription models have been created that provide altmetrics. In the future, altmetrics will be used to help identify the broader impact of research and to quickly identify high-impact research. Altmetrics and Article-Level Metrics Washington as an academic research project to rank journals based on a vast network of citations The term “altmetrics” was coined by Jason Priem, (Eigenfactor.org, http://www.eigenfactor.org/ a PhD candidate at the School of Information and whyeigenfactor.php). -
Eigenfactor: Ranking and Mapping Scientific Knowledge
Eigenfactor: ranking and mapping scientific knowledge Jevin D. West A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Washington 2010 Program Authorized to Offer Degree: Department of Biology University of Washington Graduate School This is to certify that I have examined this copy of a doctoral dissertation by Jevin D. West and have found that it is complete and satisfactory in all respects, and that any and all revisions required by the final examining committee have been made. Chair of the Supervisory Committee: Carl T. Bergstrom Reading Committee: Carl T. Bergstrom Benjamin B. Kerr Thomas L. Daniel Date: c Copyright 2010 Jevin D. West In presenting this dissertation in partial fulfillment of the requirements for the doctoral degree at the University of Washington, I agree that the Library shall make its copies freely available for inspection. I further agree that extensive copying of the dissertation is allowable only for scholarly purposes, consistent with fair use as prescribed in the U.S. Copyright Law. Requests for copying or reproduction of this dissertation may be referred to ProQuest Information and Learning, 300 North Zeeb Road, Ann Arbor, MI 48106- 1346, 1-800-521-0600, to whom the author has granted the right to reproduce and sell (a) copies of the manuscript in microform and/or (b) printed copies of the manuscript made from microform. Signature Date University of Washington Abstract Eigenfactor: ranking and mapping the scholarly literature Jevin D. West Chair of the Supervisory Committee: Professor Carl T. Bergstrom Department of Biology Each year, tens of thousands of scholarly journals publish hundreds of thou- sands of scholarly papers, collectively containing tens of millions of citations.