Eigenfactor: Ranking and Mapping Scientific Knowledge
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Citation Analysis for the Modern Instructor: an Integrated Review of Emerging Research
CITATION ANALYSIS FOR THE MODERN INSTRUCTOR: AN INTEGRATED REVIEW OF EMERGING RESEARCH Chris Piotrowski University of West Florida USA Abstract While online instructors may be versed in conducting e-Research (Hung, 2012; Thelwall, 2009), today’s faculty are probably less familiarized with the rapidly advancing fields of bibliometrics and informetrics. One key feature of research in these areas is Citation Analysis, a rather intricate operational feature available in modern indexes such as Web of Science, Scopus, Google Scholar, and PsycINFO. This paper reviews the recent extant research on bibliometrics within the context of citation analysis. Particular focus is on empirical studies, review essays, and critical commentaries on citation-based metrics across interdisciplinary academic areas. Research that relates to the interface between citation analysis and applications in higher education is discussed. Some of the attributes and limitations of citation operations of contemporary databases that offer citation searching or cited reference data are presented. This review concludes that: a) citation-based results can vary largely and contingent on academic discipline or specialty area, b) databases, that offer citation options, rely on idiosyncratic methods, coverage, and transparency of functions, c) despite initial concerns, research from open access journals is being cited in traditional periodicals, and d) the field of bibliometrics is rather perplex with regard to functionality and research is advancing at an exponential pace. Based on these findings, online instructors would be well served to stay abreast of developments in the field. Keywords: Bibliometrics, informetrics, citation analysis, information technology, Open resource and electronic journals INTRODUCTION In an ever increasing manner, the educational field is irreparably linked to advances in information technology (Plomp, 2013). -
The Publish Or Perish Book
The Publish or Perish Book Your guide to effective and responsible citation analysis Anne-Wil Harzing Edition: September 2010 For inquiries about this book, refer to the book's web page: http://www.harzing.com/popbook.htm ISBN 978-0-9808485-0-2 (PDF) ISBN 978-0-9808485-1-9 (paperback, colour) ISBN 978-0-9808485-2-6 (paperback, black & white) © 2010 by Anne-Wil Harzing All rights reserved. No part of this book may be reproduced in any form or by any electronic or mechanical means (including electronic mail, photocopying, recording, or information sto- rage and retrieval) without permission in writing from the publisher. As the SOLE exception to the above if you purchased this book in its PDF edition, then you are allowed to print 1 (one) hard copy for your own use only for each licence that you pur- chased. Published by Tarma Software Research Pty Ltd, Melbourne, Australia. National Library of Australia Cataloguing-in-Publication entry: Author Harzing, Anne-Wil. Title The publish or perish book [electronic resource]: Your guide to effective and responsible citation analysis / Anne-Wil Harzing. Edition 1st ed. ISBN 9780980848502 (pdf) Notes Includes bibliographical references. Subjects Publish or perish (Computer program), Authorship, Academic writing, Scholarly publishing. Dewey Number 808.02 TABLE OF CONTENTS PREFACE .................................................................................................................................... VII CHAPTER 1: INTRODUCTION TO CITATION ANALYSIS ................................................................... -
Is Sci-Hub Increasing Visibility of Indian Research Papers? an Analytical Evaluation Vivek Kumar Singh1,*, Satya Swarup Srichandan1, Sujit Bhattacharya2
Journal of Scientometric Res. 2021; 10(1):130-134 http://www.jscires.org Perspective Paper Is Sci-Hub Increasing Visibility of Indian Research Papers? An Analytical Evaluation Vivek Kumar Singh1,*, Satya Swarup Srichandan1, Sujit Bhattacharya2 1Department of Computer Science, Banaras Hindu University, Varanasi, Uttar Pradesh, INDIA. 2CSIR-National Institute of Science Technology and Development Studies, New Delhi, INDIA. ABSTRACT Sci-Hub, founded by Alexandra Elbakyan in 2011 in Kazakhstan has, over the years, Correspondence emerged as a very popular source for researchers to download scientific papers. It is Vivek Kumar Singh believed that Sci-Hub contains more than 76 million academic articles. However, recently Department of Computer Science, three foreign academic publishers (Elsevier, Wiley and American Chemical Society) have Banaras Hindu University, filed a lawsuit against Sci-Hub and LibGen before the Delhi High Court and prayed for Varanasi-221005, INDIA. complete blocking these websites in India. It is in this context, that this paper attempts to Email id: [email protected] find out how many Indian research papers are available in Sci-Hub and who downloads them. The citation advantage of Indian research papers available on Sci-Hub is analysed, Received: 16-03-2021 with results confirming that such an advantage do exist. Revised: 29-03-2021 Accepted: 25-04-2021 Keywords: Indian Research, Indian Science, Black Open Access, Open Access, Sci-Hub. DOI: 10.5530/jscires.10.1.16 INTRODUCTION access publishing of their research output, and at the same time encouraging their researchers to publish in openly Responsible Research and Innovation (RRI) has become one accessible forms. -
Publish Or Perish
Publish or Perish Maximizing the impact of your next publication Andrew Plume, Director – Scientometrics & Market Analysis City Graduate School Welcome Event - 18th November 2013 The research life: Enable, do, share Recruit/evaluate Secure Establish Manage Develop researchers funding partnerships facilities Strategy ? Search, read, Collaborate & Experiment Analyze & review network synthesize ! Have Manage data Publish and Commercialise Promote impact disseminate 2 The research life: Publish or Perish 3 Publish or Perish: Origin 4 Publish or Perish: Evidence of published authors agree/strongly agree*: “My career depends on a history of publishing research 81% articles in peer reviewed journals” Institutional: Career advancement and funding Reasons for National: Research assessment exercises agreeing Global: Research dissemination is a goal of research Articles in peer-reviewed journals make the most At my institution, there are defined thresholds of important contribution to my career in terms of status, publications for academic promotions at least during merit pay, and marketability, vs. teaching or service. early career. Engineering & Technology, UK (36-45) Social Science, USA (36-45) If I publish well (Impact Factor, h-index) I have more Because the primary role of my job is to produce chance to get a better position and to have grants. research which is of no use if it does not get into the Medicine & Allied Health, Italy (46-55) public domain. Earth & Planetary Sciences, UK (56-65) * Survey of 3,090 published authors in November 2012 -
Open Access Availability of Scientific Publications
Analytical Support for Bibliometrics Indicators Open access availability of scientific publications Analytical Support for Bibliometrics Indicators Open access availability of scientific publications* Final Report January 2018 By: Science-Metrix Inc. 1335 Mont-Royal E. ▪ Montréal ▪ Québec ▪ Canada ▪ H2J 1Y6 1.514.495.6505 ▪ 1.800.994.4761 [email protected] ▪ www.science-metrix.com *This work was funded by the National Science Foundation’s (NSF) National Center for Science and Engineering Statistics (NCSES). Any opinions, findings, conclusions or recommendations expressed in this report do not necessarily reflect the views of NCSES or the NSF. The analysis for this research was conducted by SRI International on behalf of NSF’s NCSES under contract number NSFDACS1063289. Analytical Support for Bibliometrics Indicators Open access availability of scientific publications Contents Contents .............................................................................................................................................................. i Tables ................................................................................................................................................................. ii Figures ................................................................................................................................................................ ii Abstract ............................................................................................................................................................ -
The Eigenfactor Metricstm: a Network Approach to Assessing Scholarly Journals
The Eigenfactor MetricsTM: A network approach to assessing scholarly journals Jevin D. West1 Theodore C. Bergstrom2 Carl T. Bergstrom1 July 16, 2009 1Department of Biology, University of Washington, Seattle, WA 2Department of Economics, University of California, Santa Barbara, CA *The authors are the founders of the Eigenfactor Project. All of the rank- ings, algorithms, visual tools and maps of science described here are freely available at http://www.eigenfactor.org/. Correspondence can be sent to Jevin D. West at [email protected]. Keywords: EigenfactorTM Score, Article InfluenceTM Score, Impact Factor, Bibliometrics, Citation Networks 1 Abstract Limited time and budgets have created a legitimate need for quan- titative measures of scholarly work. The well-known journal impact factor is the leading measure of this sort; here we describe an alter- native approach based on the full structure of the scholarly citation network. The Eigenfactor Metrics | Eigenfactor Score and Article Influence Score | use an iterative ranking scheme similar to Google's PageRank algorithm. By this approach, citations from top journals are weighted more heavily than citations from lower-tier publications. Here we describe these metrics and the rankings that they provide. 2 1 The Need for Alternative Metrics There is only one adequate approach to evaluating the quality of an individ- ual paper: read it carefully, or talk to others who have done so. The same is largely true when it comes to evaluating any small collection of papers, such as the publications of an individual scholar. But as one moves toward assessment challenges that involve larger bodies of work across broader seg- ments of scholarship, reading individual papers becomes infeasible and a legitimate need arises for quantitative metrics for research evaluation. -
Journal Rankings Paper
NBER WORKING PAPER SERIES NEOPHILIA RANKING OF SCIENTIFIC JOURNALS Mikko Packalen Jay Bhattacharya Working Paper 21579 http://www.nber.org/papers/w21579 NATIONAL BUREAU OF ECONOMIC RESEARCH 1050 Massachusetts Avenue Cambridge, MA 02138 September 2015 We thank Bruce Weinberg, Vetla Torvik, Neil Smalheiser, Partha Bhattacharyya, Walter Schaeffer, Katy Borner, Robert Kaestner, Donna Ginther, Joel Blit and Joseph De Juan for comments. We also thank seminar participants at the University of Illinois at Chicago Institute of Government and Public Affairs, at the Research in Progress Seminar at Stanford Medical School, and at the National Bureau of Economic Research working group on Invention in an Aging Society for helpful feedback. Finally, we thank the National Institute of Aging for funding for this research through grant P01-AG039347. We are solely responsible for the content and errors in the paper. The views expressed herein are those of the authors and do not necessarily reflect the views of the National Bureau of Economic Research. NBER working papers are circulated for discussion and comment purposes. They have not been peer- reviewed or been subject to the review by the NBER Board of Directors that accompanies official NBER publications. © 2015 by Mikko Packalen and Jay Bhattacharya. All rights reserved. Short sections of text, not to exceed two paragraphs, may be quoted without explicit permission provided that full credit, including © notice, is given to the source. Neophilia Ranking of Scientific Journals Mikko Packalen and Jay Bhattacharya NBER Working Paper No. 21579 September 2015 JEL No. I1,O3,O31 ABSTRACT The ranking of scientific journals is important because of the signal it sends to scientists about what is considered most vital for scientific progress. -
Google Scholar, Sci-Hub and Libgen: Could They Be Our New Partners?
Purdue University Purdue e-Pubs Proceedings of the IATUL Conferences 2017 IATUL Proceedings Google Scholar, Sci-Hub and LibGen: Could they be our New Partners? Louis Houle McGill University, [email protected] Louis Houle, "Google Scholar, Sci-Hub and LibGen: Could they be our New Partners?." Proceedings of the IATUL Conferences. Paper 3. https://docs.lib.purdue.edu/iatul/2017/partnership/3 This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] for additional information. GOOGLE SCHOLAR, SCI-HUB AND LIBGEN: COULD THEY BE OUR NEW PARTNERS? Louis Houle McGill University Canada [email protected] Abstract Since its debut I November 2004, librarians have raised several criticisms at Google Scholar (GS) such as its inconsistency of coverage and its currency and scope of coverage. It may have been true in the early years of Google Scholar but is this still through twelve years after? Is this sufficient to ignore it totally either in our information literacy programs or evaluate its value against the values of subscription-based abstracts and indexes? In this era of severe budget constraints that libraries are facing, can we imagine of substituting most or all of our subject databases with the free access of Google Scholar for discoverability? How much overlap between our databases and Google Scholar? How reliable is Google Scholar? How stable is its content over time? Open Access is getting to be the predominant form of getting access to peer reviewed articles. Many new non-traditional tools (institutional repositories, social media and peer to peer sites) are available out there to retrieve the full-text of peer reviewed articles. -
Exploratory Analysis of Publons Metrics and Their Relationship with Bibliometric and Altmetric Impact
Exploratory analysis of Publons metrics and their relationship with bibliometric and altmetric impact José Luis Ortega Institute for Advanced Social Studies (IESA-CSIC), Córdoba, Spain, [email protected] Abstract Purpose: This study aims to analyse the metrics provided by Publons about the scoring of publications and their relationship with impact measurements (bibliometric and altmetric indicators). Design/methodology/approach: In January 2018, 45,819 research articles were extracted from Publons, including all their metrics (scores, number of pre and post reviews, reviewers, etc.). Using the DOI identifier, other metrics from altmetric providers were gathered to compare the scores of those publications in Publons with their bibliometric and altmetric impact in PlumX, Altmetric.com and Crossref Event Data (CED). Findings: The results show that (1) there are important biases in the coverage of Publons according to disciplines and publishers; (2) metrics from Publons present several problems as research evaluation indicators; and (3) correlations between bibliometric and altmetric counts and the Publons metrics are very weak (r<.2) and not significant. Originality/value: This is the first study about the Publons metrics at article level and their relationship with other quantitative measures such as bibliometric and altmetric indicators. Keywords: Publons, Altmetrics, Bibliometrics, Peer-review 1. Introduction Traditionally, peer-review has been the most appropriate way to validate scientific advances. Since the first beginning of the scientific revolution, scientific theories and discoveries were discussed and agreed by the research community, as a way to confirm and accept new knowledge. This validation process has arrived until our days as a suitable tool for accepting the most relevant manuscripts to academic journals, allocating research funds or selecting and promoting scientific staff. -
Fumero Alfonso2017.Pdf (4.146Mb)
This thesis has been submitted in fulfilment of the requirements for a postgraduate degree (e.g. PhD, MPhil, DClinPsychol) at the University of Edinburgh. Please note the following terms and conditions of use: This work is protected by copyright and other intellectual property rights, which are retained by the thesis author, unless otherwise stated. A copy can be downloaded for personal non-commercial research or study, without prior permission or charge. This thesis cannot be reproduced or quoted extensively from without first obtaining permission in writing from the author. The content must not be changed in any way or sold commercially in any format or medium without the formal permission of the author. When referring to this work, full bibliographic details including the author, title, awarding institution and date of the thesis must be given. Accelerating Interpreted Programming Languages on GPUs with Just-In-Time Compilation and Runtime Optimisations Juan José Fumero Alfonso I V N E R U S E I T H Y T O H F G E R D I N B U Doctor of Philosophy Institute of Computing Systems Architecture School of Informatics University of Edinburgh 2017 Abstract Nowadays, most computer systems are equipped with powerful parallel devices such as Graphics Processing Units (GPUs). They are present in almost every com- puter system including mobile devices, tablets, desktop computers and servers. These parallel systems have unlocked the possibility for many scientists and companies to process significant amounts of data in shorter time. But the usage of these parallel systems is very challenging due to their programming complexity. -
On the Relation Between the Wos Impact Factor, the Eigenfactor, the Scimago Journal Rank, the Article Influence Score and the Journal H-Index
1 On the relation between the WoS impact factor, the Eigenfactor, the SCImago Journal Rank, the Article Influence Score and the journal h-index Ronald ROUSSEAU 1 and the STIMULATE 8 GROUP 2 1 KHBO, Dept. Industrial Sciences and Technology, Oostende, Belgium [email protected] 2Vrije Universiteit Brussel (VUB), Brussels, Belgium The STIMULATE 8 Group consists of: Anne Sylvia ACHOM (Uganda), Helen Hagos BERHE (Ethiopia), Sangeeta Namdev DHAMDHERE (India), Alicia ESGUERRA (The Philippines), Nguyen Thi Ngoc HOAN (Vietnam), John KIYAGA (Uganda), Sheldon Miti MAPONGA (Zimbabwe), Yohannis MARTÍ- LAHERA (Cuba), Kelefa Tende MWANTIMWA (Tanzania), Marlon G. OMPOC (The Philippines), A.I.M. Jakaria RAHMAN (Bangladesh), Bahiru Shifaw YIMER (Ethiopia). Abstract Four alternatives to the journal Impact Factor (IF) indicator are compared to find out their similarities. Together with the IF, the SCImago Journal Rank indicator (SJR), the EigenfactorTM score, the Article InfluenceTM score and the journal h- index of 77 journals from more than ten fields were collected. Results show that although those indicators are calculated with different methods and even use different databases, they are strongly correlated with the WoS IF and among each other. These findings corroborate results published by several colleagues and show the feasibility of using free alternatives to the Web of Science for evaluating scientific journals. Keywords: WoS impact factor, Eigenfactor, SCImago Journal Rank (SJR), Article Influence Score, journal h-index, correlations Introduction STIMULATE stands for Scientific and Technological Information Management in Universities and Libraries: an Active Training Environment. It is an international training programme in information management, supported by the Flemish Interuniversity Council (VLIR), aiming at young scientists and professionals from 2 developing countries. -
Do You Speak Open Science? Resources and Tips to Learn the Language
Do You Speak Open Science? Resources and Tips to Learn the Language. Paola Masuzzo1, 2 - ORCID: 0000-0003-3699-1195, Lennart Martens1,2 - ORCID: 0000- 0003-4277-658X Author Affiliation 1 Medical Biotechnology Center, VIB, Ghent, Belgium 2 Department of Biochemistry, Ghent University, Ghent, Belgium Abstract The internet era, large-scale computing and storage resources, mobile devices, social media, and their high uptake among different groups of people, have all deeply changed the way knowledge is created, communicated, and further deployed. These advances have enabled a radical transformation of the practice of science, which is now more open, more global and collaborative, and closer to society than ever. Open science has therefore become an increasingly important topic. Moreover, as open science is actively pursued by several high-profile funders and institutions, it has fast become a crucial matter to all researchers. However, because this widespread interest in open science has emerged relatively recently, its definition and implementation are constantly shifting and evolving, sometimes leaving researchers in doubt about how to adopt open science, and which are the best practices to follow. This article therefore aims to be a field guide for scientists who want to perform science in the open, offering resources and tips to make open science happen in the four key areas of data, code, publications and peer-review. The Rationale for Open Science: Standing on the Shoulders of Giants One of the most widely used definitions of open science originates from Michael Nielsen [1]: “Open science is the idea that scientific knowledge of all kinds should be openly shared as early as is practical in the discovery process”.