Impact Factor, H-Index, I10-Index and I20-Index of Webology
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Citation Analysis for the Modern Instructor: an Integrated Review of Emerging Research
CITATION ANALYSIS FOR THE MODERN INSTRUCTOR: AN INTEGRATED REVIEW OF EMERGING RESEARCH Chris Piotrowski University of West Florida USA Abstract While online instructors may be versed in conducting e-Research (Hung, 2012; Thelwall, 2009), today’s faculty are probably less familiarized with the rapidly advancing fields of bibliometrics and informetrics. One key feature of research in these areas is Citation Analysis, a rather intricate operational feature available in modern indexes such as Web of Science, Scopus, Google Scholar, and PsycINFO. This paper reviews the recent extant research on bibliometrics within the context of citation analysis. Particular focus is on empirical studies, review essays, and critical commentaries on citation-based metrics across interdisciplinary academic areas. Research that relates to the interface between citation analysis and applications in higher education is discussed. Some of the attributes and limitations of citation operations of contemporary databases that offer citation searching or cited reference data are presented. This review concludes that: a) citation-based results can vary largely and contingent on academic discipline or specialty area, b) databases, that offer citation options, rely on idiosyncratic methods, coverage, and transparency of functions, c) despite initial concerns, research from open access journals is being cited in traditional periodicals, and d) the field of bibliometrics is rather perplex with regard to functionality and research is advancing at an exponential pace. Based on these findings, online instructors would be well served to stay abreast of developments in the field. Keywords: Bibliometrics, informetrics, citation analysis, information technology, Open resource and electronic journals INTRODUCTION In an ever increasing manner, the educational field is irreparably linked to advances in information technology (Plomp, 2013). -
How Can Citation Impact in Bibliometrics Be Normalized?
RESEARCH ARTICLE How can citation impact in bibliometrics be normalized? A new approach combining citing-side normalization and citation percentiles an open access journal Lutz Bornmann Division for Science and Innovation Studies, Administrative Headquarters of the Max Planck Society, Hofgartenstr. 8, 80539 Munich, Germany Downloaded from http://direct.mit.edu/qss/article-pdf/1/4/1553/1871000/qss_a_00089.pdf by guest on 01 October 2021 Keywords: bibliometrics, citation analysis, citation percentiles, citing-side normalization Citation: Bornmann, L. (2020). How can citation impact in bibliometrics be normalized? A new approach ABSTRACT combining citing-side normalization and citation percentiles. Quantitative Since the 1980s, many different methods have been proposed to field-normalize citations. In this Science Studies, 1(4), 1553–1569. https://doi.org/10.1162/qss_a_00089 study, an approach is introduced that combines two previously introduced methods: citing-side DOI: normalization and citation percentiles. The advantage of combining two methods is that their https://doi.org/10.1162/qss_a_00089 advantages can be integrated in one solution. Based on citing-side normalization, each citation Received: 8 May 2020 is field weighted and, therefore, contextualized in its field. The most important advantage of Accepted: 30 July 2020 citing-side normalization is that it is not necessary to work with a specific field categorization scheme for the normalization procedure. The disadvantages of citing-side normalization—the Corresponding Author: Lutz Bornmann calculation is complex and the numbers are elusive—can be compensated for by calculating [email protected] percentiles based on weighted citations that result from citing-side normalization. On the one Handling Editor: hand, percentiles are easy to understand: They are the percentage of papers published in the Ludo Waltman same year with a lower citation impact. -
Research Performance of Top Universities in Karnataka: Based on Scopus Citation Index Kodanda Rama PES College of Engineering, [email protected]
University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Library Philosophy and Practice (e-journal) Libraries at University of Nebraska-Lincoln September 2019 Research Performance of Top Universities in Karnataka: Based on Scopus Citation Index Kodanda Rama PES College of Engineering, [email protected] C. P. Ramasesh [email protected] Follow this and additional works at: https://digitalcommons.unl.edu/libphilprac Part of the Scholarly Communication Commons, and the Scholarly Publishing Commons Rama, Kodanda and Ramasesh, C. P., "Research Performance of Top Universities in Karnataka: Based on Scopus Citation Index" (2019). Library Philosophy and Practice (e-journal). 2889. https://digitalcommons.unl.edu/libphilprac/2889 Research Performance of Top Universities in Karnataka: Based on Scopus Citation Index 1 2 Kodandarama and C.P. Ramasesh ABSTRACT: [Paper furnishes the results of the analysis of citations of research papers covered by Scopus database of Elsevier, USA. The coverage of the database is complete; citations depicted by Scopus upto June 2019 are considered. Study projects the research performance of six well established top universities in the state of Karnataka with regard the number of research papers covered by scholarly journals and number of scholars who have cited these research papers. Also projected is the average citations per research paper and h-Index of authors. Paper also projects the performance of top faculty members who are involved in contributing research papers. Collaboration with authors of foreign countries in doing research work and publishing papers are also comprehended in the study, including the trends in publishing research papers which depict the decreasing and increasing trends of research work.] INTRODUCTION: Now-a-days, there is emphasis on improving the quality of research papers on the whole. -
A Quick Guide to Scholarly Publishing
A QUICK GUIDE TO SCHOLARLY PUBLISHING GRADUATE WRITING CENTER • GRADUATE DIVISION UNIVERSITY OF CALIFORNIA • BERKELEY Belcher, Wendy Laura. Writing Your Journal Article in 12 Weeks: A Guide to Academic Publishing Success. Thousand Oaks, CA: SAGE Publications, 2009. Benson, Philippa J., and Susan C. Silver. What Editors Want: An Author’s Guide to Scientific Journal Publishing. Chicago: University of Chicago Press, 2013. Derricourt, Robin. An Author’s Guide to Scholarly Publishing. Princeton, NJ: Princeton University Press, 1996. Germano, William. From Dissertation to Book. 2nd ed. Chicago Guides to Writing, Editing, and Publishing. Chicago: University of Chicago Press, 2013. ———. Getting It Published: A Guide for Scholars and Anyone Else Serious about Serious Books. 3rd ed. Chicago: University of Chicago Press, 2016. Goldbort, Robert. Writing for Science. New Haven & London: Yale University Press, 2006. Harman, Eleanor, Ian Montagnes, Siobhan McMenemy, and Chris Bucci, eds. The Thesis and the Book: A Guide for First- Time Academic Authors. 2nd ed. Toronto: University of Toronto Press, 2003. Harmon, Joseph E., and Alan G. Gross. The Craft of Scientific Communication. Chicago: University of Chicago Press, 2010. Huff, Anne Sigismund. Writing for Scholarly Publication. Thousand Oaks, CA: SAGE Publications, 1999. Luey, Beth. Handbook for Academic Authors. 5th ed. Cambridge, UK: Cambridge University Press, 2009. Luey, Beth, ed. Revising Your Dissertation: Advice from Leading Editors. Updated ed. Berkeley: University of California Press, 2007. Matthews, Janice R., John M. Bowen, and Robert W. Matthews. Successful Scientific Writing: A Step-by-Step Guide for the Biological and Medical Sciences. 3rd ed. Cambridge, UK: Cambridge University Press, 2008. Moxley, Joseph M. PUBLISH, Don’t Perish: The Scholar’s Guide to Academic Writing and Publishing. -
Scientific Journal Article: Writing Summaries and Critiques Definition of Genre Summaries and Critiques Are Two Ways to Write a Review of a Scientific Journal Article
How to Read and Review a Scientific Journal Article: Writing Summaries and Critiques Definition of Genre Summaries and critiques are two ways to write a review of a scientific journal article. Both types of writing ask you first to read and understand an article from the primary literature about your topic. The summary involves briefly but accurately stating the key points of the article for a reader who has not read the original article. The critique begins by summarizing the article and then analyzes and evaluates the author’s research. Summaries and critiques help you learn to synthesize information from different sources and are usually limited to two pages maximum. Actions to Take 1. Skim the article without taking notes: cture.” 2. Re-read the article more carefully: is at the very end of the paper) and “Results” sections. Was the study repeated? What was the sample size? Is this representative of the larger population? What variables were held constant? Was there a control? What factors might affect the outcome? 3. Read the “Materials and Methods” and “Results” sections multiple times: raphs, tables, and diagrams. 4. Before you begin the first draft of your summary: Try to describe the article in your own words first. Try to distill the article down to its “scientific essence.” -written summary: The egg capsules of the marine snails Nucella lamellosa and N. lima protect developing embryos against low-salinity stress, even though the solute concentration within the capsules falls to near that of the surrounding water within about 1 h. 5. Write a draft of your summary: and avoid unintentional plagiarism. -
An Initiative to Track Sentiments in Altmetrics
Halevi, G., & Schimming, L. (2018). An Initiative to Track Sentiments in Altmetrics. Journal of Altmetrics, 1(1): 2. DOI: https://doi.org/10.29024/joa.1 PERSPECTIVE An Initiative to Track Sentiments in Altmetrics Gali Halevi and Laura Schimming Icahn School of Medicine at Mount Sinai, US Corresponding author: Gali Halevi ([email protected]) A recent survey from Pew Research Center (NW, Washington & Inquiries 2018) found that over 44 million people receive science-related information from social media channels to which they subscribe. These include a variety of topics such as new discoveries in health sciences as well as “news you can use” information with practical tips (p. 3). Social and news media attention to scientific publications has been tracked for almost a decade by several platforms which aggregate all public mentions of and interactions with scientific publications. Since the amount of comments, shares, and discussions of scientific publications can reach an audience of thousands, understanding the overall “sentiment” towards the published research is a mammoth task and typically involves merely reading as many posts, shares, and comments as possible. This paper describes an initiative to track and provide sentiment analysis to large social and news media mentions and label them as “positive”, “negative” or “neutral”. Such labels will enable an overall understanding of the content that lies within these social and news media mentions of scientific publications. Keywords: sentiment analysis; opinion mining; scholarly output Background Traditionally, the impact of research has been measured through citations (Cronin 1984; Garfield 1964; Halevi & Moed 2015; Moed 2006, 2015; Seglen 1997). The number of citations a paper receives over time has been directly correlated to the perceived impact of the article. -
On the Relation Between the Wos Impact Factor, the Eigenfactor, the Scimago Journal Rank, the Article Influence Score and the Journal H-Index
1 On the relation between the WoS impact factor, the Eigenfactor, the SCImago Journal Rank, the Article Influence Score and the journal h-index Ronald ROUSSEAU 1 and the STIMULATE 8 GROUP 2 1 KHBO, Dept. Industrial Sciences and Technology, Oostende, Belgium [email protected] 2Vrije Universiteit Brussel (VUB), Brussels, Belgium The STIMULATE 8 Group consists of: Anne Sylvia ACHOM (Uganda), Helen Hagos BERHE (Ethiopia), Sangeeta Namdev DHAMDHERE (India), Alicia ESGUERRA (The Philippines), Nguyen Thi Ngoc HOAN (Vietnam), John KIYAGA (Uganda), Sheldon Miti MAPONGA (Zimbabwe), Yohannis MARTÍ- LAHERA (Cuba), Kelefa Tende MWANTIMWA (Tanzania), Marlon G. OMPOC (The Philippines), A.I.M. Jakaria RAHMAN (Bangladesh), Bahiru Shifaw YIMER (Ethiopia). Abstract Four alternatives to the journal Impact Factor (IF) indicator are compared to find out their similarities. Together with the IF, the SCImago Journal Rank indicator (SJR), the EigenfactorTM score, the Article InfluenceTM score and the journal h- index of 77 journals from more than ten fields were collected. Results show that although those indicators are calculated with different methods and even use different databases, they are strongly correlated with the WoS IF and among each other. These findings corroborate results published by several colleagues and show the feasibility of using free alternatives to the Web of Science for evaluating scientific journals. Keywords: WoS impact factor, Eigenfactor, SCImago Journal Rank (SJR), Article Influence Score, journal h-index, correlations Introduction STIMULATE stands for Scientific and Technological Information Management in Universities and Libraries: an Active Training Environment. It is an international training programme in information management, supported by the Flemish Interuniversity Council (VLIR), aiming at young scientists and professionals from 2 developing countries. -
— Emerging Sources Citation Index
WEB OF SCIENCE™ CORE COLLECTION — EMERGING SOURCES CITATION INDEX THE VALUE OF COVERAGE IN INTRODUCING THE EMERGING SOURCES WEB OF SCIENCE CITATION INDEX This year, Clarivate Analytics is launching the Emerging INCREASING THE VISIBILITY OF Sources Citation Index (ESCI), which will extend the SCHOLARLY RESEARCH universe of publications in Web of Science to include With content from more than 12,700 top-tier high-quality, peer-reviewed publications of regional international and regional journals in the Science importance and in emerging scientific fields. ESCI will Citation Index Expanded™ (SCIE),the Social Sciences also make content important to funders, key opinion Citation Index® (SSCI), and the Arts & Humanities leaders, and evaluators visible in Web of Science Core Citation Index® (AHCI); more than 160,000 proceedings Collection even if it has not yet demonstrated citation in the Conference Proceedings Citation Index; and impact on an international audience. more than 68,000 books in the Book Citation Index; Journals in ESCI have passed an initial editorial Web of Science Core Collection sets the benchmark evaluation and can continue to be considered for for information on research across the sciences, social inclusion in products such as SCIE, SSCI, and AHCI, sciences, and arts and humanities. which have rigorous evaluation processes and selection Only Web of Science Core Collection indexes every paper criteria. All ESCI journals will be indexed according in the journals it covers and captures all contributing to the same data standards, including cover-to-cover institutions, no matter how many there are. To be indexing, cited reference indexing, subject category included for coverage in the flagship indexes in the assignment, and indexing all authors and addresses. -
Amplifying the Impact of Open Access: Wikipedia and the Diffusion of Science
(forthcoming in the Journal of the Association for Information Science and Technology) Amplifying the Impact of Open Access: Wikipedia and the Diffusion of Science Misha Teplitskiy Grace Lu Eamon Duede Dept. of Sociology and KnowledgeLab Computation Institute and KnowledgeLab University of Chicago KnowledgeLab University of Chicago [email protected] University of Chicago [email protected] (773) 834-4787 [email protected] (773) 834-4787 5735 South Ellis Avenue (773) 834-4787 5735 South Ellis Avenue Chicago, Illinois 60637 5735 South Ellis Avenue Chicago, Illinois 60637 Chicago, Illinois 60637 Abstract With the rise of Wikipedia as a first-stop source for scientific knowledge, it is important to compare its representation of that knowledge to that of the academic literature. Here we identify the 250 most heavi- ly used journals in each of 26 research fields (4,721 journals, 19.4M articles in total) indexed by the Scopus database, and test whether topic, academic status, and accessibility make articles from these journals more or less likely to be referenced on Wikipedia. We find that a journal’s academic status (im- pact factor) and accessibility (open access policy) both strongly increase the probability of its being ref- erenced on Wikipedia. Controlling for field and impact factor, the odds that an open access journal is referenced on the English Wikipedia are 47% higher compared to paywall journals. One of the implica- tions of this study is that a major consequence of open access policies is to significantly amplify the dif- fusion of science, through an intermediary like Wikipedia, to a broad audience. Word count: 7894 Introduction Wikipedia, one of the most visited websites in the world1, has become a destination for information of all kinds, including information about science (Heilman & West, 2015; Laurent & Vickers, 2009; Okoli, Mehdi, Mesgari, Nielsen, & Lanamäki, 2014; Spoerri, 2007). -
The Thomson Scientific Journal Selection Process
PERSPECTIVES INTERNATIONAL MICROBIOLOGY (2006) 9:135-138 www.im.microbios.org James Testa The Thomson Scientific Director of Editorial journal selection process Development, Thomson Scientific, Philadelphia, Pennsylvania, USA E-mail: [email protected] several multidisciplinary citation indexes. In addition to its Introduction three major components, the Science Citation Index Expanded, the Social Sciences Citation Index, and the Arts The Web of Knowledge is an electronic platform that consists of and Humanities Citation Index, it includes Index Chemicus important searchable bibliographic databases and key analytical and Current Chemical Reactions. Since significant resources tools, such as the Journal Citation Reports (JCR). Several of are dedicated to the selection of the most important and influ- these bibliographic databases pro- ential journals for inclusion in the WoS, we will examine the vide extensive coverage of the lit- journal selection process used erature of specific subjects, such to achieve this goal, particu- as the Life Sciences, Agriculture larly with regard to citation and Forestry, Food Science, Phy- analysis and the role of the sics and Engineering, Medicine, impact factor. Behavioral Sciences, and Animal The journal selection Biology (Fig. 1). The internation- process has been applied con- al patent literature and conference sistently for more than 45 proceedings are also key compo- years. Every journal included nents of the Web of Knowledge. in the WoS has been evaluated In all, this environment covers according to the highest stan- over 22,000 unique journals, more dards, which were originally than 60,000 conferences from defined by Eugene Garfield, 1990 to the present, and approxi- the founder and president mately one and a half million Fig. -
Implementing the RDA Research Data Policy Framework in Slovenian Scientific Journals.Data Science Journal, 19: 49 Pp
Štebe, J, et al. 2020. Implementing the RDA Research Data Policy Framework in Slovenian Scientific Journals. Data Science Journal, 19: 49 pp. 1–15. DOI: https://doi.org/10.5334/dsj-2020-049 RESEARCH PAPER Implementing the RDA Research Data Policy Framework in Slovenian Scientific Journals Janez Štebe, Maja Dolinar, Sonja Bezjak and Ana Inkret Slovenian Social Science Data Archives (ADP – Arhiv družboslovnih podatkov), University of Ljubljana, Ljubljana, SI Corresponding author: Janez Štebe ([email protected]) The paper aims to present the implementation of the RDA research data policy framework in Slovenian scientific journals within the project RDA Node Slovenia. The activity aimed to implement the practice of data sharing and data citation in Slovenian scientific journals and was based on internationally renowned practices and policies, particularly the Research Data Policy Framework of the RDA Data Policy Standardization and Implementation Interest Group. Following this, the RDA Node Slovenia coordination prepared a guidance document that allowed the four pilot participating journals (from fields of archaeology, history, linguistics and social sciences) to adjust their journal policies regarding data sharing, data citation, adapted the definitions of research data and suggested appropriate data repositories that suit their dis- ciplinary specifics. The comparison of results underlines how discipline-specific the aspects of data-sharing are. The pilot proved that a grass-root approach in advancing open science can be successful and well-received in the research community, however, it also pointed out several issues in scientific publishing that would benefit from a planned action on a national level. The context of an underdeveloped data sharing culture, slow implementation of open data strategy by the national research funder and sparse national data service infrastructure creates a unique environment for this study, the result of which can be used in similar contexts worldwide. -
Ranking Forestry Journals Using the H-Index
Ranking forestry journals using the h-index Journal of Informetrics, in press Jerome K Vanclay Southern Cross University PO Box 157, Lismore NSW 2480, Australia [email protected] Abstract An expert ranking of forestry journals was compared with journal impact factors and h -indices computed from the ISI Web of Science and internet-based data. Citations reported by Google Scholar offer an efficient way to rank all journals objectively, in a manner consistent with other indicators. This h-index exhibited a high correlation with the journal impact factor (r=0.92), but is not confined to journals selected by any particular commercial provider. A ranking of 180 forestry journals is presented, on the basis of this index. Keywords : Hirsch index, Research quality framework, Journal impact factor, journal ranking, forestry Introduction The Thomson Scientific (TS) Journal Impact Factor (JIF; Garfield, 1955) has been the dominant measure of journal impact, and is often used to rank journals and gauge relative importance, despite several recognised limitations (Hecht et al., 1998; Moed et al., 1999; van Leeuwen et al., 1999; Saha et al., 2003; Dong et al., 2005; Moed, 2005; Dellavalle et al., 2007). Other providers offer alternative journal rankings (e.g., Lim et al., 2007), but most deal with a small subset of the literature in any discipline. Hirsch’s h-index (Hirsch, 2005; van Raan, 2006; Bornmann & Daniel, 2007a) has been suggested as an alternative that is reliable, robust and easily computed (Braun et al., 2006; Chapron and Husté, 2006; Olden, 2007; Rousseau, 2007; Schubert and Glänzel, 2007; Vanclay, 2007; 2008).