Advanced Scientometrics Workshop

Total Page:16

File Type:pdf, Size:1020Kb

Advanced Scientometrics Workshop Advanced Scientometrics Workshop Nilofar Barahmand 1 Phd Student of Library and Information Sciences [email protected] 2 Author Level Metrics in Scopus Productivity Impact Document Count Citation Count With out self citation H- index Field-weighted Citation Impact (FWCI) h-index 3 The h-index is an author-level metric that attempts to measure both the productivity and citation impact of the publications of a scientist or scholar. “A scientist has index h if h of his or her Np papers have at least h citations each and the other (Np – h) papers have ≤h citations each.” 4 h- index: 5 Scopus Merge Author 6 Field-weighted Citation Impact (FWCI) Entity Subject Area Citation/Paper Researcher A Clinical Medicine 3 Researcher B Mathematics 1 Which one’s Performance is Better? 7 Field-weighted Citation Impact (FWCI) Field-Weighted Citation Impact is a Ratio that takes into account the differences in research behavior across disciplines. Field-Weighted Citation Impact is the ratio of the total citations actually received by the denominator’s output, and the total citations that would be expected based on the average of the subject field, year & document type. 8 Field-weighted Citation Impact (FWCI) Entity Subject Area Citation/Paper Researcher A Clinical Medicine 3 Researcher B Mathematics 1 Which one’s Performance is Better? Field-Weighted Entity Subject Area Citation Impact Researcher A Clinical Medicine 1.5 Researcher B Mathematics 3 Article Level Metrics in Scopus 9 Citation count and percentile benchmark Field-Weighted Citation Impact (FWCI) Altmetrics Altmetrics is a quantitative measure of Scholarly Activity Downloads and posts in common research tools the quality and quantity of attention such as Mendeley and CiteULike that a scholarly work is receiving through social media, citations, and Mentions characterized by rapid, brief engagement article downloads. Social Activity on platforms used by the public, such as Twitter, Facebook and Google+ Altmetrics are not meant to replace Scholarly Reviews, articles and blogs by experts and citation counts or the h-index, but scholars, such as F1000 Prime, research blogs and instead compliment the metrics with Commentary Wikipedia additional data. Coverage of research output in the mass media Mass Media (e.g., coverage in top tier media media) Citation count and percentile benchmark 10 Citation count: The Citation Count shows how many times a publication has been cited. Citation benchmarking: Citation benchmarking shows how citations received by this article compare with the average for similar articles. 99th percentile is high, and indicates an article in the top 1% globally. Citation benchmarking takes into account: The date of publication The document type (number of articles allowing) Disciplines associated with its source Citation benchmarking compares articles within an 18 month window and is computed separately for each of its sources' disciplines. 11 Field-weighted Citation Impact (FWCI) Entity Subject Area Citation Article A Clinical Medicine 3 Article B Mathematics 1 Which one’s Performance is Better? 12 Field-weighted Citation Impact (FWCI) Field-Weighted Citation Impact is a Ratio that takes into account the differences in research behavior across disciplines. Field-Weighted Citation Impact is the ratio of the total citations actually received by the denominator’s output, and the total citations that would be expected based on the average of the subject field, year & document type. 13 Field-weighted Citation Impact (FWCI) FWCI=1 means that the output performs just as expected for the global average. FWCI>1 means that the output is more cited than expected according to the global average; for example, 1.48 means 48% more cited than expected. FWCI<1 means that the output is cited less than expected according to the global average. 14 Field-weighted Citation Impact (FWCI) Entity Subject Area Citation Article A Clinical Medicine 3 Article B Mathematics 1 Which one’s Performance is Better? Field-Weighted Entity Subject Area Citation Impact Article A Clinical Medicine 1.5 Article B Mathematics 3 15 Journal Metrics Journal metrics provide information on the comparative performance of academic journals. Journal metrics are statistics, usually based on citation counts. Journal metrics can also be used to rank journals. Commonly used journal metrics are: • Impact factor • CiteScore • h- index • SJR • SNIP • Quartile 16 Impact Factor The Journal Impact Factor reflects the average number of citations to recent articles published in that journal. It’s one measure of the relative importance of a journal within its field. High impact factor ~ more important A = the number of times articles published in 2012 and 2013 were cited by indexed journals during 2014. B = the total number of "citable items" published in 2012 and 2013. A/B = 2014 impact factor 17 CiteScore CiteScore metrics calculate the citations from all documents in year one to all documents published in the prior three years for a title. As an example, to calculate a 2015 value, CiteScore counts the citations received in 2015 to documents published in 2012, 2013 or 2014. This number is divided by the number of documents indexed on Scopus published in 2012, 2013 and 2014. 18 H-index 19 CiteScore rank and trend 20 Snip: Source Normalized Impact per Paper SNIP measures a source’s contextual citation impact by weighting citations based on the total number of citations in a subject field. It helps you make a direct comparison of sources in different subject fields. SNIP takes into account characteristics of the source's subject field, which is the set of documents citing that source. SNIP is the ratio of a source's average citation count per paper and the citation potential of its subject field. 21 SJR (SCImago Journal Rank) SJR is weighted by the prestige of a journal. Subject field, quality, and reputation of the journal have a direct effect on the value of a citation. SJR assigns relative scores to all of the sources in a citation network. Its methodology is inspired by the Google PageRank algorithm, in that not all citations are equal. A source transfers its own 'prestige', or status, to another source through the act of citing it. A citation from a source with a relatively high SJR is worth more than a citation from a source with a lower SJR. 22 Scimago Journal & Country Rank http://www.scimagojr.com/ 23 Scimago Journal & Country Rank 24 Journal Citation Ranking and Quartile Scores (Q1-Q4) In ISI: Based on Impact Factor (IF) data, the Journal Citation Reports published by Thomson Reuters (see also here) provides yearly rankings of science and social science journals, in the subject categories relevant for the journal (in fact, there may be more than one). Quartile rankings are therefore derived for each journal in each of its subject categories according to which quartile of the IF distribution the journal occupies for that subject category. Q1 denotes the top 25% of the IF distribution, Q2 for middle-high position (between top 50% and top 25%), Q3 middle-low position (top 75% to top 50%), and Q4 the lowest position (bottom 25% of the IF distribution). 25 Journal Citation Ranking and Quartile Scores (Q1-Q4) In Scopus: Based on CiteScore and SJR data Scopus provides yearly rankings of journals in the subject categories relevant for the journal (in fact, there may be more than one). Quartile rankings are therefore derived for each journal in each of its subject categories according to which quartile of the on CiteScore and SJR distribution the journal occupies for that subject category. Q1 denotes the top 25% of the on CiteScore and SJR distribution, Q2 for middle-high position (between top 50% and top 25%), Q3 middle-low position (top 75% to top 50%), and Q4 the lowest position (bottom 25% of the on CiteScore and SJR distribution). 26 Journal Citation Ranking and Quartile Scores (Q1-Q4) 27 Journal Finders Elsevier Springer Ulrich 28 Thank you for your attention!.
Recommended publications
  • Australian Business Deans Council 2019 Journal Quality List Review Final Report 6 December 2019
    Australian Business Deans Council 2019 Journal Quality List Review Final Report 6 December 2019 1 About the Australian Business Deans Council The Australian Business Deans Council (ABDC) is the peak body of Australian university business schools. Our 38 members graduate one-third of all Australian university students and more than half of the nation’s international tertiary students. ABDC’s mission is to make Australian business schools better, and to foster the national and global impact of Australian business education and research. ABDC does this by: • Being the collective and collegial voice of university business schools • Providing opportunities for members to share knowledge and best practice • Creating and maintaining strong, collaborative relationships with affiliated national and international peak industry, higher education, professional and government bodies • Engaging in strategic initiatives and activities that further ABDC’s mission. Australian Business Deans Council Inc. UNSW Business School, Deans Unit, Level 6, West Lobby, College Road, Kensington, NSW, Australia 2052 T: +61 (0)2 6162 2970 E: [email protected] 2 Table of Contents Acknowledgements 4 Background and Context 4 Method and Approach 7 Outcomes 10 Beyond the 2019 Review 13 Appendix 1 – Individual Panel Reports 14 Information Systems 15 Economics 20 Accounting 37 Finance including Actuarial Studies 57 Management, Commercial Services and Transport and Logistics 63 (and Other, covering 1599) Marketing and Tourism 78 Business and Taxation Law 85 Appendix 2 – Terms
    [Show full text]
  • New Perspectives Welcome to the First Research Trends Maga- Zine, Which Accompanies Our 15Th Issue of Research Trends Online
    r ReseaRch TRendst magazine New perspectives Welcome to the first Research Trends maga- zine, which accompanies our 15th issue of Research Trends online. Research Trends 15 is a special issue devoted to the role of bibliometric indicators in journal evaluation and research assessment. Over the past 40 years, academic evaluation has changed radically, in both its objectives and methods. Bibliometrics has grown to meet emerging challenges, and new metrics are frequently introduced to measure different aspects of academic performance and improve upon earlier, pioneering work. it is becoming increasingly apparent that assessment and benchmarking are here to stay, and bibliometrics are an intrinsic aspect of today’s evaluation landscape, whether of journals, researchers or insti- tutes. This is not uncontroversial, and we have featured many critics of bibliometric analysis in general and of specific tools or their applications in past issues (please visit www.researchtrends.com). The bibliometric community acknowl- edges these shortcomings and is tirelessly campaigning for better understanding and careful use of the metrics they produce. Therefore, in this issue, we speak to three of the world’s leading researchers in this area, who all agree that bibliometric indi- cators should always be used with exper- tise and care, and that they should never completely supercede more traditional methods of evaluation, such as peer review. in fact, they all advocate a wider choice of more carefully calibrated metrics as the only way to ensure fair assessment. if you would like to read any of our back issues or subscribe to our bi-monthly e-newsletter, please visit www.researchtrends.com Kind regards, The Research Trends editorial Board Research Trends is a bi-monthly online newsletter providing objective, up-to-the-minute insights into scientific trends based on bibliometric analysis.
    [Show full text]
  • Research Metrics Definitions
    University Library Research Metrics Definitions Research Metrics 1. Author metrics h-index The h-index was developed by J.E. Hirsch and published in PNAS in 2005. The h-index is method of measuring the productivity and impact of a researcher's work and is calculated using the number of publications with a citation number greater than or equal to h. Example: An author with 20 publications that have at least 20 citations each will have a h-index of 20 Your h-index will vary, depending on which research outputs are included in the calculation. Tools: Aurora, Google Scholar, Scopus, Web of Science i10-index The number of publications with at least 10 citations. Tool: Google Scholar Total citations Tools: Google Scholar, Scopus, Web of Science Altmetrics Altmetrics are alternative metrics such as social media mentions, downloads etc. Tools: Altmetric Explorer (aggregated and article-level) 2. Article metrics Citation Count The number of times an article has been cited Tools: Aurora, Google Scholar, Scopus, Web of Science Field Weighted Citation Impact (FWCI) This shows how well cited this article is when compared to similar articles. A FWCI greater than 1.00 means the article is more cited than expected according to the world average. FWCI takes into account the year of publication, document type, and the disciplines associated with its source. Tools: Scopus Altmetrics Altmetrics are alternative metrics such as social media mentions, downloads etc. Tools: Altmetric Explorer, PlumX Metrics (Scopus), Journal web pages Last updated April 2019 Research Metrics 3. Journal metrics Journal Impact Factor (JIF) Definition The Journal Impact Factor is the average number of times articles from the journal published in the past two years have been cited in the JCR year.
    [Show full text]
  • Citation Metrics Overview
    UNIVERSITY LIBRARY Contents ARTICLE-LEVEL CITATION METRICS ........................................................................................................................................................................................................................ 1 AUTHOR-LEVEL METRICS ....................................................................................................................................................................................................................................... 2 JOURNAL-LEVEL METRICS ...................................................................................................................................................................................................................................... 4 Article-Level Citation Metrics These metrics can be used as indicators of the performance of individual articles or publications. Data Field- Metric Brief Description Access via Timeframe Source(s) Weighted* Citation Count Number of citations to the document captured within the data Scopus, Web Scopus, Web source. of Science, of Science, No Customisable Google Google Scholar Scholar Field-Weighted Citation Measure of citations received by a document relative to the world Impact (FWCI) average of citations received by documents of similar type, year of publication, and research field over a three year period. An FWCI Scopus Scopus Yes 3 years of 1.00 indicates that the document has been cited as expected based on the global average. Citation benchmarking Measure
    [Show full text]
  • 2017​ ​Metrics​ ​Tools​ ​-​ ​Current​ ​Strengths​ ​And
    2017 Metrics Tools - Current Strengths and Issues ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ Author evaluation ​ ​ Strengths Known Issues ​ ​ h-index ● Combines measure of quantity ● Different for different fields of research – ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ (publications) and impact (citations) may be more useful for ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ● Provides high objectivity Science/Engineering, less suitable for ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ● Easily obtained Education and the Social Sciences ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ● Measures impact of researcher rather ● Depends on duration of researcher’s ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ than of any particular publication career ​ ​ ​ ​ ​ ​ ​ ​ ● Does not highlight highly cited papers ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ● Difficult to obtain for complete output ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ including all material types ​ ​ ​ ​ ​ ​ ● Different sources may give different ​ ​ ​ ​ ​ ​ ​ ​ values e.g. Google Scholar vs Web of ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ Science ● Authors need to check that publications ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ counted belong to them ​ ​ ​ ​ ​ ​ i10-index ● Simple to calculate ● Used only in Google Scholar ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ● Authors need to check that publications ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ counted belong to them ​ ​ ​ ​ ​ ​ Publication count ​ ​ ● Easily obtained - most databases and ● Does not distinguish between different ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ Google Scholar give publication counts document types and duration of ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ publication process ​ ​ ● Dependent on material types indexed by ​ ​ ​ ​ ​ ​ ​ ​ ​ ​ particular
    [Show full text]
  • Citescore Metrics: Creating Journal Metrics from the Scopus Citation Index
    CiteScore metrics: Creating journal metrics from the Scopus citation index Authors: Chris James 0000-0003-1619-5204, Lisa Colledge 0000-0002-8921-7147, Wim Meester 0000-0001-9350-3448, Norman Azoulay 0000-0003-1453-4882, Andrew Plume 0000-0002-4942-1426 Affiliations: Chris James, Lisa Colledge and Wim Meester: Elsevier, Radarweg 29, 1043 NX Amsterdam, Netherlands Norman Azoulay: Elsevier, Suite 800, 230 Park Ave, New York, NY 10169, United States Andrew Plume: Elsevier, The Boulevard, Langford Lane, Kidlington OX5 1GB, UK Corresponding author: Andrew Plume ([email protected]) Competing interests statement: At the time of this work being conducted all authors were employees of Elsevier, which owns the Scopus database and produces the CiteScore metrics. NA is now no longer employed by Elsevier. Highlights: • Journal citation metrics remain important in the age of responsible metrics • CiteScore metrics were designed on the principles of transparency, comprehensiveness and currency • CiteScore metrics are calculated from snapshots of a dynamic database • CiteScore Tracker is updated monthly to show progress towards the static CiteScore value as it builds • Feedback after the initial release of CiteScore is acknowledged and resulting developments are discussed Abstract In December 2016, after several years of development, Elsevier launched a set of transparent, comprehensive, current, and freely-available journal citation metrics called CiteScore metrics. Most of the CiteScore metrics are static, annual indicators calculated from the dynamic Scopus citation index. In the spirit of recent public statements on the responsible use of metrics, we outline the desirable characteristics of journal citation metrics, discuss how we decided on the cited and citing publications years and document types to be used for CiteScore metrics, and detail the precise method of calculation of each metric.
    [Show full text]
  • <User Training Session> the World Largest Abstract and Citation Database
    <User training session> The world largest abstract and citation database www.scopus.com December, 2019 @ Gifu University Elsevier Japan KK 2 Agenda • What is Scopus? • Find the latest articles Basic search workflow about your research topic Find important articles • Find high-impact articles based on citations Check the author profile of • Search for authors a top researcher Select a journal to publish • Check journal metrics with • Personal Log-in Alert What is Scopus︖- Comparison 1 3 Indexed journal information. (All Scopus indexed titles are peer-review titles.) What is Scopus? - Comparison 2 Abstract & Citation Database 4 • Title • Authors • Abstract • Keywords • Main text −Introduction −Methods −Results −Discussions (Conclusion) • References What is Scopus? - Comparison 2 Abstract & Citation Database 5 Journal Title Article Title Article Title Article Title Bibliographic Info Bibliographic Info Bibliographic Info Abstract Abstract Abstract Fields Keywords Keywords Keywords Can rank and sort Full Text articles by the number of citations ReferencesReferences References DB Type E-journal list, Abstract DB Abstract & Citation DB Full Text DB OPAC Know what (who) Purpose Check availability Know what the is citing the article Read and search for of the title article is about and by what full-text articles (whom) the article is cited Examples Covers more Covers a particular than 5,000 publisherʼs full-text publishers articles worldwide Basic search workflow 6 ① Start searching Select search mode Search terms Search fields Date range and
    [Show full text]
  • Citescore Faqs 2
    FAQs December 2016 Contents 1. About CiteScore and its derivative metrics 4 1.1 What is CiteScore? 5 1.2 Why don’t you include articles-in-press in CiteScore? 5 1.3 Why don’t you include abstracts in CiteScore? 6 1.4 Are documents that were published in supplements and/or special issues included in the CiteScore calculation? 6 1.5 How did you calculate the static CiteScore metrics from Scopus.com? 6 1.6 What is CiteScore Tracker? 6 1.7 Which database are CiteScore metrics calculated from? 7 1.8 How do I feed back to Scopus if I find an error in the data? 7 1.9 Why is the CiteScore value that I’ve calculated from Scopus different from the one that’s published? 7 1.10 How much do CiteScore metrics cost? 8 1.11 Can I export CiteScore metrics? 8 1.12 Which titles are CiteScore metrics available for? 8 1.13 When do new serials get CiteScore metrics? 8 1.14 Why doesn’t a serial title have CiteScore metrics? 8 1.15 How is the CiteScore calculation different from the Impact Factor calculation? 9 1.16 Why does this serial title have a lower CiteScore than Impact Factor? 10 1.17 Why does this serial title have a higher CiteScore than Impact Factor? 10 1.18 Why is this serial title’s rank different than its rank with the Impact Factor? 10 1.19 Why does CiteScore have a 3 year publication window? 10 1.20 Why do you include all types of documents in the CiteScore calculation? 11 Recognizing the total influence of a serial title on the scholarly community 11 Reducing the ease of manipulation of the metric 11 Disagreements over document type classification 11 1.21 How do I compare between subject fields using CiteScore? 12 1.22 What is the baseline so I know whether a CiteScore of 3.64 is good in the serial title’s subject field? 12 1.23 What are the benefits of CiteScore Tracker? 12 1.24 How are self-citations handled? 13 1.25 How should I refer to CiteScore metrics? 13 2 CiteScore FAQs 2.
    [Show full text]
  • Relationship Between Journal-Ranking Metrics for a Multidisciplinary Set of Journals
    Upeksha Perera and ManjulaWijewickrema 35 Relationship between Journal-Ranking Metrics for 18.1. a Multidisciplinary Set of portal Journals Upeksha Perera and ManjulaWijewickrema publication, for abstract: Ranking of scholarly journals is important to many parties. Studying the relationships among various ranking metrics is key to understanding the significance of one metric based on another. This research investigates the relationship among four major journal-ranking indicators: the impact factor (IF), the Eigenfactor score (ES), the acceptedh-index (hI), and SCImago Journal Rank (SJR). The authors used 519 journals from a diverse range of subjects and executed a correlation analysis. The results show that a higher correlationand exists between impact factor and SCImago Journal Rank (0.796). However, a very strong positive correlation between two metrics does not necessarily imply that journals are ranked according to the same pattern. edited, copy Introduction t is important for a researcher to know the rank of the journal to which he or she intends to submit a manuscript for numerous reasons, including to identify the most appropriate journals for publishing, to form an idea about the level of journals, and to Iidentify the reviewed,publishing outlets that could advance the author’s career.1 Bibliometrics, the statistical analysis of journals and similar publications, has introduced journal-ranking indicatorspeer for evaluating publication outlets. These metrics are defined by how they reflectis the properties of the considered journals. Having an idea about the relationships among different ranking indicators is important to predict the behavior of one metric mss.based on another. For instance, some metrics emphasize popularity, while others indicate prestige.2 Hence, a relationship between the metrics of popularity and prestige could be used to illustrate the performance of one metric in comparison to another.
    [Show full text]
  • The Bibliometric Literature on Scopus and Wos: the Medicine and Environmental Sciences Categories As Case of Study
    International Journal of Environmental Research and Public Health Article The Bibliometric Literature on Scopus and WoS: The Medicine and Environmental Sciences Categories as Case of Study Mila Cascajares 1 , Alfredo Alcayde 1 , Esther Salmerón-Manzano 2,* and Francisco Manzano-Agugliaro 1 1 Department of Engineering, University of Almeria, ceiA3, 04120 Almeria, Spain; [email protected] (M.C.); [email protected] (A.A.); [email protected] (F.M.-A.) 2 Faculty of Law, Universidad Internacional de La Rioja (UNIR), Av. de la Paz, 137, 26006 Logroño, Spain * Correspondence: [email protected] Abstract: In a broad sense, science can be understood as the knowledge contained in scientific manuscripts published in scientific journals. Scientific databases index only those journals that reach certain quality standards. Therefore, research and dissemination of scientific knowledge are essential activities for the growth of science itself. The aim of this manuscript is to assess the situation of medicine and environmental sciences among the bibliometric literature and to put it in perspective with the overall bibliometric publications in all scientific fields. The main countries publishing bibliometric manuscripts are China, USA and Spain. The latter country is ranked three out of the top five institutions according to the Scopus and WoS databases. In both databases, the average scientific collaboration of the top 20 institutions offers the same result, 41%. According to Scopus, Citation: Cascajares, M.; Alcayde, A.; the main subject categories in which this research falls are social sciences (38%), computer science Salmerón-Manzano, E.; (26%) and medicine (23%), while the environmental sciences category has 8%. In the analysis of Manzano-Agugliaro, F.
    [Show full text]
  • Research Intelligence Research Metrics Guidebook Table of Contents
    Research Intelligence Research Metrics Guidebook Table of contents 1.0 Introduction: Elsevier’s approach to research metrics.............04 3.2 Factors besides performance that affect the value of a metric..................................................................11 1.1 Journal metrics....................................................................05 3.2.1 Size...........................................................................13 1.2 Article-level metrics.............................................................05 3.2.2 Discipline..................................................................13 3.2.3 Publication-type.......................................................14 1.3 Author and institutional metrics.........................................05 3.2.4 Database coverage...................................................16 2.0 Scopus: the primary data source for 3.2.5 Manipulation............................................................18 Elsevier’s research metrics.......................................................06 3.2.6 Time.........................................................................18 2.1 Scopus content...................................................................07 4.0 SciVal and research metrics......................................................20 2.1.1 Scopus content and SciVal......................................07 2.1.2 Books and SciVal.....................................................07 4.1 Groups of metrics in SciVal.................................................21
    [Show full text]
  • Scholarly Metrics Qianjin (Marina) Zhang Engineering & Informatics Librarian Lichtenberger Engineering Library [email protected] Learning Objectives
    Scholarly Metrics Qianjin (Marina) Zhang Engineering & Informatics Librarian Lichtenberger Engineering Library [email protected] Learning Objectives Be able to interpret some standard citation metrics with caution and know where to find them Explore three major sources of citation data including Web of Science (JCR), Scopus and Google Scholar Identify best practices on broadening your scholarly impact h-index • J. E. Hirsch, 2005 • # of papers (h) published that have each been cited at least h times. • Can be applied to any level of aggregation such as author, institution and journal, etc. Image from Hirsch, J. E. “An Index to Quantify an Individual’s Scientific Research Output.” Proceedings of the National Academy of Sciences 102.46 (2005): 16569–16572. http://web.cs.ucla.edu/~palsberg/hirsch05.pdf h-index (Cont’d) An author has 8 papers that have been cited 34, 29, 20, 15, 7, 6, 5 and 4 times. What is the author’s h-index? Paper # # of Citations 1 34 2 29 3 20 4 15 5 7 6 6 = h-index 7 5 Example from 8 4 https://subjectguides.uwaterloo.ca/calc ulate-academic-footprint/YourHIndex John’s h-index = 10 Sara’s h-index = 5 h-index Issues Paper # # of Citations # of Citations 1 34 62 • Is not influenced by 2 28 50 citation counts of papers 3 25 40 that have been cited at least h times 4 23 32 • Distinctive. Its value can 5 20 5 only increase, never 6 17 4 decrease, over time. 7 16 3 • Advantage senior researchers but 8 14 2 disadvantage early-career 9 13 1 researchers 10 10 1 Finding Your h-index in Web of Science Finding Your h-index in Scopus • i10-index: # of papers with at least 10 citations.
    [Show full text]