The Coverage of Microsoft Academic: Analyzing the Publication Output of a University
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
MAJ 04/2021, Cf. Feuilletage Ci-Dessous
Recherche d’informations sur Internet (perfectionnement) méthodologie et outils disponibles A. Bouchard 04/2021 Pour commencer Principes Moteurs de recherche Sites internet Bases de données bibliographiques Autres bases de données textuelles Images et multimédia Web social Actualités et temps réel Quelques outils complémentaires Veille automatisée Exercices de synthèse Bibliographie Principes Internet ? web ? • internet • réseau de réseaux • fin des années 1960 • protocole TCP/IP • applications et services divers : courrier électronique (mail), messagerie instantanée (IM), forums de discussion, transfert de fichiers (FTP), pair à pair (P2P), web (www)… • données • utilisateurs : 5,1 MM. dans le monde (Internet World Stats, 2020) • langues : 61 % du contenu en anglais, 2,8 % en français (W3Techs, 2021) • voir également Internet live stats Internet ? web ? • Web • World Wide Web (www) • milieu des années 1990 • ensemble de pages HTML (textes, images, liens…) avec une URL et accessibles avec le protocole HTTP • web visible / web invisible • web invisible ou web profond (deep web) : partie du web non indexée et qui ne peut être trouvée par les moteurs de recherche (pages protégées par un mot de passe, pages générées dynamiquement à la suite d’une requête…), voire dark web (web illégal) : 95 % du total ? • taille • 1,2 MM. de sites (Netcraft) • web indexé : au moins 5,3 milliards de pages (Worldwidewebsize) • taille du web identifié (URL connues) ? web général ? Internet ? web ? Ascodocpsy ConceptArt multimédia, 2010 Les âges du web du web âges Les Méthodologie • DEBUSQUER l’information Différents outils Esprit critique Bookmark organisé URL significative Syntaxe de recherche Questions préalables Utilisation réfléchie Evaluation Règles à respecter d’après Eduscol. Rechercher sur internet Méthodologie 1° définir le sujet (contexte de la recherche et mots-clés) Questions Prendre du temps au départ pour en gagner par la suite.. -
Mapping the Future of Scholarly Publishing
THE OPEN SCIENCE INITIATIVE WORKING GROUP Mapping the Future of Scholarly Publishing The Open Science Initiative (OSI) is a working group convened by the National Science Communi- cation Institute (nSCI) in October 2014 to discuss the issues regarding improving open access for the betterment of science and to recommend possible solutions. The following document summa- rizes the wide range of issues, perspectives and recommendations from this group’s online conver- sation during November and December 2014 and January 2015. The 112 participants who signed up to participate in this conversation were drawn mostly from the academic, research, and library communities. Most of these 112 were not active in this conversa- tion, but a healthy diversity of key perspectives was still represented. Individual participants may not agree with all of the viewpoints described herein, but participants agree that this document reflects the spirit and content of the conversation. This main body of this document was written by Glenn Hampson and edited by Joyce Ogburn and Laura Ada Emmett. Additional editorial input was provided by many members of the OSI working group. Kathleen Shearer is the author of Annex 5, with editing by Dominque Bambini and Richard Poynder. CC-BY 2015 National Science Communication Institute (nSCI) www.nationalscience.org [email protected] nSCI is a US-based 501(c)(3) nonprofit organization First edition, January 2015 Final version, April 2015 Recommended citation: Open Science Initiative Working Group, Mapping the Future of Scholarly -
Do You Speak Open Science? Resources and Tips to Learn the Language
Do You Speak Open Science? Resources and Tips to Learn the Language. Paola Masuzzo1, 2 - ORCID: 0000-0003-3699-1195, Lennart Martens1,2 - ORCID: 0000- 0003-4277-658X Author Affiliation 1 Medical Biotechnology Center, VIB, Ghent, Belgium 2 Department of Biochemistry, Ghent University, Ghent, Belgium Abstract The internet era, large-scale computing and storage resources, mobile devices, social media, and their high uptake among different groups of people, have all deeply changed the way knowledge is created, communicated, and further deployed. These advances have enabled a radical transformation of the practice of science, which is now more open, more global and collaborative, and closer to society than ever. Open science has therefore become an increasingly important topic. Moreover, as open science is actively pursued by several high-profile funders and institutions, it has fast become a crucial matter to all researchers. However, because this widespread interest in open science has emerged relatively recently, its definition and implementation are constantly shifting and evolving, sometimes leaving researchers in doubt about how to adopt open science, and which are the best practices to follow. This article therefore aims to be a field guide for scientists who want to perform science in the open, offering resources and tips to make open science happen in the four key areas of data, code, publications and peer-review. The Rationale for Open Science: Standing on the Shoulders of Giants One of the most widely used definitions of open science originates from Michael Nielsen [1]: “Open science is the idea that scientific knowledge of all kinds should be openly shared as early as is practical in the discovery process”. -
Tipping Points: Cancelling Journals When Arxiv Access Is Good Enough
Tipping points: cancelling journals when arXiv access is good enough Tony Aponte Sciences Collection Coordinator UCLA Library ASEE ELD Lightning Talk June 17, 2019 Preprint explosion! Brian Resnick and Julia Belluz. (2019). The war to free science. Vox https://www.vox.com/the-highlight/2019/6/3/18271538/open- access-elsevier-california-sci-hub-academic-paywalls Preprint explosion! arXiv. (2019). arXiv submission rate statistics https://arxiv.org/help/stats/2018_by_area/index 2018 Case Study: two physics journals and arXiv ● UCLA: heavy users of arXiv. Not so heavy users of version of record ● Decent UC authorship ● No UC editorial board members 2017 Usage Annual cost Cost per use 2017 Impact Factor Journal A 103 $8,315 ~$80 1.291 Journal B 72 $6,344 ~$88 0.769 Just how many of these articles are OA? OAISSN.py - Enter a Journal ISSN and a year and this python program will tell you how many DOIs from that year have an open access version2 Ryan Regier. (2018). OAISSN.py https://github.com/ryregier/OAcounts. Just how many of these articles are OA? Ryan Regier. (2018). OAISSN.py https://github.com/ryregier/OAcounts. Just how many of these articles are OA? % OA articles from 2017 % OA articles from 2018 Journal A 68% 64% Journal B 11% 8% Ryan Regier. (2018). OAISSN.py https://github.com/ryregier/OAcounts. arXiv e-prints becoming closer to publisher versions of record according to UCLA similarity study of arXiv articles vs versions of record Martin Klein, Peter Broadwell, Sharon E. Farb, Todd Grappone. 2018. Comparing Published Scientific Journal Articles to Their Pre-Print Versions -- Extended Version. -
Microsoft Academic Search and Google Scholar Citations
Microsoft Academic Search and Google Scholar Citations Microsoft Academic Search and Google Scholar Citations: A comparative analysis of author profiles José Luis Ortega1 VICYT-CSIC, Serrano, 113 28006 Madrid, Spain, [email protected] Isidro F. Aguillo Cybermetrics Lab, CCHS-CSIC, Albasanz, 26-28 28037 Madrid, Spain, [email protected] Abstract This paper aims to make a comparative analysis between the personal profiling capabilities of the two most important free citation-based academic search engines, namely Microsoft Academic Search (MAS) and Google Scholar Citations (GSC). Author profiles can be very useful for evaluation purposes once the advantages and the shortcomings of these services are described and taken into consideration. A total of 771 personal profiles appearing in both the MAS and the GSC databases are analysed. Results show that the GSC profiles include more documents and citations than those in MAS, but with a strong bias towards the Information and Computing sciences, while the MAS profiles are disciplinarily better balanced. MAS shows technical problems such as a higher number of duplicated profiles and a lower updating rate than GSC. It is concluded that both services could be used for evaluation proposes only if they are applied along with other citation indexes as a way to supplement that information. Keywords: Microsoft Academic Search; Google Scholar Citations; Web bibliometrics; Academic profiles; Research evaluation; Search engines Introduction In November 2009, Microsoft Research Asia started a new web search service specialized in scientific information. Even though Google (Google Scholar) already introduced an academic search engine in 2004, the proposal of Microsoft Academic Search (MAS) went beyond a mere document retrieval service that counts citations. -
Aggregating Research Papers from Publishers' Systems to Support Text
Aggregating Research Papers from Publishers’ Systems to Support Text and Data Mining: Deliberate Lack of Interoperability or Not? Petr Knoth, Nancy Pontika The Open University Walton Drive, Milton Keynes, United Kingdom [email protected], [email protected] Abstract In the current technology dominated world, interoperability of systems managed by different organisations is an essential property enabling the provision of services at a global scale. In the Text and Data Mining field (TDM), interoperability of systems offering access to text corpora offers the opportunity of increasing the uptake and impact of TDM applications. The global corpus of all research papers, i.e. the collection of human knowledge so large no one can ever read in their lifetime, represents one of the most exciting opportunities for TDM. Although the Open Access movement, which has been advocating for free availability and reuse rights to TDM from research papers, has achieved some major successes on the legal front, the technical interoperability of systems offering free access to research papers continues to be a challenge. COnnecting REpositories (CORE) (Knoth and Zdrahal, 2012) aggregates the world’s open access full-text scientific manuscripts from repositories, journals and publisher systems. One of the main goals of CORE is to harmonise and pre-process these data to lower the barrier for TDM. In this paper, we report on the preliminary results of an interoperability survey of systems provided by journal publishers, both open access and toll access. This helps us to assess the current level of systems’ interoperability and suggest ways forward. Keywords: Interoperability, publishers, standardisation 1. -
Scholarly Metrics Recommendations for Research Libraries: Deciphering the Trees in the Forest Table of Contents
EUROPE’S RESEARCH LIBRARY NETWORK Scholarly Metrics Recommendations for Research Libraries: Deciphering the Trees in the Forest Table of Contents 01 3D. Learn About Platforms Before Implementing Them In Services & 15 INTRODUCTION 03 Homogenize Different Sources ABOUT LIBER 3E. Work With Researchers To Build Awareness Of Benefits But Educate 16 1. DISCOVERY & DISCOVERABILITY 05 About Weaknesses of Scholarly Metrics 1A. Provide Contextual Information To Allow the Discovery of Related 05 3F. Expand Your Perspective When Developing Services; Avoid Single- 17 Work & Users Purpose Approaches 1B. Exploit Rich Network Structures & Implement Bibliometric Methods To 07 3G. Teach Colleagues New Skills & Library Goals & Services 17 Enable Discovery 1C. Encourage Sharing Of Library Collections Under Open Licenses 08 4. RESEARCH ASSESSMENT 19 4A. Establish Appropriate Goals for Assessment Exercises Before Selecting 19 2. SHOWCASING ACHIEVEMENTS 09 Databases & Metrics; Be Transparent About Use & Interpretation 2A. Incentivize Researchers To Share Scholarly Works, Promote Achievements 09 4B. Use Different Data Sources to Include Various Scholarly Works & 20 Online & Engage With Audiences Disciplinary Communication & Publication Cultures 2B. Encourage Researchers To Showcase Scientific Contributions & Monitor 11 4C. Rely on Objective, Independent & Commonly-Accepted Data Sources 21 Impact To Provide Sound & Transparent Scholarly Metrics 4D. Avoid Using Composite Indicators & Conflating Different Aspects 21 3. SERVICE DEVELOPMENT 13 Of Scholarly Works & Impact; Don’t Lose The Multifaceted Nature of 3A. Join Forces With Stakeholders To Provide Services Reusing Existing 13 Metrics & Distort Interpretation Resources, Tools, Methods & Data 3B. Value Various Levels of Engagement; Favour Standardized, Well-Established 15 23 CONCLUSION & CALL FOR ACTION Practices & Easy-To-Use Tools 25 REFERENCES 3C. -
Exposing Dmps: Use Cases, Workflows and Guidelines
Exposing DMPs: Use cases, workflows and guidelines Case statement for community review July 2017 Contributors David Carr - The Wellcome Trust John Chodacki, California Digital Library John Faundeen, Earth Resources Observation Center, USGS Bev Jones University of Lincoln Natalie Meyers, Centre for Open Science/ University of Notre Dame Paul Millar, DESY Fiona Murphy, MMC Ltd Kathryn Unsworth, Australian National Data Service Angus Whyte, Digital Curation Centre (Editor) Elena Zudilova-Seinstra, Elsevier Working Group Charter A variety of stakeholders are showing growing interest in exposing data management plans (*) to other actors (human/machine) in the research lifecycle, beyond their creator and the funder or institution that mandates their production. Interested stakeholders include researchers themselves, funders, institutions, and a variety of service providers and community organisations including repositories, institutions, journals, publishers, and providers of tools for writing and maintaining plans. Implementation and adoption is currently hampered by two problems: ● A lack of standards for expression and interchange of DMPs ● Insufficient understanding of the needs of users and the benefits and risks of different modes of action This proposed working group will address both of these issues; the issue of a standardised form of expression for DMPs is the concern of the proposed DMP Common Standards Working Group. The group’s output will include a reference model and alternative strategies for exposing plans, to best serve community interests in meeting FAIR principles,1 based on shared experience of ‘early adopters’ in test implementations. It will be supported by work to gauge user needs and motivations for exposing DMPs as well as perceived risks and disbenefits. -
Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and Opencitations’ COCI: a Multidisciplinary Comparison of Coverage Via Citations
✅ The content of this manuscript has been peer-reviewed and accepted for publication at Scientometrics (ISSN 1588-2861) Cite as: Martín-Martín, A., Thelwall, M., Orduna-Malea, E., & Delgado López-Cózar, E. (2021). Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations’ COCI: A multidisciplinary comparison of coverage via citations. Scientometrics, 126(1), 871-906. https://doi.org/10.1007/s11192-020-03690-4 Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations’ COCI: a multidisciplinary comparison of coverage via citations Alberto Martín-Martín 1 , Mike Thelwall 2 , Enrique Orduna- 3 1 Malea , Emilio Delgado López-Cózar Abstract (also in Spanish, Chinese) Introduction New sources of citation data have recently become available, such as Microsoft Academic, Dimensions, and the OpenCitations Index of CrossRef open DOI-to-DOI citations (COCI). Although these have been compared to the Web of Science (WoS), Scopus, or Google Scholar, there is no systematic evidence of their differences across subject categories. Methods In response, this paper investigates 3,073,351 citations found by these six data sources to 2,515 English-language highly-cited documents published in 2006 from 252 subject categories, expanding and updating the largest previous study. Results Google Scholar found 88% of all citations, many of which were not found by the other sources, and nearly all citations found by the remaining sources (89%-94%). A similar pattern held within most subject categories. Microsoft Academic is the second largest overall (60% of all citations), including 82% of Scopus citations and 86% of Web of Science citations. In most categories, Microsoft Academic found more citations than Scopus and WoS (182 and 223 subject categories, respectively), but had coverage gaps in some areas, such as Physics and some Humanities categories. -
Citeseerx: 20 Years of Service to Scholarly Big Data
CiteSeerX: 20 Years of Service to Scholarly Big Data Jian Wu Kunho Kim C. Lee Giles Old Dominion University Pennsylvania State University Pennsylvania State University Norfolk, VA University Park, PA University Park, PA [email protected] [email protected] [email protected] ABSTRACT access to a growing number of researchers. Mass digitization par- We overview CiteSeerX, the pioneer digital library search engine, tially solved the problem by storing document collections in digital that has been serving academic communities for more than 20 years repositories. The advent of modern information retrieval methods (first released in 1998), from three perspectives. The system per- significantly expedited the process of relevant search. However, spective summarizes its architecture evolution in three phases over documents are still saved individually by many users. In 1997, three the past 20 years. The data perspective describes how CiteSeerX computer scientists at the NEC Research Institute (now NEC Labs), has created searchable scholarly big datasets and made them freely New Jersey, United States – Steven Lawrence, Kurt Bollacker, and available for multiple purposes. In order to be scalable and effective, C. Lee Giles, conceived an idea to create a network of computer AI technologies are employed in all essential modules. To effectively science research papers through citations, which was to be imple- train these models, a sufficient amount of data has been labeled, mented by a search engine, the prototype CiteSeer. Their intuitive which can then be reused for training future models. Finally, we idea, automated citation indexing [8], changed the way researchers discuss the future of CiteSeerX. -
Overview of the Altmetrics Landscape
Purdue University Purdue e-Pubs Charleston Library Conference Overview of the Altmetrics Landscape Richard Cave Public Library of Science, [email protected] Follow this and additional works at: https://docs.lib.purdue.edu/charleston Part of the Library and Information Science Commons An indexed, print copy of the Proceedings is also available for purchase at: http://www.thepress.purdue.edu/series/charleston. You may also be interested in the new series, Charleston Insights in Library, Archival, and Information Sciences. Find out more at: http://www.thepress.purdue.edu/series/charleston-insights-library-archival- and-information-sciences. Richard Cave, "Overview of the Altmetrics Landscape" (2012). Proceedings of the Charleston Library Conference. http://dx.doi.org/10.5703/1288284315124 This document has been made available through Purdue e-Pubs, a service of the Purdue University Libraries. Please contact [email protected] for additional information. Overview of the Altmetrics Landscape Richard Cave, Director of IT and Computer Operations, Public Library of Science Abstract While the impact of article citations has been examined for decades, the “altmetrics” movement has exploded in the past year. Altmetrics tracks the activity on the Social Web and looks at research outputs besides research articles. Publishers of scientific research have enabled altmetrics on their articles, open source applications are available for platforms to display altmetrics on scientific research, and subscription models have been created that provide altmetrics. In the future, altmetrics will be used to help identify the broader impact of research and to quickly identify high-impact research. Altmetrics and Article-Level Metrics Washington as an academic research project to rank journals based on a vast network of citations The term “altmetrics” was coined by Jason Priem, (Eigenfactor.org, http://www.eigenfactor.org/ a PhD candidate at the School of Information and whyeigenfactor.php). -
The Visibility of Authority Records, Researcher Identifiers, Academic Social Networking
James Madison University JMU Scholarly Commons Libraries Libraries 4-23-2019 The iV sibility of Authority Records, Researcher Identifiers, Academic Social Networking Profiles, and Related Faculty Publications in Search Engine Results Rebecca B. French James Madison University, [email protected] Jody Condit Fagan James Madison University, [email protected] Follow this and additional works at: https://commons.lib.jmu.edu/letfspubs Part of the Library and Information Science Commons Recommended Citation French, Rebecca B. and Fagan, Jody Condit, "The iV sibility of Authority Records, Researcher Identifiers, Academic Social Networking Profiles, and Related Faculty Publications in Search Engine Results" (2019). Libraries. 148. https://commons.lib.jmu.edu/letfspubs/148 This Article is brought to you for free and open access by the Libraries at JMU Scholarly Commons. It has been accepted for inclusion in Libraries by an authorized administrator of JMU Scholarly Commons. For more information, please contact [email protected]. The visibility of authority records, researcher identifiers, academic social networking profiles, and related faculty publications in search engine results Rebecca B. French and Jody Condit Fagan James Madison University Author Note JMU Libraries, Carrier Library 1704, James Madison University, Harrisonburg, VA 22807 [email protected] [email protected] This is an Accepted Manuscript of an article published by Taylor & Francis in the Journal of Web Librarianship on April 23, 2019, available online: http://www.tandfonline.com/10.1080/19322909.2019.1591324. 1 Abstract Information about faculty and their publications can be found in library databases such as the Library of Congress Name Authority File, VIAF, WorldCat, and institutional repositories; in identifier registries like ORCID and ISNI; and on academic social networking sites like Academia, Google Scholar, and ResearchGate, but the way search engines use such identifiers and profiles is unclear.