Machine Learning in Scientometrics

Total Page:16

File Type:pdf, Size:1020Kb

Machine Learning in Scientometrics DEPARTAMENTO DE INTELIGENCIA ARTIFICIAL Escuela Tecnica´ Superior de Ingenieros Informaticos´ Universidad Politecnica´ de Madrid PhD THESIS Machine Learning in Scientometrics Author Alfonso Iba´nez˜ MS Computer Science MS Artificial Intelligence PhD supervisors Concha Bielza PhD Computer Science Pedro Larranaga˜ PhD Computer Science 2015 Thesis Committee President: C´esarHerv´as Member: Jos´eRam´onDorronsoro Member: Enrique Herrera Member: Irene Rodr´ıguez Secretary: Florian Leitner There are no secrets to success. It is the result of preparation, hard work, and learning from failure. Acknowledgements Ph.D. research often appears a solitary undertaking. However, it is impossible to maintain the degree of focus and dedication required for its completion without the help and support of many people. It has been a difficult long journey to finish my Ph.D. research and it is of justice to cite here all of them. First and foremost, I would like to thank Concha Bielza and Pedro Larra~nagafor being my supervisors and mentors. Without your unfailing support, recommendations and patient, this thesis would not have been the same. You have been role models who not only guided my research but also demonstrated your enthusiastic research attitudes. I owe you so much. Whatever research path I do take, I will be prepared because of you. I would also like to express my thanks to all my friends and colleagues at the Computa- tional Intelligence Group who provided me with not only an excellent working atmosphere and stimulating discussions but also friendships, care and assistance when I needed. My special thank-you goes to Rub´enArma~nanzas,Roberto Santana, Diego Vidaurre, Hanen Borchani, Pedro L. L´opez-Cruz, Hossein Karshenas, Luis Guerra, Bojan Mihaljevic and Laura Ant´on- S´anchez. I have learned so much from all of you. Your constructive recommendations and collaboration have been tremendous assets throughout my Ph.D. research. This dissertation would not have been possible without the financial support offered by the Spanish Ministry of Economy and Competitiveness, projects TIN2008-04528-E and Con- solider Ingenio 2010-CSD-2007-00018. I would also like to thank TIN2007-62626, TIN2010- 20900-C04-04 and TIN2011-14083-E projects that supported my research during these years. Finally, my deepest gratitude goes to my wife for her understanding and faithful support as well as her patience and unconditional love. Thanks also for believing in me, being a never-ending fount of moral support and standing by me, when the going is difficult. For all of this, I thank you. Last but of course not least, I wish to thank my parents and sister for their unending encouragement and love from childhood to now. Without them, I would not be where I am. Thank you. Abstract Machine learning and scientometrics are the scientific disciplines which are covered in this dissertation. Machine learning deals with the construction and study of algorithms that can learn from data, whereas scientometrics is mainly concerned with the analysis of science from a quantitative perspective. Nowadays, advances in machine learning provide the mathemat- ical and statistical tools for properly working with the vast amount of scientometrics data stored in bibliographic databases. In this context, the use of novel machine learning methods in scientometrics applications is the focus of attention of this dissertation. This dissertation proposes new machine learning contributions which would shed light on the scientometrics area. These contributions are divided in three parts: Several supervised cost-(in)sensitive models are learned to predict the scientific success of articles and researchers. Cost-sensitive models are not interested in maximizing classification accuracy, but in minimizing the expected total cost of the error derived from mistakes in the classification process. In this context, publishers of scientific journals could have a tool capa- ble of predicting the citation count of an article in the future before it is published, whereas promotion committees could predict the annual increase of the h-index of researchers within the first few years. These predictive models would pave the way for new assessment systems. Several probabilistic graphical models are learned to exploit and discover new relationships among the vast number of existing bibliometric indices. In this context, scientific community could measure how some indices influence others in probabilistic terms and perform evidence propagation and abduction inference for answering bibliometric questions. Also, scientific community could uncover which bibliometric indices have a higher predictive power. This is a multi-output regression problem where the role of each variable, predictive or response, is unknown beforehand. The resulting indices could be very useful for prediction purposes, that is, when their index values are known, knowledge of any index value provides no information on the prediction of other bibliometric indices. A scientometric study of the Spanish computer science research is performed under the publish-or-perish culture. This study is based on a cluster analysis methodology which char- acterizes the research activity in terms of productivity, visibility, quality, prestige and inter- national collaboration. This study also analyzes the effects of collaboration on productivity and visibility under different circumstances. Resumen El aprendizaje autom´aticoy la cienciometr´ıason las disciplinas cient´ıficasque se tratan en esta tesis. El aprendizaje autom´aticotrata sobre la construcci´ony el estudio de algoritmos que puedan aprender a partir de datos, mientras que la cienciometr´ıase ocupa principalmente del an´alisisde la ciencia desde una perspectiva cuantitativa. Hoy en d´ıa,los avances en el aprendizaje autom´aticoproporcionan las herramientas matem´aticasy estad´ısticaspara tra- bajar correctamente con la gran cantidad de datos cienciom´etricosalmacenados en bases de datos bibliogr´aficas.En este contexto, el uso de nuevos m´etodos de aprendizaje autom´atico en aplicaciones de cienciometr´ıaes el foco de atenci´onde esta tesis doctoral. Esta tesis propone nuevas contribuciones en el aprendizaje autom´aticoque podr´ıanarro- jar luz sobre el ´areade la cienciometr´ıa.Estas contribuciones est´andivididas en tres partes: Varios modelos supervisados (in)sensibles al coste son aprendidos para predecir el ´exito cient´ıficode los art´ıculosy los investigadores. Los modelos sensibles al coste no est´anin- teresados en maximizar la precisi´onde clasificaci´on,sino en la minimizaci´ondel coste total esperado derivado de los errores ocasionados. En este contexto, los editores de revistas cient´ıficaspodr´ıandisponer de una herramienta capaz de predecir el n´umerode citas de un art´ıculoen el fututo antes de ser publicado, mientras que los comit´esde promoci´onpodr´ıan predecir el incremento anual del ´ındiceh de los investigadores en los primeros a~nos. Estos modelos predictivos podr´ıanallanar el camino hacia nuevos sistemas de evaluaci´on. Varios modelos gr´aficosprobabil´ısticosson aprendidos para explotar y descubrir nuevas relaciones entre el gran n´umerode ´ındicesbibliom´etricosexistentes. En este contexto, la comunidad cient´ıficapodr´ıamedir c´omoalgunos ´ındicesinfluyen en otros en t´erminospro- babil´ısticosy realizar propagaci´onde la evidencia e inferencia abductiva para responder a preguntas bibliom´etricas. Adem´as,la comunidad cient´ıficapodr´ıadescubrir qu´e´ındicesbi- bliom´etricostienen mayor poder predictivo. Este es un problema de regresi´onmulti-respuesta en el que el papel de cada variable, predictiva o respuesta, es desconocido de antemano. Los ´ındicesresultantes podr´ıanser muy ´utilespara la predicci´on,es decir, cuando se conocen sus valores, el conocimiento de cualquier valor no proporciona informaci´onsobre la predicci´onde otros ´ındicesbibliom´etricos. Un estudio bibliom´etricosobre la investigaci´onespa~nolaen inform´aticaha sido realizado bajo la cultura de publicar o morir. Este estudio se basa en una metodolog´ıade an´alisis de clusters que caracteriza la actividad en la investigaci´onen t´erminosde productividad, visibilidad, calidad, prestigio y colaboraci´oninternacional. Este estudio tambi´enanaliza los efectos de la colaboraci´onen la productividad y la visibilidad bajo diferentes circunstancias. Contents Contents xvi I Preliminaries 1 1 Introduction 3 1.1 Contributions of the dissertation . 4 1.2 Overview of the dissertation . 6 II Background 9 2 Machine Learning 11 2.1 Introduction . 11 2.2 Supervised learning . 13 2.2.1 Supervised learning approaches . 13 2.2.2 Classification validation . 17 2.3 Unsupervised learning . 19 2.3.1 Unsupervised learning approaches . 20 2.3.2 Clustering validation . 23 3 Probabilistic Graphical Models 27 3.1 Introduction . 27 3.2 Bayesian networks . 28 3.2.1 Discrete Bayesian networks . 28 3.2.2 Gaussian Bayesian networks . 29 3.3 Learning Bayesian networks . 29 3.3.1 Structural learning . 30 3.3.2 Parametric learning . 32 3.4 Inference in Bayesian networks . 32 3.5 Bayesian networks classifiers . 33 xiii xiv CONTENTS 4 Scientometrics 37 4.1 Introduction . 37 4.2 Citation analysis in research evaluation . 38 4.3 The h-index . 40 4.4 Improvements of the h-index . 42 4.4.1 Bibliometric measures that complement the h-index . 42 4.4.2 Bibliometric measures that take time into account . 47 4.4.3 Bibliometric measures that allow for co-authorship . 49 4.4.4 Bibliometric measures that consider other variables . 51 4.5 Journal-based measures . 52 4.5.1 Impact factor . 53 4.5.2 Bibliometric measures that assess the quality of citations . 54 4.5.3 Bibliometric measures that correct for differences among fields . 55 4.6 Bibliographic databases . 57 4.6.1 Web of Science . 58 4.6.2 Scopus . 61 4.6.3 Google Scholar . 62 III Data Mining in Research Evaluation 65 5 Predicting citation counts using supervised algorithms 67 5.1 Introduction . 67 5.2 Predicting citation count of Bioinformatics papers . 69 5.2.1 Dataset compilation . 69 5.2.2 Data distribution . 70 5.2.3 Predictive models . 71 5.2.4 Exploiting the best models . 75 5.3 Discussion and conclusions . 77 6 Predicting the h-index using cost-sensitive algorithms 79 6.1 Introduction .
Recommended publications
  • Newsletternewsletter Volume 1 ■ Number 41 ■ Decemberapril 2003 2003
    newsletternewsletter Volume 1 ■ Number 41 ■ DecemberApril 2003 2003 Welcome In this Issue FEATURES Dear Colleagues, Scholarly Publishing: 12 Observations on the In the transition from the print to the electronic world, technology has Current Situation and Challenges for descended upon the library like a tsunami. Librarians have labored long the Future ..............................................................2-3 and hard to create and be steered by a clear sense of mission in this new age, rather than simply being swept along by the technological wave. Library Vision of the Future: University of Rochester Libraries — There have been enormous challenges in understanding the potential of Geoff Adams the new technology and how to harness it to a needs-driven environment. Focusing on Users, Hiding Technology ............4 Fundamental questions have even been raised about the very value of traditional libraries in this The Future of Librarianship: A View From a new environment, and librarians continue to wrestle with what their future role in the scientific and School of Library and Information Science ......5-6 academic communities will be. Libraries Look to the Future: As a member of these communities Elsevier is equally affected. This highly dynamic environment Place and Space, Physical and Virtual, presents both great challenges and great opportunities. In the course of the last few years we have Where are Libraries Headed? ................................7 seen production, pricing and packaging models from the print journal world change out of all recognition to accommodate the needs of the new digital environment. We have seen the emergence of e-publishing CENTER OF ATTENTION environments such as PUBMED, JSTOR, and most recently the Open Access movement.
    [Show full text]
  • A Philosophical Treatise on the Connection of Scientific Reasoning
    mathematics Review A Philosophical Treatise on the Connection of Scientific Reasoning with Fuzzy Logic Evangelos Athanassopoulos 1 and Michael Gr. Voskoglou 2,* 1 Independent Researcher, Giannakopoulou 39, 27300 Gastouni, Greece; [email protected] 2 Department of Applied Mathematics, Graduate Technological Educational Institute of Western Greece, 22334 Patras, Greece * Correspondence: [email protected] Received: 4 May 2020; Accepted: 19 May 2020; Published:1 June 2020 Abstract: The present article studies the connection of scientific reasoning with fuzzy logic. Induction and deduction are the two main types of human reasoning. Although deduction is the basis of the scientific method, almost all the scientific progress (with pure mathematics being probably the unique exception) has its roots to inductive reasoning. Fuzzy logic gives to the disdainful by the classical/bivalent logic induction its proper place and importance as a fundamental component of the scientific reasoning. The error of induction is transferred to deductive reasoning through its premises. Consequently, although deduction is always a valid process, it is not an infallible method. Thus, there is a need of quantifying the degree of truth not only of the inductive, but also of the deductive arguments. In the former case, probability and statistics and of course fuzzy logic in cases of imprecision are the tools available for this purpose. In the latter case, the Bayesian probabilities play a dominant role. As many specialists argue nowadays, the whole science could be viewed as a Bayesian process. A timely example, concerning the validity of the viruses’ tests, is presented, illustrating the importance of the Bayesian processes for scientific reasoning.
    [Show full text]
  • Scientometrics1
    Scientometrics1 Loet Leydesdorff a and Staša Milojević b a Amsterdam School of Communication Research (ASCoR), University of Amsterdam, Kloveniersburgwal 48, 1012 CX Amsterdam, The Netherlands; [email protected] b School of Informatics and Computing, Indiana University, Bloomington 47405-1901, United States; [email protected]. Abstract The paper provides an overview of the field of scientometrics, that is: the study of science, technology, and innovation from a quantitative perspective. We cover major historical milestones in the development of this specialism from the 1960s to today and discuss its relationship with the sociology of scientific knowledge, the library and information sciences, and science policy issues such as indicator development. The disciplinary organization of scientometrics is analyzed both conceptually and empirically. A state-of-the-art review of five major research threads is provided. Keywords: scientometrics, bibliometrics, citation, indicator, impact, library, science policy, research management, sociology of science, science studies, mapping, visualization Cross References: Communication: Electronic Networks and Publications; History of Science; Libraries; Networks, Social; Merton, Robert K.; Peer Review and Quality Control; Science and Technology, Social Study of: Computers and Information Technology; Science and Technology Studies: Experts and Expertise; Social network algorithms and software; Statistical Models for Social Networks, Overview; 1 Forthcoming in: Micheal Lynch (Editor), International
    [Show full text]
  • Steering Science Through Output Indicators & Data Capitalism
    University of Nebraska - Lincoln DigitalCommons@University of Nebraska - Lincoln Copyright, Fair Use, Scholarly Communication, etc. Libraries at University of Nebraska-Lincoln 9-2019 Steering science through Output Indicators & Data Capitalism Ulrich Herb Saarland University & State Library Follow this and additional works at: https://digitalcommons.unl.edu/scholcom Part of the Intellectual Property Law Commons, Scholarly Communication Commons, and the Scholarly Publishing Commons Herb, Ulrich, "Steering science through Output Indicators & Data Capitalism" (2019). Copyright, Fair Use, Scholarly Communication, etc.. 125. https://digitalcommons.unl.edu/scholcom/125 This Article is brought to you for free and open access by the Libraries at University of Nebraska-Lincoln at DigitalCommons@University of Nebraska - Lincoln. It has been accepted for inclusion in Copyright, Fair Use, Scholarly Communication, etc. by an authorized administrator of DigitalCommons@University of Nebraska - Lincoln. Steering science through Output Indicators & Data Capitalism Published in: Proceedings of the 23rd Congress of the European Society of Veterinary and Comparative Nutrition (ESVCN 2019), Turin/ Italy, 17-20 September 2019. Author: Ulrich Herb, ORCID: 0000-0002-3500-3119 Affiliations: Saarland University & State Library ./scidecode science consulting & research Please cite as: Herb, Ulrich (2019). Steering science through Output Indicators & Data Capitalism. Proceedings of the 23rd Congress of the European Society of Veterinary and Comparative Nutrition, Turin 2019. DOI:10.5281/zenodo.3333395 Performance tracking in science Since the mid-1920s, a scientist's performance was tracked by using bibliometric information such as the number of publications or their citations. Today, there are unprecedented possibilities for controlling science by analysing data on production and use of scientific information, so that citations only play a subordinate role in the evaluation of science.
    [Show full text]
  • 1.12 News Feat 4 Searches MH
    NEWS FEATURE NATURE|Vol 438|1 December 2005 conference proceedings and institutional repositories, often locating free versions of Start your engines articles on author websites. This ‘grey litera- ture’ is growing in importance but remains poorly defined. It is widely assumed that Google has launched another challenge to commercial search Google considers a source scholarly if it is cited services — this time aimed at scientists. But is the new engine by another scholarly resource — but as online publishing evolves, so may this definition. running as smoothly as its fans hope? Jim Gilesinvestigates. Advocates of greater access to the scientific lit- erature hope that Scholar will encourage more s an undergraduate in India in the Science had a monopoly on citation tracking. researchers to deposit their articles in free mid-1980s, Anurag Acharya had to Citation counts allow researchers, institutes online repositories. write letters to scientists when he and journals to follow the impact of individual But how well does Scholar actually work? Acould not find the papers he wanted. articles through time, leading to metrics, such Librarians who have It is a memory that makes the softly spoken as journal impact factors, that are the bane and run systematic computer engineer laugh. Now working at blessing of many academic careers. Google, Acharya is creating a search tool that But unlike Scopus and Web of aims to be the first choice for everyone from Science, Scholar does more Indian students to Iranian professors. “I want than just search the peer- to make it the one place to go to for scholarly reviewed literature.
    [Show full text]
  • Application of Fuzzy Query Based on Relation Database Dongmei Wei Liangzhong Yi Zheng Pei
    Application of Fuzzy Query Based on Relation Database Dongmei Wei Liangzhong Yi Zheng Pei School of Mathematics & Computer Engineering, Xihua University, Chengdu 610039, China Abstract SELECT <list of ¯elds> FROM <list of tables> ; The traditional query in relation database is unable WHERE <attribute> in <multi-valued attribute> to satisfy the needs for dealing with fuzzy linguis- tic values. In this paper, a new data query tech- SELECT <list of ¯elds> FROM <list of tables>; nique combined fuzzy theory and SQL is provided, WHERE NOT <condition> and the query can be implemented for fuzzy linguis- tic values query via a interface to Microsoft Visual SELECT <list of ¯elds> FROM <list of tables>; Foxpro. Here, we applied it to an realism instance, WHERE <subcondition> AND <subcondition> questions could be expressed by fuzzy linguistic val- ues such as young, high salary, etc, in Employee SELECT <list of ¯elds> FROM <list of tables>; relation database. This could be widely used to WHERE <subcondition> OR <subcondition> realize the other fuzzy query based on database. However, the Complexity is limited in precise data processing and is unable to directly express Keywords: Fuzzy query, Relation database, Fuzzy fuzzy concepts of natural language. For instance, theory, SQL, Microsoft visual foxpro in employees relation database, to deal with a query statement like "younger, well quali¯ed or better 1. Introduction performance ", it is di±cult to construct SQL be- cause the query words are fuzzy expressions. In Database management systems(DBMS) are ex- order to obtain query results, there are two basic tremely useful software products which have been methods of research in the use of SQL Combined used in many kinds of systems [1]-[3].
    [Show full text]
  • On Fuzzy and Rough Sets Approaches to Vagueness Modeling⋆ Extended Abstract
    From Free Will Debate to Embodiment of Fuzzy Logic into Washing Machines: On Fuzzy and Rough Sets Approaches to Vagueness Modeling⋆ Extended Abstract Piotr Wasilewski Faculty of Mathematics, Informatics and Mechanics University of Warsaw Banacha 2, 02-097 Warsaw, Poland [email protected] Deep scientific ideas have at least one distinctive property: they can be applied both by philosophers in abstract fundamental debates and by engineers in concrete practi- cal applications. Mathematical approaches to modeling of vagueness also possess this property. Problems connected with vagueness have been discussed at the beginning of XXth century by philosophers, logicians and mathematicians in developing foundations of mathematics leading to clarification of logical semantics and establishing of math- ematical logic and set theory. Those investigations led also to big step in the history of logic: introduction of three-valued logic. In the second half of XXth century some mathematical theories based on vagueness idea and suitable for modeling vague con- cepts were introduced, including fuzzy set theory proposed by Lotfi Zadeh in 1965 [16] and rough set theory proposed by Zdzis¸saw Pawlak in 1982 [4] having many practical applications in various areas from engineering and computer science such as control theory, data mining, machine learning, knowledge discovery, artificial intelligence. Concepts in classical philosophy and in mathematics are not vague. Classical the- ory of concepts requires that definition of concept C hast to provide exact rules of the following form: if object x belongs to concept C, then x possess properties P1,P2,...,Pn; if object x possess properties P1,P2,...,Pn, then x belongs to concept C.
    [Show full text]
  • Research Hotspots and Frontiers of Product R&D Management
    applied sciences Article Research Hotspots and Frontiers of Product R&D Management under the Background of the Digital Intelligence Era—Bibliometrics Based on Citespace and Histcite Hongda Liu 1,2, Yuxi Luo 2, Jiejun Geng 3 and Pinbo Yao 2,* 1 School of Economics & Management, Tongji University, Shanghai 200092, China; [email protected] or [email protected] 2 School of Management, Shanghai University, Shanghai 200444, China; [email protected] 3 School of Management, Shanghai International Studies University, Shanghai 200333, China; [email protected] * Correspondence: [email protected] Abstract: The rise of “cloud-computing, mobile-Internet, Internet of things, big-data, and smart-data” digital technology has brought a subversive revolution to enterprises and consumers’ traditional patterns. Product research and development has become the main battlefield of enterprise compe- Citation: Liu, H.; Luo, Y.; Geng, J.; tition, facing an environment where challenges and opportunities coexist. Regarding the concepts Yao, P. Research Hotspots and and methods of product R&D projects, the domestic start was later than the international ones, and Frontiers of Product R&D many domestic companies have also used successful foreign cases as benchmarks to innovate their Management under the Background management methods in practice. “Workers must first sharpen their tools if they want to do their of the Digital Intelligence jobs well”. This article will start from the relevant concepts of product R&D projects and summarize Era—Bibliometrics Based on current R&D management ideas and methods. We combined the bibliometric analysis software Citespace and Histcite. Appl. Sci. Histcite and Citespace to sort out the content of domestic and foreign literature and explore the 2021, 11, 6759.
    [Show full text]
  • Download (Accessed on 30 July 2016)
    data Article Earth Observation for Citizen Science Validation, or Citizen Science for Earth Observation Validation? The Role of Quality Assurance of Volunteered Observations Didier G. Leibovici 1,* ID , Jamie Williams 2 ID , Julian F. Rosser 1, Crona Hodges 3, Colin Chapman 4, Chris Higgins 5 and Mike J. Jackson 1 1 Nottingham Geospatial Science, University of Nottingham, Nottingham NG7 2TU, UK; [email protected] (J.F.R.); [email protected] (M.J.J.) 2 Environment Systems Ltd., Aberystwyth SY23 3AH, UK; [email protected] 3 Earth Observation Group, Aberystwyth University Penglais, Aberystwyth SY23 3JG, UK; [email protected] 4 Welsh Government, Aberystwyth SY23 3UR, UK; [email protected] 5 EDINA, University of Edinburgh, Edinburgh EH3 9DR, UK; [email protected] * Correspondence: [email protected] Received: 28 August 2017; Accepted: 19 October 2017; Published: 23 October 2017 Abstract: Environmental policy involving citizen science (CS) is of growing interest. In support of this open data stream of information, validation or quality assessment of the CS geo-located data to their appropriate usage for evidence-based policy making needs a flexible and easily adaptable data curation process ensuring transparency. Addressing these needs, this paper describes an approach for automatic quality assurance as proposed by the Citizen OBservatory WEB (COBWEB) FP7 project. This approach is based upon a workflow composition that combines different quality controls, each belonging to seven categories or “pillars”. Each pillar focuses on a specific dimension in the types of reasoning algorithms for CS data qualification.
    [Show full text]
  • A Framework for Software Modelling in Social Science Re- Search
    A FRAMEWORK FOR SOFTWARE MODELLING IN SOCIAL SCIENCE RESEARCH by Piper J. Jackson B.Sc., Simon Fraser University, 2005 B.A. (Hons.), McGill University, 1996 a Thesis submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the School of Computing Science Faculty of Applied Sciences c Piper J. Jackson 2013 SIMON FRASER UNIVERSITY Summer 2013 All rights reserved. However, in accordance with the Copyright Act of Canada, this work may be reproduced without authorization under the conditions for \Fair Dealing." Therefore, limited reproduction of this work for the purposes of private study, research, criticism, review and news reporting is likely to be in accordance with the law, particularly if cited appropriately. APPROVAL Name: Piper J. Jackson Degree: Doctor of Philosophy Title of Thesis: A Framework for Software Modelling in Social Science Re- search Examining Committee: Dr. Steven Pearce Chair Dr. Uwe Gl¨asser Senior Supervisor Professor Dr. Vahid Dabbaghian Supervisor Adjunct Professor, Mathematics Associate Member, Computing Science Dr. Lou Hafer Internal Examiner Associate Professor Dr. Nathaniel Osgood External Examiner Associate Professor, University of Saskatchewan Date Approved: April 29, 2013 ii Partial Copyright Licence iii Abstract Social science is critical to decision making at the policy level. Software modelling and sim- ulation are innovative computational methods that provide alternative means of developing and testing theory relevant to policy decisions. Software modelling is capable of dealing with obstacles often encountered in traditional social science research, such as the difficulty of performing real-world experimentation. As a relatively new science, computational research in the social sciences faces significant challenges, both in terms of methodology and accep- tance.
    [Show full text]
  • Get Noticed Promoting Your Article for Maximum Impact Get Noticed 2 GET NOTICED
    Get Noticed Promoting your article for maximum impact GET NOTICED 2 GET NOTICED More than one million scientific articles are published each year, and that number is rising. So it’s increasingly important for you to find ways to make your article stand out. While there is much that publishers and editors can do to help, as the paper’s author you are often best placed to explain why your findings are so important or novel. This brochure shows you what Elsevier does and what you can do yourself to ensure that your article gets the attention it deserves. GET NOTICED 3 1 PREPARING YOUR ARTICLE SEO Optimizing your article for search engines – Search Engine Optimization (SEO) – helps to ensure it appears higher in the results returned by search engines such as Google and Google Scholar, Elsevier’s Scirus, IEEE Xplore, Pubmed, and SciPlore.org. This helps you attract more readers, gain higher visibility in the academic community and potentially increase citations. Below are a few SEO guidelines: • Use keywords, especially in the title and abstract. • Add captions with keywords to all photographs, images, graphs and tables. • Add titles or subheadings (with keywords) to the different sections of your article. For more detailed information on how to use SEO, see our guideline: elsevier.com/earlycareer/guides GIVE your researcH THE IMpact it deserVes Thanks to advances in technology, there are many ways to move beyond publishing a flat PDF article and achieve greater impact. You can take advantage of the technologies available on ScienceDirect – Elsevier’s full-text article database – to enhance your article’s value for readers.
    [Show full text]
  • Scientometrics: the Project for a Science of Science Transformed Into an Industry of Measurements
    scientiæ zudia, São Paulo, v. 12, special issue, p. 147-59, 2014 Scientometrics: the project for a science of science transformed into an industry of measurements Renato Rodrigues Kinouchi abstract This paper discusses the intellectual justification of scientometrics through the claim that it is part of the quest for a quantitative science of science. Initially, I will make a brief description of scientometrics’ his- torical background. Next, I will explain that those disciplines that have been satisfactorily mathematized always contain two distinct basic components: an axiomatic, defining the operations that can be realized with the available data, and an interpretation of their meaning. Counting papers and citations is a way to collect statistical data about scientific activities, and therefore the axiomatic basis of scientometrics comes from statistics. Regarding the interpretation of scientometrics, I will argue that the meanings attributed to their key concepts are usually borrowed from economics. Then I discuss how the promise of a science of science becomes a too well adjusted historical narrative that apparently justifies the economic con- cerns of governments and private corporations. Keywords ● Scientometrics. Citation analysis. Quantitative methods. Commoditization. The significance of science and technology for economic development is currently taken for granted as an indisputable premise of any science policy. A wide range of authors from the nineteenth century already had emphasized the connections between sci- ence, technology,
    [Show full text]