Is Simple Wikipedia Simple? – a Study of Readability and Guidelines

Total Page:16

File Type:pdf, Size:1020Kb

Is Simple Wikipedia Simple? – a Study of Readability and Guidelines Linköping University | Department of Computer and Information Science Bachelor thesis, 18 ECTS | Cognitive science 2018 | LIU-IDA/KOGVET-G--18/029--SE Is Simple Wikipedia simple? – A study of readability and guidelines Fabian Isaksson Supervisor : Arne Jönsson Examiner : Henrik Danielsson Linköpings universitet SE–581 83 Linköping +46 13 28 10 00 , www.liu.se Upphovsrätt Detta dokument hålls tillgängligt på Internet – eller dess framtida ersättare – under 25 år från publiceringsdatum under förutsättning att inga extraordinära omständigheter uppstår. Tillgång till dokumentet innebär tillstånd för var och en att läsa, ladda ner, skriva ut enstaka kopior för enskilt bruk och att använda det oförändrat för ickekommersiell forskning och för undervisning. Överföring av upphovsrätten vid en senare tidpunkt kan inte upphäva detta tillstånd. All annan användning av dokumentet kräver upphovsmannens medgivande. För att garantera äktheten, säkerheten och tillgängligheten finns lösningar av teknisk och admin- istrativ art. Upphovsmannens ideella rätt innefattar rätt att bli nämnd som upphovsman i den omfattning som god sed kräver vid användning av dokumentet på ovan beskrivna sätt samt skydd mot att dokumentet ändras eller presenteras i sådan form eller i sådant sam- manhang som är kränkande för upphovsmannens litterära eller konstnärliga anseende eller egenart. För ytterligare information om Linköping University Electronic Press se förlagets hemsida http://www.ep.liu.se/. Copyright The publishers will keep this document online on the Internet – or its possible replacement – for a period of 25 years starting from the date of publication barring exceptional circum- stances. The online availability of the document implies permanent permission for anyone to read, to download, or to print out single copies for his/hers own use and to use it unchanged for non-commercial research and educational purpose. Subsequent transfers of copyright cannot revoke this permission. All other uses of the document are conditional upon the con- sent of the copyright owner. The publisher has taken technical and administrative measures to assure authenticity, security and accessibility. According to intellectual property law the author has the right to be mentioned when his/her work is accessed as described above and to be protected against infringement. For additional information about the Linköping Uni- versity Electronic Press and its procedures for publication and for assurance of document integrity, please refer to its www home page: http://www.ep.liu.se/. c Fabian Isaksson Abstract Creating easy-to-read text is an issue that has traditionally been solved with manual work. But with advancing research in natural language processing, automatic systems for text simplification are being developed. These systems often need training data that is par- allel aligned. For several years, simple Wikipedia has been the main source for this data. In the current study, several readability measures has been tested on a popular simplification corpus. A selection of guidelines from simple Wikipedia has also been operationalized and tested. The results imply that the following of guidelines are not greater in simple Wikipedia than in standard Wikipedia. There are however differences in the readability measures. The syntactical structures of simple Wikipedia seems to be less complex than those of standard Wikipedia. A continuation of this study would be to examine other readability measures and evaluate the guidelines not covered within the current work. keywords: corpus, readability, Wikipedia, automatic text simplification Acknowledgments There are no words that can describe my gratitude for the endless support from my best friend and partner Nora. And the selfless dedication to my happiness from my mother Elis- abeth. Without you this paper would not exist. iv Contents Abstract iii Acknowledgments iv Contents v List of Tables vii 1 Introduction 1 1.1 Aim............................................ 2 1.2 Research questions . 2 1.3 Delimitations . 2 1.4 Structure . 2 2 Theory 3 2.1 Concerning the authorship of simple Wikipedia . 4 2.2 The basis for text simplification . 4 2.2.1 Measuring Readability . 5 2.3 Material . 6 2.3.1 Dataset . 6 2.3.2 Preprocessing tool . 7 2.3.3 The architecture of spaCy . 7 3 Method 9 3.1 Surface features . 9 3.2 Dependency based Features . 10 3.3 Operationalization of Simple Wikipedia Guidelines . 11 3.4 Statistical analysis . 13 4 Results 14 4.1 Surface features and word list comparsion . 14 4.2 Dependency based Features . 15 4.3 Features from Simple Wikipedia Guidelines . 16 5 Discussion 17 5.1 Results . 17 5.1.1 Sentence complexity . 17 5.1.2 Were the guidelines followed? . 18 5.2 Method . 19 5.2.1 Parallelism matters . 20 5.2.2 Concerning the parser . 21 5.2.3 Studying the guidelines . 21 5.2.4 Social aspects . 22 v 6 Conclusion 23 Bibliography 24 vi List of Tables 4.1 Surface features and Word list comparison . 15 4.2 Dependency based features . 15 4.3 Statistics pertaining to simple wiki Guidelines . 16 vii 1 Introduction Linguists have been working on ways to concretely assess the readability of text prior to the dawn of computer science. Early approaches were aimed at determining the years of formal education needed to understand a text. With the practical motivation of providing children with reading material appropriate to their language abilities. In contemporary readability research, both the motivations and ambitions have expanded. Natural language processing has allowed for more sophisticated ways of measuring text complexity. With statistical ma- chine learning we can now create tools that automatically distinguish easy-to-read language from other language. Using that knowledge to let a computer simplify texts is known as automatic text simplification. Several groups of our population are in need of easy-to-read material. Including second language learners, aphasic patients and children. Since manual simplification is a tedious and time consuming task, the interest is growing in automating the process. With data driven machine learning, new systems can automatically learn how to simplify texts. But in order to develop such systems, a large amount of simple language data is needed. Simple Wikipedia is a large collection of encyclopaedic articles aimed at people learning english. There is research suggesting that these articles can be said to be lower in readability (Yasseri, Kornal, and Kertész 2012)(Kauchak 2013)(Hwang, Hajishirzi, Ostendorf, and Wu 2015). For several years the Simple English Wikipedia corpus (Kauchak and Coster 2011) has been a primary source of data in automatic text simplification (ATS)(Napoles and Dredze 2010). It has been used as a resource for creating language models, developing clas- sifiers and studying readability. The corpus consists of parallel aligned sentences between the standard part of Wikipedia (standard Wikipedia) and simple english Wikipedia (simple Wikipedia). In creating parallel aligned corpora from unaligned texts, a lot of text from the source material is left out. The results of readability studies on the whole of standard and simple Wikipedia do not necessarily apply to any subsection of the data. It would be useful to see if the aligned corpus follow the same patterns as has been found in the whole of sim- ple Wikipedia. At this point in time there is also a lack of descriptive information about the readability differences in the corpus. And no studies have been made about whether author follow the simple Wikipedia guidelines for writing. These points of concern is what will be addressed in this study. 1 1.1. Aim 1.1 Aim The aim of this study is to evaluate the language complexity differences between simple and standard Wikipedia. This study also aims to test the extent to which the guidelines for simple english Wikipedia has been followed. The analysis will specifically pertain to the corpus prepared by Kauchak (Kauchak and Coster 2011). This means that only a subset of all available articles, that has been sentence aligned with standard Wikipedia, will be the subject of this study. The results found here will hopefully provide insights that can be used in developing new text simplification systems based on the corpus. 1.2 Research questions There are two research questions that I intend to answer in this thesis: 1. Are there surface level or dependency based features that indicate differences in readability between parallel aligned sentences from simple Wikipedia and standard Wikipedia? 2. Do these parallel aligned sentences follow the guidelines for simple English Wikipedia to the same extent? 1.3 Delimitations Not all Simple Wikipedia guidelines have been operationalized and tested. Constituency based readability features have not been used. 1.4 Structure This thesis consists of six different chapters. Chapter 2 gives an account of previous research on text simplification, presents the guidelines, how the corpus was created and the NLP tools used in this project. Chapter 3 contains motivations and descriptions of the readability mea- sures and guidelines that have been used to analyse the corpus. In chapter 4, results of the measures are presented. Chapter 5 is a discussion on how to interpret the results and method- ological alternatives. Chapter 6 presents the conclusions drawn from the study and proposi- tions for future research. 2 2 Theory Simple English Wikipedia is a collection of simplified encyclopaedic articles available for free online. The database contains 133,641 articles at this time (Wikipedia 2018a). Most of which have corresponding standard articles on the same subjects. This is a resource cre- ated to share knowledge. And specifically to facilitate an easy-to-read option to standard Wikipedia for children and adults who are learning English. Even though this resource is unique in many ways, addressing the need for simple language is not a novel concept. Before the website Wikipedia launched, guidelines for writing plain English were being developed. These guidelines specify instructions for how to limit a vocabulary, change the syntactic struc- tures of sentences, structure a text in logical order and prioritize the core information needed for comprehension (PLAIN 2011)1.
Recommended publications
  • A Topic-Aligned Multilingual Corpus of Wikipedia Articles for Studying Information Asymmetry in Low Resource Languages
    Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020), pages 2373–2380 Marseille, 11–16 May 2020 c European Language Resources Association (ELRA), licensed under CC-BY-NC A Topic-Aligned Multilingual Corpus of Wikipedia Articles for Studying Information Asymmetry in Low Resource Languages Dwaipayan Roy, Sumit Bhatia, Prateek Jain GESIS - Cologne, IBM Research - Delhi, IIIT - Delhi [email protected], [email protected], [email protected] Abstract Wikipedia is the largest web-based open encyclopedia covering more than three hundred languages. However, different language editions of Wikipedia differ significantly in terms of their information coverage. We present a systematic comparison of information coverage in English Wikipedia (most exhaustive) and Wikipedias in eight other widely spoken languages (Arabic, German, Hindi, Korean, Portuguese, Russian, Spanish and Turkish). We analyze the content present in the respective Wikipedias in terms of the coverage of topics as well as the depth of coverage of topics included in these Wikipedias. Our analysis quantifies and provides useful insights about the information gap that exists between different language editions of Wikipedia and offers a roadmap for the Information Retrieval (IR) community to bridge this gap. Keywords: Wikipedia, Knowledge base, Information gap 1. Introduction other with respect to the coverage of topics as well as Wikipedia is the largest web-based encyclopedia covering the amount of information about overlapping topics.
    [Show full text]
  • Measuring Text Simplification with the Crowd
    Measuring Text Simplification with the Crowd Walter S. Lasecki Luz Rello Jeffrey P. Bigham ROC HCI HCI Institute HCI and LT Institutes Dept. of Computer Science School of Computer Science School of Computer Science University of Rochester Carnegie Mellon University Carnegie Mellon University [email protected] [email protected] [email protected] ABSTRACT the U.S. [41]1 and the European Guidelines for the Production of Text can often be complex and difficult to read, especially for peo­ Easy-to-Read Information in Europe [23]. ple with cognitive impairments or low literacy skills. Text simplifi­ Text simplification is an area of Natural Language Processing cation is a process that reduces the complexity of both wording and (NLP) that attempts to solve this problem automatically by reduc­ structure in a sentence, while retaining its meaning. However, this ing the complexity of both wording (lexicon) and structure (syn­ is currently a challenging task for machines, and thus, providing tax) in a sentence [46]. Previous efforts on automatic text simpli­ effective on-demand text simplification to those who need it re­ fication have focused on people with autism [20, 39], people with mains an unsolved problem. Even evaluating the simplicity of text Down syndrome [48], people with dyslexia [42, 43], and people remains a challenging problem for both computers, which cannot with aphasia [11, 18]. understand the meaning of text, and humans, who often struggle to However, automatic text simplification is in an early stage and agree on what constitutes a good simplification. still is not useful for the target users [20, 43, 48].
    [Show full text]
  • The Culture of Wikipedia
    Good Faith Collaboration: The Culture of Wikipedia Good Faith Collaboration The Culture of Wikipedia Joseph Michael Reagle Jr. Foreword by Lawrence Lessig The MIT Press, Cambridge, MA. Web edition, Copyright © 2011 by Joseph Michael Reagle Jr. CC-NC-SA 3.0 Purchase at Amazon.com | Barnes and Noble | IndieBound | MIT Press Wikipedia's style of collaborative production has been lauded, lambasted, and satirized. Despite unease over its implications for the character (and quality) of knowledge, Wikipedia has brought us closer than ever to a realization of the centuries-old Author Bio & Research Blog pursuit of a universal encyclopedia. Good Faith Collaboration: The Culture of Wikipedia is a rich ethnographic portrayal of Wikipedia's historical roots, collaborative culture, and much debated legacy. Foreword Preface to the Web Edition Praise for Good Faith Collaboration Preface Extended Table of Contents "Reagle offers a compelling case that Wikipedia's most fascinating and unprecedented aspect isn't the encyclopedia itself — rather, it's the collaborative culture that underpins it: brawling, self-reflexive, funny, serious, and full-tilt committed to the 1. Nazis and Norms project, even if it means setting aside personal differences. Reagle's position as a scholar and a member of the community 2. The Pursuit of the Universal makes him uniquely situated to describe this culture." —Cory Doctorow , Boing Boing Encyclopedia "Reagle provides ample data regarding the everyday practices and cultural norms of the community which collaborates to 3. Good Faith Collaboration produce Wikipedia. His rich research and nuanced appreciation of the complexities of cultural digital media research are 4. The Puzzle of Openness well presented.
    [Show full text]
  • Proceedings of the 1St Workshop on Tools and Resources to Empower
    LREC 2020 Workshop Language Resources and Evaluation Conference 11–16 May 2020 1st Workshop on Tools and Resources to Empower People with REAding DIfficulties (READI) PROCEEDINGS Edited by Nuria´ Gala and Rodrigo Wilkens Proceedings of the LREC 2020 first workshop on Tools and Resources to Empower People with REAding DIfficulties (READI) Edited by: Núria Gala and Rodrigo Wilkens ISBN: 979-10-95546-44-3 EAN: 9791095546443 For more information: European Language Resources Association (ELRA) 9 rue des Cordelières 75013, Paris France http://www.elra.info Email: [email protected] c European Language Resources Association (ELRA) These workshop proceedings are licensed under a Creative Commons Attribution-NonCommercial 4.0 International License ii Preface Recent studies show that the number of children and adults facing difficulties in reading and understanding written texts is steadily growing. Reading challenges can show up early on and may include reading accuracy, speed, or comprehension to the extent that the impairment interferes with academic achievement or activities of daily life. Various technologies (text customization, text simplification, text to speech devices, screening for readers through games and web applications, to name a few) have been developed to help poor readers to get better access to information as well as to support reading development. Among those technologies, text simplification is a powerful way to leverage document accessibility by using NLP techniques. The “First Workshop on Tools and Resources to Empower People with REAding DIfficulties” (READI), collocated with the “International Conference on Language Resources and Evaluation” (LREC 2020), aims at presenting current state-of-the-art techniques and achievements for text simplification together with existing reading aids and resources for lifelong learning, addressing a variety of domains and languages, including natural language processing, linguistics, psycholinguistics, psychophysics of vision and education.
    [Show full text]
  • How to Contribute Climate Change Information to Wikipedia : a Guide
    HOW TO CONTRIBUTE CLIMATE CHANGE INFORMATION TO WIKIPEDIA Emma Baker, Lisa McNamara, Beth Mackay, Katharine Vincent; ; © 2021, CDKN This work is licensed under the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/legalcode), which permits unrestricted use, distribution, and reproduction, provided the original work is properly credited. Cette œuvre est mise à disposition selon les termes de la licence Creative Commons Attribution (https://creativecommons.org/licenses/by/4.0/legalcode), qui permet l’utilisation, la distribution et la reproduction sans restriction, pourvu que le mérite de la création originale soit adéquatement reconnu. IDRC Grant/ Subvention du CRDI: 108754-001-CDKN knowledge accelerator for climate compatible development How to contribute climate change information to Wikipedia A guide for researchers, practitioners and communicators Contents About this guide .................................................................................................................................................... 5 1 Why Wikipedia is an important tool to communicate climate change information .................................................................................................................................. 7 1.1 Enhancing the quality of online climate change information ............................................. 8 1.2 Sharing your work more widely ......................................................................................................8 1.3 Why researchers should
    [Show full text]
  • Semantically Annotated Snapshot of the English Wikipedia
    Semantically Annotated Snapshot of the English Wikipedia Jordi Atserias, Hugo Zaragoza, Massimiliano Ciaramita, Giuseppe Attardi Yahoo! Research Barcelona, U. Pisa, on sabbatical at Yahoo! Research C/Ocata 1 Barcelona 08003 Spain {jordi, hugoz, massi}@yahoo-inc.com, [email protected] Abstract This paper describes SW1, the first version of a semantically annotated snapshot of the English Wikipedia. In recent years Wikipedia has become a valuable resource for both the Natural Language Processing (NLP) community and the Information Retrieval (IR) community. Although NLP technology for processing Wikipedia already exists, not all researchers and developers have the computational resources to process such a volume of information. Moreover, the use of different versions of Wikipedia processed differently might make it difficult to compare results. The aim of this work is to provide easy access to syntactic and semantic annotations for researchers of both NLP and IR communities by building a reference corpus to homogenize experiments and make results comparable. These resources, a semantically annotated corpus and a “entity containment” derived graph, are licensed under the GNU Free Documentation License and available from http://www.yr-bcn.es/semanticWikipedia. 1. Introduction 2. Processing Wikipedia1, the largest electronic encyclopedia, has be- Starting from the XML Wikipedia source we carried out a come a widely used resource for different Natural Lan- number of data processing steps: guage Processing tasks, e.g. Word Sense Disambiguation (Mihalcea, 2007), Semantic Relatedness (Gabrilovich and • Basic preprocessing: Stripping the text from the Markovitch, 2007) or in the Multilingual Question Answer- XML tags and dividing the obtained text into sen- ing task at Cross-Language Evaluation Forum (CLEF)2.
    [Show full text]
  • Data-Driven Text Simplification
    Data-Driven Text Simplification Sanja Štajner and Horacio Saggion [email protected] http://web.informatik.uni-mannheim.de/sstajner/index.html University of Mannheim, Germany [email protected] #TextSimplification2018 https://www.upf.edu/web/horacio-saggion University Pompeu Fabra, Spain COLING 2018 Tutorial Presenters • Sanja Stajner • Horacio Saggion • http://web.informatik.uni- • http://www.dtic.upf.edu/~hsaggion mannheim.de/sstajner • https://www.linkedin.com/pub/horacio- • https://www.linkedin.com/in/sanja- saggion/16/9b9/174 stajner-a6904738 • https://twitter.com/h_saggion • Data and Web Science Group • Large Scale Text Understanding • https://dws.informatik.uni - Systems Lab / TALN group mannheim.de/en/home/ • http://www.taln.upf.edu • University of Mannheim, Germany • Department of Information & Communication Technologies • Universitat Pompeu Fabra, Barcelona, Spain © 2018 by S. Štajner & H. Saggion 2 Tutorial antecedents • Previous tutorials on the topic given at: • IJCNLP 2013 and RANLP 2015 (H. Saggion) • RANLP 2017 (S. Štajner) • Automatic Text Simplification. H. Saggion. 2017. Morgan & Claypool. Synthesis Lectures on Human Language Technologies Series. https://www.morganclaypool.com/doi/abs/10.2200/S00700ED1V01Y201602 HLT032 © 2018 by S. Štajner & H. Saggion 3 Outline • Motivation for ATS • Automatic text simplification • TS projects • TS resources • Neural text simplification © 2018 by S. Štajner & H. Saggion 4 PART 1 Motivation for Text Simplification © 2018 by S. Štajner & H. Saggion 5 Text Simplification (TS) The process of transforming a text into an equivalent which is more readable and/or understandable by a target audience • During simplification, complex sentences are split into simple ones and uncommon vocabulary is replaced by more common expressions • TS is a complex task which encompasses a number of operations applied at different linguistic levels: • Lexical • Syntactic • Discourse • Started to attract the attention of natural language processing some years ago (1996) mainly as a pre-processing step © 2018 by S.
    [Show full text]
  • Classifier-Based Text Simplification for Improved Machine Translation
    Classifier-Based Text Simplification for Improved Machine Translation Shruti Tyagi Deepti Chopra Iti Mathur Nisheeth Joshi Department of Computer Department of Computer Department of Computer Department of Computer Science Science Science Science Banasthali University Banasthali University Banasthali University Banasthali University Rajasthan, India Rajasthan, India Rajasthan, India Rajasthan, India [email protected] [email protected] [email protected] [email protected] Abstract— Machine Translation is one of the research fields E. Text Summarization: Splitting of sentence or conversion of of Computational Linguistics. The objective of many MT long sentence into smaller ones helps in text summarization. Researchers is to develop an MT System that produce good quality and high accuracy output translations and which also There are also different approaches through which we can covers maximum language pairs. As internet and Globalization is apply text simplification. These are: increasing day by day, we need a way that improves the quality of translation. For this reason, we have developed a Classifier based Text Simplification Model for English-Hindi Machine A. Lexical Simplification: In this approach, identification of Translation Systems. We have used support vector machines and complex words takes place and generation of Naïve Bayes Classifier to develop this model. We have also evaluated the performance of these classifiers. substitutes/synonyms takes place accordingly. For example, in the example below, we have highlighted the complex words Keywords— Machine Translation, Text Simplification, Naïve and their respective substitutes. Bayes Classifier, Support Vector Machine Classifier Original Sentence: Audible word was originally named audire in Latin. I. INTRODUCTION Simplified Sentence: Audible word was first called audire in Text Simplification is the process that reduces the Latin.
    [Show full text]
  • Natural Language Processing Tools for Reading Level Assessment and Text Simplification for Bilingual Education
    Natural Language Processing Tools for Reading Level Assessment and Text Simplification for Bilingual Education Sarah E. Petersen A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Washington 2007 Program Authorized to Offer Degree: Computer Science & Engineering University of Washington Graduate School This is to certify that I have examined this copy of a doctoral dissertation by Sarah E. Petersen and have found that it is complete and satisfactory in all respects, and that any and all revisions required by the final examining committee have been made. Chair of the Supervisory Committee: Mari Ostendorf Reading Committee: Mari Ostendorf Oren Etzioni Henry Kautz Date: In presenting this dissertation in partial fulfillment of the requirements for the doctoral degree at the University of Washington, I agree that the Library shall make its copies freely available for inspection. I further agree that extensive copying of this dissertation is allowable only for scholarly purposes, consistent with “fair use” as prescribed in the U.S. Copyright Law. Requests for copying or reproduction of this dissertation may be referred to Proquest Information and Learning, 300 North Zeeb Road, Ann Arbor, MI 48106-1346, 1-800-521-0600, to whom the author has granted “the right to reproduce and sell (a) copies of the manuscript in microform and/or (b) printed copies of the manuscript made from microform.” Signature Date University of Washington Abstract Natural Language Processing Tools for Reading Level Assessment and Text Simplification for Bilingual Education Sarah E. Petersen Chair of the Supervisory Committee: Professor Mari Ostendorf Electrical Engineering Reading proficiency is a fundamental component of language competency.
    [Show full text]
  • Lessons from Citizendium
    Lessons from Citizendium Wikimania 2009, Buenos Aires, 28 August 2009 HaeB [[de:Benutzer:HaeB]], [[en:User:HaeB]] Please don't take photos during this talk. Citizendium Timeline ● September 2006: Citizendium announced. Sole founder: Larry Sanger, known as former editor-in-chief of Nupedia, chief organizer of Wikipedia (2001-2002), and later as Wikipedia critic ● October 2006: Started non-public pilot phase ● January 2007: “Big Unfork”: All unmodified copies of Wikipedia articles deleted ● March 2007: Public launch ● December 2007: Decision to use CC-BY-3.0, after debate about commercial reuse and compatibility with Wikipedia ● Mid-2009: Sanger largely inactive on Citizendium, focuses on WatchKnow ● August 2009: Larry Sanger announces he will step down as editor-in-chief soon (as committed to in 2006) Citizendium and Wikipedia: Similarities and differences ● Encyclopedia ● Strict real names ● Free license policy ● ● Open (anyone can Special role for contribute) experts: “editors” can issue content ● Created by amateurs decisions, binding to ● MediaWiki-based non-editors collaboration ● Governance: Social ● Non-profit contract, elements of a constitutional republic Wikipedian views of Citizendium ● Competitor for readers, contributions ● Ally, common goal of creating free encyclopedic content ● “Who?” ● In this talk: A long-time experiment testing several fundamental policy changes, in a framework which is still similar enough to that of Wikipedia to generate valuable evidence as to what their effect might be on WP Active editors: Waiting to explode ● Sanger (October 2007): ”At some point, possibly very soon, the Citizendium will grow explosively-- say, quadruple the number of its active contributors, or even grow by an order of magnitude ....“ © Aleksander Stos, CC-BY 3.0 Number of users that made at least one edit in each month Article creation rate: Still muddling Sanger (October 2007): “It's still possible that the project will, from here until eternity, muddle on creating 14 articles per day.
    [Show full text]
  • Wikipedia Matters∗
    Wikipedia Matters∗ Marit Hinnosaar† Toomas Hinnosaar‡ Michael Kummer§ Olga Slivko¶ September 29, 2017 Abstract We document a causal impact of online user-generated information on real-world economic outcomes. In particular, we conduct a randomized field experiment to test whether additional content on Wikipedia pages about cities affects tourists’ choices of overnight visits. Our treatment of adding information to Wikipedia increases overnight visits by 9% during the tourist season. The impact comes mostly from improving the shorter and incomplete pages on Wikipedia. These findings highlight the value of content in digital public goods for informing individual choices. JEL: C93, H41, L17, L82, L83, L86 Keywords: field experiment, user-generated content, Wikipedia, tourism industry 1 Introduction Asymmetric information can hinder efficient economic activity. In recent decades, the Internet and new media have enabled greater access to information than ever before. However, the digital divide, language barriers, Internet censorship, and technological con- straints still create inequalities in the amount of accessible information. How much does it matter for economic outcomes? In this paper, we analyze the causal impact of online information on real-world eco- nomic outcomes. In particular, we measure the impact of information on one of the primary economic decisions—consumption. As the source of information, we focus on Wikipedia. It is one of the most important online sources of reference. It is the fifth most ∗We are grateful to Irene Bertschek, Avi Goldfarb, Shane Greenstein, Tobias Kretschmer, Thomas Niebel, Marianne Saam, Greg Veramendi, Joel Waldfogel, and Michael Zhang as well as seminar audiences at the Economics of Network Industries conference in Paris, ZEW Conference on the Economics of ICT, and Advances with Field Experiments 2017 Conference at the University of Chicago for valuable comments.
    [Show full text]
  • Experimenting Sentence Split-And-Rephrase Using Part-Of-Speech Labels
    Experimenting Sentence Split-and-Rephrase Using Part-of-Speech Labels P. Berlanga Neto and E. Y. Okano and E. E. S. Ruiz Departamento de Computação e Matemática, FFCLRP Universidade de São Paulo (USP). Av. Bandeirantes, 3900, Monte Alegre. 14040-901, Ribeirão Preto, SP – Brazil [pauloberlanga, okano700, evandro]@usp.br Abstract. Text simplification (TS) is a natural language transformation process that reduces linguistic complexity while preserving semantics and retaining its original meaning. This work aims to present a research proposal for automatic simplification of texts, precisely a split-and-rephrase approach based on an encoder-decoder neural network model. The proposed method was trained against the WikiSplit English corpus with the help of a part-of-speech tagger and obtained a BLEU score validation of 74.72%. We also experimented with this trained model to split-and-rephrase sentences written in Portuguese with relative success, showing the method’s potential. CCS Concepts: • Computing methodologies → Natural language processing. Keywords: natural language processing, neural networks, sentence simplification 1. INTRODUCTION Text Simplification (TS) is the process of modifying natural language to reduce complexity and im- prove both readability and understandability [Shardlow 2014]. A simplified vocabulary or a simplified text structure can benefit different publics, such as people with low education levels, children, non- native individuals, people who have learning disorders (such as autism, dyslexia, and aphasia), among others [Štajner et al. 2015]. Although the simplicity of a text seems intuitively obvious, it does not have a precise definition in technical terms. Traditional assessment metrics consider factors such as sentence length, syllable count, and other linguistic features of the text to identify it as elaborate or not [Shardlow 2014].
    [Show full text]