Un Approccio Basato Su Dbpedia Per La Sistematizzazione Della Conoscenza Sul Web

Total Page:16

File Type:pdf, Size:1020Kb

Un Approccio Basato Su Dbpedia Per La Sistematizzazione Della Conoscenza Sul Web POLITECNICO DI TORINO SCUOLA DI DOTTORATO Dottorato in Beni Culturali - XXV Ciclo Tesi di Dottorato Un approccio basato su DBpedia per la sistematizzazione della conoscenza sul Web Federico Cairo Tutore Coordinatore del corso di dottorato prof. Mario Ricciardi prof. Costanza Roggero Marzo 2013 2 Indice Ringraziamenti ............................................................................................... 7 Introduzione ................................................................................................... 9 Capitolo 1 - Wikipedia e l’intelligenza collettiva .......................................... 13 1.1 Il successo di Wikipedia ................................................................... 13 1.2 Intelligenza collettiva e cultura partecipativa in Wikipedia ............. 21 1.3 Criticità in Wikipedia ...................................................................... 34 1.3.1 Inaccuratezza e inaffidabilità ................................................... 35 1.3.2 Copertura diseguale dei diversi ambiti del sapere .................... 39 1.3.3 Parzialità ed esposizione a pregiudizi e interessi ...................... 42 1.3.4 Volatilità ................................................................................. 45 1.4 Conclusioni ....................................................................................... 45 Capitolo 2 - DBpedia, Linked Open Data e classificazione semantica dei contenuti ............................................................................................................ 52 2.1 Web Semantico e Linked Open Data ............................................... 52 2.2 DBpedia: il punto di riferimento per il Web dei Dati ...................... 66 2.3 Annotazione e classificazione semantica del testo ............................ 78 2.3.1 Che cosa si intende per classificazione ed annotazione semantica ................................................................................................................ 78 2.3.2 Vantaggi nell’uso di DBpedia per il Natural Language Processing ............................................................................................... 87 2.3.3 Stato dell’arte: una comparazione tra i software di annotazione semantica ................................................................................................ 88 Capitolo 3 - TellMeFirst: un sistema per l’annotazione, la classificazione e l’arricchimento dei documenti attraverso DBpedia ........................................... 107 3 3.1 L’approccio di TellMeFirst all’annotazione e alla classificazione semantica ........................................................................................ 107 3.2 Componenti del sistema ................................................................. 111 3.3 Interazione tra componenti e artefatti ........................................... 114 3.3.1 Dataset iniziali ...................................................................... 114 3.3.2 Dataset elaborati ................................................................... 116 3.4 Il modulo di disambiguazione di TellMeFirst ................................. 120 3.4.1 Corpus-based disambiguation ................................................ 122 3.4.2 Knowledge-based disambiguation .......................................... 125 3.4.3 First Sense Heuristic disambiguation .................................... 127 3.4.4 Default disambiguation ......................................................... 129 3.5 Modulo di classificazione ................................................................ 132 3.5.1 Input e output ....................................................................... 132 3.5.2 Funzionamento ...................................................................... 133 3.5.3 Esempio di chiamata e risposta ............................................. 136 3.6. Modulo di enhancement ................................................................ 138 3.6.1 GetImage ............................................................................... 138 3.6.2 GetText ................................................................................. 141 3.6.3 GetNews ................................................................................ 142 3.6.4 GetMap ................................................................................. 144 3.6.5 GetVideo ............................................................................... 145 3.7 Interfaccia Web .............................................................................. 146 3.7.1 Modulo di acquisizione del testo ............................................ 147 3.7.2 Griglia degli argomenti .......................................................... 148 3.7.3 Box dei contenuti .................................................................. 149 3.7.4 Flusso di esecuzione ............................................................... 154 4 3.8 Test e prestazioni ........................................................................... 163 Capitolo 4 - DBpedia Gateways contro l’information overload sul Web ..... 171 4.1 Il problema dell’information overload in Rete ................................ 171 4.2 DBpedia Gateways ......................................................................... 184 4.3 Un progetto di esempio: Future Internet Gateway ........................ 192 4.3.1 Introduzione .......................................................................... 192 4.3.2 I progetti di CSA in campo FIRE ......................................... 193 4.3.3 FIG – Future Internet Gateway ............................................ 195 Conclusioni ................................................................................................. 201 Bibliografia ................................................................................................. 205 5 6 Ringraziamenti Desidero innanzitutto ringraziare il Prof. Mario Ricciardi, tutore del mio per- corso di dottorato, per il sostegno e l’aiuto fornitomi nella scelta e nello sviluppo del tema di questa tesi. La mia gratitudine va inoltre al Centro Nexa su Internet & Società del Politecnico di Torino, dove ho svolto buona parte del mio lavoro su TellMeFirst e dove ho avuto occasione di approfondire le tematiche relative ai Linked Open Data in un ambiente di grande competenza e cordialità. Voglio an- che esprimere la mia gratitudine nei confronti di Telecom Italia (in particolare Walter Goix, Fabio Mondin, Carmen Crimisi e Carlo Alberto Licciardi) per aver creduto nel progetto TellMeFirst tanto da finanziarlo e seguirne lo sviluppo in tutte le sue fasi. Grazie a CSI Piemonte e a CELI (in particolare Giuliana Bonel- lo, Nathalie Coué, Andrea Muraca, Alessandro Trombotto e Vittorio Di Tomaso) per avermi formato “sul campo” nei primi tempi della mia esperienza dottorale. Un ultimo grande ringraziamento va a Giuseppe Futia e a Federico Benedetto, sviluppatori del front-end di TellMeFirst, brillanti collaboratori e soprattutto a- mici, senza i quali non ce l’avrei mai fatta. 7 8 Introduzione Nel 2004, quando ho scritto la mia tesi di laurea in Filosofia all’Università di Bologna, il Web 2.0 cominciava a muovere i primi passi incerti e il compito di un tesista era più semplice. Si raccoglievano i testi più rilevanti su di un argomento, per lo più cartacei, in base alle bibliografie contenute in altri testi cartacei sullo stesso tema. Inizialmente i titoli sconosciuti erano parecchi, ma ben presto co- minciavano a ripetersi, le novità diminuivano di numero e si giungeva alla piace- vole situazione in cui i documenti citati erano sempre gli stessi: in quel momento ci si sentiva di padroneggiare una materia, per quanto piccola e insignificante ri- spetto alla totalità della conoscenza umana. Chi deve scrivere una tesi oggi, invece, in un’epoca di maggiore maturità del Web, si trova nella complicata posizione di dover porre dei limiti estremamente soggettivi alle proprie letture. Il rapporto numerico tra i testi digitali reperibili online e quelli pubblicati su carta, soprattutto in merito a un argomento di re- cente interesse, è incredibilmente alto, ed è in costante crescita. Al 3 ottobre 2012, digitando “Semantic Web” sul motore di ricerca del Catalogo del Servizio Bibliotecario Nazionale 1, otteniamo 98 risultati, ma digitando le stesse keyword su una nota piattaforma di social sharing per paper scientifici, Mendeley 2, il nu- mero sale fino a 8.770 (circa 90 volte maggiore). I risultati della stessa ricerca su Google, si assestano invece sui 8.840.000 (un numero circa 90.000 volte maggio- re). Si tratta di un muro quasi minaccioso di documenti, una situazione che fa subito venire in mente parole come sovraccarico informativo o «infobesity» (Jo- hnson, 2012). Siamo arrivati al punto in cui la quantità di informazioni reperibili su Internet ci impedisce di fatto di conoscere un argomento? Alcuni sono di quest’avviso. Nicolas Carr per esempio, nel suo famoso saggio The shallows: what the Internet is doing to our brains , mette in guardia sulla facilità di ottenere a portata di mano enormi quantità di dati e informazioni senza la minima capacità 1 URL: http://www.sbn.it/opacsbn/opac/iccu/free.jsp 2 URL: http://www.mendeley.com/ 9 di saperli discernere (Carr, 2010). Oppure Umberto Eco, che sempre più spesso in interviste e articoli manifesta il suo pessimismo nei confronti del Web come strumento conoscitivo, a causa della «censura per eccesso di rumore» (Eco, 2012) che caratterizza l’ information overload 3. Combattere il sovraccarico
Recommended publications
  • Wikimedia Conferentie Nederland 2012 Conferentieboek
    http://www.wikimediaconferentie.nl WCN 2012 Conferentieboek CC-BY-SA 9:30–9:40 Opening 9:45–10:30 Lydia Pintscher: Introduction to Wikidata — the next big thing for Wikipedia and the world Wikipedia in het onderwijs Technische kant van wiki’s Wiki-gemeenschappen 10:45–11:30 Jos Punie Ralf Lämmel Sarah Morassi & Ziko van Dijk Het gebruik van Wikimedia Commons en Wikibooks in Community and ontology support for the Wikimedia Nederland, wat is dat precies? interactieve onderwijsvormen voor het secundaire onderwijs 101wiki 11:45–12:30 Tim Ruijters Sandra Fauconnier Een passie voor leren, de Nederlandse wikiversiteit Projecten van Wikimedia Nederland in 2012 en verder Bliksemsessie …+discussie …+vragensessie Lunch 13:15–14:15 Jimmy Wales 14:30–14:50 Wim Muskee Amit Bronner Lotte Belice Baltussen Wikipedia in Edurep Bridging the Gap of Multilingual Diversity Open Cultuur Data: Een bottom-up initiatief vanuit de erfgoedsector 14:55–15:15 Teun Lucassen Gerard Kuys Finne Boonen Scholieren op Wikipedia Onderwerpen vinden met DBpedia Blijf je of ga je weg? 15:30–15:50 Laura van Broekhoven & Jan Auke Brink Jeroen De Dauw Jan-Bart de Vreede 15:55–16:15 Wetenschappelijke stagiairs vertalen onderzoek naar Structured Data in MediaWiki Wikiwijs in vergelijking tot Wikiversity en Wikibooks Wikipedia–lemma 16:20–17:15 Prijsuitreiking van Wiki Loves Monuments Nederland 17:20–17:30 Afsluiting 17:30–18:30 Borrel Inhoudsopgave Organisatie 2 Voorwoord 3 09:45{10:30: Lydia Pintscher 4 13:15{14:15: Jimmy Wales 4 Wikipedia in het onderwijs 5 11:00{11:45: Jos Punie
    [Show full text]
  • Towards a Korean Dbpedia and an Approach for Complementing the Korean Wikipedia Based on Dbpedia
    Towards a Korean DBpedia and an Approach for Complementing the Korean Wikipedia based on DBpedia Eun-kyung Kim1, Matthias Weidl2, Key-Sun Choi1, S¨orenAuer2 1 Semantic Web Research Center, CS Department, KAIST, Korea, 305-701 2 Universit¨at Leipzig, Department of Computer Science, Johannisgasse 26, D-04103 Leipzig, Germany [email protected], [email protected] [email protected], [email protected] Abstract. In the first part of this paper we report about experiences when applying the DBpedia extraction framework to the Korean Wikipedia. We improved the extraction of non-Latin characters and extended the framework with pluggable internationalization components in order to fa- cilitate the extraction of localized information. With these improvements we almost doubled the amount of extracted triples. We also will present the results of the extraction for Korean. In the second part, we present a conceptual study aimed at understanding the impact of international resource synchronization in DBpedia. In the absence of any informa- tion synchronization, each country would construct its own datasets and manage it from its users. Moreover the cooperation across the various countries is adversely affected. Keywords: Synchronization, Wikipedia, DBpedia, Multi-lingual 1 Introduction Wikipedia is the largest encyclopedia of mankind and is written collaboratively by people all around the world. Everybody can access this knowledge as well as add and edit articles. Right now Wikipedia is available in 260 languages and the quality of the articles reached a high level [1]. However, Wikipedia only offers full-text search for this textual information. For that reason, different projects have been started to convert this information into structured knowledge, which can be used by Semantic Web technologies to ask sophisticated queries against Wikipedia.
    [Show full text]
  • The Culture of Wikipedia
    Good Faith Collaboration: The Culture of Wikipedia Good Faith Collaboration The Culture of Wikipedia Joseph Michael Reagle Jr. Foreword by Lawrence Lessig The MIT Press, Cambridge, MA. Web edition, Copyright © 2011 by Joseph Michael Reagle Jr. CC-NC-SA 3.0 Purchase at Amazon.com | Barnes and Noble | IndieBound | MIT Press Wikipedia's style of collaborative production has been lauded, lambasted, and satirized. Despite unease over its implications for the character (and quality) of knowledge, Wikipedia has brought us closer than ever to a realization of the centuries-old Author Bio & Research Blog pursuit of a universal encyclopedia. Good Faith Collaboration: The Culture of Wikipedia is a rich ethnographic portrayal of Wikipedia's historical roots, collaborative culture, and much debated legacy. Foreword Preface to the Web Edition Praise for Good Faith Collaboration Preface Extended Table of Contents "Reagle offers a compelling case that Wikipedia's most fascinating and unprecedented aspect isn't the encyclopedia itself — rather, it's the collaborative culture that underpins it: brawling, self-reflexive, funny, serious, and full-tilt committed to the 1. Nazis and Norms project, even if it means setting aside personal differences. Reagle's position as a scholar and a member of the community 2. The Pursuit of the Universal makes him uniquely situated to describe this culture." —Cory Doctorow , Boing Boing Encyclopedia "Reagle provides ample data regarding the everyday practices and cultural norms of the community which collaborates to 3. Good Faith Collaboration produce Wikipedia. His rich research and nuanced appreciation of the complexities of cultural digital media research are 4. The Puzzle of Openness well presented.
    [Show full text]
  • Name of the Tool Citizendium Home Page Logo URL En.Citizendium
    Name of the Tool Citizendium Home Page Logo URL en.citizendium.org Subject Encyclopedias Accessibility Free Language English Publisher CZ: Media Assets Workgroup Brief History It is an English-language Wiki-based free encyclopedia project launched by Larry Sanger, who had previously co - founded Wikipedia in 2001. It had launched on 23 October 2006 (as pilot project) and on 25 March 2007 (publicly). Scope and Coverage It serves all over the world and provides the access and modifications of information related to the “parent topics” or main topics like Natural Sciences, Social Sciences, Humanities, Arts, Applied arts and sciences, Recreation. Under each main topic or parent topic, there is hyperlinked list of sub topics and other related topics. As for example, under the main topic “ Natural Sciences” there is a following list of “Subtopics” and “Other related topics”: Subtopics Physics Chemistry Biology Astronomy Earth science Mathematics Other related topics Natural philosophy Natural history Applied science Health science Geology Under the main topic “Humanities” there is a list of following topics: Subtopics Classics History Literature Philosophy Religion Theology Other related topics Art Applied arts Education Law Music Science Social science Scholarship Society Theatre The encyclopedia includes total number of 16,891 articles when last accessed. Kind of Information Every article under subtopics and other related topics is provided with “Talk”, “Related Articles”, “Biography”, “External Links”, “Citable Version”, “Video” related to that article. Brief description, history of a topic etc. are present in the articles. Coloured images on topics, charts, graphs etc. are available where applicable. Notes and references are also found after the articles.
    [Show full text]
  • Chaudron: Extending Dbpedia with Measurement Julien Subercaze
    Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue.
    [Show full text]
  • SI 410 Ethics and Information Technology
    Author(s): Paul Conway, PhD, 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/ We have reviewed this material in accordance with U.S. Copyright Law and have tried to maximize your ability to use, share, and adapt it. The citation key on the following slide provides information about how you may share and adapt this material. Copyright holders of content included in this material should contact [email protected] with any questions, corrections, or clarification regarding the use of content. For more information about how to cite these materials visit http://open.umich.edu/privacy-and-terms-use. Any medical information in this material is intended to inform and educate and is not a tool for self-diagnosis or a replacement for medical evaluation, advice, diagnosis or treatment by a healthcare professional. Please speak to your physician if you have questions about your medical condition. Viewer discretion is advised: Some medical content is graphic and may not be suitable for all viewers. Citation Key for more information see: http://open.umich.edu/wiki/CitationPolicy Use + Share + Adapt { Content the copyright holder, author, or law permits you to use, share and adapt. } Public Domain – Government: Works that are produced by the U.S. Government. (17 USC § 105) Public Domain – Expired: Works that are no longer protected due to an expired copyright term. Public Domain – Self Dedicated: Works that a copyright holder has dedicated to the public domain.
    [Show full text]
  • Wiki-Metasemantik: a Wikipedia-Derived Query Expansion Approach Based on Network Properties
    Wiki-MetaSemantik: A Wikipedia-derived Query Expansion Approach based on Network Properties D. Puspitaningrum1, G. Yulianti2, I.S.W.B. Prasetya3 1,2Department of Computer Science, The University of Bengkulu WR Supratman St., Kandang Limun, Bengkulu 38371, Indonesia 3Department of Information and Computing Sciences, Utrecht University PO Box 80.089, 3508 TB Utrecht, The Netherlands E-mails: [email protected], [email protected], [email protected] Pseudo-Relevance Feedback (PRF) query expansions suffer Abstract- This paper discusses the use of Wikipedia for building from several drawbacks such as query-topic drift [8][10] and semantic ontologies to do Query Expansion (QE) in order to inefficiency [21]. Al-Shboul and Myaeng [1] proposed a improve the search results of search engines. In this technique, technique to alleviate topic drift caused by words ambiguity selecting related Wikipedia concepts becomes important. We and synonymous uses of words by utilizing semantic propose the use of network properties (degree, closeness, and pageRank) to build an ontology graph of user query concept annotations in Wikipedia pages, and enrich queries with which is derived directly from Wikipedia structures. The context disambiguating phrases. Also, in order to avoid resulting expansion system is called Wiki-MetaSemantik. We expansion of mistranslated words, a query expansion method tested this system against other online thesauruses and ontology using link texts of a Wikipedia page has been proposed [9]. based QE in both individual and meta-search engines setups. Furthermore, since not all hyperlinks are helpful for QE task Despite that our system has to build a Wikipedia ontology graph though (e.g.
    [Show full text]
  • Télécharger Le Texte Intégral En Format
    ANNUAIRE FRANÇAIS DE RELATIONS INTERNATIONALES 2019 Volume XX PUBLICATION COURONNÉE PAR L’ACADÉMIE DES SCIENCES MORALES ET POLITIQUES (Prix de la Fondation Edouard Bonnefous, 2008) Université Panthéon-Assas Centre Thucydide AFRI_2019_v3_1124p.indd 3 24/04/2019 11:44 WIKIPÉDIA PAR VALÉRIE NICOLAS (*) Wikipédia (1) est un site Internet collaboratif qui se décrit lui-même comme « encyclopédie en libre accès, en lecture comme en écriture » (2). Multilingue, ce site est servi par un logiciel identique le Mediawiki. Il est adossé à une fondation à but non lucratif Wikimedia, qui assure son fonctionnement et gère d’autres projets frères. Le contenu de Wikipédia est disponible sous licence libre (3). Ainsi chacun peut le recopier, le modifier et l’utiliser. Le projet encyclopédique est alimenté par chaque utilisateur par une écriture collaborative, participative et bénévole. Crée en 2001 par deux ressortissants américains (4), Wikipédia (WP) est en 2014 le 5e site le plus fréquenté au monde (5). 500 millions de visiteurs le consultent chaque mois. Il offre aux lecteurs plus de 30 millions d’articles dans plus de 300 versions linguistiques. La version en anglais – matrice du projet –, compte plus de 5 millions de contributions. Les chiffres sont évocateurs du formidable recueil de connaissances que WP constitue. Internet est le moteur et le vecteur du succès du projet Wikipédia. Internet est un réseau de réseaux informatiques international organisé grâce à un protocole unique de communication (TCP/IP). Outil de communication, le World Wide Web (Web), un des services fournis par le réseau (6), a bouleversé les échanges entre les individus. Le réseau permet leur multiplication sans considération de frontières, ni de temps.
    [Show full text]
  • Navigating Dbpedia by Topic Tanguy Raynaud, Julien Subercaze, Delphine Boucard, Vincent Battu, Frederique Laforest
    Fouilla: Navigating DBpedia by Topic Tanguy Raynaud, Julien Subercaze, Delphine Boucard, Vincent Battu, Frederique Laforest To cite this version: Tanguy Raynaud, Julien Subercaze, Delphine Boucard, Vincent Battu, Frederique Laforest. Fouilla: Navigating DBpedia by Topic. CIKM 2018, Oct 2018, Turin, Italy. hal-01860672 HAL Id: hal-01860672 https://hal.archives-ouvertes.fr/hal-01860672 Submitted on 23 Aug 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Fouilla: Navigating DBpedia by Topic Tanguy Raynaud, Julien Subercaze, Delphine Boucard, Vincent Battu, Frédérique Laforest Univ Lyon, UJM Saint-Etienne, CNRS, Laboratoire Hubert Curien UMR 5516 Saint-Etienne, France [email protected] ABSTRACT only the triples that concern this topic. For example, a user is inter- Navigating large knowledge bases made of billions of triples is very ested in Italy through the prism of Sports while another through the challenging. In this demonstration, we showcase Fouilla, a topical prism of Word War II. For each of these topics, the relevant triples Knowledge Base browser that offers a seamless navigational expe- of the Italy entity differ. In such circumstances, faceted browsing rience of DBpedia. We propose an original approach that leverages offers no solution to retrieve the entities relative to a defined topic both structural and semantic contents of Wikipedia to enable a if the knowledge graph does not explicitly contain an adequate topic-oriented filter on DBpedia entities.
    [Show full text]
  • Wikipedia Editing History in Dbpedia Fabien Gandon, Raphael Boyer, Olivier Corby, Alexandre Monnin
    Wikipedia editing history in DBpedia Fabien Gandon, Raphael Boyer, Olivier Corby, Alexandre Monnin To cite this version: Fabien Gandon, Raphael Boyer, Olivier Corby, Alexandre Monnin. Wikipedia editing history in DB- pedia : extracting and publishing the encyclopedia editing activity as linked data. IEEE/WIC/ACM International Joint Conference on Web Intelligence (WI’ 16), Oct 2016, Omaha, United States. hal- 01359575 HAL Id: hal-01359575 https://hal.inria.fr/hal-01359575 Submitted on 2 Sep 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Wikipedia editing history in DBpedia extracting and publishing the encyclopedia editing activity as linked data Fabien Gandon, Raphael Boyer, Olivier Corby, Alexandre Monnin Université Côte d’Azur, Inria, CNRS, I3S, France Wimmics, Sophia Antipolis, France [email protected] Abstract— DBpedia is a huge dataset essentially extracted example, the French editing history dump represents 2TB of from the content and structure of Wikipedia. We present a new uncompressed data. This data extraction is performed by extraction producing a linked data representation of the editing stream in Node.js with a MongoDB instance. It takes 4 days to history of Wikipedia pages. This supports custom querying and extract 55 GB of RDF in turtle on 8 Intel(R) Xeon(R) CPU E5- combining with other data providing new indicators and insights.
    [Show full text]
  • Lessons from Citizendium
    Lessons from Citizendium Wikimania 2009, Buenos Aires, 28 August 2009 HaeB [[de:Benutzer:HaeB]], [[en:User:HaeB]] Please don't take photos during this talk. Citizendium Timeline ● September 2006: Citizendium announced. Sole founder: Larry Sanger, known as former editor-in-chief of Nupedia, chief organizer of Wikipedia (2001-2002), and later as Wikipedia critic ● October 2006: Started non-public pilot phase ● January 2007: “Big Unfork”: All unmodified copies of Wikipedia articles deleted ● March 2007: Public launch ● December 2007: Decision to use CC-BY-3.0, after debate about commercial reuse and compatibility with Wikipedia ● Mid-2009: Sanger largely inactive on Citizendium, focuses on WatchKnow ● August 2009: Larry Sanger announces he will step down as editor-in-chief soon (as committed to in 2006) Citizendium and Wikipedia: Similarities and differences ● Encyclopedia ● Strict real names ● Free license policy ● ● Open (anyone can Special role for contribute) experts: “editors” can issue content ● Created by amateurs decisions, binding to ● MediaWiki-based non-editors collaboration ● Governance: Social ● Non-profit contract, elements of a constitutional republic Wikipedian views of Citizendium ● Competitor for readers, contributions ● Ally, common goal of creating free encyclopedic content ● “Who?” ● In this talk: A long-time experiment testing several fundamental policy changes, in a framework which is still similar enough to that of Wikipedia to generate valuable evidence as to what their effect might be on WP Active editors: Waiting to explode ● Sanger (October 2007): ”At some point, possibly very soon, the Citizendium will grow explosively-- say, quadruple the number of its active contributors, or even grow by an order of magnitude ....“ © Aleksander Stos, CC-BY 3.0 Number of users that made at least one edit in each month Article creation rate: Still muddling Sanger (October 2007): “It's still possible that the project will, from here until eternity, muddle on creating 14 articles per day.
    [Show full text]
  • Reliability of Wikipedia 1 Reliability of Wikipedia
    Reliability of Wikipedia 1 Reliability of Wikipedia The reliability of Wikipedia (primarily of the English language version), compared to other encyclopedias and more specialized sources, is assessed in many ways, including statistically, through comparative review, analysis of the historical patterns, and strengths and weaknesses inherent in the editing process unique to Wikipedia. [1] Because Wikipedia is open to anonymous and collaborative editing, assessments of its Vandalism of a Wikipedia article reliability usually include examinations of how quickly false or misleading information is removed. An early study conducted by IBM researchers in 2003—two years following Wikipedia's establishment—found that "vandalism is usually repaired extremely quickly — so quickly that most users will never see its effects"[2] and concluded that Wikipedia had "surprisingly effective self-healing capabilities".[3] A 2007 peer-reviewed study stated that "42% of damage is repaired almost immediately... Nonetheless, there are still hundreds of millions of damaged views."[4] Several studies have been done to assess the reliability of Wikipedia. A notable early study in the journal Nature suggested that in 2005, Wikipedia scientific articles came close to the level of accuracy in Encyclopædia Britannica and had a similar rate of "serious errors".[5] This study was disputed by Encyclopædia Britannica.[6] By 2010 reviewers in medical and scientific fields such as toxicology, cancer research and drug information reviewing Wikipedia against professional and peer-reviewed sources found that Wikipedia's depth and coverage were of a very high standard, often comparable in coverage to physician databases and considerably better than well known reputable national media outlets.
    [Show full text]