Browsing Dbpedia Entities with Summaries
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Report on Requirements for Usage and Reuse Statistics for GLAM Content
Report on Requirements for Usage and Reuse Statistics for GLAM Content Author: Maarten Zeinstra, Kennisland 1/19 Report on Requirements for Usage and Reuse Statistics for GLAM Content 1. Table of Contents 1. Table of Contents 2 2. GLAMTools project 3 3. Executive Summary 4 4. Introduction 5 5. Problem definition 6 6. Current status 7 7. Research overview 8 7.1. Questionnaire 8 7.2. Qualitative research on the technical possibilities of GLAM stats 12 7.3. Research conclusions 13 8. Requirements for usage and reuse statistics for GLAM content 14 8.1. High-level requirements 14 8.2. High-level technical requirements 14 8.3. Medium-level technical requirements 15 8.4. Non-functional requirements 15 8.5. Non-requirements / out of scope 15 8.6. Data model 15 9. Next steps 19 2/19 2. GLAMTools project The “Report on requirements for usage and reuse statistics for GLAM content” is part of the Europeana GLAMTools project, a joint Wikimedia Chapters and Europeana project. The purpose of the project is to develop a scalable and maintainable system for mass uploading (open) content from galleries, libraries, archives and museums (henceforth GLAMs) to Wikimedia Commons and to create GLAM-specific requirements for usage statistics. The GLAMTools project is initiated and made possible by financial support from Wikimedia France, Wikimedia UK, Wikimedia Netherlands, and Wikimedia Switzerland. Maarten Zeinstra from Kennisland wrote the research behind this report and the report itself. Kennisland is a Dutch think tank active in the cultural sector, working especially with open content, dissemination of cultural heritage and copyright law and is also a partner in the project. -
Brion Vibber 2011-08-05 Haifa Part I: Editing So Pretty!
Editing 2.0 MediaWiki's upcoming visual editor and the future of templates Brion Vibber 2011-08-05 Haifa Part I: Editing So pretty! Wikipedia articles can include rich formatting, beyond simple links and images to complex templates to generate tables, pronunciation guides, and all sorts of details. So icky! But when you push that “edit” button, you often come face to face with a giant blob of markup that’s very hard to read. Here we can’t even see the first paragraph of the article until we scroll past several pages of infobox setup. Even fairly straightforward paragraphs start to blur together when you break out the source. The markup is often non-obvious; features that are clearly visible in the rendered view like links and images don’t stand out in the source view, and long inline citations and such can make it harder to find the actual body text you wanted to edit. RTL WTF? As many of you here today will be well aware, the way our markup displays in a raw text editor can also be really problematic for right-to-left scripts like Hebrew and Arabic. It’s very easy to get lost about whether you’ve opened or closed some tag, or whether your list item is starting at the beginning of a line. Without control over the individual pieces, we can’t give any hints to the text editor. RTL A-OK The same text rendered as structured HTML doesn’t have these problems; bullets stay on the right side of their text, and reference citations are distinct entities. -
Chaudron: Extending Dbpedia with Measurement Julien Subercaze
Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue. -
Genedb and Wikidata[Version 1; Peer Review: 1 Approved, 1 Approved with Reservations]
Wellcome Open Research 2019, 4:114 Last updated: 20 OCT 2020 SOFTWARE TOOL ARTICLE GeneDB and Wikidata [version 1; peer review: 1 approved, 1 approved with reservations] Magnus Manske , Ulrike Böhme , Christoph Püthe , Matt Berriman Parasites and Microbes, Wellcome Trust Sanger Institute, Cambridge, CB10 1SA, UK v1 First published: 01 Aug 2019, 4:114 Open Peer Review https://doi.org/10.12688/wellcomeopenres.15355.1 Latest published: 14 Oct 2019, 4:114 https://doi.org/10.12688/wellcomeopenres.15355.2 Reviewer Status Abstract Invited Reviewers Publishing authoritative genomic annotation data, keeping it up to date, linking it to related information, and allowing community 1 2 annotation is difficult and hard to support with limited resources. Here, we show how importing GeneDB annotation data into Wikidata version 2 allows for leveraging existing resources, integrating volunteer and (revision) report report scientific communities, and enriching the original information. 14 Oct 2019 Keywords GeneDB, Wikidata, MediaWiki, Wikibase, genome, reference, version 1 annotation, curation 01 Aug 2019 report report 1. Sebastian Burgstaller-Muehlbacher , This article is included in the Wellcome Sanger Medical University of Vienna, Vienna, Austria Institute gateway. 2. Andra Waagmeester , Micelio, Antwerp, Belgium Any reports and responses or comments on the article can be found at the end of the article. Corresponding author: Magnus Manske ([email protected]) Author roles: Manske M: Conceptualization, Methodology, Software, Writing – Original Draft Preparation; Böhme U: Data Curation, Validation, Writing – Review & Editing; Püthe C: Project Administration, Writing – Review & Editing; Berriman M: Resources, Writing – Review & Editing Competing interests: No competing interests were disclosed. Grant information: This work was supported by the Wellcome Trust through a Biomedical Resources Grant to MB [108443]. -
Easychair Preprint Digital Curation with Semantic Mediawiki And
EasyChair Preprint № 1853 Digital Curation with Semantic MediaWiki and Wikidata in numismatic research Bernhard Krabina EasyChair preprints are intended for rapid dissemination of research results and are integrated with the rest of EasyChair. November 5, 2019 Digital Curation with Semantic MediaWiki and Wikidata in numismatic research Bernhard Krabina1[0000-0002-6871-3037] 1 KDZ – Centre for Public Administration Research, Guglgasse 13, 1110 Vienna, Austria [email protected] Abstract. The FINA wiki is a Semantic MediaWiki (SMW) supporting numis- matic research showing how digital curation can be facilitated. It focuses on cu- ration of non-published material such as letters or manuscript sources scattered over archives and libraries across the world and previously neglected by scholars. A vast number of knowledge visualisations like maps, timelines, charts, word clouds or flow charts provide researchers with better support for handling large amount of content. With a simple mechanism, SMW can query Wikidata and users adding content can decide to use the suggested Wikidata IDs as point of reference. By leveraging Semantic Web standards, SMW instances can ensure long-term viability of valuable digital content. Keywords: Digital Curation, Semantic MediaWiki, Wikis, Wikidata, Semantic Web, Knowledge Visualisation, Collaboration, Metadata, Controlled Vocabu- laries, Numismatics. 1 Introduction 1.1 Digital curation Librarians and archivists collect valuable information, organize it, keep it in usable condition and provide access to users. Though the forms of content have changed throughout the years from handwritten items and published books now to digital con- tent, the purpose of librarianship and archiving has always been to collect and provide effective access to curated information. -
The History of Mediawiki and the Prospective for the Future
The history of MediaWiki and the prospective for the future Bern, 2017-02-04 Magnus Manske https://commons.wikimedia.org/wiki/File:MediaWiki_talk,_Bern,_2017-02-04,_Magnus_Manske.pdf Overview ● History of MediaWiki, with a focus on Wikimedia ● Usage, extensions and scripts ● Wikidata ● Of course, a few tools Pre-history UseModWiki by Ward Cunningham ● Single file Perl script ● All data stored in file system ● Does not scale well ● No “special pages” beyond Recent Changes ● Default: delete revisions older than two weeks, to save disk space… ● Used for Wikipedia since 2001-01 Phase 2 ● Mea maxima culpa ● Switch on 2002-01-25 ● PHP (for better or worse) ● Using MySQL to store data ● Based on UseModWiki syntax for backwards compatibility Brion Vibber Photo by JayWalsh CC-BY-SA 3.0 Tim Starling Photo by Lane Hartwell Nostalgia Wikipedia screenshot CC-BY-SA 3.0 Phase 3 / MediaWiki ● Massive refactoring/rewrite by Lee Daniel Crocker ● Proper OO, profiling functions ● Known as “Phase 3”, subsequently renamed “MediaWiki” (July 2003) Lee Daniel Crocker Picture by User:Cowtung, CC-BY-SA 3.0 New features since UseModWiki (for Wikipedia) ● Namespaces ● Special Pages with advanced functionality ● Skins ● Local (later: Commons) file upload ● Categories ● Templates ● Parser functions ● Table syntax ● Media viewer ● Visual editor ● Kartograph MediaWiki today ● Most widely used Wiki software ● Over 2,200 extensions ● Customizable via JavaScript gadgets and user scripts ● Comprehensive API Multiple caching mechanisms for large installations (e.g. Wikipedia) -
End-User Story Telling with a CIDOC CRM- Based Semantic Wiki
End-user storytelling with a CIDOC CRM-based semantic wiki Vincent Ribaud (reviewed by Patrick Le Bœuf) Département d'Informatique - U. B. O. Brest, France [email protected] Abstract: This paper presents the current state of an experiment intended to use the CIDOC CRM as a knowledge representation language. STEM freshers freely constitute groups of 2 to 4 members and choose a theme; groups have to model, structure, write and present a story within a web-hosted semantic wiki. The main part of the CIDOC CRM is used as an ontological core where students are hanging up classes and properties of the domain related to the story. The hypothesis is made that once the entry ticket has been paid, the CRM guides the end-user in a fairly natural manner for reading - and writing - the story. The intermediary assessment of the wikis allowed us to detect confusion between immaterial work and (physical) realisation of the work; and difficulty in having event-centred modelling. Final assessment results are satisfactory but may be improved. Some groups did not acquire modelling abilities - although this is a central issue in a semantic web course. Results also indicate that the scope of the course (semantic web) is somewhat too ambitious. This experience was performed in order to attract students to computer science studies but it did not produce the expected results. It did however succeed in arousing student interest, and it may contribute to the dissemination of ontologies and to making CIDOC CRM widespread. CIDOC 2010 ICOM General Conference Shanghai, China 2010-11-08 – 11-10 Introduction This paper presents the current state of an experiment intended to use the CIDOC CRM as a knowledge representation language inside a semantic wiki. -
Semantic Mediawiki in Operation: Experiences with Building a Semantic Portal
Semantic MediaWiki in Operation: Experiences with Building a Semantic Portal Daniel M. Herzig and Basil Ell Institute AIFB Karlsruhe Institute of Technology 76128 Karlsruhe, Germany fherzig, [email protected] http://www.aifb.kit.edu Abstract. Wikis allow users to collaboratively create and maintain con- tent. Semantic wikis, which provide the additional means to annotate the content semantically and thereby allow to structure it, experience an enormous increase in popularity, because structured data is more us- able and thus more valuable than unstructured data. As an illustration of leveraging the advantages of semantic wikis for semantic portals, we report on the experience with building the AIFB portal based on Seman- tic MediaWiki. We discuss the design, in particular how free, wiki-style semantic annotations and guided input along a predefined schema can be combined to create a flexible, extensible, and structured knowledge representation. How this structured data evolved over time and its flex- ibility regarding changes are subsequently discussed and illustrated by statistics based on actual operational data of the portal. Further, the features exploiting the structured data and the benefits they provide are presented. Since all benefits have its costs, we conducted a performance study of the Semantic MediaWiki and compare it to MediaWiki, the non- semantic base platform. Finally we show how existing caching techniques can be applied to increase the performance. 1 Introduction Web portals are entry points for information presentation and exchange over the Internet about a certain topic or organization, usually powered by a community. Leveraging semantic technologies for portals and exploiting semantic content has been proven useful in the past [1] and especially the aspect of providing seman- tic data got a lot of attention lately due to the Linked Open Data initiative. -
Customized Edit Interfaces for Wikis Via Semantic Annotations
Workshop on Adaptation and Personalization for Web 2.0, UMAP'09, June 22-26, 2009 Customized Edit Interfaces for Wikis via Semantic Annotations Angelo Di Iorio1, Silvia Duca1, Alberto Musetti1, Silvia Righini1, Davide Rossi1, Fabio Vitali1 1 Department of Computer Science, University of Bologna, Mura Anteo Zamboni 7, 40127 Bologna, Italy { [email protected], [email protected], [email protected], [email protected], [email protected], [email protected]} Abstract. Authoring support for semantic annotations represent the wiki way of the Semantic Web, ultimately leading to the wiki version of the Semantic Web's eternal dilemma: why should authors correctly annotate their content? The obvious solution is to make the ratio between the needed effort and the acquired advantages as small as possible. Two are, at least, the specificities that set wikis apart from other Web-accessible content in this respect: social aspects (wikis are often the expression of a community) and technical issues (wikis are edited "on-line"). Being related to a community, wikis are intrinsically associated to the model of knowledge of that community, making the relation between wiki content and ontologies the result of a natural process. Being edited on-line, wikis can benefit from a synergy of Web technologies that support all the information sharing process, from authoring to delivery. In this paper we present an approach to reduce the authoring effort by providing ontology-based tools to integrate models of knowledge with authoring-support technologies, using a functional approach to content fragment creation that plays nicely with the "wiki way" of managing information. -
Generating Semantic Media Wiki Content from Domain Ontologies
Generating Semantic Media Wiki Content from Domain Ontologies Dominik Filipiak1;2 and Agnieszka Ławrynowicz1 Institute of Computing Science, Poznan University of Technology, Poland Business Information Systems Institute Ltd., Poland Abstract. There is a growing interest in holistic ontology engineering approaches that involve multidisciplinary teams consisting of ontology engineers and domain experts. For the latter, who often lack ontology engineering expertise, tools such as Web forms, spreadsheet like templates or semantic wikis have been devel- oped that hide or decrease the complexity of logical axiomatisation in expressive ontology languages. This paper describes a prototype solution for an automatic OWL ontology conversion to articles in Semantic Media Wiki system. Our im- plemented prototype converts a branch of an ontology rooted at the user defined class into Wiki articles and categories. The final result is defined by a template expressed in the Wiki markup language. We describe tests on two domain ontolo- gies with different characteristics: DMOP and DMRO. The tests show that our solution can be used for fast bootstrapping of Semantic Media Wiki content from OWL files. 1 Introduction There is a growing interest in holistic ontology engineering approaches [1]. Those ap- proaches use various ontological as well as non-ontological resources (such as thesauri, lexica and relational DBs) [2, 3] . They also involve multidisciplinary teams consist- ing of ontology engineers as well as non-conversant in ontology construction domain experts. Whilst an active, direct involvement of the domain experts in the construction of quality domain ontologies within the teams appears beneficial, there are barriers to overcome for such involvement to be effective. -
Working-With-Mediawiki-Yaron-Koren.Pdf
Working with MediaWiki Yaron Koren 2 Working with MediaWiki by Yaron Koren Published by WikiWorks Press. Copyright ©2012 by Yaron Koren, except where otherwise noted. Chapter 17, “Semantic Forms”, includes significant content from the Semantic Forms homepage (https://www. mediawiki.org/wiki/Extension:Semantic_Forms), available under the Creative Commons BY-SA 3.0 license. All rights reserved. Library of Congress Control Number: 2012952489 ISBN: 978-0615720302 First edition, second printing: 2014 Ordering information for this book can be found at: http://workingwithmediawiki.com All printing of this book is handled by CreateSpace (https://createspace.com), a subsidiary of Amazon.com. Cover design by Grace Cheong (http://gracecheong.com). Contents 1 About MediaWiki 1 History of MediaWiki . 1 Community and support . 3 Available hosts . 4 2 Setting up MediaWiki 7 The MediaWiki environment . 7 Download . 7 Installing . 8 Setting the logo . 8 Changing the URL structure . 9 Updating MediaWiki . 9 3 Editing in MediaWiki 11 Tabs........................................................... 11 Creating and editing pages . 12 Page history . 14 Page diffs . 15 Undoing . 16 Blocking and rollbacks . 17 Deleting revisions . 17 Moving pages . 18 Deleting pages . 19 Edit conflicts . 20 4 MediaWiki syntax 21 Wikitext . 21 Interwiki links . 26 Including HTML . 26 Templates . 27 3 4 Contents Parser and tag functions . 30 Variables . 33 Behavior switches . 33 5 Content organization 35 Categories . 35 Namespaces . 38 Redirects . 41 Subpages and super-pages . 42 Special pages . 43 6 Communication 45 Talk pages . 45 LiquidThreads . 47 Echo & Flow . 48 Handling reader comments . 48 Chat........................................................... 49 Emailing users . 49 7 Images and files 51 Uploading . 51 Displaying images . 55 Image galleries . -
Wembedder: Wikidata Entity Embedding Web Service
Wembedder: Wikidata entity embedding web service Finn Årup Nielsen Cognitive Systems, DTU Compute Technical University of Denmark Kongens Lyngby, Denmark [email protected] ABSTRACT as the SPARQL-based Wikidata Query Service (WDQS). General I present a web service for querying an embedding of entities in functions for finding related items or generating fixed-length fea- the Wikidata knowledge graph. The embedding is trained on the tures for machine learning are to my knowledge not available. Wikidata dump using Gensim’s Word2Vec implementation and a simple graph walk. A REST API is implemented. Together with There is some research that combines machine learning and Wiki- the Wikidata API the web service exposes a multilingual resource data [9, 14], e.g., Mousselly-Sergieh and Gurevych have presented for over 600’000 Wikidata items and properties. a method for aligning Wikidata items with FrameNet based on Wikidata labels and aliases [9]. Keywords Wikidata, embedding, RDF, web service. Property suggestion is running live in the Wikidata editing inter- face, where it helps editors recall appropriate properties for items during manual editing, and as such a form of recommender sys- 1. INTRODUCTION tem. Researchers have investigated various methods for this pro- The Word2Vec model [7] spawned an interest in dense word repre- cess [17]. sentation in a low-dimensional space, and there are now a consider- 1 able number of “2vec” models beyond the word level. One recent Scholia at https://tools.wmflabs.org/scholia/ is avenue of research in the “2vec” domain uses knowledge graphs our web service that presents scholarly profiles based on data in [13].