Patterns of Content Transclusion in Wikipedia
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Improving Wikimedia Projects Content Through Collaborations Waray Wikipedia Experience, 2014 - 2017
Improving Wikimedia Projects Content through Collaborations Waray Wikipedia Experience, 2014 - 2017 Harvey Fiji • Michael Glen U. Ong Jojit F. Ballesteros • Bel B. Ballesteros ESEAP Conference 2018 • 5 – 6 May 2018 • Bali, Indonesia In 8 November 2013, typhoon Haiyan devastated the Eastern Visayas region of the Philippines when it made landfall in Guiuan, Eastern Samar and Tolosa, Leyte. The typhoon affected about 16 million individuals in the Philippines and Tacloban City in Leyte was one of [1] the worst affected areas. Philippines Eastern Visayas Eastern Visayas, specifically the provinces of Biliran, Leyte, Northern Samar, Samar and Eastern Samar, is home for the Waray speakers in the Philippines. [3] [2] Outline of the Presentation I. Background of Waray Wikipedia II. Collaborations made by Sinirangan Bisaya Wikimedia Community III. Problems encountered IV. Lessons learned V. Future plans References Photo Credits I. Background of Waray Wikipedia (https://war.wikipedia.org) Proposed on or about June 23, 2005 and created on or about September 24, 2005 Deployed lsjbot from Feb 2013 to Nov 2015 creating 1,143,071 Waray articles about flora and fauna As of 24 April 2018, it has a total of 1,262,945 articles created by lsjbot (90.5%) and by humans (9.5%) As of 31 March 2018, it has 401 views per hour Sinirangan Bisaya (Eastern Visayas) Wikimedia Community is the (offline) community that continuously improves Waray Wikipedia and related Wikimedia projects I. Background of Waray Wikipedia (https://war.wikipedia.org) II. Collaborations made by Sinirangan Bisaya Wikimedia Community A. Collaborations with private* and national government** Introductory letter organizations Series of meetings and communications B. -
Approved DITA 2.0 Proposals
DITA Technical Committee DITA 2.0 proposals DITA TC work product Page 1 of 189 Table of contents 1 Overview....................................................................................................................................................3 2 DITA 2.0: Stage two proposals.................................................................................................................. 3 2.1 Stage two: #08 <include> element.................................................................................................... 3 2.2 Stage two: #15 Relax specialization rules......................................................................................... 7 2.3 Stage two: #17 Make @outputclass universal...................................................................................9 2.4 Stage two: #18 Make audience, platform, product, otherprops into specializations........................12 2.5 Stage two: #27 Multimedia domain..................................................................................................16 2.6 Stage two: #29 Update bookmap.................................................................................................... 20 2.7 Stage two: #36 Remove deprecated elements and attributes.........................................................23 2.8 Stage two: #46: Remove @xtrf and @xtrc...................................................................................... 31 2.9 Stage 2: #73 Remove delayed conref domain.................................................................................36 -
XRI 2.0 FAQ 1 December 2005
XRI 2.0 FAQ 1 December 2005 This document is a comprehensive FAQ on the XRI 2.0 suite of specifications, with a particular emphasis on the XRI Syntax 2.0 Committee Specification which was submitted for consideration as an OASIS Standard on November 14, 2005. 1 General..................................................................................... 3 1.1 What does the acronym XRI stand for? ................................................................3 1.2 What is the relationship of XRI to URI and IRI? ....................................................3 1.3 Why was XRI needed?..........................................................................................3 1.4 Who is involved in the XRI specification effort? ....................................................4 1.5 What is the XRI 2.0 specification suite? ................................................................4 1.6 Are there any intellectual property restrictions on XRI? ........................................4 2 Uses of XRI .............................................................................. 5 2.1 What things do XRIs identify? ...............................................................................5 2.2 What are some example uses of XRI?..................................................................5 2.3 What are some applications that use XRI? ...........................................................5 3 Features of XRI Syntax ........................................................... 6 3.1 What were some of the design requirements -
Włodzimierz Lewoniewski Metoda Porównywania I Wzbogacania
Włodzimierz Lewoniewski Metoda porównywania i wzbogacania informacji w wielojęzycznych serwisach wiki na podstawie analizy ich jakości The method of comparing and enriching informa- on in mullingual wikis based on the analysis of their quality Praca doktorska Promotor: Prof. dr hab. Witold Abramowicz Promotor pomocniczy: dr Krzysztof Węcel Pracę przyjęto dnia: podpis Promotora Kierunek: Specjalność: Poznań 2018 Spis treści 1 Wstęp 1 1.1 Motywacja .................................... 1 1.2 Cel badawczy i teza pracy ............................. 6 1.3 Źródła informacji i metody badawcze ....................... 8 1.4 Struktura rozprawy ................................ 10 2 Jakość danych i informacji 12 2.1 Wprowadzenie .................................. 12 2.2 Jakość danych ................................... 13 2.3 Jakość informacji ................................. 15 2.4 Podsumowanie .................................. 19 3 Serwisy wiki oraz semantyczne bazy wiedzy 20 3.1 Wprowadzenie .................................. 20 3.2 Serwisy wiki .................................... 21 3.3 Wikipedia jako przykład serwisu wiki ....................... 24 3.4 Infoboksy ..................................... 25 3.5 DBpedia ...................................... 27 3.6 Podsumowanie .................................. 28 4 Metody określenia jakości artykułów Wikipedii 29 4.1 Wprowadzenie .................................. 29 4.2 Wymiary jakości serwisów wiki .......................... 30 4.3 Problemy jakości Wikipedii ............................ 31 4.4 -
List of Different Digital Practices 3
Categories of Digital Poetics Practices (from the Electronic Literature Collection) http://collection.eliterature.org/1/ (Electronic Literature Collection, Vol 1) http://collection.eliterature.org/2/ (Electronic Literature Collection, Vol 2) Ambient: Work that plays by itself, meant to evoke or engage intermittent attention, as a painting or scrolling feed would; in John Cayley’s words, “a dynamic linguistic wall- hanging.” Such work does not require or particularly invite a focused reading session. Kinetic (Animated): Kinetic work is composed with moving images and/or text but is rarely an actual animated cartoon. Transclusion, Mash-Up, or Appropriation: When the supply text for a piece is not composed by the authors, but rather collected or mined from online or print sources, it is appropriated. The result of appropriation may be a “mashup,” a website or other piece of digital media that uses content from more than one source in a new configuration. Audio: Any work with an audio component, including speech, music, or sound effects. CAVE: An immersive, shared virtual reality environment created using goggles and several pairs of projectors, each pair pointing to the wall of a small room. Chatterbot/Conversational Character: A chatterbot is a computer program designed to simulate a conversation with one or more human users, usually in text. Chatterbots sometimes seem to offer intelligent responses by matching keywords in input, using statistical methods, or building models of conversation topic and emotional state. Early systems such as Eliza and Parry demonstrated that simple programs could be effective in many ways. Chatterbots may also be referred to as talk bots, chat bots, simply “bots,” or chatterboxes. -
Web Boot Camp Week 1 (Custom for Liberty)
"Charting the Course ... ... to Your Success!" Web Boot Camp Week 1 (Custom for Liberty) Course Summary Description This class is designed for students that have experience with HTML5 and CSS3 and wish to learn about, JavaScript and jQuery and AngularJS. Topics Introduction to the Course Designing Responsive Web Applications Introduction to JavaScript Data Types and Assignment Operators Flow Control JavaScript Events JavaScript Objects JavaScript Arrays JavaScript Functions The JavaScript Window and Document Objects JavaScript and CSS Introduction to jQuery Introduction to jQuery UI jQuery and Ajax Introduction to AngularJS Additional AngularJS Topics The Angular $http Service AngularJS Filters AngularJS Directives AngularJS Forms Testing JavaScript Introduction to AngularJS 2 Prerequisites Prior knowledge of HTML5 and CSS3 is required. Duration Five days Due to the nature of this material, this document refers to numerous hardware and software products by their trade names. References to other companies and their products are for informational purposes only, and all trademarks are the properties of their respective companies. It is not the intent of ProTech Professional Technical Services, Inc. to use any of these names generically "Charting the Course ... ... to Your Success!" TDP Web Week 1: HTML5/CSS3/JavaScript Programming Course Outline I. Introduction to the Course O. Client-Side JavaScript Objects A. TDP Web Bootcamp Week 1, 2016 P. Embedding JavaScript in HTML B. Legal Information Q. Using the script Tag C. TDP Web Bootcamp Week 1, 2016 R. Using an External File D. Introductions S. Defining Functions E. Course Description T. Modifying Page Elements F. Course Objectives U. The Form Submission Event G. Course Logistics V. -
Autonomous Agents on the Web
Report from Dagstuhl Seminar 21072 Autonomous Agents on the Web Edited by Olivier Boissier1, Andrei Ciortea2, Andreas Harth3, and Alessandro Ricci4 1 Ecole des Mines – St. Etienne, FR, [email protected] 2 Universität St. Gallen, CH, [email protected] 3 Fraunhofer IIS – Nürnberg, DE, [email protected] 4 Università di Bologna, IT, [email protected] Abstract The World Wide Web has emerged as the middleware of choice for most distributed systems. Recent standardization efforts for the Web of Things and Linked Data are now turning hypermedia into a homogeneous information fabric that interconnects everything – devices, information resources, abstract concepts, etc. The latest standards allow clients not only to browse and query, but also to observe and act on this hypermedia fabric. Researchers and practitioners are already looking for means to build more sophisticated clients able to meet their design objectives through flexible autonomous use of this hypermedia fabric. Such autonomous agents have been studied to large extent in research on distributed artificial intelligence and, in particular, multi-agent systems. These recent developments thus motivate the need for a broader perspective that can only be achieved through a concerted effort of the research communities on the Web Architecture and the Web of Things, Semantic Web and Linked Data, and Autonomous Agents and Multi-Agent Systems. The Dagstuhl Seminar 21072 on “Autonomous Agents on the Web” brought together leading scholars and practitioners across these research areas in order to support the transfer of knowledge and results – and to discuss new opportunities for research on Web-based autonomous systems. This report documents the seminar’s program and outcomes. -
Freenet-Like Guids for Implementing Xanalogical Hypertext
Freenet-like GUIDs for Implementing Xanalogical Hypertext Tuomas J. Lukka Benja Fallenstein Hyperstructure Group Oberstufen-Kolleg Dept. of Mathematical Information Technology University of Bielefeld, PO. Box 100131 University of Jyvaskyl¨ a,¨ PO. Box 35 D-33501 Bielefeld FIN-40351 Jyvaskyl¨ a¨ Germany Finland [email protected] lukka@iki.fi ABSTRACT For example, an email quoting another email would be automati- We discuss the use of Freenet-like content hash GUIDs as a prim- cally and implicitly connected to the original via the transclusion. itive for implementing the Xanadu model in a peer-to-peer frame- Bidirectional, non-breaking external links (content linking) can be work. Our current prototype is able to display the implicit con- resolved through the same mechanism. Nelson[9] argues that con- nection (transclusion) between two different references to the same ventional software, unable to reflect such interconnectivity of doc- permanent ID. We discuss the next layers required in the implemen- uments, is unsuited to most human thinking and creative work. tation of the Xanadu model on a world-wide peer-to-peer network. In order to implement the Xanadu model, it must be possible to efficiently search for references to permanent IDs on a large scale. The original Xanadu design organized content IDs in a DNS-like Categories and Subject Descriptors hierarchical structure (tumblers), making content references arbi- H.5.4 [Information Interfaces and Presentation]: Hypertext/Hy- trary intervals (spans) in the hierarchy. Advanced tree-like data permedia—architectures; H.3.4 [Information Storage and Retrie- structures[6] were used to retrieve the content efficiently. -
Angularjs Workbook
AngularJS Workbook Exercises take you to mastery By Nicholas Johnson version 0.1.0 - beta Image credit: San Diego Air and Space Museum Archives Welcome to the Angular Workbook This little book accompanies the Angular course taught over 3 days. Each exercise builds on the last, so it's best to work through these exercises in order. We start out from the front end with templates, moving back to controllers. We'll tick off AJAX, and then get into building custom components, filters, services and directives. After covering directives in some depth we'll build a real app backed by an open API. By the end you should have a pretty decent idea of what is what and what goes where. Template Exercises The goal of this exercise is just to get some Angular running in a browser. Getting Angular We download Angular from angularjs.org Alternately we can use a CDN such as the Google Angular CDN. Activating the compiler Angular is driven by the template. This is different from other MVC frameworks where the template is driven by the app. In Angular we modify our app by modifying our template. The JavaScript we write simply supports this process. We use the ng-app attribute (directive) to tell Angular to begin compiling our DOM. We can attach this attribute to any DOM node, typically the html or body tags: 1 <body ng-app> 2 Hello! 3 </body> All the HTML5 within this directive is an Angular template and will be compiled as such. Linking from a CDN Delivering common libraries from a shared CDN can be a good idea. -
Language-Agnostic Relation Extraction from Abstracts in Wikis
information Article Language-Agnostic Relation Extraction from Abstracts in Wikis Nicolas Heist, Sven Hertling and Heiko Paulheim * ID Data and Web Science Group, University of Mannheim, Mannheim 68131, Germany; [email protected] (N.H.); [email protected] (S.H.) * Correspondence: [email protected] Received: 5 February 2018; Accepted: 28 March 2018; Published: 29 March 2018 Abstract: Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extraction from text, using the data in the knowledge graph as training data, i.e., using distant supervision. While most existing approaches use language-specific methods (usually for English), we present a language-agnostic approach that exploits background knowledge from the graph instead of language-specific techniques and builds machine learning models only from language-independent features. We demonstrate the extraction of relations from Wikipedia abstracts, using the twelve largest language editions of Wikipedia. From those, we can extract 1.6 M new relations in DBpedia at a level of precision of 95%, using a RandomForest classifier trained only on language-independent features. We furthermore investigate the similarity of models for different languages and show an exemplary geographical breakdown of the information extracted. In a second series of experiments, we show how the approach can be transferred to DBkWik, a knowledge graph extracted from thousands of Wikis. We discuss the challenges and first results of extracting relations from a larger set of Wikis, using a less formalized knowledge graph. Keywords: relation extraction; knowledge graphs; Wikipedia; DBpedia; DBkWik; Wiki farms 1. Introduction Large-scale knowledge graphs, such as DBpedia [1], Freebase [2], Wikidata [3], or YAGO [4], are usually built using heuristic extraction methods, by exploiting crowd-sourcing processes, or both [5]. -
Connecting Topincs Using Transclusion to Connect Proxy Spaces
Connecting Topincs Using transclusion to connect proxy spaces Robert Cerny Anton-Kubernat-Straße 15, A-2512 Oeynhausen, Austria [email protected] http://www.cerny-online.com Abstract. Topincs is a software system for agile and distributed knowledge management on top of the common web infrastructure and the Topic Maps Data Model. It segments knowledge into stores and offers users a way to establish a temporary connection between stores through transclusion and merging. This allows to easily copy topics and statements. Through this mechanism, later acts of integration are simplified, due to matching item identifiers. Furthermore, transclusion can be used to create permanent connections between stores. 1 Introduction A Topincs store [5, 6, 10] is a knowledge repository for an individual or a group to make statements about subjects. It uses the Topic Maps Data Model [7] for representing knowledge. Throughout this paper we will illustrate crucial aspects by the example of a small fictional software company, called Modestsoft. It employs around 10 people and has various departments, including development, testing, and sales & services. Modestsoft uses several shared Topincs stores to manage its organizational knowledge. Those stores deal with Modestsoft’s staff, products, issues, clients, and their configurations, to name only a few. Additionally, every employee has his own store. While the shared stores have a clear boundary around their domain, the personal stores may contain topics and statements of a wider variety and are considered a repository for note taking, and as such they function as a memory extender [4]. Figure 1 illustrates Modestsoft’s setup and the Topincs stores they use to manage their spontaneous knowledge recording demands. -
What Can Software Learn from Hypermedia?
What Can Software Learn From Hypermedia? Philip Tchernavskij Clemens Nylandsted Klokmose Michel Beaudouin-Lafon LRI, Univ. Paris-Sud, CNRS, Digital Deisgn & Information Studies, LRI, Univ. Paris-Sud, CNRS, Inria, Université Paris-Saclay Aarhus University Inria, Université Paris-Saclay Orsay 43017-6221, France Aarhus N 8200, Denmark Orsay 43017-6221, France [email protected] [email protected] [email protected] ABSTRACT procedures, or working materials in the physical world, these dimen- Most of our interactions with the digital world are mediated by sions of flexibility are limited or nonexistent in computer-mediated apps: desktop, web, or mobile applications. Apps impose artificial activity. These limitations are especially evident as software has limitations on collaboration among users, distribution across de- become ubiquitous in our social and professional lives. vices, and the changing procedures that constantly occur in real In general, apps model procedures, in the sense that they couple work. These limitations are partially due to the engineering prin- what users can do, e.g., changing color, with what they can do it ciples of encapsulation and program-data separation. By contrast, to, e.g., text. Some apps supports collaborating on a particular task, the field of hypermedia envisions collaboration, distribution and but limit the available tools and working materials. Some apps can flexible practices as fundamental features of software. We discuss be extended with new tools, but are highly specialized, e.g., for shareable dynamic media, an alternative model for software that illustration or programming. Apps can only be combined in limited unifies hypermedia and interactive systems, and Webstrates, an ways, e.g., it is possible to work on an image in two image treatment experimental implementation of that model.