11€€€€The Most Important Laboratory for Social Scientific and Computing
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Copyrighted Material
1 The Duality of Peer Production Infrastructure for the Digital Commons, Free Labor for Free‐Riding Firms Mathieu O’Neil, Sophie Toupin, and Christian Pentzold 1 Introduction There never was a “tragedy of the commons”: Garrett Hardin’s overgrazing farmers were victims of a tragedy of self‐management, as they failed to collectively regulate, as equals, their common pasture. When Elinor Ostrom was awarded the Nobel Prize in Economics in 2009, the immemorial notion that there are only two types of goods in the world – private and public, coordinated by markets or the state – was finally put to rest. In the most general terms, peer producers are people who create and manage common‐pool resources together. It sometimes seems as if “peer production” and “digital commons” can be used interchangeably. Digital commons are non‐rivalrous (they can be reproduced at little or no cost) and non‐excludable (no‐one can prevent others from using them, through prop- erty rights for example). So, practically speaking, proprietary objects could be produced by equal “peers,” however we argue that peer production has a normative dimension, so that what chiefly char- acterizes this mode of production is that “the output is orientated towards the further expansion of the commons; while the commons, recursively, is the chief resource in this mode of production” (Söderberg & O’Neil, 2014, p. 2). Though there are many historical antecedents, the term “peer pro- duction,” as an object of public and scientific interest, is historically situated in the early 2000s.1 The meanings associated with a term that is deeply connected to the Internet as it was 20 years ago are bound to change. -
1 Wikipedia: an Effective Anarchy Dariusz Jemielniak, Ph.D
Wikipedia: An Effective Anarchy Dariusz Jemielniak, Ph.D. Kozminski University [email protected] Paper presented at the Society for Applied Anthropology conference in Baltimore, MD (USA), 27-31 March, 2012 (work in progress) This paper is the first report from a virtual ethnographic study (Hine, 2000; Kozinets, 2010) of Wikipedia community conducted 2006-2012, by the use of participative methods, and relying on an narrative analysis of Wikipedia organization (Czarniawska, 2000; Boje, 2001; Jemielniak & Kostera, 2010). It serves as a general introduction to Wikipedia community, and is also a basis for a discussion of a book in progress, which is going to address the topic. Contrarily to a common misconception, Wikipedia was not the first “wiki” in the world. “Wiki” (originated from Hawaiian word for “quick” or “fast”, and named after “Wiki Wiki Shuttle” on Honolulu International Airport) is a website technology based on a philosophy of tracking changes added by the users, with a simplified markup language (allowing easy additions of, e.g. bold, italics, or tables, without the need to learn full HTML syntax), and was originally created and made public in 1995 by Ward Cunningam, as WikiWikiWeb. WikiWikiWeb was an attractive choice among enterprises and was used for communication, collaborative ideas development, documentation, intranet, knowledge management, etc. It grew steadily in popularity, when Jimmy “Jimbo” Wales, then the CEO of Bomis Inc., started up his encyclopedic project in 2000: Nupedia. Nupedia was meant to be an online encyclopedia, with free content, and written by experts. In an attempt to meet the standards set by professional encyclopedias, the creators of Nupedia based it on a peer-review process, and not a wiki-type software. -
Proceedings of the 11Th International Symposium on Open Collaboration
Proceedings of the 11th International Symposium on Open Collaboration August 19-21, 2015 San Francisco, California, U.S.A. General chair: Dirk Riehle, Friedrich-Alexander University Erlangen-Nürnberg Research track chairs: Kevin Crowston, Syracuse University Carlos Jensen, Oregon State University Carl Lagoze, University of Michigan Ann Majchrzak, University of Southern California Arvind Malhotra, University of North Carolina at Chapel Hill Claudia Müller-Birn, Freie Universität Berlin Gregorio Robles, Universidad Rey Juan Carlos Aaron Shaw, Northwestern University Sponsors: Wikimedia Foundation Google Inc. University of California Berkeley In-cooperation: ACM SIGSOFT ACM SIGWEB Fiscal sponsor: The John Ernest Foundation The Association for Computing Machinery 2 Penn Plaza, Suite 701 New York New York 10121-0701 ACM COPYRIGHT NOTICE. Copyright © 2014 by the Association for Computing Machin- ery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be hon- ored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from Publications Dept., ACM, Inc., fax +1 (212) 869-0481, or [email protected]. For other copying of articles that carry a code at the bottom of the first or last page, copying is permitted provided that the per-copy fee indicated in the code is paid through the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923, +1-978-750-8400, +1-978-750- 4470 (fax). -
A Topic-Aligned Multilingual Corpus of Wikipedia Articles for Studying Information Asymmetry in Low Resource Languages
Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020), pages 2373–2380 Marseille, 11–16 May 2020 c European Language Resources Association (ELRA), licensed under CC-BY-NC A Topic-Aligned Multilingual Corpus of Wikipedia Articles for Studying Information Asymmetry in Low Resource Languages Dwaipayan Roy, Sumit Bhatia, Prateek Jain GESIS - Cologne, IBM Research - Delhi, IIIT - Delhi [email protected], [email protected], [email protected] Abstract Wikipedia is the largest web-based open encyclopedia covering more than three hundred languages. However, different language editions of Wikipedia differ significantly in terms of their information coverage. We present a systematic comparison of information coverage in English Wikipedia (most exhaustive) and Wikipedias in eight other widely spoken languages (Arabic, German, Hindi, Korean, Portuguese, Russian, Spanish and Turkish). We analyze the content present in the respective Wikipedias in terms of the coverage of topics as well as the depth of coverage of topics included in these Wikipedias. Our analysis quantifies and provides useful insights about the information gap that exists between different language editions of Wikipedia and offers a roadmap for the Information Retrieval (IR) community to bridge this gap. Keywords: Wikipedia, Knowledge base, Information gap 1. Introduction other with respect to the coverage of topics as well as Wikipedia is the largest web-based encyclopedia covering the amount of information about overlapping topics. -
Understanding the Challenges of Collaborative Evidence-Based
Do you have a source for that? Understanding the Challenges of Collaborative Evidence-based Journalism Sheila O’Riordan Gaye Kiely Bill Emerson Joseph Feller Business Information Business Information Business Information Business Information Systems, University Systems, University Systems, University Systems, University College Cork, Ireland College Cork, Ireland College Cork, Ireland College Cork, Ireland [email protected] [email protected] [email protected] [email protected] ABSTRACT consumption has evolved [25]. There has been a move WikiTribune is a pilot news service, where evidence-based towards digital news with increasing user involvement as articles are co-created by professional journalists and a well as the use of social media platforms for accessing and community of volunteers using an open and collaborative discussing current affairs [14,39,43]. As such, the boundaries digital platform. The WikiTribune project is set within an are shifting between professional and amateur contributions. evolving and dynamic media landscape, operating under Traditional news organizations are adding interactive principles of openness and transparency. It combines a features as participatory journalism practices rise (see [8,42]) commercial for-profit business model with an open and the technologies that allow citizens to interact en masse collaborative mode of production with contributions from provide new avenues for engaging in democratic both paid professionals and unpaid volunteers. This deliberation [19]; the “process of reaching reasoned descriptive case study captures the first 12-months of agreement among free and equal citizens” [6:322]. WikiTribune’s operations to understand the challenges and opportunities within this hybrid model of production. We use With these changes, a number of challenges have arisen. -
Omnipedia: Bridging the Wikipedia Language
Omnipedia: Bridging the Wikipedia Language Gap Patti Bao*†, Brent Hecht†, Samuel Carton†, Mahmood Quaderi†, Michael Horn†§, Darren Gergle*† *Communication Studies, †Electrical Engineering & Computer Science, §Learning Sciences Northwestern University {patti,brent,sam.carton,quaderi}@u.northwestern.edu, {michael-horn,dgergle}@northwestern.edu ABSTRACT language edition contains its own cultural viewpoints on a We present Omnipedia, a system that allows Wikipedia large number of topics [7, 14, 15, 27]. On the other hand, readers to gain insight from up to 25 language editions of the language barrier serves to silo knowledge [2, 4, 33], Wikipedia simultaneously. Omnipedia highlights the slowing the transfer of less culturally imbued information similarities and differences that exist among Wikipedia between language editions and preventing Wikipedia’s 422 language editions, and makes salient information that is million monthly visitors [12] from accessing most of the unique to each language as well as that which is shared information on the site. more widely. We detail solutions to numerous front-end and algorithmic challenges inherent to providing users with In this paper, we present Omnipedia, a system that attempts a multilingual Wikipedia experience. These include to remedy this situation at a large scale. It reduces the silo visualizing content in a language-neutral way and aligning effect by providing users with structured access in their data in the face of diverse information organization native language to over 7.5 million concepts from up to 25 strategies. We present a study of Omnipedia that language editions of Wikipedia. At the same time, it characterizes how people interact with information using a highlights similarities and differences between each of the multilingual lens. -
From Faculty Enemy to Faculty Enabler
Wikipedia @ 20 • ::Wikipedia @ 20 9 The First Twenty Years of Teaching with Wikipedia: From Faculty Enemy to Faculty Enabler Robert E. Cummings Published on: Oct 15, 2020 Updated on: Nov 16, 2020 License: Creative Commons Attribution 4.0 International License (CC-BY 4.0) Wikipedia @ 20 • ::Wikipedia @ 20 9 The First Twenty Years of Teaching with Wikipedia: From Faculty Enemy to Faculty Enabler Wikipedia jumbles the faculty roles of teaching, researching, and service by challenging traditional notions of faculty expertise, but a more integrated approach for these roles is also possible. I have never been able to see Wikipedia without the lens of a faculty member. On the one hand, I subconsciously carry with me perpetual concerns about accuracy and reliability. I am often trying to prove to myself that Wikipedia is legitimate work: it is, after all, the world’s largest encyclopedia. Writing to Wikipedia follows a set of complicated rules. So if it is a serious project with rules and enforcers and the world relies on its information, it ought to be universally accepted. And on the other hand? I want Wikipedia to be a lark. I also want Wikipedia to be a place where I wander freely and learn lots and lots of information. A guilty pleasure, not unlike pulling a book off a library shelf and simply reading it. I suspect that these anxieties I experience might speak to inherent conflict in being a faculty member while engaging Wikipedia. Reading Wikipedia, contributing to Wikipedia, and certainly teaching with Wikipedia jumble and reconfigure the faculty identities of teacher and researcher because they recontextualize our relationships with expertise itself. -
Arxiv:1212.0018V2 [Cs.SI] 18 Dec 2012 Etn Ucin Ffe-Akteoois[ Economies Free-Market of Functions Setting E.[ Ref
Evidence for Non-Finite-State Computation in a Human Social System Simon DeDeo∗ Santa Fe Institute, Santa Fe, NM 87501, USA (Dated: December 19, 2012) We investigate the computational structure of a paradigmatic example of distributed so- cial interaction: that of the open-source Wikipedia community. The typical computational approach to modeling such a system is to rely on finite-state machines. However, we find strong evidence in this system for the emergence of processing powers over and above the finite-state. Thus, Wikipedia, understood as an information processing system, must have access to (at least one) effectively unbounded resource. The nature of this resource is such that one observes far longer runs of cooperative behavior than one would expect using finite- state models. We provide evidence that the emergence of this non-finite-state computation is driven by collective interaction effects. Social systems—particularly human social systems—process information. From the price- setting functions of free-market economies [1, 2] to resource management in traditional communi- ties [3], from deliberations in large-scale democracies [4, 5] to the formation of opinions and spread of reputational information in organizations [6] and social groups [7, 8], it has been recognized that such groups can perform functions analogous to (and often better than) engineered systems. Such functional roles are found in groups in addition to their contingent historical aspects and, when described mathematically, may be compared across cultures and times. The computational phenomena implicit in social systems are only now, with the advent of large, high-resolution data-sets, coming under systematic, empirical study at large scales. -
Book Review – Yochai Benkler, the Wealth of Networks: How Social Production Transforms Markets and Freedom
CLPE Research Paper 6/2007 Vol. 03 No. 02 (2007) James Brink Book Review – Yochai Benkler, The Wealth of Networks: How Social Production Transforms Markets and Freedom Keywords: networked economy, internet regulation, peer production JEL Classification: K10, K29, K42 Osgoode Hall Law School, York University 4700 Keele St., Toronto, ON, Canada M3J 1P3 Email: [email protected] Published in (2006) 7 German Law Journal 853, online: http://www.germanlawjournal.com/article.php?id=765. 967505 CLPE Research Paper 6/2007 Vol. 03 No. 02 (2007) James Brink BOOK REVIEW – YOCHAI BENKLER, THE WEALTH OF NETWORKS: HOW SOCIAL PRODUCTION TRANSFORMS MARKETS AND FREEDOM Benkler argues that the West is engaged in an escalating culture war between the industrial information economy—a one-way, capital- intensive, and professionally-produced model that has held sway for 150 years—and the networked information economy (NIE)—a many- to-many, low-capital, and cooperative model that has been emerging in the last 15 years. The NIE is built on the infrastructure of the internet and is characterized by characterized by (1) non-proprietary strategies, (2) rising non-market production, and (3) more effective, large-scale cooperative efforts; in other words, “peer production of information, knowledge, and culture.” These aspects challenge our economies and our polities, and hold significant promise for enhancing personal autonomy; however, Benkler, in writing what amounts to a manifesto for the internet, pays little attention to the way in which the NIE is vulnerable to technical capture in the same way the industrial information economy is vulnerable to capital capture, taking a “wait-and-see” approach to regulatory intervention. -
Customizing Debian Benjamin Mako Hill
Customizing Debian “Fork Yours with Debian GNU/Linux” Benjamin Mako Hill [email protected] http://mako.yukidoke.org Ubuntu Debian Project Software in the Public Interest Benjamin Mako Hill LCA - Debian MiniConf4 http://mako.yukidoke.org The World of Debian Customizers There are 115 distributions derived from Debian. AbulÉdu • Adamantix • AGNULA GNU/Linux Audio Distribution • ANTEMIUM Linux • Arabbix • ARMA aka Omoikane GNU/Linux • ASLinux • Auditor Security Linux • Augustux • B2D Linux • BEERnix • Biadix • BIG LINUX • Bioknoppix • BlackRhino • Bluewall GNU/Linux • Bonzai Linux • BrlSpeak • Càtix • CensorNet • Clusterix • ClusterKNOPPIX • Condorux • Damn Small Linux • Danix • DebXPde • eduKnoppix • ERPOSS • ESware • Euronode • FAMELIX • Feather Linux • Flonix • Vital Data Forensic or Rescue Kit (FoRK) • Freeduc-cd • GEOLivre Linux • Gibraltar Firewall • GNIX-Vivo • Gnoppix Linux • gnuLinEx • GNU/Linux Kinneret • GNUstep Live CD • grml • Guadalinex • Helix • Hiweed Linux • Impi Linux • Julex • K-DEMar • Kaella • Knoppix Linux Azur • Kalango Linux • KANOTIX • KlusTriX • knopILS • Knoppel • Knoppix • Knoppix 64 • Knoppix STD • KnoppiXMAME • KnoppMyth • Kurumin Linux • LAMPPIX • Libranet GNU/Linux • LIIS Linux • LinEspa • Linspire • Linux Live Game Project • Linux Loco • LinuxDefender Live! CD • Linuxin • LiVux • Local Area Security Linux (L.A.S.) • Luinux • Luit Linux • MAX: Madrid_Linux • Mediainlinux • MEPIS Linux • Metadistro-Pequelin • MIKO GNYO/Linux • Morphix • Munjoy Linux • Nature's Linux • NordisKnoppix • OGo Knoppix • Oralux • Overclockix -
Using Shape Expressions (Shex) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repositorio Institucional de la Universidad de Oviedo Using Shape Expressions (ShEx) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S. Stupp3, Jose Emilio Labra Gayo4, Daniel Mietchen5, Eric Prud’hommeaux6, and Andra Waagmeester7 1 Yale University, New Haven, CT, USA [email protected] 2 Johns Hopkins University, Baltimore, MD, USA [email protected] 3 The Scripps Research Institute, San Diego, CA, USA [email protected] 4 University of Oviedo, Oviedo, Spain [email protected] 5 Data Science Institute, University of Virginia, Charlottesville, VA, USA [email protected] 6 World Wide Web Consortium (W3C), MIT, Cambridge, MA, USA [email protected] 7 Micelio, Antwerpen, Belgium [email protected] Abstract. We discuss Shape Expressions (ShEx), a concise, formal, modeling and validation language for RDF structures. For instance, a Shape Expression could prescribe that subjects in a given RDF graph that fall into the shape “Paper” are expected to have a section called “Abstract”, and any ShEx implementation can confirm whether that is indeed the case for all such subjects within a given graph or subgraph. There are currently five actively maintained ShEx implementations. We discuss how we use the JavaScript, Scala and Python implementa- tions in RDF data validation workflows in distinct, applied contexts. We present examples of how ShEx can be used to model and validate data from two different sources, the domain-specific Fast Healthcare Interop- erability Resources (FHIR) and the domain-generic Wikidata knowledge base, which is the linked database built and maintained by the Wikimedia Foundation as a sister project to Wikipedia. -
An Analysis of Contributions to Wikipedia from Tor
Are anonymity-seekers just like everybody else? An analysis of contributions to Wikipedia from Tor Chau Tran Kaylea Champion Andrea Forte Department of Computer Science & Engineering Department of Communication College of Computing & Informatics New York University University of Washington Drexel University New York, USA Seatle, USA Philadelphia, USA [email protected] [email protected] [email protected] Benjamin Mako Hill Rachel Greenstadt Department of Communication Department of Computer Science & Engineering University of Washington New York University Seatle, USA New York, USA [email protected] [email protected] Abstract—User-generated content sites routinely block contri- butions from users of privacy-enhancing proxies like Tor because of a perception that proxies are a source of vandalism, spam, and abuse. Although these blocks might be effective, collateral damage in the form of unrealized valuable contributions from anonymity seekers is invisible. One of the largest and most important user-generated content sites, Wikipedia, has attempted to block contributions from Tor users since as early as 2005. We demonstrate that these blocks have been imperfect and that thousands of attempts to edit on Wikipedia through Tor have been successful. We draw upon several data sources and analytical techniques to measure and describe the history of Tor editing on Wikipedia over time and to compare contributions from Tor users to those from other groups of Wikipedia users. Fig. 1. Screenshot of the page a user is shown when they attempt to edit the Our analysis suggests that although Tor users who slip through Wikipedia article on “Privacy” while using Tor. Wikipedia’s ban contribute content that is more likely to be reverted and to revert others, their contributions are otherwise similar in quality to those from other unregistered participants and to the initial contributions of registered users.