Wikibrain: Democratizing Computation on Wikipedia

Total Page:16

File Type:pdf, Size:1020Kb

Wikibrain: Democratizing Computation on Wikipedia WikiBrain: Democratizing computation on Wikipedia Shilad Sen Toby Jia-Jun Li Macalester College University of Minnesota St. Paul, Minnesota Minneapolis, Minnesota [email protected] [email protected] ∗ WikiBrain Team Brent Hecht Matt Lesicko, Ari Weiland, Rebecca Gold, University of Minnesota Yulun Li, Benjamin Hillmann Minneapolis, Minnesota Macalester College [email protected] St. Paul, Minnesota ABSTRACT has raced towards this goal, growing in articles, revisions, 1 Wikipedia is known for serving humans' informational needs. and gigabytes by factor of 20. The planet has taken notice. Over the past decade, the encyclopedic knowledge encoded Humans turn to Wikipedia's 31 million articles more than in Wikipedia has also powerfully served computer systems. any other online reference. The online encylopedia's 287 languages attract a global audience that make it the sixth Leading algorithms in artificial intelligence, natural language 2 processing, data mining, geographic information science, and most visited site on the web. many other fields analyze the text and structure of articles However, outside the public eye, computers have also come to build computational models of the world. to rely on Wikipedia. Algorithms across many fields extract Many software packages extract knowledge from Wiki- meaning from the encyclopedia by analyzing articles' text, pedia. However, existing tools either (1) provide Wikipedia links, titles, and category structure. Search engines such as Google and Bing respond to user queries by extracting struc- data, but not well-known Wikipedia-based algorithms or (2) 3 narrowly focus on one such algorithm. tured knowledge from Wikipedia . Leading semantic relat- This paper presents the WikiBrain software framework, edness (SR) algorithms estimate the strength of association an extensible Java-based platform that democratizes access between words by mining Wikipedia's text, link structure to a range of Wikipedia-based algorithms and technologies. [32], and concept structure [5]. Geospatial technologies ex- WikiBrain provides simple access to the diverse Wikipedia tract geographic information from Wikipedia to understand data needed for semantic algorithms and technologies, rang- the world around us in entirely new ways (e.g. [22, 20]). ing from page views to Wikidata. In a few lines of code, Though powerful, these Wikipedia-centric algorithms pre- a developer can use WikiBrain to access Wikipedia data sent serious engineering obstacles. Downloading, parsing, and state-of-the-art algorithms. WikiBrain also enables re- and saving hundreds of gigabytes of database content re- searchers to extend Wikipedia-based algorithms and evalu- quires careful software engineering. Extracting structure ate their extensions. WikiBrain promotes a new vision of the from Wikipedia requires computers to reconcile the messy, Wikipedia software ecosystem: every researcher and devel- human process of article creation and revision. Finally, de- oper should have access to state-of-the-art Wikipedia-based termining exactly what information an application needs, technologies. whether Wikipedia provides that information, and how that information can be efficiently delivered taxes software de- signers. 1. INTRODUCTION These engineering challenges stand as a barrier to develop- In 2004, Jimmy Wales articulated his audacious vision for ers of intelligent applications and researchers of Wikipedia- Wikipedia: \Imagine a world in which every single person on based algorithms. While large software companies such as the planet is given free access to the sum of all human knowl- Google and Microsoft extensively use Wikipedia-mined knowl- edge [31]." In the decade since Wales' declaration, Wikipedia edge1, smaller companies with limited engineering resources have a much harder time leveraging it. Researchers mining ∗[email protected], [email protected], rebecca- [email protected], [email protected], bhillh- Wikipedia face similar engineering challenges. Instead of [email protected] focusing on innovative contributions, they frequently spend Permission to make digital or hard copies of all or part of this work for per- hours reinventing the Wikipedia data processing wheel or sonal or classroom use is granted without fee provided that copies are not wrangling a complex Wikipedia analysis framework to meet made or distributed for profit or commercial advantage and that copies bear their needs. This situation has created a fragmented Wiki- this notice and the full citation on the first page. Copyrights for components pedia software ecosystem in which it is extremely difficult to of this work owned by others than the author(s) must be honored. Abstract- reproduce results and extend Wikipedia-based technologies. ing with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a 1http://en.wikipedia.org/wiki/Wikipedia:Size_of_Wikipedia fee. Request permissions from [email protected]. 2http://www.alexa.com/topsites OpenSym ’14, August 27 - 29 2014, Berlin, Germany. Copyright is held 3 by the owner/author(s). Publication rights licensed to ACM. ACM 978-1- http://googleblog.blogspot.com/2012/05/ 4503-3016-9/14/08 $15.00. http://dx.doi.org/10.1145/2641580.2641615 introducing-knowledge-graph-things-not.html This paper introduces WikiBrain,4 a software library that software packages and datasets for processing Wikipedia data. democratizes access to state-of-the-art Wikipedia-based al- We divide this discussion into two parts: the first is targeted gorithms and techniques. In a few minutes, a developer at software and datasets related to basic Wikipedia struc- can add the WikiBrain Java library and run a single pro- tures and the second is focused on software that is specific gram that downloads, parses, and saves Wikipedia data on to a single Wikipedia-based technology. As the WikiBrain commodity hardware. WikiBrain provides a robust, fast team has the strict policy of not reinventing the wheel when- API to access and query article text and structured meta- ever possible, WikiBrain leverages some of this existing work data such as links, categories, pageview analytics, and Wiki- in select portions of its functionality. Where this is the case, data.[49]. WikiBrain also makes three state-of-the-art re- we describe how WikiBrain was aided by a particular soft- search advances available to every developer and researcher ware package or dataset. as API\primitives:" a multilingual concept network (Section 4), semantic relatedness algorithms (Section 5), and geospa- 2.1 Access to basic data structures tial data integration (Section 6). In doing so, WikiBrain • Wikipedia API: The Wikimedia Foundation provides ac- empowers developers to create rich intelligent applications cess to an impressive array of data through a REST API. and enables researchers to focus their efforts on creating and Accessing this API is substantially slower than accessing evaluating robust, innovative, reproducible software. well-engineered locally stored versions of the data. How- This paper describes WikiBrain, compares it to related ever, the Wikipedia API has one important advantage over software libraries, and provides examples of its use. Sec- all other means of accessing Wikipedia data: it is always tion two begins by reviewing research algorithms that mine 100% up to date. In order to support real-time access to Wikipedia and software APIs that support those algorithms. Wikipedia data, WikiBrain has wrappers for the Wikipedia Section three discusses the overall design and usage of Wik- API that allow developers to access the live Wikipedia API iBrain. We then delve into details on the three 'primitives' just as easily as they can locally stored Wikipedia data (the (Sections four, five, and six). Finally, section seven consists switch is a single line of code per data structure). They can of a case study that uses WikiBrain to quantitatively evalu- even easily integrate data from the two sources. ate Tobler's First Law of Geography { \everything is related to everything else, but near things are more related than • Java Wikipedia Library (JWPL): JWPL provides ac- distant things" [48] { replicating the results of a study by cess to basic data structures like links and categories. Wik- Hecht and Moxley [20] in a few lines of code. iBrain uses some libraries from JWPL to faciliate the pars- ing of wiki markup. 2. RELATED WORK • DBpedia: DBpedia [1] is a highly influential project that extracts structured information from Wikipedia pages. Its Wikipedia has attracted the attention of researchers in a widely-used datasets include basic structures like links and wide range of fields. Broadly speaking, Wikipedia research categories, but also an inferred general-purpose ontology, can be split into two categories: work that studies Wikipedia mappings to other data sources (e.g. Freebase, YAGO), and work that uses Wikipedia as a source of world knowl- and other enriched corpora. WikiBrain connects to DBpe- edge. While the former body of literature is large (e.g. [23, dia's dataset of geographic coordinates as described below. 41, 25, 34]), the latter is likely at least an order of magnitude larger. • bliki: Bliki is a \parser library for converting Wikipedia Researchers have leveraged the unprecedented structured wikitext notation to HTML". Bliki allows for structured repository of world knowledge in Wikipedia for an enormous access to a variety of wiki markup data structures (e.g. variety of tasks. For instance,
Recommended publications
  • Proceedings of the 11Th International Symposium on Open Collaboration
    Proceedings of the 11th International Symposium on Open Collaboration August 19-21, 2015 San Francisco, California, U.S.A. General chair: Dirk Riehle, Friedrich-Alexander University Erlangen-Nürnberg Research track chairs: Kevin Crowston, Syracuse University Carlos Jensen, Oregon State University Carl Lagoze, University of Michigan Ann Majchrzak, University of Southern California Arvind Malhotra, University of North Carolina at Chapel Hill Claudia Müller-Birn, Freie Universität Berlin Gregorio Robles, Universidad Rey Juan Carlos Aaron Shaw, Northwestern University Sponsors: Wikimedia Foundation Google Inc. University of California Berkeley In-cooperation: ACM SIGSOFT ACM SIGWEB Fiscal sponsor: The John Ernest Foundation The Association for Computing Machinery 2 Penn Plaza, Suite 701 New York New York 10121-0701 ACM COPYRIGHT NOTICE. Copyright © 2014 by the Association for Computing Machin- ery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be hon- ored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from Publications Dept., ACM, Inc., fax +1 (212) 869-0481, or [email protected]. For other copying of articles that carry a code at the bottom of the first or last page, copying is permitted provided that the per-copy fee indicated in the code is paid through the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923, +1-978-750-8400, +1-978-750- 4470 (fax).
    [Show full text]
  • Causal Inference in the Presence of Interference in Sponsored Search Advertising
    Causal Inference in the Presence of Interference in Sponsored Search Advertising Razieh Nabi∗ Joel Pfeiffer Murat Ali Bayir Johns Hopkins University Microsoft Bing Ads Microsoft Bing Ads [email protected] [email protected] [email protected] Denis Charles Emre Kıcıman Microsoft Bing Ads Microsoft Research [email protected] [email protected] Abstract In classical causal inference, inferring cause-effect relations from data relies on the assumption that units are independent and identically distributed. This assumption is violated in settings where units are related through a network of dependencies. An example of such a setting is ad placement in sponsored search advertising, where the clickability of a particular ad is potentially influenced by where it is placed and where other ads are placed on the search result page. In such scenarios, confounding arises due to not only the individual ad-level covariates but also the placements and covariates of other ads in the system. In this paper, we leverage the language of causal inference in the presence of interference to model interactions among the ads. Quantification of such interactions allows us to better understand the click behavior of users, which in turn impacts the revenue of the host search engine and enhances user satisfaction. We illustrate the utility of our formalization through experiments carried out on the ad placement system of the Bing search engine. 1 Introduction In recent years, advertisers have increasingly shifted their ad expenditures online. One of the most effective platforms for online advertising is search engine result pages. Given a user query, the search engine allocates a few ad slots (e.g., above or below its organic search results) and runs an auction among advertisers who are bidding and competing for these slots.
    [Show full text]
  • Wikimedia Conferentie Nederland 2012 Conferentieboek
    http://www.wikimediaconferentie.nl WCN 2012 Conferentieboek CC-BY-SA 9:30–9:40 Opening 9:45–10:30 Lydia Pintscher: Introduction to Wikidata — the next big thing for Wikipedia and the world Wikipedia in het onderwijs Technische kant van wiki’s Wiki-gemeenschappen 10:45–11:30 Jos Punie Ralf Lämmel Sarah Morassi & Ziko van Dijk Het gebruik van Wikimedia Commons en Wikibooks in Community and ontology support for the Wikimedia Nederland, wat is dat precies? interactieve onderwijsvormen voor het secundaire onderwijs 101wiki 11:45–12:30 Tim Ruijters Sandra Fauconnier Een passie voor leren, de Nederlandse wikiversiteit Projecten van Wikimedia Nederland in 2012 en verder Bliksemsessie …+discussie …+vragensessie Lunch 13:15–14:15 Jimmy Wales 14:30–14:50 Wim Muskee Amit Bronner Lotte Belice Baltussen Wikipedia in Edurep Bridging the Gap of Multilingual Diversity Open Cultuur Data: Een bottom-up initiatief vanuit de erfgoedsector 14:55–15:15 Teun Lucassen Gerard Kuys Finne Boonen Scholieren op Wikipedia Onderwerpen vinden met DBpedia Blijf je of ga je weg? 15:30–15:50 Laura van Broekhoven & Jan Auke Brink Jeroen De Dauw Jan-Bart de Vreede 15:55–16:15 Wetenschappelijke stagiairs vertalen onderzoek naar Structured Data in MediaWiki Wikiwijs in vergelijking tot Wikiversity en Wikibooks Wikipedia–lemma 16:20–17:15 Prijsuitreiking van Wiki Loves Monuments Nederland 17:20–17:30 Afsluiting 17:30–18:30 Borrel Inhoudsopgave Organisatie 2 Voorwoord 3 09:45{10:30: Lydia Pintscher 4 13:15{14:15: Jimmy Wales 4 Wikipedia in het onderwijs 5 11:00{11:45: Jos Punie
    [Show full text]
  • Understanding the Challenges of Collaborative Evidence-Based
    Do you have a source for that? Understanding the Challenges of Collaborative Evidence-based Journalism Sheila O’Riordan Gaye Kiely Bill Emerson Joseph Feller Business Information Business Information Business Information Business Information Systems, University Systems, University Systems, University Systems, University College Cork, Ireland College Cork, Ireland College Cork, Ireland College Cork, Ireland [email protected] [email protected] [email protected] [email protected] ABSTRACT consumption has evolved [25]. There has been a move WikiTribune is a pilot news service, where evidence-based towards digital news with increasing user involvement as articles are co-created by professional journalists and a well as the use of social media platforms for accessing and community of volunteers using an open and collaborative discussing current affairs [14,39,43]. As such, the boundaries digital platform. The WikiTribune project is set within an are shifting between professional and amateur contributions. evolving and dynamic media landscape, operating under Traditional news organizations are adding interactive principles of openness and transparency. It combines a features as participatory journalism practices rise (see [8,42]) commercial for-profit business model with an open and the technologies that allow citizens to interact en masse collaborative mode of production with contributions from provide new avenues for engaging in democratic both paid professionals and unpaid volunteers. This deliberation [19]; the “process of reaching reasoned descriptive case study captures the first 12-months of agreement among free and equal citizens” [6:322]. WikiTribune’s operations to understand the challenges and opportunities within this hybrid model of production. We use With these changes, a number of challenges have arisen.
    [Show full text]
  • Towards a Korean Dbpedia and an Approach for Complementing the Korean Wikipedia Based on Dbpedia
    Towards a Korean DBpedia and an Approach for Complementing the Korean Wikipedia based on DBpedia Eun-kyung Kim1, Matthias Weidl2, Key-Sun Choi1, S¨orenAuer2 1 Semantic Web Research Center, CS Department, KAIST, Korea, 305-701 2 Universit¨at Leipzig, Department of Computer Science, Johannisgasse 26, D-04103 Leipzig, Germany [email protected], [email protected] [email protected], [email protected] Abstract. In the first part of this paper we report about experiences when applying the DBpedia extraction framework to the Korean Wikipedia. We improved the extraction of non-Latin characters and extended the framework with pluggable internationalization components in order to fa- cilitate the extraction of localized information. With these improvements we almost doubled the amount of extracted triples. We also will present the results of the extraction for Korean. In the second part, we present a conceptual study aimed at understanding the impact of international resource synchronization in DBpedia. In the absence of any informa- tion synchronization, each country would construct its own datasets and manage it from its users. Moreover the cooperation across the various countries is adversely affected. Keywords: Synchronization, Wikipedia, DBpedia, Multi-lingual 1 Introduction Wikipedia is the largest encyclopedia of mankind and is written collaboratively by people all around the world. Everybody can access this knowledge as well as add and edit articles. Right now Wikipedia is available in 260 languages and the quality of the articles reached a high level [1]. However, Wikipedia only offers full-text search for this textual information. For that reason, different projects have been started to convert this information into structured knowledge, which can be used by Semantic Web technologies to ask sophisticated queries against Wikipedia.
    [Show full text]
  • Understanding Attention Metrics
    Understanding Attention Metrics sponsored by: 01 introduction For years and years we’ve been told that if the click isn’t on its deathbed, it should be shivved to death immediately. While early display ads in the mid-90s could boast CTRs above 40%, the likely number you’ll see attached to a campaign these days is below 0.1%. But for some reason, even in 2015, click-through rate remains as vital to digital ad measurement although its importance is widely ridiculed. Reliance on the click means delivering unfathomable amounts of ad impressions to users, which has hyper-inflated the value of the pageview. This has led to a broken advertising system that rewards quantity over quality. Instead of building audiences, digital publishers are chasing traffic, trying to lure users to sites via “clickbait” headlines where the user is assaulted with an array of intrusive ad units. The end effect is overwhelming, ineffective ads adjacent to increasingly shoddy content. However, the rise of digital video and viewability have injected linear TV measurement capabilities into the digital space. By including time in its calculation, viewability has opened a conversation into the value of user attention. It has introduced metrics with the potential to assist premium publishers in building attractive audiences; better quantify exposure for advertisers; and ultimately financially reward publishers with more engaged audiences. This is a new and fast developing area, but one where a little education can go a long way. In addition to describing current attention metrics, this playbook will dive into what you should look for in an advanced metrics partner and how to use these metrics in selling inventory and optimizing campaign performance.
    [Show full text]
  • Using Shape Expressions (Shex) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repositorio Institucional de la Universidad de Oviedo Using Shape Expressions (ShEx) to Share RDF Data Models and to Guide Curation with Rigorous Validation B Katherine Thornton1( ), Harold Solbrig2, Gregory S. Stupp3, Jose Emilio Labra Gayo4, Daniel Mietchen5, Eric Prud’hommeaux6, and Andra Waagmeester7 1 Yale University, New Haven, CT, USA [email protected] 2 Johns Hopkins University, Baltimore, MD, USA [email protected] 3 The Scripps Research Institute, San Diego, CA, USA [email protected] 4 University of Oviedo, Oviedo, Spain [email protected] 5 Data Science Institute, University of Virginia, Charlottesville, VA, USA [email protected] 6 World Wide Web Consortium (W3C), MIT, Cambridge, MA, USA [email protected] 7 Micelio, Antwerpen, Belgium [email protected] Abstract. We discuss Shape Expressions (ShEx), a concise, formal, modeling and validation language for RDF structures. For instance, a Shape Expression could prescribe that subjects in a given RDF graph that fall into the shape “Paper” are expected to have a section called “Abstract”, and any ShEx implementation can confirm whether that is indeed the case for all such subjects within a given graph or subgraph. There are currently five actively maintained ShEx implementations. We discuss how we use the JavaScript, Scala and Python implementa- tions in RDF data validation workflows in distinct, applied contexts. We present examples of how ShEx can be used to model and validate data from two different sources, the domain-specific Fast Healthcare Interop- erability Resources (FHIR) and the domain-generic Wikidata knowledge base, which is the linked database built and maintained by the Wikimedia Foundation as a sister project to Wikipedia.
    [Show full text]
  • Chaudron: Extending Dbpedia with Measurement Julien Subercaze
    Chaudron: Extending DBpedia with measurement Julien Subercaze To cite this version: Julien Subercaze. Chaudron: Extending DBpedia with measurement. 14th European Semantic Web Conference, Eva Blomqvist, Diana Maynard, Aldo Gangemi, May 2017, Portoroz, Slovenia. hal- 01477214 HAL Id: hal-01477214 https://hal.archives-ouvertes.fr/hal-01477214 Submitted on 27 Feb 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Chaudron: Extending DBpedia with measurement Julien Subercaze1 Univ Lyon, UJM-Saint-Etienne, CNRS Laboratoire Hubert Curien UMR 5516, F-42023, SAINT-ETIENNE, France [email protected] Abstract. Wikipedia is the largest collaborative encyclopedia and is used as the source for DBpedia, a central dataset of the LOD cloud. Wikipedia contains numerous numerical measures on the entities it describes, as per the general character of the data it encompasses. The DBpedia In- formation Extraction Framework transforms semi-structured data from Wikipedia into structured RDF. However this extraction framework of- fers a limited support to handle measurement in Wikipedia. In this paper, we describe the automated process that enables the creation of the Chaudron dataset. We propose an alternative extraction to the tra- ditional mapping creation from Wikipedia dump, by also using the ren- dered HTML to avoid the template transclusion issue.
    [Show full text]
  • Classifying Wikipedia Article Quality with Revision History Networks
    Classifying Wikipedia Article Quality With Revision History Networks Narun Raman∗ Nathaniel Sauerberg∗ Carleton College Carleton College [email protected] [email protected] Jonah Fisher Sneha Narayan Carleton College Carleton College [email protected] [email protected] ABSTRACT long been interested in maintaining and investigating the quality We present a novel model for classifying the quality of Wikipedia of its content [4][6][12]. articles based on structural properties of a network representation Editors and WikiProjects typically rely on assessments of article of the article’s revision history. We create revision history networks quality to focus volunteer attention on improving lower quality (an adaptation of Keegan et. al’s article trajectory networks [7]), articles. This has led to multiple efforts to create classifiers that can where nodes correspond to individual editors of an article, and edges predict the quality of a given article [3][4][18]. These classifiers can join the authors of consecutive revisions. Using descriptive statistics assist in providing assessments of article quality at scale, and help generated from these networks, along with general properties like further our understanding of the features that distinguish high and the number of edits and article size, we predict which of six quality low quality Wikipedia articles. classes (Start, Stub, C-Class, B-Class, Good, Featured) articles belong While many Wikipedia article quality classifiers have focused to, attaining a classification accuracy of 49.35% on a stratified sample on assessing quality based on the content of the latest version of of articles. These results suggest that structures of collaboration an article [1, 4, 18], prior work has suggested that high quality arti- underlying the creation of articles, and not just the content of the cles are associated with more intense collaboration among editors article, should be considered for accurate quality classification.
    [Show full text]
  • Wiki-Metasemantik: a Wikipedia-Derived Query Expansion Approach Based on Network Properties
    Wiki-MetaSemantik: A Wikipedia-derived Query Expansion Approach based on Network Properties D. Puspitaningrum1, G. Yulianti2, I.S.W.B. Prasetya3 1,2Department of Computer Science, The University of Bengkulu WR Supratman St., Kandang Limun, Bengkulu 38371, Indonesia 3Department of Information and Computing Sciences, Utrecht University PO Box 80.089, 3508 TB Utrecht, The Netherlands E-mails: [email protected], [email protected], [email protected] Pseudo-Relevance Feedback (PRF) query expansions suffer Abstract- This paper discusses the use of Wikipedia for building from several drawbacks such as query-topic drift [8][10] and semantic ontologies to do Query Expansion (QE) in order to inefficiency [21]. Al-Shboul and Myaeng [1] proposed a improve the search results of search engines. In this technique, technique to alleviate topic drift caused by words ambiguity selecting related Wikipedia concepts becomes important. We and synonymous uses of words by utilizing semantic propose the use of network properties (degree, closeness, and pageRank) to build an ontology graph of user query concept annotations in Wikipedia pages, and enrich queries with which is derived directly from Wikipedia structures. The context disambiguating phrases. Also, in order to avoid resulting expansion system is called Wiki-MetaSemantik. We expansion of mistranslated words, a query expansion method tested this system against other online thesauruses and ontology using link texts of a Wikipedia page has been proposed [9]. based QE in both individual and meta-search engines setups. Furthermore, since not all hyperlinks are helpful for QE task Despite that our system has to build a Wikipedia ontology graph though (e.g.
    [Show full text]
  • WTF IS NATIVE ADVERTISING? Introduction
    SPONSORED BY IS NATIVE ADVERTISING? Table of Contents Introduction WTF is Native 3 14 Programmatic? Nomenclature The Native Ad 4 16 Triumvirate 5 Decision tree 17 The Ad Man 18 The Publisher Issues Still Plaguing 19 The Platform 6 Native Advertising Glossary 7 Scale 20 8 Metrics 10 Labelling 11 The Church/State Divide 12 Credibility 3 / WTF IS NATIVE ADVERTISING? Introduction In 2013, native advertising galloped onto the scene like a masked hero, poised to hoist publishers atop a white horse, rescuing them from the twin menaces of programmatic advertising and sagging CPMs. But who’s really there when you peel back the mask? Native advertising is a murky business. Ad executives may not consider it advertising. Editorial departments certainly don’t consider it editorial. Even among its practitioners there is debate — is it a format or is it a function? Publishers who have invested in the studio model position native advertising as the perfect storm of context, creative capital and digital strategy. For platforms, it may be the same old banner advertising refitted for the social stream. Digiday created the WTF series to parse murky digital marketing concepts just like these. WTF is Native Advertising? Keep reading to find out..... DIGIDAY 4 / WTF IS NATIVE ADVERTISING? Nomenclature Native advertising An advertising message designed Branded content Content created to promote a Content-recommendation widgets Another form to mimic the form and function of its environment brand’s products or values. Branded content can take of native advertising often used by publishers, these a variety of formats, not all of them technically “native.” appear to consumers most often at the bottom of a web Content marketing Any marketing messages that do Branded content placed on third-party publishing page with lines like “From around the web,” or “You not fit within traditional formats like TV and radio spots, sites or platforms can be considered native advertising, may also like.” print ads or banner messaging.
    [Show full text]
  • E-Posters in Academic/Scientific Conferεnces – Guidelines, Comparative Study & New Suggestions
    University of the Aegean Department of Product and Systems Design Engineering DISSERTATION: E-POSTERS IN ACADEMIC/SCIENTIFIC CONFERΕNCES – GUIDELINES, COMPARATIVE STUDY & NEW SUGGESTIONS Pissaridi Maria Anastasia 511/2004041 Supervisor: Paraskevas Papanikos Committee Members: Nikolaos Zacharopoulos Maria Simosi Syros, June 2015 DISSERTATION: E-POSTERS IN ACADEMIC/SCIENTIFIC CONFERENCES – GUIDELINES, COMPARATIVE STUDY & NEW SUGGESTIONS Supervisor: Paraskevas Papanikos Committee Members: Nikolaos Zacharopoulos Maria Simosi Pissaridi Maria Anastasia Syros, June 2015 3 ABSTRACT Conferences play a key role in getting people interested in a field together to network and exchange knowledge. The poster presentation is a commonly used format for communicating information within the academic scientific conference sector. Paper posters were the beginning but as technology and the way people work changes, posters have to be developed and implemented in order to achieve successful knowledge transfer. Incorporating aspects of information technology into poster presentations can promote an interactive learning environment for users and counter the current passive nature of poster design as an integrated approach with supplemental material is required to achieve changes in user knowledge, attitude and behaviour. After conducting literature review, research of existing e-poster providers, interviews with 5 of them and a personal evaluation in a real time conference environment results show a gradual turn towards e-posters with the medical sector pioneering. Authors, viewers and organisers embrace this new format, and the features and functions it offers, although objections exist since people have different preferences and the e-poster sector is relatively young, an average of 5 years. Pissaridi Maria Anastasia dpsd04041 4 SUMMARY Η ανάγκη των ανθρώπων να συγκεντρώνονται για να ανταλλάσσουν ιδέες, ευρήματα, έρευνες και απόψεις υπήρχε από τη δημιουργία των πρώτων πόλεων όπου φτιάχνονται με χώρους συγκέντρωσης.
    [Show full text]