Using Social Annotations to Improve Web Search

Total Page:16

File Type:pdf, Size:1020Kb

Using Social Annotations to Improve Web Search USING SOCIAL ANNOTATIONS TO IMPROVE WEB SEARCH by Worasit Choochaiwattana B.B.A. (Marketing), Chulalongkorn University, Thailand, 1996 M.Sc. (Tech of Info Sys Mgt), Mahidol University, Thailand, 2000 M.Sc. (Computer Science), Chulalongkorn University, Thailand, 2003 Submitted to the Graduate Faculty of School of Information Sciences in partial fulfillment of the requirements for the degree of Doctor of Philosophy University of Pittsburgh 20081 UNIVERSITY OF PITTSBURGH SCHOOL OF INFORMATION SCIENCES This dissertation was presented by Worasit Choochaiwattana It was defended on April 17, 2008 and approved by Peter Brusilovsky, PhD, Associate Professor, School of Information Sciences Daqing He, PhD, Assistance Professor, School of Information Sciences Stephen Hirtle, PhD, Professor, School of Information Sciences Brian Butler, PhD, Associate Professor, Joseph M. Katz Graduate School of Business and College of Business Administration Dissertation Advisor: Michael Spring, PhD, Associate Professor, School of Information Sciences 2 Copyright © by Worasit Choochaiwattana 2008 3 USING SOCIAL ANNOTATIONS TO IMPROVE WEB SEARCH Worasit Choochaiwattana, PhD University of Pittsburgh, 2008 Web-based tagging systems, which include social bookmarking systems such as Delicious, have become increasingly popular. These systems allow participants to annotate or tag web resources. This research examined the use of social annotations to improve the quality of web searches. The research involved three components. First, social annotations were used to index resources. Two annotation-based indexing methods were proposed: annotation based indexing and full text with annotation indexing. Second, social annotations were used to improve search result ranking. Six annotation based ranking methods were proposed: Popularity Count, Propagate Popularity Count, Query Weighted Popularity Count, Query Weighted Propagate Popularity Count, Match Tag Count and Normalized Match Tag Count. Third, social annotations were used to both index and rank resources. The result from the first experiment suggested that both static feature and similarity feature should be considered when using social annotations to re-rank search result. The result of the second experiment showed that using only annotation as an index of resources may not be a good idea. Since social Annotations could be viewed as a high level concept of the content, combining them to the content of resource could add some more important concepts to the resources. Last but not least, the result from the third experiment confirmed that the combination of using social annotations to rank the search result and using social annotations as resource index augmentation provided a promising rank of search results. It showed that social annotations could benefit web search. 4 TABLE OF CONTENTS PREFACE .................................................................................................................................... 16 1.0 INTRODUCTION ...................................................................................................... 18 1.1 DELICIOUS ....................................................................................................... 19 1.2 MOTIVATION .................................................................................................. 21 1.2.1 Using Social Annotations as Resource Indexes ........................................... 21 1.2.2 Search Result Ranking with Social Annotation .......................................... 22 1.3 RESEARCH OBJECTIVES ............................................................................. 23 1.4 ORGANIZATION OF THE PAPER ............................................................... 24 1.5 DEFINTIONS OF TERMS ............................................................................... 24 1.5.1 Annotation ...................................................................................................... 24 1.5.2 Anchor and Anchor text ............................................................................... 25 1.5.3 Link ................................................................................................................. 25 1.5.4 Metadata ......................................................................................................... 26 1.5.5 Collaboration ................................................................................................. 26 1.5.6 Folksonomy .................................................................................................... 26 1.5.7 Social Tagging/Social Annotation ................................................................ 26 1.5.8 Synonymy and Polysemy .............................................................................. 27 1.5.9 Noise ................................................................................................................ 27 5 1.5.10 Similarity Ranking and Static Ranking ..................................................... 27 1.5.11 Resource and URL ....................................................................................... 27 2.0 RELATED WORK .................................................................................................... 28 2.1 BACKGROUND ................................................................................................ 28 2.1.1 Annotation Forms and Functions ................................................................ 32 2.1.2 Purposes of Annotation ................................................................................. 34 2.1.2.1 Annotation for Memory ...................................................................... 34 2.1.2.2 Annotation for Communication ......................................................... 35 2.1.2.3 Annotation for Collaboration ............................................................ 36 2.1.2.4 Annotation for Description ................................................................ 37 2.2 DIGITAL ANNOTATION SYSTEMS ............................................................ 38 2.2.1 Systems that Used Annotations for Collaborative Purposes ..................... 39 2.2.1.1 NoteCard .............................................................................................. 39 2.2.1.2 InterNote .............................................................................................. 41 2.2.1.3 CASCADE ........................................................................................... 43 2.2.1.4 Collaborative Clinical Trial Protocol Writing System .................... 45 2.2.1.5 Col•laboració ....................................................................................... 46 2.2.2 Systems that Used Annotations for Memory and Communication Purposes ...................................................................................................................... 48 2.2.2.1 WebAnn ............................................................................................... 48 2.2.2.2 EDUCOSM .......................................................................................... 50 2.2.2.3 CritLink ............................................................................................... 51 2.2.2.4 Annotea ................................................................................................ 53 6 2.2.2.5 Microsoft Office ................................................................................... 55 2.2.2.6 Microsoft Research Annotation System ............................................ 56 2.2.3 Systems that Used Annotations for Description Purposes ......................... 57 2.2.3.1 YAWAS ................................................................................................ 57 2.2.3.2 Delicious ............................................................................................... 60 2.2.3.3 OntoAnnotate ...................................................................................... 61 2.2.3.4 Ont-O-Mat ........................................................................................... 62 2.2.3.5 Flickr .................................................................................................... 64 2.2.3.6 PhotoFinder ......................................................................................... 65 2.2.3.7 HBP Image Annotation Service ......................................................... 66 2.2.3.8 MADCOW ........................................................................................... 67 2.2.3.9 AntV ..................................................................................................... 68 2.2.3.10 VAnnotator ........................................................................................ 69 2.3 USING ANNOTATIONS TO IMPROVE SEARCH RESULTS .................. 70 3.0 PRELIMINARY ANALYSIS AND RESEARCH DESIGN .................................. 86 3.1 PROBLEM AND RESEARCH QUESTIONS ................................................ 86 3.2 PRELIMINARY ANALYSIS ........................................................................... 87 3.3 RESEARCH DESIGN ....................................................................................... 95 3.3.1 Evaluation metric .......................................................................................... 95 3.3.2 Experiment 1: Re-ranking search results .................................................... 96 3.3.2.1 Variables and Expected Results ......................................................... 96 3.3.2.2 The 1st
Recommended publications
  • Understanding Job-Skill Relationships Using Big Data and Neural Networks
    Understanding Job-Skill Relationships using Big Data and Neural Networks Abhinav Maurya Carnegie Mellon University Pittsburgh, PA – 15213 [email protected] ABSTRACT Our proposed model is inspired by topic modeling where Nonlinear vector space embeddings have shown great promise latent topics occur throughout the documents of a corpus in many applications such as similarity search, analogy map- in varying proportions, which enables their identification to pings, dataset visualization, etc. In this paper, we propose characterize the text corpus. Similarly, in our model, vari- a simple distribution-to-distribution regression model based ation in the co-occurrence of jobs and skills can be used on neural networks which provides (i) job2vec: interpretable to disentangle which skills are associated with which jobs. embeddings of jobs in terms of associated skills, (ii) skill2vec: However, unlike unsupervised topic models such as LDA [1], converse embeddings of skills in terms of the jobs that re- we cast our problem of identifying job-skill associations as quire them, and (iii) SkillRank: a mechanism for ranking a supervised distribution-to-distribution regression, where skills associated with each job that can be recommended the input is the empirical distribution over job titles and to members aspiring for that job. Due to the simplicity of the output is the corresponding empirical distribution over our model, it has no hyperparameters that need tuning ex- skills for that person. Moreover, unlike unsupervised topic cept for the number of stochastic gradient descent iterations models such as LDA and supervised variants such as [4], our which is easily determined using the early stopping criterion.
    [Show full text]
  • Devpartner Java Edition Getting Started Guide
    DevPartner Java Edition Getting Started Guide Release 4.5 Copyright © 2001–2009 Micro Focus (IP) Ltd. All rights reserved. Micro Focus (IP) Ltd. has made every effort to ensure that this book is correct and accurate, but reserves the right to make changes without notice at its sole discretion at any time. The software described in this document is supplied under a license and may be used or copied only in accordance with the terms of such license, and in particular any warranty of fitness of Micro Focus software products for any particular purpose is expressly excluded and in no event will Micro Focus be liable for any consequential loss. Animator®, COBOLWorkbench®, EnterpriseLink®, Mainframe Express®, Micro Focus®, Net Express®, REQL® and Revolve® are registered trademarks, and AAI™, Analyzer™, Application Quality Workbench™, Application Server™, Application to Application Interface™, AddPack™, AppTrack™, AssetMiner™, BoundsChecker™, CARS™, CCI™, DataConnect™, DevPartner™, DevPartnerDB™, DevPartner Fault Simulator™, DevPartner SecurityChecker™,Dialog System™, Driver:Studio™, Enterprise Server™, Enterprise View™, EuroSmart™, FixPack™, LEVEL II COBOL™, License Server™, Mainframe Access™, Mainframe Manager™, Micro Focus COBOL™, Micro Focus Studio™, Micro Focus Server™, Object COBOL™, OpenESQL™, Optimal Trace™,Personal COBOL™, Professional COBOL™, QACenter™, QADirector™, QALoad™, QARun™, Quality Maturity Model™, Server Express™, SmartFind™, SmartFind Plus™, SmartFix™, SoftICE™, SourceConnect™, SupportLine™, TestPartner™, Toolbox™, TrackRecord™, WebCheck™, WebSync™, and Xilerator™ are trademarks of Micro Focus (IP) Ltd. All other trademarks are the property of their respective owners. No part of this publication, with the exception of the software product user documentation contained on a CD-ROM, may be copied, photocopied, reproduced, transmitted, transcribed, or reduced to any electronic medium or machine-readable form without prior written consent of Micro Focus (IP) Ltd.
    [Show full text]
  • SDL Contenta S1000D and SDL Livecontent S1000D Cross-Product Graphics and Multimedia Support
    SDL Contenta S1000D and SDL LiveContent S1000D Cross-Product Graphics and Multimedia Support SDL Contenta S1000D and SDL LiveContent S1000D 5.9 December 2020 Legal notice Copyright and trademark information relating to this product release. Copyright © 2009–2020 SDL Group. SDL Group means SDL PLC. and its subsidiaries and affiliates. All intellectual property rights contained herein are the sole and exclusive rights of SDL Group. All references to SDL or SDL Group shall mean SDL PLC. and its subsidiaries and affiliates details of which can be obtained upon written request. All rights reserved. Unless explicitly stated otherwise, all intellectual property rights including those in copyright in the content of this website and documentation are owned by or controlled for these purposes by SDL Group. Except as otherwise expressly permitted hereunder or in accordance with copyright legislation, the content of this site, and/or the documentation may not be copied, reproduced, republished, downloaded, posted, broadcast or transmitted in any way without the express written permission of SDL. Contenta S1000D is a registered trademark of SDL Group. All other trademarks are the property of their respective owners. The names of other companies and products mentioned herein may be the trade- marks of their respective owners. Unless stated to the contrary, no association with any other company or product is intended or should be inferred. This product may include open source or similar third-party software, details of which can be found by clicking the following link: “Acknowledgments” on page 7. Although SDL Group takes all reasonable measures to provide accurate and comprehensive information about the product, this information is provided as-is and all warranties, conditions or other terms concerning the documentation whether express or implied by statute, common law or otherwise (including those relating to satisfactory quality and fitness for purposes) are excluded to the extent permitted by law.
    [Show full text]
  • Peiwen Zhu. Web Annotation Systems: a Literature Review and Case Study
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Carolina Digital Repository Peiwen Zhu. Web Annotation Systems: A Literature Review and Case Study. A Master’s Paper for the M.S. in I.S. degree. April, 2008. 35 pages. Advisor: Bradley M.Hemminger. Web annotation has been a popular research topic since the appearance of Internet and its supplementary technologies. This paper provides a literature review in the related research areas, and introduces some currently available web annotation systems with the comparison of these systems in seven aspects. In addition, this paper incorporates a description of the ongoing NeoNote project, and identifies limitations of this study for future work. Headings: Scholarly Communication Web annotations Information Systems/Design WEB ANNOTATION SYSTEMS: A LITERATURE REVIEW AND CASE STUDY by Peiwen Zhu A Master's paper submitted to the faculty of the School of Information and Library Science of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements for the degree of Master of Science in Information Science. Chapel Hill, North Carolina April, 2008 Approved by: ___________________________ Bradley M.Hemminger 1 Table of Contents Introduction………………………………………………………………………………..2 Literature Review………………………………………………………………………….4 Current Systems………………………………………………………………………….15 Case Study……………………………………………………………………………….26 Conclusion……………………………………………………………………………….30 References………………………………………………………………………………..32 2 Introduction Web browsing plays an important role nowadays in people’s daily life, study, and work. Since documents exist mostly in digital format on the web, people may spend a large part of their time on browsing or searching on the web to look for useful information. However, this used to be a one-way interaction with users having few options to mark texts or to highlight important sections in a web document; what’s more, it is difficult to add extra information as reference on web pages, which is useful for further reference or sharing with friends.
    [Show full text]
  • Javaedge Setup and Installation
    APPENDIX A ■ ■ ■ JavaEdge Setup and Installation Throughout the book, we have used the example application, JavaEdge, to provide a practical demonstration of all the features discussed. In this appendix, we will walk you through setting up the tools and applications required to build and run JavaEdge, as well as take you through the steps needed to get the JavaEdge application running on your platform. Environment Setup Before you can get started with the JavaEdge application, you need to configure your platform to be able to build and run JavaEdge. Specifically, you need to configure Apache Ant in order to build the JavaEdge application and package it up for deployment. In addition, the JavaEdge application is designed to run on a J2EE application server and to use MySQL as the back-end database. You also need to have a current JDK installed; the JavaEdge application relies on JVM version 1.5 or higher, so make sure your JDK is compatible. We haven’t included instruc- tions for this here, since we are certain that you will already have a JDK installed if you are reading this book. However, if you do need to download one, you can find it at http://java. sun.com/j2se/1.5.0/download.jsp. Installing MySQL The JavaEdge application uses MySQL as the data store for all user, story, and comment data. If you don’t already have the MySQL database server, then you need to obtain the version applicable to your platform. You can obtain the latest production binary release of MySQL for your platform at http://www.mysql.com.
    [Show full text]
  • Appwave Studio User Guide
    Product Documentation Embarcadero® AppWave™ Studio User Guide Version 3.0 Published November, 2011 © 2011 Embarcadero Technologies, Inc. Embarcadero, the Embarcadero Technologies logos, and all other Embarcadero Technologies product or service names are trademarks or registered trademarks of Embarcadero Technologies, Inc. All other trademarks are property of their respective owners. Embarcadero Technologies, Inc. is a leading provider of award-winning tools for application developers and database professionals so they can design systems right, build them faster and run them better, regardless of their platform or programming language. Ninety of the Fortune 100 and an active community of more than three million users worldwide rely on Embarcadero products to increase productivity, reduce costs, simplify change management and compliance and accelerate innovation. The company's flagship tools include: Embarcadero® Change Manager™, RAD Studio, DBArtisan®, Delphi®, ER/Studio®, JBuilder® and Rapid SQL®. Founded in 1993, Embarcadero is headquartered in San Francisco, with offices located around the world. Embarcadero is online at www.embarcadero.com. November, 2011 Contents Welcome to Embarcadero AppWave Studio . 7 AppWave Studio Users . .7 Technical Requirements . .9 Mastering Apps . .9 Benefits of Using Apps. .11 Private vs. Public AppWave . .11 Using AppWave Studio . 13 Studio Access . .13 Best Mastering Practices . .14 Create an App with One Click. .15 Licensing . 16 Preparation . 16 Setup . 17 Recording . 20 Running the App. 28 Create AppWave Supported or Custom App from Install . .30 Start . 31 Launch & Brand. 33 App . 35 Test . 41 Broadcast. 43 Actions taken in Each Step . 47 Install Using a Zip File. 49 Create AppWave Supported or Custom App from Scratch . .52 Start .
    [Show full text]
  • Semantic Annotation Framework for Web Extracted Data
    International Journal of Computer Engineering & Technology (IJCET) Volume 7, Issue 5, Sep–Oct 2016, pp. 65–76, Article ID: IJCET_07_05_008 Available online at http://www.iaeme.com/ijcet/issues.asp?JType=IJCET&VType=7&IType=6 Journal Impact Factor (2016): 9.3590(Calculated by GISI) www.jifactor.com ISSN Print: 0976-6367 and ISSN Online: 0976–6375 © IAEME Publication FIOBODA - SEMANTIC ANNOTATION FRAMEWORK FOR WEB EXTRACTED DATA C. Gnana Chithra Equity Research Consultant, Angeeras Securities, Chennai, Tamilnadu, India Dr. E. Ramaraj Professor, Department of Computer Science and Engineering, Alagappa University, Karaikudi, Tamilnadu, India ABSTRACT Semantic annotation of web pages is the state of art technology for achieving the unified objective of attaining Semantic web Universe, which enables sharing, and reusing the document content beyond the boundaries and applications. Web is a treasury of knowledge and efficient tools should be designed to explore the structured and unstructured data. Annotating million of web pages manually is an impossible task. For high information retrieval rates, automatic annotation of documents is mandatory. Metadata is added to the web pages to make it intelligent for processing in content based intelligent applications. This paper analyses the problems with the current Semantic annotation systems and proposes a new Ontology based Automatic annotation system Framework. Ontology based semantic annotation is one of the best methods for extracting data from the Knowledge Base. The integration of Modified Manning’s Sentence boundary detection algorithm and Noun Phrase Collocation algorithm and classification using machine learning techiques in the Information Extraction module, and developing a new data model and ontology for Structured Ontology engineering model is contributed in this paper.
    [Show full text]
  • Starteam 16.2
    StarTeam 16.2 Release Notes Micro Focus The Lawn 22-30 Old Bath Road Newbury, Berkshire RG14 1QN UK http://www.microfocus.com Copyright © Micro Focus 2017. All rights reserved. MICRO FOCUS, the Micro Focus logo and StarTeam are trademarks or registered trademarks of Micro Focus IP Development Limited or its subsidiaries or affiliated companies in the United States, United Kingdom and other countries. All other marks are the property of their respective owners. 2017-11-02 ii Contents StarTeam Release Notes ....................................................................................5 What's New ........................................................................................................ 6 16.2 ..................................................................................................................................... 6 StarTeam Command Line Tools .............................................................................. 6 StarTeam Cross-Platform Client ...............................................................................6 StarTeam Git Command Line Utility. .........................................................................7 StarTeam Server ...................................................................................................... 7 Workflow Extensions ................................................................................................ 8 StarTeam Web Client ................................................................................................8 16.1 Update 1 ......................................................................................................................9
    [Show full text]
  • GDA Release Notes Release 9.21
    GDA Release Notes Release 9.21 Diamond Light Source Jul 16, 2021 CONTENTS 1 GDA 9.0 and GDA 8.52 (expected release: Mar/2016)3 2 GDA 8.50 (Oct/2015) 5 3 GDA 8.48 (Aug/2015) 7 4 GDA 8.46 (Jun/2015) 9 5 GDA 8.44 (Mar/2015) 11 6 GDA 8.42 (Aug/2014) 29 7 GDA 8.40 (Jun/2014) 31 8 GDA 8.38 (Mar/2014) 35 9 GDA 8.36 (Nov/2013) 45 10 GDA 8.34 (Aug/2013) 47 11 GDA 8.32 (Jun/2013) 49 12 GDA 8.30 (Apr/2013) 51 13 GDA 8.28 (Jan/2013) 53 14 GDA 8.26 (Aug/2012) 55 15 GDA 8.24 (June/2012) 57 16 GDA 8.20 (Jan/2012) 61 17 GDA 8.18 (Nov/2011) 63 18 GDA 8.14 (Jun/2011) 65 19 GDA 8.12 (not externally released) 67 20 GDA 8.10 (released 6/Dec/2010) 69 21 GDA 8.8 (released 21/Sep/2010) 71 22 GDA 8.6 (released 11/Jun/2010) 73 i 23 GDA 8.4 (released 30/Mar/2010) 75 24 GDA 8.2 (released 1/Mar/2010) 77 25 GDA 8.0 79 26 GDA 7 and earlier 85 27 Contributors to the GDA project 123 ii GDA Release Notes, Release 9.21 These release notes describe the major changes in each release of the GDA. This includes both changes relevant to developers, and changes relevant to users of the GDA. CONTENTS 1 GDA Release Notes, Release 9.21 2 CONTENTS CHAPTER ONE GDA 9.0 AND GDA 8.52 (EXPECTED RELEASE: MAR/2016) 1.1 User Highlights Development of GDA has been split into 2 streams.
    [Show full text]
  • A Semantic Web Annotation Tool for a Web-Based Audio Sequencer
    A semantic web annotation tool for a web-based audio sequencer Luca Restagno1, Vincent Akkermans2, Giuseppe Rizzo1, and Antonio Servetti1 1 Dipartimento di Automatica e Informatica, Politecnico di Torino, Torino, Italy, [email protected], [email protected], [email protected], 2 Music Technology Group, Universitat Pompeu Fabra, Barcelona, Spain, [email protected] Abstract. Music and sound have a rich semantic structure which is so clear to the composer and the listener, but that remains mostly hidden to computing machinery. Nevertheless, in recent years, the introduction of software tools for music production have enabled new opportunities for migrating this knowledge from humans to machines. A new generation of these tools may exploit sound samples and semantic information coupling for the creation not only of a musical, but also of a “semantic” compo- sition. In this paper we describe an ontology driven content annotation framework for a web-based audio editing tool. In a supervised approach, during the editing process, the graphical web interface allows the user to annotate any part of the composition with concepts from publicly avail- able ontologies. As a test case, we developed a collaborative web-based audio sequencer that provides users with the functionality to remix the audio samples from the Freesound website and subsequently annotate them. The annotation tool can load any ontology and thus gives users the opportunity to augment the work with annotations on the structure of the composition, the musical materials, and the creator’s reasoning and intentions. We believe this approach will provide several novel ways to make not only the final audio product, but also the creative process, first class citizens of the Semantic Web.
    [Show full text]
  • 2017-Networkthousand
    1 Annotated Bibliography, and Summaries I. Annotating Text II. Annotating Media III. Annotation Theory and Practice IV. Group Dynamics and Social Annotation V. Bibliographic Reference/Metadata/Tagging Note to Reader: The categories of this bibliography have been separated into ‘Tools’ and ‘Articles’ for ease of reading. Cross-posted entries have been marked with an asterisk (*) after their first appearance. Category I: Annotating Text This category provides an environmental scan of the current state of text-based annotation practices and the foundational tools in the discipline. Many of the publications focus on situating annotation in the field of digital humanities by drafting definitions for annotation practice; specifying a general annotation framework for commentary across mediums; or brainstorming platforms that would better support user interaction with objects and with each other. Authors remark that the evolution of scholarship brought about by interactive Web 2.0 practices shifted the focus from learner-content interaction to learner-learner interaction, and that this behavioural shift necessitates a redesign of tools (Agosti et al. 2005; Gao 2013). It is widely acknowledged that annotation practices are beneficial for learning, archiving, clarifying, sharing, and expanding; current Web architecture, however, struggles to facilitate these advantages (Agosti et al. 2012; Bottoni 2003; Farzan et al. 2008). Several publications introduce tools that bridge the gap between tool design and user needs, specifically AnnotatEd and CommentPress (Farzan et al. 2008; Fitzpatrick 2007). These tools, along with alternative platforms that gamify annotation, allow for interactive reading and support user engagement with resources in a customizable way. The tools treat documents as mutable objects that can be tagged, highlighted, and underlined.
    [Show full text]
  • Collaborative Annotation for Reliable Natural Language Processing
    Collaborative Annotation for Reliable Natural Language Processing Kar¨enFort February 12, 2016 14:47 2 Contents 1 Introduction 9 1.1 Natural Language Processing and Manual Annotation: Dr Jekyll and Mr Hyjide? . .9 1.1.1 Where Linguistics Hides . .9 1.1.2 What is Annotation? . 10 1.1.3 New Forms, Old Issues . 11 1.2 Rediscovering Annotation . 13 1.2.1 A Rise in Diversity and Complexity . 13 1.2.2 Redefining Manual Annotation Costs . 15 2 Annotating Collaboratively 17 2.1 The Annotation Process (Re)visited . 17 2.1.1 Building Consensus . 17 2.1.2 Existing Methodologies . 18 2.1.3 Preparatory Work . 20 2.1.4 Pre-campaign . 24 2.1.5 Annotation . 27 2.1.6 Finalization . 29 2.2 Annotation Complexity . 31 2.2.1 Examples Overview . 31 2.2.2 What to Annotate? . 33 2.2.3 How to Annotate? . 35 2.2.4 The Weight of the Context . 38 2.2.5 Visualization . 39 2.2.6 Elementary Annotation Tasks . 41 2.3 Annotation Tools . 42 2.3.1 To be or not to be an Annotation Tool . 43 2.3.2 Much more than Prototypes . 44 2.3.3 Addressing the new Annotation Challenges . 46 2.3.4 The Impossible Dream Tool . 49 2.4 Evaluating the Annotation Quality . 50 2.4.1 What is Annotation Quality? . 50 2.4.2 Understanding the Basics . 50 2.4.3 Beyond Kappas . 55 2.4.4 Giving Meaning to the Metrics . 57 3 4 CONTENTS 3 Crowdsourcing Annotation 63 3.1 What is Crowdsourcing and Why Should we be Interested in it? 63 3.1.1 A Moving Target .
    [Show full text]