Crowd Truth and the Seven Myths of Human Annotation Aroyo, Lora; Welty, Chris

Total Page:16

File Type:pdf, Size:1020Kb

Crowd Truth and the Seven Myths of Human Annotation Aroyo, Lora; Welty, Chris VU Research Portal Truth Is a Lie: Crowd Truth and the Seven Myths of Human Annotation Aroyo, Lora; Welty, Chris published in The AI Magazine 2015 DOI (link to publisher) 10.1609/aimag.v36i1.2564 document version Publisher's PDF, also known as Version of record Link to publication in VU Research Portal citation for published version (APA) Aroyo, L., & Welty, C. (2015). Truth Is a Lie: Crowd Truth and the Seven Myths of Human Annotation. The AI Magazine, 36(1), 15-24. https://doi.org/10.1609/aimag.v36i1.2564 General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal ? Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. E-mail address: [email protected] Download date: 30. Sep. 2021 Articles Truth Is a Lie: Crowd Truth and the Seven Myths of Human Annotation Lora Aroyo, Chris Welty n Big data is having a disruptive impact amar prestar aen … “The world is changed.” In the past across the sciences. Human annotation of decade the amount of data and the scale of computation semantic interpretation tasks is a critical Iavailable has increased by a previously inconceivable part of big data semantics, but it is based amount. Computer science, and AI along with it, has moved on an antiquated ideal of a single correct truth that needs to be similarly disrupted. solidly out of the realm of thought problems and into an We expose seven myths about human empirical science. However, many of the methods we use annotation, most of which derive from predate this fundamental shift, including the ideal of truth. that antiquated ideal of truth, and dispel Our central purpose is to revisit this ideal in computer sci - these myths with examples from our ence and AI, expose it as a fallacy, and begin to form a new research. We propose a new theory of theory of truth that is more appropriate for big data seman - truth, crowd truth, that is based on the tics. We base this new theory on the claim that, outside intuition that human interpretation is subjective, and that measuring annota - mathematics, truth is entirely relative and is most closely tions on the same objects of interpretation related to agreement and consensus. (in our examples, sentences) across a Our theories arise from experimental data that has been crowd will provide a useful representation published previously (Aroyo and Welty 2013a, Soberon et al. of their subjectivity and the range of rea - 2013, Inel et al. 2013, Dumitrache et al. 2013), and we use sonable interpretations. throughout the article examples from our natural language Copyright © 2015, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602 SPRING 2015 15 Articles processing (NLP) work in medical relation extraction; The Seven Myths however, the ideas generalize across all of semantics, including semantics from images, audio, video, sen - The need for human annotation of data is very real. sor networks, medical data, and others. We begin by We need to be able to measure machine performance on tasks that require empirical analysis. As the need looking at the current fallacy of truth in AI and where for machines to handle the scale of the data increas - it has led us, and then examine how big data, crowd - es, so will the need for human annotated gold stan - sourcing, and semantics can be combined to correct dards. We have found that, just as the sciences and the fallacy and improve big data analytic systems. humanities are reinventing themselves in the pres - ence of data, so too must the collection of human Human Annotation annotated data. We have discovered the following myths that In the age of big data, machine assistance for empir - directly influence the practice of collecting human ical analysis is required in all the sciences. Empirical annotated data. Like most myths, they are based in analysis is very much akin to semantic interpreta - fact but have grown well beyond it, and need to be tion: data is abstracted into categories, and meaning - revisited in the context of the new changing world: ful patterns, correlations, associations, and implica - Myth One: One Truth tions are extracted. For our purposes, we consider all Most data collection efforts assume that there is one semantic interpretation of data to be a form of empir - correct interpretation for every input example. ical analysis. Myth Two: Disagreement Is Bad In NLP, which has been a big data science for more To increase the quality of annotation data, disagree - than 20 years, numerous methods for empirical ment among the annotators should be avoided or analysis have been developed and are more or less reduced. standards. One of these methods is human annota - Myth Three: Detailed Guidelines Help tion to create a gold standard or ground truth. This When specific cases continuously cause disagree - method is founded on the ideal that there is a single, ment, more instructions are added to limit interpre - universally constant truth. But we all know this is a tations. fallacy, as there is not only one universally constant Myth Four: One Is Enough truth. Most annotated examples are evaluated by one per - Gold standards exist in order to train, test, and eval - son. uate algorithms that do empirical analysis. Humans Myth Five: Experts Are Better perform the same analysis on small amounts of exam - Human annotators with domain knowledge provide ple data to provide annotations that establish the better annotated data. truth. This truth specifies for each example what the Myth Six: All Examples Are Created Equal correct output of the analysis should be. Machines The mathematics of using ground truth treats every can learn (in the machine-learning sense) from these example the same; either you match the correct examples, or human programmers can develop algo - result or not. rithms by looking at them, and the correctness of Myth Seven: Once Done, Forever Valid their performance can be measured on annotated Once human annotated data is collected for a task, it examples that weren’t seen during training. is used over and over with no update. New annotat - The quality of annotation gold standards is estab - ed data is not aligned with previous data. lished by measuring the interannotator agreement, which is roughly the average pairwise probability Debunking the Myths that two people agree, adjusted for chance (Cohen 1960). This follows from the ideal of truth: a higher- Now let us revisit these myths in the context of the quality ground truth is one in which multiple new changing world, and thus in the face of a new humans provide the same annotation for the same theory of truth. examples. Again, we all know this is a fallacy, as there One Truth is more than one truth for every example. In the Our basic premise is that the ideal of truth is a falla - extreme case, if we want to interpret music, or a cy for semantic interpretation and needs to be poem, we would not expect all human annotations changed. All analytics are grounded in this fallacy, of it to be the same. In our experiments we have and human annotation efforts proceed from the found that we don’t need extreme cases to see clear - assumption that for each task, every example has a ly multiple human perspectives reflected in annota - “correct” interpretation and all others are incorrect. tion; this has revealed the fallacy of truth and helped With the widespread use of classifiers as a tool, this us to identify a number of myths in the process of “one truth” myth has become so pervasive that is is collecting human annotated data. assumed even in cases where it obviously does not 16 AI MAGAZINE Articles No. Sentence ex1 [GADOLINIUM AGENTS] used for patients with severe renal failure show signs of [NEPHROGENIC SYSTEMIC FIBROSIS]. ex2 He was the rst physician to identify the relationship between [HEMOPHILIA] and [HEMOPHILIC ARTHROPATHY]. ex3 [Antibiotics] are the rst line treatment for indications of [TYPHUS]. ex4 With [Antibiotics] in short supply, DDT was used during World War II to control the insect vectors of [TYPHUS]. ex5 [Monica Lewinsky] came here to get away from the chaos in [the nation’s capital]. ex6 [Osama bin Laden] used money from his own construction company to suport the [Muhajadeen] in Afganistan against Soviet forces. def1 MANIFESTATION links disorders to the observations that are closely associated with them; for example, abdominal distension is a manifestation of liver failure def2 CONTRAINDICATES refers to a condition that indicates that drug or treatment SHOULD NOT BE USED, for example, patients with obesity should avoid using danazol def3 ASSOCIATED WITH refers to signs, symptoms, or ndings that often appear together; for example, patients who smoke often have yellow teeth Table 1. Example Sentences and Definitions hold, like analyzing a segment of music for its mood erate the same ground truth. Rather than accepting (Lee and Hu 2012). In our research we have found this as a natural property of semantic interpretation, countless counterexamples to the one truth myth.
Recommended publications
  • Lecture Notes in Computer Science 7032 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan Van Leeuwen
    Lecture Notes in Computer Science 7032 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen Editorial Board David Hutchison Lancaster University, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Alfred Kobsa University of California, Irvine, CA, USA Friedemann Mattern ETH Zurich, Switzerland John C. Mitchell Stanford University, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel Oscar Nierstrasz University of Bern, Switzerland C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen TU Dortmund University, Germany Madhu Sudan Microsoft Research, Cambridge, MA, USA Demetri Terzopoulos University of California, Los Angeles, CA, USA Doug Tygar University of California, Berkeley, CA, USA Gerhard Weikum Max Planck Institute for Informatics, Saarbruecken, Germany Lora Aroyo Chris Welty Harith Alani Jamie Taylor Abraham Bernstein Lalana Kagal Natasha Noy Eva Blomqvist (Eds.) The Semantic Web – ISWC 2011 10th International Semantic Web Conference Bonn, Germany, October 23-27, 2011 Proceedings, Part II 13 Volume Editors Lora Aroyo VU University Amsterdam, The Netherlands; [email protected] Chris Welty IBM Research, Yorktown Heights, NY, USA; [email protected] Harith Alani The Open University, Milton Keynes, UK; [email protected] Jamie Taylor Google, Mountain View, CA, USA; [email protected] Abraham Bernstein University of Zurich,
    [Show full text]
  • An Overview of Ontoclean
    8 An Overview of OntoClean Nieola Guarino1 and Christopher A. Welty lLaboratory for Applied Ontology (ISTC-CNR) Polo Tecnologico, Via Solteri 38, 38100 Trento, ITAL Y [email protected] 2IBM Watson Research Center 19 Skyline Dr., Hawthome, NY 10532, USA [email protected] Summary. OntoClean is a methodology for validating the ontologie al adequaey of taxonomie relationships. It is based on highly general ontologieal notions drawn from philosophy, like essence, identity, and unity, whieh are used to ehar­ aeterize relevant aspeets of the intended meaning of the properties, classes, and relations that make up an ontology. These aspeets are represented by formal metaproperties, whieh impose several eonstraints on the taxonomie strueture of an ontology. The analysis of these eonstraints helps in evaluating and validating the ehoiees made. In this ehapter we present an informal overview ofthe philoso­ phieal notions involved and their role in OntoClean, review some eommon onto­ logieal pitfalls, and walk through the example that has appeared in pieees in pre­ vious papers and has been the basis of numerous tutorials and talks. 8.1 Introduction The OntoClean methodology was first introdueed in aseries of eonferenee-Iength papers in 2000 [Guarino and Welty, 2000a-e; Welty and Guarino, 2001], and re­ eeived mueh attention and use in subsequent years. The main eontribution of On­ toClean was the beginning of a formal foundation for ontologieal analysis. Alan Reetor, a seasoned veteran at ontologieal analysis in the medieal domain, said of OntoClean, " ... what you have done is reduee the amount of time I spend arguing with doetors that the way I want to model the world is right..." [Reetor, 2002].
    [Show full text]
  • Cristobalthesis
    Open Personalization: Involving Third Parties in Improving the User Experience of Websites Dissertation presented to the Department of Computer Languages and Systems of the University of the Basque Country in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy (“international” mention) Cristóbal Arellano Bartolomé Supervisors: Prof. Dr. Oscar Díaz García Dr. Jon Iturrioz Sánchez San Sebastián, Spain, 2013 This work was hosted by the University of the Basque Country (Faculty of Computer Sciences). The author enjoyed a doctoral grant under de FPI (Formacion de Personal Investigador) from the Spanish Ministry of Science & Education during the years 2007 to 2011. The work was was co- supported by the Spanish Ministry of Education, and the European Social Fund under contracts (TIN2005-05610), MODELINE (TIN2008-06507- C02-01) and Scriptongue (TIN2011-23839). Summary Traditional software development captures the user needs during the requirement analysis. The Web makes this endeavour even harder due to the difficulty to determine who these users are. In an attempt to tackle the heterogeneity of the user base, Web Personalization techniques are proposed to guide the users’ experience. In addition, Open Innovation allows organisations to look beyond their internal resources to develop new products or improve existing processes. This thesis sits in between by introducing Open Personalization as a means to incorporate actors other than webmasters in the personalization of web applications. The aim is to provide the technological basis that builds up a trusty environment for webmasters and companion actors to collaborate, i.e. "an architecture of participation". Such architecture very much depends on these actors’ profile.
    [Show full text]
  • Semantic Web: a Review of the Field Pascal Hitzler [email protected] Kansas State University Manhattan, Kansas, USA
    Semantic Web: A Review Of The Field Pascal Hitzler [email protected] Kansas State University Manhattan, Kansas, USA ABSTRACT which would probably produce a rather different narrative of the We review two decades of Semantic Web research and applica- history and the current state of the art of the field. I therefore do tions, discuss relationships to some other disciplines, and current not strive to achieve the impossible task of presenting something challenges in the field. close to a consensus – such a thing seems still elusive. However I do point out here, and sometimes within the narrative, that there CCS CONCEPTS are a good number of alternative perspectives. The review is also necessarily very selective, because Semantic • Information systems → Graph-based database models; In- Web is a rich field of diverse research and applications, borrowing formation integration; Semantic web description languages; from many disciplines within or adjacent to computer science, Ontologies; • Computing methodologies → Description log- and a brief review like this one cannot possibly be exhaustive or ics; Ontology engineering. give due credit to all important individual contributions. I do hope KEYWORDS that I have captured what many would consider key areas of the Semantic Web field. For the reader interested in obtaining amore Semantic Web, ontology, knowledge graph, linked data detailed overview, I recommend perusing the major publication ACM Reference Format: outlets in the field: The Semantic Web journal,1 the Journal of Pascal Hitzler. 2020. Semantic Web: A Review Of The Field. In Proceedings Web Semantics,2 and the proceedings of the annual International of . ACM, New York, NY, USA, 7 pages.
    [Show full text]
  • Introduction to Semantic Web Technologies & Linked Data
    IntroductionIntroduction toto SemanticSemantic WebWeb TechnologiesTechnologies && LinkedLinked DataData OktieOktie HassanzadehHassanzadeh UniversityUniversity ofof TorontoToronto March 2011 CS 443: Database Management Systems - Winter 2011 Outline 2 Introduction Semantic Web Technologies Resource Description Framework (RDF) Querying RDF data (SPARQL) Linked Data Linked Data Principles Linking Open Data Community Project Example Data Sources Example Applications 3 Introduction Web of Documents vs. Web of Data Web of Documents 4 Untyped Untyped Untyped Links Links Links API/ HTML HTML HTML XML A B C D Primary objects: documents Links between documents (or parts of them) Degree of structure in data: fairly low Implicit semantics of contents Designed for: human consumption Based on presentations by Chris Bizer, Richard Cyganiak, Tom Heath, available at http://linkeddata.org/guides-and-tutorials Web of Documents: Problem 5 ? thing thing Are two documents talking about the same ? ? “thing”? ? ? ? Untyped Untyped Untyped Links Links Links API/ HTML HTML HTML XML A B C D Based on presentations by Chris Bizer, Richard Cyganiak, Tom Heath, available at http://linkeddata.org/guides-and-tutorials Example Query 6 Elvis Presley 1935 - 1977 Will there ever be someone like him again? Based on presentation by Lauw, Schenkel, Suchanek, Theobald and Weikum, available at http://www.mpi-inf.mpg.de/yago-naga/CIKM10-tutorial/ Example Query 7 Another Elvis Elvis Presley: The Early Years Elvis spent more weeks at the top of the charts than any other artist. www.fiftiesweb.com/elvis.htm Based on presentation by Lauw, Schenkel, Suchanek, Theobald and Weikum, available at http://www.mpi-inf.mpg.de/yago-naga/CIKM10-tutorial/ Example Query 8 Another singer called Elvis, young Personal relationships of Elvis Presley – Wikipedia ...when Elvis was a young teen...
    [Show full text]
  • Semantics Driven Human-Machine Computation Framework for Linked Islamic Knowledge Engineering
    Semantics Driven Human-Machine Computation Framework for Linked Islamic Knowledge Engineering by Amna Basharat (Under the Direction of Khaled Rasheed and I. Budak Arpinar) Abstract Formalized knowledge engineering activities including semantic annotation and linked data management tasks in specialized domains suffer from considerable knowledge acquisition bottleneck - owing to the lack of availability of experts and in-efficacy of automated approaches. Human Computation & Crowdsourcing (HC&C) methods successfully advocate leveraging the human intelligence and pro- cessing power to solve problems that are still difficult to be solved computationally. Contextualized to the domain of Islamic Knowledge, this research investigates the synergistic interplay of these HC&C methods and the semantic web and proposes a semantics driven human-machine computation framework for knowledge engineer- ing in specialized and knowledge intensive domains. The overall objective is to augment the process of automated knowledge extraction and text mining methods using a hybrid approach for combining collective intelligence of the crowds with that of experts to facilitate activities in formalized knowledge engineering - thus overcoming the so-called knowledge acquisition bottleneck. As part of this framework, we design and implement formal and scalable knowl- edge acquisition workflows through the application of semantics driven crowdsourc- ing methodology and its specialized derivative, called learnersourcing. We evaluate these methods and workflows for a range of knowledge engineering tasks including thematic classification, thematic disambiguation, thematic annotation and contex- tual interlinking for two primary Islamic texts, namely the Qur’an and the books of Prophetic narrations called the Hadith. This is done at various levels of granu- larity, including atomic and composite task workflows, that existing research fails to address.
    [Show full text]
  • An Overview of Ontoclean
    8 An Overview of OntoClean Nicola Guarino1 and Christopher A. Welty2 1Laboratory for Applied Ontology (ISTC-CNR) Polo Tecnologico, Via Solteri 38, 38100 Trento, ITALY [email protected] 2IBM Watson Research Center 19 Skyline Dr., Hawthorne, NY 10532, USA [email protected] Summary. OntoClean is a methodology for validating the ontological adequacy of taxonomic relationships. It is based on highly general ontological notions drawn from philosophy, like essence, identity, and unity, which are used to characterize relevant aspects of the intended meaning of the properties, classes, and relations that make up an ontology. These aspects are represented by formal metaproperties, which impose several constraints on the taxonomic structure of an ontology. The analysis of these constraints helps in evaluating and validating the choices made. In this chapter we present an informal overview of the philosophical notions in- volved and their role in OntoClean, review some common ontological pitfalls, and walk through the example that has appeared in pieces in previous papers and has been the basis of numerous tutorials and talks. 8.1 Introduction The OntoClean methodology was first introduced in a series of conference-length papers in 2000 [Guarino and Welty, 2000a-c; Welty and Guarino, 2001], and re- ceived much attention and use in subsequent years. The main contribution of On- toClean was the beginning of a formal foundation for ontological analysis. Alan Rector, a seasoned veteran at ontological analysis in the medical domain, said of OntoClean, “…what you have done is reduce the amount of time I spend arguing with doctors that the way I want to model the world is right…” [Rector, 2002].
    [Show full text]
  • Ontology-Driven Conceptual Modeling
    Ontology-Driven Conceptual Modeling Chris Welty IBM Watson Research Center CAISE-02 Acknowledgements People Organizations Vassar College, USA Nicola Guarino LADSEB-CNR, Padova Cladio Masolo CNR Cognitive Science Aldo Gangemi Institute, Trento Alessandro Oltramari Bill Andersen OntologyWorks, Inc. CAISE-02 2 1 Outline • Setting the record straight • Motivation • Formal foundation • “Upper Level” distinctions • Common pitfalls CAISE-02 3 What is Ontology? • A discipline of Philosophy – Meta-physics dates back to Artistotle • Meta (after) + physica (physical, real) – Ontology dates back to 17th century • Ontos (that which exists) + logos (knowledge of) • As in TorONTO, ONTario, ON TOp – The science of what is (in the universe) – “One universe, One ontology” • Quine, 1969: “To exist is to be the value of a quantified variable” CAISE-02 4 2 What is Ontology? • Borrowed by AI community – McCarthy (1980) calls for “a list of things that exist” – Specify all the kinds of things that can be the values of variables • Evolution of meaning in CS – Now refers to domain modeling, conceptual modeling, knowledge engineering, etc. • Note: not a “new name for an old thing” CAISE-02 5 What is an Ontology? • Poor definition: “Specification of a conceptualization” [Gruber, 1993] • Better: “Description of the kinds of entities there are and how they are related.” • Good ontologies should provide: – Meaning –Agreement – Organization – Common Understanding – Taxonomy – Vocabulary – Connection to the “real world” CAISE-02 6 3 What is an Ontology? a set of a collection
    [Show full text]
  • IBM Research Report Harnessing Disagreement in Crowdsourcing A
    RC25371 (WAT1304-058) April 23, 2013 Computer Science IBM Research Report Harnessing Disagreement in Crowdsourcing a Relation Extraction Gold Standard Lora Aroyo VU University Amsterdam, The Netherlands Chris Welty IBM Research Division Thomas J. Watson Research Center P.O. Box 208 Yorktown Heights, NY 10598 USA Research Division Almaden - Austin - Beijing - Cambridge - Haifa - India - T. J. Watson - Tokyo - Zurich LIMITED DISTRIBUTION NOTICE: This report has been submitted for publication outside of IBM and will probably be copyrighted if accepted for publication. It has been issued as a Research Report for early dissemination of its contents. In view of the transfer of copyright to the outside publisher, its distribution outside of IBM prior to publication should be limited to peer communications and specific requests. After outside publication, requests should be filled only by reprints or legally obtained copies of the article (e.g. , payment of royalties). Copies may be requested from IBM T. J. Watson Research Center , P. O. Box 218, Yorktown Heights, NY 10598 USA (email: [email protected]). Some reports are available on the internet at http://domino.watson.ibm.com/library/CyberDig.nsf/home . Harnessing disagreement in crowdsourcing a relation extraction gold standard Lora Aroyo Chris Welty VU University Amsterdam IBM Watson Research Center The Netherlands USA [email protected] [email protected] ABSTRACT are described in natural language. The importance of rela- One of the first steps in any kind of web data analytics tions and their interpretation is widely recognized in NLP, is creating a human annotated gold standard. These gold but whereas NLP technology for detecting entities (such as standards are created based on the assumption that for each people, places, organizations, etc.) in text can be expected annotated instance there is a single right answer.
    [Show full text]
  • Crowdsourcing,The,Semantic,Web
    ! ! ! ! ! ! Proceedings! of!the! First!International!Workshop! on! ! Crowdsourcing,the,Semantic,Web, held%in%conjunction%with! the!12th!International!Semantic!Web!Conference! ! Sydney,!Australia! ! Editors:! Maribel!Acosta! Lora!Aroyo! Abraham!Bernstein! Jens!Lehmann! Natasha!Noy! Elena!Simperl! ! Copyright!©!2013!for!the!individual!papers!by!the!papers’!authors.!Copying! permitted!only!for!private!and!academic!purposes.!This!volume!is!published!and! copyrighted!by!its!editors.! Preface This volume contains the papers presented at the 1st International Workshop on ”Crowdsourcing the Semantic Web” that was held in conjunction with the 12th International Semantic Web Conference (ISWC 2013), 21-25 October 2013, in Sydney, Australia. This interactive workshop takes stock of the emergent work and chart the research agenda with interactive sessions to brainstorm ideas and potential applications of collective intelligence to solving AI hard semantic web problems. There were 12 submissions. Each submission was reviewed by at least 2, and on the average 3, program committee members. The committee decided to accept 9 papers. Our special thanks goes to the reviewers who diligently reviewed the papers within this volume. September 3, 2013 Maribel Acosta Lora Aroyo Abraham Bernstein Jens Lehmann Natasha Noy Elena Simperl v Table of Contents Full Papers Crowdsourced Semantics with Semantic Tagging: “Don’t just tag it, LexiTag it!” ...................................................... 1 Csaba Veres ”Dr. Detective”: combining gamification techniques and crowdsourcing to create a gold standard in medical text ............................. 16 Anca Dumitrache, Lora Aroyo, Chris Welty, Robert-Jan Sips and An- thony Levas SLUA: Towards Semantic Linking of Users with Actions in Crowdsourcing 32 Umair Ul Hassan, Sean O’Riain and Edward Curry Content and Behaviour Based Metrics for Crowd Truth ...............
    [Show full text]
  • The Disagreement Deconvolution: Bringing Machine Learning Performance Metrics in Line with Reality
    The Disagreement Deconvolution: Bringing Machine Learning Performance Metrics In Line With Reality Mitchell L. Gordon Kaitlyn Zhou Kayur Patel Stanford University Stanford University Apple Inc. [email protected] [email protected] [email protected] Tatsunori Hashimoto Michael S. Bernstein Stanford University Stanford University [email protected] [email protected] ABSTRACT 1 INTRODUCTION Machine learning classifiers for human-facing tasks such as com- The machine learning classifiers that underpin modern social com- ment toxicity and misinformation often score highly on metrics puting systems such as Facebook, Wikipedia, and Twitter are a such as ROC AUC but are received poorly in practice. Why this gap? study in contrasts. On one hand, current performance metrics for Today, metrics such as ROC AUC, precision, and recall are used to popular tasks such as comment toxicity classification and disin- measure technical performance; however, human-computer inter- formation detection are extremely high, featuring up to .95 ROC action observes that evaluation of human-facing systems should AUC [11, 72], making the problem appear nearly solved. On the account for people’s reactions to the system. In this paper, we intro- other hand, audits suggest that these algorithms’ performance is duce a transformation that more closely aligns machine learning in reality much poorer than advertised [56]. Even well-intentioned classification metrics with the values and methods of user-facing platforms will continue to make highly-publicized mistakes [16] performance measures. The disagreement deconvolution takes in until they can better align metrics with reality. any multi-annotator (e.g., crowdsourced) dataset, disentangles sta- This disconnect between classifier performance and user-facing ble opinions from noise by estimating intra-annotator consistency, performance is indicative of a larger disconnect between how ma- and compares each test set prediction to the individual stable opin- chine learning (ML) and human-computer interaction (HCI) re- ions from each annotator.
    [Show full text]
  • Evaluating Ontology Cleaning
    Evaluating Ontology Cleaning Christopher Welty, Ruchi Mahindru, and Jennifer Chu-Carroll IBM T.J. Watson Research Center 19 Skyline Dr. Hawthorne, NY 10532 {welty,jencc}@us.ibm.com, [email protected] Abstract very little evidence that it can have impact on Ontology as a discipline of Computer Science has made knowledge-based systems. In fact, there appears many claims about its usefulness, however to date there has to be a significant obstacle in understanding the been very little evaluation of those claims. We present the methodology, and even without this “learning results of an experiment using a hybrid search system with a curve”, significant manual effort must be significant knowledge-based component to measure, using expended to employ the methodology to develop precision and recall, the impact of improving the quality of actual “clean” ontologies. Furthermore, there an ontology on overall performance. We demonstrate that has been no clear argument that such an improving the ontology using OntoClean (Guarino and expenditure will pay for itself in the long run. Welty, 2002), does positively impact performance, and that Indeed, “Why does it matter?” has been the most having knowledge of the search domain is more effective than domain-knowledge-free search techniques such as link frequent criticism of the OntoClean approach. analysis. We report here on a series of experiments using a Content Areas: Ontologies hybrid search system with a significant knowledge-based component to test the impact of improving the quality of ontologies on system Introduction performance. The use of search as a test system provides a well understood framework for Ontologies have been proposed as a solution to empirical evaluation, and gives an excellent numerous problems in areas such as search, opportunity to address the “Why does it matter?” semantic integration, agents, human-computer question.
    [Show full text]