Essential Coding Theory Book

Total Page:16

File Type:pdf, Size:1020Kb

Essential Coding Theory Book Essential Coding Theory Venkatesan Guruswami Atri Rudra1 Madhu Sudan March 15, 2019 1Department of Computer Science and Engineering, University at Buffalo, SUNY. Work supported by NSF CAREER grant CCF-0844796. 2 Foreword This book is based on lecture notes from coding theory courses taught by Venkatesan Gu- ruswami at University at Washington and CMU; by Atri Rudra at University at Buffalo, SUNY and by Madhu Sudan at Harvard and MIT. This version is dated March 15, 2019. For the latest version, please go to http://www.cse.buffalo.edu/faculty/atri/courses/coding-theory/book/ The material in this book is supported in part by the National Science Foundation under CA- REER grant CCF-0844796. Any opinions, findings and conclusions or recomendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation (NSF). ©Venkatesan Guruswami, Atri Rudra, Madhu Sudan, 2018. This work is licensed under the Creative Commons Attribution-NonCommercial- NoDerivs 3.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc-nd/3.0/ or send a letter to Creative Commons, 444 Castro Street, Suite 900, Mountain View, California, 94041, USA. 3 4 Contents I The Basics 17 1 The Fundamental Question 19 1.1 Overview ............................................ 19 1.2 Some definitions and codes ................................. 21 1.3 Error correction ........................................ 23 1.4 Distance of a code ...................................... 26 1.5 Hamming Code ........................................ 30 1.6 Hamming Bound ....................................... 33 1.7 Generalized Hamming Bound ............................... 34 1.8 Exercises ............................................ 36 1.9 Bibliographic Notes ...................................... 38 2 A Look at Some Nicely Behaved Codes: Linear Codes 39 2.1 Finite Fields .......................................... 39 2.2 Linear Subspaces ....................................... 41 2.3 Properties of Linear Codes .................................. 44 2.4 Hamming Codes ....................................... 46 2.5 Family of codes ........................................ 47 2.6 Efficient Decoding of Hamming codes ........................... 49 2.7 Dual of a Linear Code .................................... 50 2.8 Exercises ............................................ 52 2.9 Bibliographic Notes ...................................... 58 3 Probability as Fancy Counting and the q-ary Entropy Function 59 3.1 A Crash Course on Probability ............................... 59 3.2 The Probabilistic Method .................................. 65 3.3 The q-ary Entropy Function ................................. 65 3.4 Exercises ............................................ 72 3.5 Bibliographic Notes ...................................... 72 II The Combinatorics 73 4 WhatCanandCannotBeDone-I 75 5 4.1 Asymptotic Version of the Hamming Bound ....................... 75 4.2 Gilbert-Varshamov Bound .................................. 76 4.3 Singleton Bound ....................................... 80 4.4 Plotkin Bound ......................................... 82 4.5 Exercises ............................................ 87 4.6 Bibliographic Notes ...................................... 91 5 The Greatest Code of Them All: Reed-Solomon Codes 93 5.1 Polynomials and Finite Fields ................................ 93 5.2 Reed-Solomon Codes ..................................... 96 5.3 A Property of MDS Codes .................................. 99 5.4 Exercises ............................................100 5.5 Bibliographic Notes ......................................108 6 What Happens When the Noise is Stochastic: Shannon’s Theorem 109 6.1 Overview of Shannon’s Result ................................109 6.2 Shannon’s Noise Model ...................................110 6.3 Shannon’s Result for BSCp ..................................113 6.4 Hamming vs. Shannon ....................................121 6.5 Exercises ............................................122 6.6 Bibliographic Notes ......................................126 7 Bridging the Gap Between Shannon and Hamming: List Decoding 127 7.1 Hamming versus Shannon: part II .............................127 7.2 List Decoding .........................................129 7.3 Johnson Bound ........................................131 7.4 List-Decoding Capacity ...................................134 7.5 List Decoding from Random Errors ............................138 7.6 Exercises ............................................141 7.7 Bibliographic Notes ......................................146 8 What Cannot be Done-II 147 8.1 Elias-Bassalygo bound ....................................147 8.2 The MRRW bound: A better upper bound .........................149 8.3 A Breather ...........................................149 8.4 Bibliographic Notes ......................................150 III The Codes 151 9 When Polynomials Save the Day: Polynomial Based Codes 153 9.1 The generic construction ..................................154 9.2 The low degree case .....................................155 9.3 The case of the binary field .................................157 6 9.4 The general case ........................................158 9.5 Exercises ............................................165 9.6 Bibliographic Notes ......................................166 10 From Large to Small Alphabets: Code Concatenation 167 10.1 Code Concatenation .....................................168 10.2 Zyablov Bound ........................................169 10.3 Strongly Explicit Construction ...............................171 10.4 Exercises ............................................174 10.5 Bibliographic Notes ......................................174 11 Information Theory Strikes Back: Polar Codes 175 11.1 Achieving Gap to Capacity ..................................176 11.2 Reduction to Linear Compression .............................177 11.3 The Polarization Phenomenon ...............................178 11.4 Polar codes, Encoder and Decoder .............................184 11.5 Analysis: Speed of Polarization ...............................189 11.6 Entropic Calculations ....................................201 11.7 Summary and additional information ...........................204 11.8 Exercises ............................................205 11.9 Bibliographic Notes ......................................206 IV The Algorithms 209 12 Decoding Concatenated Codes 211 12.1 A Natural Decoding Algorithm ...............................211 12.2 Decoding From Errors and Erasures ............................214 12.3 Generalized Minimum Distance Decoding ........................215 12.4 Bibliographic Notes ......................................219 13 Efficiently Achieving the Capacity of the BSCp 221 13.1 Achieving capacity of BSCp .................................221 13.2 Decoding Error Probability .................................224 13.3 The Inner Code ........................................224 13.4 The Outer Code ........................................225 13.5 Discussion and Bibliographic Notes ............................227 14 Decoding Reed-Muller Codes 229 14.1 A natural decoding algorithm ................................229 14.2 Majority Logic Decoding ...................................236 14.3 Decoding by reduction to Reed-Solomon decoding ...................238 14.4 Exercises ............................................244 14.5 Bibliographic Notes ......................................246 7 15 Efficient Decoding of Reed-Solomon Codes 247 15.1 Unique decoding of Reed-Solomon codes ........................247 15.2 List Decoding Reed-Solomon Codes ............................252 15.3 Extensions ...........................................267 15.4 Bibliographic Notes ......................................269 16 Efficiently Achieving List Decoding Capacity 271 16.1 Folded Reed-Solomon Codes ................................271 16.2 List Decoding Folded Reed-Solomon Codes: I ......................275 16.3 List Decoding Folded Reed-Solomon Codes: II ......................278 16.4 Bibliographic Notes and Discussion ............................288 V The Applications 293 17 Cutting Data Down to Size: Hashing 295 17.1 Why Should You Care About Hashing? ...........................295 17.2 Avoiding Hash Collisions ...................................297 17.3 Almost Universal Hash Function Families and Codes ..................300 17.4 Data Possession Problem ..................................301 17.5 Bibliographic Notes ......................................305 18 Securing Your Fingerprints: Fuzzy Vaults 307 18.1 Some quick background on fingerprints .........................307 18.2 The Fuzzy Vault Problem ..................................309 18.3 The Final Fuzzy Vault .....................................312 18.4 Bibliographic Notes ......................................314 19 Finding Defectives: Group Testing 315 19.1 Formalization of the problem ................................315 19.2 Bounds on t a(d,N) ......................................317 19.3 Bounds on t(d,N) .......................................318 19.4 Coding Theory and Disjunct Matrices ...........................322 19.5 An Application in Data Stream Algorithms ........................325 19.6 Summary of best known bounds ..............................330 19.7 Exercises ............................................331 19.8 Bibliographic Notes ......................................333
Recommended publications
  • Compilers & Translator Writing Systems
    Compilers & Translators Compilers & Translator Writing Systems Prof. R. Eigenmann ECE573, Fall 2005 http://www.ece.purdue.edu/~eigenman/ECE573 ECE573, Fall 2005 1 Compilers are Translators Fortran Machine code C Virtual machine code C++ Transformed source code Java translate Augmented source Text processing language code Low-level commands Command Language Semantic components Natural language ECE573, Fall 2005 2 ECE573, Fall 2005, R. Eigenmann 1 Compilers & Translators Compilers are Increasingly Important Specification languages Increasingly high level user interfaces for ↑ specifying a computer problem/solution High-level languages ↑ Assembly languages The compiler is the translator between these two diverging ends Non-pipelined processors Pipelined processors Increasingly complex machines Speculative processors Worldwide “Grid” ECE573, Fall 2005 3 Assembly code and Assemblers assembly machine Compiler code Assembler code Assemblers are often used at the compiler back-end. Assemblers are low-level translators. They are machine-specific, and perform mostly 1:1 translation between mnemonics and machine code, except: – symbolic names for storage locations • program locations (branch, subroutine calls) • variable names – macros ECE573, Fall 2005 4 ECE573, Fall 2005, R. Eigenmann 2 Compilers & Translators Interpreters “Execute” the source language directly. Interpreters directly produce the result of a computation, whereas compilers produce executable code that can produce this result. Each language construct executes by invoking a subroutine of the interpreter, rather than a machine instruction. Examples of interpreters? ECE573, Fall 2005 5 Properties of Interpreters “execution” is immediate elaborate error checking is possible bookkeeping is possible. E.g. for garbage collection can change program on-the-fly. E.g., switch libraries, dynamic change of data types machine independence.
    [Show full text]
  • Modern Coding Theory: the Statistical Mechanics and Computer Science Point of View
    Modern Coding Theory: The Statistical Mechanics and Computer Science Point of View Andrea Montanari1 and R¨udiger Urbanke2 ∗ 1Stanford University, [email protected], 2EPFL, ruediger.urbanke@epfl.ch February 12, 2007 Abstract These are the notes for a set of lectures delivered by the two authors at the Les Houches Summer School on ‘Complex Systems’ in July 2006. They provide an introduction to the basic concepts in modern (probabilistic) coding theory, highlighting connections with statistical mechanics. We also stress common concepts with other disciplines dealing with similar problems that can be generically referred to as ‘large graphical models’. While most of the lectures are devoted to the classical channel coding problem over simple memoryless channels, we present a discussion of more complex channel models. We conclude with an overview of the main open challenges in the field. 1 Introduction and Outline The last few years have witnessed an impressive convergence of interests between disciplines which are a priori well separated: coding and information theory, statistical inference, statistical mechanics (in particular, mean field disordered systems), as well as theoretical computer science. The underlying reason for this convergence is the importance of probabilistic models and/or probabilistic techniques in each of these domains. This has long been obvious in information theory [53], statistical mechanics [10], and statistical inference [45]. In the last few years it has also become apparent in coding theory and theoretical computer science. In the first case, the invention of Turbo codes [7] and the re-invention arXiv:0704.2857v1 [cs.IT] 22 Apr 2007 of Low-Density Parity-Check (LDPC) codes [30, 28] has motivated the use of random constructions for coding information in robust/compact ways [50].
    [Show full text]
  • Chapter 2 Basics of Scanning And
    Chapter 2 Basics of Scanning and Conventional Programming in Java In this chapter, we will introduce you to an initial set of Java features, the equivalent of which you should have seen in your CS-1 class; the separation of problem, representation, algorithm and program – four concepts you have probably seen in your CS-1 class; style rules with which you are probably familiar, and scanning - a general class of problems we see in both computer science and other fields. Each chapter is associated with an animating recorded PowerPoint presentation and a YouTube video created from the presentation. It is meant to be a transcript of the associated presentation that contains little graphics and thus can be read even on a small device. You should refer to the associated material if you feel the need for a different instruction medium. Also associated with each chapter is hyperlinked code examples presented here. References to previously presented code modules are links that can be traversed to remind you of the details. The resources for this chapter are: PowerPoint Presentation YouTube Video Code Examples Algorithms and Representation Four concepts we explicitly or implicitly encounter while programming are problems, representations, algorithms and programs. Programs, of course, are instructions executed by the computer. Problems are what we try to solve when we write programs. Usually we do not go directly from problems to programs. Two intermediate steps are creating algorithms and identifying representations. Algorithms are sequences of steps to solve problems. So are programs. Thus, all programs are algorithms but the reverse is not true.
    [Show full text]
  • Scripting: Higher- Level Programming for the 21St Century
    . John K. Ousterhout Sun Microsystems Laboratories Scripting: Higher- Cybersquare Level Programming for the 21st Century Increases in computer speed and changes in the application mix are making scripting languages more and more important for the applications of the future. Scripting languages differ from system programming languages in that they are designed for “gluing” applications together. They use typeless approaches to achieve a higher level of programming and more rapid application development than system programming languages. or the past 15 years, a fundamental change has been ated with system programming languages and glued Foccurring in the way people write computer programs. together with scripting languages. However, several The change is a transition from system programming recent trends, such as faster machines, better script- languages such as C or C++ to scripting languages such ing languages, the increasing importance of graphical as Perl or Tcl. Although many people are participat- user interfaces (GUIs) and component architectures, ing in the change, few realize that the change is occur- and the growth of the Internet, have greatly expanded ring and even fewer know why it is happening. This the applicability of scripting languages. These trends article explains why scripting languages will handle will continue over the next decade, with more and many of the programming tasks in the next century more new applications written entirely in scripting better than system programming languages. languages and system programming
    [Show full text]
  • Modern Coding Theory: the Statistical Mechanics and Computer Science Point of View
    Modern Coding Theory: The Statistical Mechanics and Computer Science Point of View Andrea Montanari1 and R¨udiger Urbanke2 ∗ 1Stanford University, [email protected], 2EPFL, ruediger.urbanke@epfl.ch February 12, 2007 Abstract These are the notes for a set of lectures delivered by the two authors at the Les Houches Summer School on ‘Complex Systems’ in July 2006. They provide an introduction to the basic concepts in modern (probabilistic) coding theory, highlighting connections with statistical mechanics. We also stress common concepts with other disciplines dealing with similar problems that can be generically referred to as ‘large graphical models’. While most of the lectures are devoted to the classical channel coding problem over simple memoryless channels, we present a discussion of more complex channel models. We conclude with an overview of the main open challenges in the field. 1 Introduction and Outline The last few years have witnessed an impressive convergence of interests between disciplines which are a priori well separated: coding and information theory, statistical inference, statistical mechanics (in particular, mean field disordered systems), as well as theoretical computer science. The underlying reason for this convergence is the importance of probabilistic models and/or probabilistic techniques in each of these domains. This has long been obvious in information theory [53], statistical mechanics [10], and statistical inference [45]. In the last few years it has also become apparent in coding theory and theoretical computer science. In the first case, the invention of Turbo codes [7] and the re-invention of Low-Density Parity-Check (LDPC) codes [30, 28] has motivated the use of random constructions for coding information in robust/compact ways [50].
    [Show full text]
  • The Packing Problem in Statistics, Coding Theory and Finite Projective Spaces: Update 2001
    The packing problem in statistics, coding theory and finite projective spaces: update 2001 J.W.P. Hirschfeld School of Mathematical Sciences University of Sussex Brighton BN1 9QH United Kingdom http://www.maths.sussex.ac.uk/Staff/JWPH [email protected] L. Storme Department of Pure Mathematics Krijgslaan 281 Ghent University B-9000 Gent Belgium http://cage.rug.ac.be/ ls ∼ [email protected] Abstract This article updates the authors’ 1998 survey [133] on the same theme that was written for the Bose Memorial Conference (Colorado, June 7-11, 1995). That article contained the principal results on the packing problem, up to 1995. Since then, considerable progress has been made on different kinds of subconfigurations. 1 Introduction 1.1 The packing problem The packing problem in statistics, coding theory and finite projective spaces regards the deter- mination of the maximal or minimal sizes of given subconfigurations of finite projective spaces. This problem is not only interesting from a geometrical point of view; it also arises when coding-theoretical problems and problems from the design of experiments are translated into equivalent geometrical problems. The geometrical interest in the packing problem and the links with problems investigated in other research fields have given this problem a central place in Galois geometries, that is, the study of finite projective spaces. In 1983, a historical survey on the packing problem was written by the first author [126] for the 9th British Combinatorial Conference. A new survey article stating the principal results up to 1995 was written by the authors for the Bose Memorial Conference [133].
    [Show full text]
  • Arxiv:1111.6055V1 [Math.CO] 25 Nov 2011 2 Definitions
    Adinkras for Mathematicians Yan X Zhang Massachusetts Institute of Technology November 28, 2011 Abstract Adinkras are graphical tools created for the study of representations in supersym- metry. Besides having inherent interest for physicists, adinkras offer many easy-to-state and accessible mathematical problems of algebraic, combinatorial, and computational nature. We use a more mathematically natural language to survey these topics, suggest new definitions, and present original results. 1 Introduction In a series of papers starting with [8], different subsets of the “DFGHILM collaboration” (Doran, Faux, Gates, Hübsch, Iga, Landweber, Miller) have built and extended the ma- chinery of adinkras. Following the spirit of Feyman diagrams, adinkras are combinatorial objects that encode information about the representation theory of supersymmetry alge- bras. Adinkras have many intricate links with other fields such as graph theory, Clifford theory, and coding theory. Each of these connections provide many problems that can be compactly communicated to a (non-specialist) mathematician. This paper is a humble attempt to bridge the language gap and generate communication. We redevelop the foundations in a self-contained manner in Sections 2 and 4, using different definitions and constructions that we consider to be more mathematically natural for our purposes. Using our new setup, we prove some original results and make new interpretations in Sections 5 and 6. We wish that these purely combinatorial discussions will equip the readers with a mental model that allows them to appreciate (or to solve!) the original representation-theoretic problems in the physics literature. We return to these problems in Section 7 and reconsider some of the foundational questions of the theory.
    [Show full text]
  • The Use of Coding Theory in Computational Complexity 1 Introduction
    The Use of Co ding Theory in Computational Complexity Joan Feigenbaum ATT Bell Lab oratories Murray Hill NJ USA jfresearchattcom Abstract The interplay of co ding theory and computational complexity theory is a rich source of results and problems This article surveys three of the ma jor themes in this area the use of co des to improve algorithmic eciency the theory of program testing and correcting which is a complexity theoretic analogue of error detection and correction the use of co des to obtain characterizations of traditional complexity classes such as NP and PSPACE these new characterizations are in turn used to show that certain combinatorial optimization problems are as hard to approximate closely as they are to solve exactly Intro duction Complexity theory is the study of ecient computation Faced with a computational problem that can b e mo delled formally a complexity theorist seeks rst to nd a solution that is provably ecient and if such a solution is not found to prove that none exists Co ding theory which provides techniques for robust representation of information is valuable b oth in designing ecient solutions and in proving that ecient solutions do not exist This article surveys the use of co des in complexity theory The improved upp er b ounds obtained with co ding techniques include b ounds on the numb er of random bits used by probabilistic algo rithms and the communication complexity of cryptographic proto cols In these constructions co des are used to design small sample spaces that approximate the b ehavior
    [Show full text]
  • Algorithmic Coding Theory ¡
    ALGORITHMIC CODING THEORY Atri Rudra State University of New York at Buffalo 1 Introduction Error-correcting codes (or just codes) are clever ways of representing data so that one can recover the original information even if parts of it are corrupted. The basic idea is to judiciously introduce redundancy so that the original information can be recovered when parts of the (redundant) data have been corrupted. Perhaps the most natural and common application of error correcting codes is for communication. For example, when packets are transmitted over the Internet, some of the packets get corrupted or dropped. To deal with this, multiple layers of the TCP/IP stack use a form of error correction called CRC Checksum [Pe- terson and Davis, 1996]. Codes are used when transmitting data over the telephone line or via cell phones. They are also used in deep space communication and in satellite broadcast. Codes also have applications in areas not directly related to communication. For example, codes are used heavily in data storage. CDs and DVDs can work even in the presence of scratches precisely because they use codes. Codes are used in Redundant Array of Inexpensive Disks (RAID) [Chen et al., 1994] and error correcting memory [Chen and Hsiao, 1984]. Codes are also deployed in other applications such as paper bar codes, for example, the bar ¡ To appear as a book chapter in the CRC Handbook on Algorithms and Complexity Theory edited by M. Atallah and M. Blanton. Please send your comments to [email protected] 1 code used by UPS called MaxiCode [Chandler et al., 1989].
    [Show full text]
  • An Introduction to Algorithmic Coding Theory
    An Introduction to Algorithmic Coding Theory M. Amin Shokrollahi Bell Laboratories 1 Part 1: Codes 1-1 A puzzle What do the following problems have in common? 2 Problem 1: Information Transmission G W D A S U N C L R V MESSAGE J K E MASSAGE ?? F T X M I H Z B Y 3 Problem 2: Football Pool Problem Bayern M¨unchen : 1. FC Kaiserslautern 0 Homburg : Hannover 96 1 1860 M¨unchen : FC Karlsruhe 2 1. FC K¨oln : Hertha BSC 0 Wolfsburg : Stuttgart 0 Bremen : Unterhaching 2 Frankfurt : Rostock 0 Bielefeld : Duisburg 1 Feiburg : Schalke 04 2 4 Problem 3: In-Memory Database Systems Database Corrupted! User memory malloc malloc malloc malloc Code memory 5 Problem 4: Bulk Data Distribution 6 What do they have in common? They all can be solved using (algorithmic) coding theory 7 Basic Idea of Coding Adding redundancy to be able to correct! MESSAGE MESSAGE MESSAGE MASSAGE Objectives: • Add as little redundancy as possible; • correct as many errors as possible. 8 Codes A code of block-length n over the alphabet GF (q) is a set of vectors in GF (q)n. If the set forms a vector space over GF (q), then the code is called linear. [n, k]q-code 9 Encoding Problem ? ? ? source ? ? ? Codewords Efficiency! 10 Linear Codes: Generator Matrix Generator matrix source codeword k n k x n O(n2) 11 Linear Codes: Parity Check Matrix 0 Parity check matrix codeword O(n2) after O(n3) preprocessing. 12 The Decoding Problem: Maximum Likelihood Decoding abcd 7→ abcd | abcd | abcd Received: a x c d | a b z d | abcd.
    [Show full text]
  • The Future of DNA Data Storage the Future of DNA Data Storage
    The Future of DNA Data Storage The Future of DNA Data Storage September 2018 A POTOMAC INSTITUTE FOR POLICY STUDIES REPORT AC INST M IT O U T B T The Future O E P F O G S R IE of DNA P D O U Data LICY ST Storage September 2018 NOTICE: This report is a product of the Potomac Institute for Policy Studies. The conclusions of this report are our own, and do not necessarily represent the views of our sponsors or participants. Many thanks to the Potomac Institute staff and experts who reviewed and provided comments on this report. © 2018 Potomac Institute for Policy Studies Cover image: Alex Taliesen POTOMAC INSTITUTE FOR POLICY STUDIES 901 North Stuart St., Suite 1200 | Arlington, VA 22203 | 703-525-0770 | www.potomacinstitute.org CONTENTS EXECUTIVE SUMMARY 4 Findings 5 BACKGROUND 7 Data Storage Crisis 7 DNA as a Data Storage Medium 9 Advantages 10 History 11 CURRENT STATE OF DNA DATA STORAGE 13 Technology of DNA Data Storage 13 Writing Data to DNA 13 Reading Data from DNA 18 Key Players in DNA Data Storage 20 Academia 20 Research Consortium 21 Industry 21 Start-ups 21 Government 22 FORECAST OF DNA DATA STORAGE 23 DNA Synthesis Cost Forecast 23 Forecast for DNA Data Storage Tech Advancement 28 Increasing Data Storage Density in DNA 29 Advanced Coding Schemes 29 DNA Sequencing Methods 30 DNA Data Retrieval 31 CONCLUSIONS 32 ENDNOTES 33 Executive Summary The demand for digital data storage is currently has been developed to support applications in outpacing the world’s storage capabilities, and the life sciences industry and not for data storage the gap is widening as the amount of digital purposes.
    [Show full text]
  • Voting Rules As Error-Correcting Codes ∗ Ariel D
    Artificial Intelligence 231 (2016) 1–16 Contents lists available at ScienceDirect Artificial Intelligence www.elsevier.com/locate/artint ✩ Voting rules as error-correcting codes ∗ Ariel D. Procaccia, Nisarg Shah , Yair Zick Computer Science Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA a r t i c l e i n f o a b s t r a c t Article history: We present the first model of optimal voting under adversarial noise. From this viewpoint, Received 21 January 2015 voting rules are seen as error-correcting codes: their goal is to correct errors in the input Received in revised form 13 October 2015 rankings and recover a ranking that is close to the ground truth. We derive worst-case Accepted 20 October 2015 bounds on the relation between the average accuracy of the input votes, and the accuracy Available online 27 October 2015 of the output ranking. Empirical results from real data show that our approach produces Keywords: significantly more accurate rankings than alternative approaches. © Social choice 2015 Elsevier B.V. All rights reserved. Voting Ground truth Adversarial noise Error-correcting codes 1. Introduction Social choice theory develops and analyzes methods for aggregating the opinions of individuals into a collective decision. The prevalent approach is motivated by situations in which opinions are subjective, such as political elections, and focuses on the design of voting rules that satisfy normative properties [1]. An alternative approach, which was proposed by the marquis de Condorcet in the 18th Century, had confounded schol- ars for centuries (due to Condorcet’s ambiguous writing) until it was finally elucidated by Young [34].
    [Show full text]