Statistical Physics and Information Theory Full Text Available At

Total Page:16

File Type:pdf, Size:1020Kb

Statistical Physics and Information Theory Full Text Available At Full text available at: http://dx.doi.org/10.1561/0100000052 Statistical Physics and Information Theory Full text available at: http://dx.doi.org/10.1561/0100000052 Statistical Physics and Information Theory Neri Merhav Department of Electrical Engineering Technion | Israel Institute of Technology Haifa 32000 Israel [email protected] Boston { Delft Full text available at: http://dx.doi.org/10.1561/0100000052 Foundations and Trends R in Communications and Information Theory Published, sold and distributed by: now Publishers Inc. PO Box 1024 Hanover, MA 02339 USA Tel. +1-781-985-4510 www.nowpublishers.com [email protected] Outside North America: now Publishers Inc. PO Box 179 2600 AD Delft The Netherlands Tel. +31-6-51115274 The preferred citation for this publication is N. Merhav, Statistical Physics and Information Theory, Foundations and Trends R in Communications and Information Theory, vol 6, nos 1{2, pp 1{212, 2009 ISBN: 978-1-60198-406-7 c 2010 N. Merhav All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, mechanical, photocopying, recording or otherwise, without prior written permission of the publishers. Photocopying. In the USA: This journal is registered at the Copyright Clearance Cen- ter, Inc., 222 Rosewood Drive, Danvers, MA 01923. Authorization to photocopy items for internal or personal use, or the internal or personal use of specific clients, is granted by now Publishers Inc for users registered with the Copyright Clearance Center (CCC). The `services' for users can be found on the internet at: www.copyright.com For those organizations that have been granted a photocopy license, a separate system of payment has been arranged. Authorization does not extend to other kinds of copy- ing, such as that for general distribution, for advertising or promotional purposes, for creating new collective works, or for resale. In the rest of the world: Permission to pho- tocopy must be obtained from the copyright owner. Please apply to now Publishers Inc., PO Box 1024, Hanover, MA 02339, USA; Tel. +1-781-871-0245; www.nowpublishers.com; [email protected] now Publishers Inc. has an exclusive license to publish this material worldwide. Permission to use this content must be obtained from the copyright license holder. Please apply to now Publishers, PO Box 179, 2600 AD Delft, The Netherlands, www.nowpublishers.com; e-mail: [email protected] Full text available at: http://dx.doi.org/10.1561/0100000052 Foundations and Trends R in Communications and Information Theory Volume 6 Issues 1{2, 2009 Editorial Board Editor-in-Chief: Sergio Verd´u Depart of Electrical Engineering Princeton University Princeton, New Jersey 08544 Editors Venkat Anantharam (UC. Berkeley) Amos Lapidoth (ETH Zurich) Ezio Biglieri (U. Torino) Bob McEliece (Caltech) Giuseppe Caire (U. Southern Neri Merhav (Technion) California) David Neuhoff (U. Michigan) Roger Cheng (U. Hong Kong) Alon Orlitsky (UC. San Diego) K.C. Chen (Taipei) Vincent Poor (Princeton) Daniel Costello (U. Notre Dame) Kannan Ramchandran (UC. Thomas Cover (Stanford) Berkeley) Anthony Ephremides (U. Maryland) Bixio Rimoldi (EPFL) Andrea Goldsmith (Stanford) Shlomo Shamai (Technion) Dave Forney (MIT) Amin Shokrollahi (EPFL) Georgios Giannakis (U. Minnesota) Gadiel Seroussi (MSRI) Joachim Hagenauer (TU Munich) Wojciech Szpankowski (Purdue) Te Sun Han (Tokyo) Vahid Tarokh (Harvard) Babak Hassibi (Caltech) David Tse (UC. Berkeley) Michael Honig (Northwestern) Ruediger Urbanke (EPFL) Johannes Huber (Erlangen) Steve Wicker (Cornell) Hideki Imai (Tokyo) Raymond Yeung (Hong Kong) Rodney Kennedy (Canberra) Bin Yu (UC. Berkeley) Sanjeev Kulkarni (Princeton) Full text available at: http://dx.doi.org/10.1561/0100000052 Editorial Scope Foundations and Trends R in Communications and Informa- tion Theory will publish survey and tutorial articles in the following topics: • Coded modulation • Multiuser detection • Coding theory and practice • Multiuser information theory • Communication complexity • Optical communication channels • Communication system design • Pattern recognition and learning • Cryptology and data security • Quantization • Data compression • Quantum information processing • Data networks • Rate-distortion theory • Demodulation and Equalization • Shannon theory • Denoising • Signal processing for • Detection and estimation communications • Information theory and statistics • Source coding • Information theory and computer • Storage and recording codes science • Speech and Image Compression • Joint source/channel coding • Wireless Communications • Modulation and signal design Information for Librarians Foundations and Trends R in Communications and Information Theory, 2009, Volume 6, 6 issues. ISSN paper version 1567-2190. ISSN online version 1567- 2328. Also available as a combined paper and online subscription. Full text available at: http://dx.doi.org/10.1561/0100000052 Foundations and Trends R in Communications and Information Theory Vol. 6, Nos. 1{2 (2009) 1{212 c 2010 N. Merhav DOI: 10.1561/0100000052 Statistical Physics and Information Theory Neri Merhav Department of Electrical Engineering, Technion | Israel Institute of Technology, Haifa 32000, Israel, [email protected] Abstract This monograph is based on lecture notes of a graduate course, which focuses on the relations between information theory and statistical physics. The course was delivered at the Technion during the Spring of 2010 for the first time, and its target audience consists of EE graduate students in the area of communications and information theory, as well as graduate students in Physics who have basic background in infor- mation theory. Strong emphasis is given to the analogy and parallelism between information theory and statistical physics, as well as to the insights, the analysis tools and techniques that can be borrowed from statistical physics and `imported' to certain problem areas in informa- tion theory. This is a research trend that has been very active in the last few decades, and the hope is that by exposing the students to the meeting points between these two disciplines, their background and perspective may be expanded and enhanced. This monograph is sub- stantially revised and expanded relative to an earlier version posted in arXiv (1006.1565v1 [cs.iT]). Full text available at: http://dx.doi.org/10.1561/0100000052 Contents 1 Introduction 1 2 Basic Background in Statistical Physics 7 2.1 What is Statistical Physics? 7 2.2 Basic Postulates and the Microcanonical Ensemble 8 2.3 The Canonical Ensemble 17 2.4 Properties of the Partition Function and the Free Energy 21 2.5 The Energy Equipartition Theorem 32 2.6 The Grand-Canonical Ensemble 34 3 Physical Interpretations of Information Measures 39 3.1 Statistical Physics of Optimum Message Distributions 40 3.2 Large Deviations and Physics of Coding Theorems 42 3.3 Gibbs' Inequality and the Second Law 57 3.4 Boltzmann's H-Theorem and the DPT 73 3.5 Generalized Temperature and Fisher Information 88 4 Analysis Tools and Asymptotic Methods 95 4.1 Introduction 95 4.2 The Laplace Method 97 4.3 The Saddle-Point Method 101 ix Full text available at: http://dx.doi.org/10.1561/0100000052 4.4 Extended Example: Capacity of a Disordered System 111 4.5 The Replica Method 116 5 Interacting Particles and Phase Transitions 123 5.1 Introduction | Sources of Interaction 124 5.2 Models of Interacting Particles 125 5.3 A Qualitative Discussion on Phase Transitions 131 5.4 Phase Transitions of the Rate{Distortion Function 135 5.5 The One-Dimensional Ising Model 139 5.6 The Curie{Weiss Model 142 5.7 Spin Glasses and Random Code Ensembles 147 6 The Random Energy Model and Random Coding 153 6.1 REM without a Magnetic Field 153 6.2 Random Code Ensembles and the REM 159 6.3 Random Coding Exponents 166 7 Extensions of the REM 177 7.1 REM Under Magnetic Field and Source{Channel Coding 178 7.2 Generalized REM (GREM) and Hierarchical Coding 187 7.3 Directed Polymers in a Random Medium and Tree Codes 198 8 Summary and Outlook 203 Acknowledgments 205 References 207 Full text available at: http://dx.doi.org/10.1561/0100000052 1 Introduction This work focuses on some of the relationships and the interplay between information theory and statistical physics | a branch of physics that deals with many-particle systems using probabilistic and statistical methods in the microscopic level. The relationships between information theory and statistical ther- modynamics are by no means new, and many researchers have been exploiting them for many years. Perhaps the first relation, or analogy, that crosses one's mind is that in both fields there is a fundamental notion of entropy. Actually, in information theory, the term entropy was coined in the footsteps of the thermodynamic entropy. The ther- modynamic entropy was first introduced by Clausius in 1850, and its probabilistic-statistical interpretation was established by Boltzmann in 1872. It is virtually impossible to miss the functional resemblance between the two notions of entropy, and indeed it was recognized by Shannon and von Neumann. The well-known anecdote on this tells that von Neumann advised Shannon to adopt this term because it would provide him with \. a great edge in debates because nobody really knows what entropy is anyway." But the relationships between the two fields go far beyond the fact that both share the notion of entropy. In fact, these relationships have 1 Full text available at: http://dx.doi.org/10.1561/0100000052
Recommended publications
  • Modern Coding Theory: the Statistical Mechanics and Computer Science Point of View
    Modern Coding Theory: The Statistical Mechanics and Computer Science Point of View Andrea Montanari1 and R¨udiger Urbanke2 ∗ 1Stanford University, [email protected], 2EPFL, ruediger.urbanke@epfl.ch February 12, 2007 Abstract These are the notes for a set of lectures delivered by the two authors at the Les Houches Summer School on ‘Complex Systems’ in July 2006. They provide an introduction to the basic concepts in modern (probabilistic) coding theory, highlighting connections with statistical mechanics. We also stress common concepts with other disciplines dealing with similar problems that can be generically referred to as ‘large graphical models’. While most of the lectures are devoted to the classical channel coding problem over simple memoryless channels, we present a discussion of more complex channel models. We conclude with an overview of the main open challenges in the field. 1 Introduction and Outline The last few years have witnessed an impressive convergence of interests between disciplines which are a priori well separated: coding and information theory, statistical inference, statistical mechanics (in particular, mean field disordered systems), as well as theoretical computer science. The underlying reason for this convergence is the importance of probabilistic models and/or probabilistic techniques in each of these domains. This has long been obvious in information theory [53], statistical mechanics [10], and statistical inference [45]. In the last few years it has also become apparent in coding theory and theoretical computer science. In the first case, the invention of Turbo codes [7] and the re-invention arXiv:0704.2857v1 [cs.IT] 22 Apr 2007 of Low-Density Parity-Check (LDPC) codes [30, 28] has motivated the use of random constructions for coding information in robust/compact ways [50].
    [Show full text]
  • Modern Coding Theory: the Statistical Mechanics and Computer Science Point of View
    Modern Coding Theory: The Statistical Mechanics and Computer Science Point of View Andrea Montanari1 and R¨udiger Urbanke2 ∗ 1Stanford University, [email protected], 2EPFL, ruediger.urbanke@epfl.ch February 12, 2007 Abstract These are the notes for a set of lectures delivered by the two authors at the Les Houches Summer School on ‘Complex Systems’ in July 2006. They provide an introduction to the basic concepts in modern (probabilistic) coding theory, highlighting connections with statistical mechanics. We also stress common concepts with other disciplines dealing with similar problems that can be generically referred to as ‘large graphical models’. While most of the lectures are devoted to the classical channel coding problem over simple memoryless channels, we present a discussion of more complex channel models. We conclude with an overview of the main open challenges in the field. 1 Introduction and Outline The last few years have witnessed an impressive convergence of interests between disciplines which are a priori well separated: coding and information theory, statistical inference, statistical mechanics (in particular, mean field disordered systems), as well as theoretical computer science. The underlying reason for this convergence is the importance of probabilistic models and/or probabilistic techniques in each of these domains. This has long been obvious in information theory [53], statistical mechanics [10], and statistical inference [45]. In the last few years it has also become apparent in coding theory and theoretical computer science. In the first case, the invention of Turbo codes [7] and the re-invention of Low-Density Parity-Check (LDPC) codes [30, 28] has motivated the use of random constructions for coding information in robust/compact ways [50].
    [Show full text]
  • The Packing Problem in Statistics, Coding Theory and Finite Projective Spaces: Update 2001
    The packing problem in statistics, coding theory and finite projective spaces: update 2001 J.W.P. Hirschfeld School of Mathematical Sciences University of Sussex Brighton BN1 9QH United Kingdom http://www.maths.sussex.ac.uk/Staff/JWPH [email protected] L. Storme Department of Pure Mathematics Krijgslaan 281 Ghent University B-9000 Gent Belgium http://cage.rug.ac.be/ ls ∼ [email protected] Abstract This article updates the authors’ 1998 survey [133] on the same theme that was written for the Bose Memorial Conference (Colorado, June 7-11, 1995). That article contained the principal results on the packing problem, up to 1995. Since then, considerable progress has been made on different kinds of subconfigurations. 1 Introduction 1.1 The packing problem The packing problem in statistics, coding theory and finite projective spaces regards the deter- mination of the maximal or minimal sizes of given subconfigurations of finite projective spaces. This problem is not only interesting from a geometrical point of view; it also arises when coding-theoretical problems and problems from the design of experiments are translated into equivalent geometrical problems. The geometrical interest in the packing problem and the links with problems investigated in other research fields have given this problem a central place in Galois geometries, that is, the study of finite projective spaces. In 1983, a historical survey on the packing problem was written by the first author [126] for the 9th British Combinatorial Conference. A new survey article stating the principal results up to 1995 was written by the authors for the Bose Memorial Conference [133].
    [Show full text]
  • Arxiv:1111.6055V1 [Math.CO] 25 Nov 2011 2 Definitions
    Adinkras for Mathematicians Yan X Zhang Massachusetts Institute of Technology November 28, 2011 Abstract Adinkras are graphical tools created for the study of representations in supersym- metry. Besides having inherent interest for physicists, adinkras offer many easy-to-state and accessible mathematical problems of algebraic, combinatorial, and computational nature. We use a more mathematically natural language to survey these topics, suggest new definitions, and present original results. 1 Introduction In a series of papers starting with [8], different subsets of the “DFGHILM collaboration” (Doran, Faux, Gates, Hübsch, Iga, Landweber, Miller) have built and extended the ma- chinery of adinkras. Following the spirit of Feyman diagrams, adinkras are combinatorial objects that encode information about the representation theory of supersymmetry alge- bras. Adinkras have many intricate links with other fields such as graph theory, Clifford theory, and coding theory. Each of these connections provide many problems that can be compactly communicated to a (non-specialist) mathematician. This paper is a humble attempt to bridge the language gap and generate communication. We redevelop the foundations in a self-contained manner in Sections 2 and 4, using different definitions and constructions that we consider to be more mathematically natural for our purposes. Using our new setup, we prove some original results and make new interpretations in Sections 5 and 6. We wish that these purely combinatorial discussions will equip the readers with a mental model that allows them to appreciate (or to solve!) the original representation-theoretic problems in the physics literature. We return to these problems in Section 7 and reconsider some of the foundational questions of the theory.
    [Show full text]
  • The Use of Coding Theory in Computational Complexity 1 Introduction
    The Use of Co ding Theory in Computational Complexity Joan Feigenbaum ATT Bell Lab oratories Murray Hill NJ USA jfresearchattcom Abstract The interplay of co ding theory and computational complexity theory is a rich source of results and problems This article surveys three of the ma jor themes in this area the use of co des to improve algorithmic eciency the theory of program testing and correcting which is a complexity theoretic analogue of error detection and correction the use of co des to obtain characterizations of traditional complexity classes such as NP and PSPACE these new characterizations are in turn used to show that certain combinatorial optimization problems are as hard to approximate closely as they are to solve exactly Intro duction Complexity theory is the study of ecient computation Faced with a computational problem that can b e mo delled formally a complexity theorist seeks rst to nd a solution that is provably ecient and if such a solution is not found to prove that none exists Co ding theory which provides techniques for robust representation of information is valuable b oth in designing ecient solutions and in proving that ecient solutions do not exist This article surveys the use of co des in complexity theory The improved upp er b ounds obtained with co ding techniques include b ounds on the numb er of random bits used by probabilistic algo rithms and the communication complexity of cryptographic proto cols In these constructions co des are used to design small sample spaces that approximate the b ehavior
    [Show full text]
  • Algorithmic Coding Theory ¡
    ALGORITHMIC CODING THEORY Atri Rudra State University of New York at Buffalo 1 Introduction Error-correcting codes (or just codes) are clever ways of representing data so that one can recover the original information even if parts of it are corrupted. The basic idea is to judiciously introduce redundancy so that the original information can be recovered when parts of the (redundant) data have been corrupted. Perhaps the most natural and common application of error correcting codes is for communication. For example, when packets are transmitted over the Internet, some of the packets get corrupted or dropped. To deal with this, multiple layers of the TCP/IP stack use a form of error correction called CRC Checksum [Pe- terson and Davis, 1996]. Codes are used when transmitting data over the telephone line or via cell phones. They are also used in deep space communication and in satellite broadcast. Codes also have applications in areas not directly related to communication. For example, codes are used heavily in data storage. CDs and DVDs can work even in the presence of scratches precisely because they use codes. Codes are used in Redundant Array of Inexpensive Disks (RAID) [Chen et al., 1994] and error correcting memory [Chen and Hsiao, 1984]. Codes are also deployed in other applications such as paper bar codes, for example, the bar ¡ To appear as a book chapter in the CRC Handbook on Algorithms and Complexity Theory edited by M. Atallah and M. Blanton. Please send your comments to [email protected] 1 code used by UPS called MaxiCode [Chandler et al., 1989].
    [Show full text]
  • An Introduction to Algorithmic Coding Theory
    An Introduction to Algorithmic Coding Theory M. Amin Shokrollahi Bell Laboratories 1 Part 1: Codes 1-1 A puzzle What do the following problems have in common? 2 Problem 1: Information Transmission G W D A S U N C L R V MESSAGE J K E MASSAGE ?? F T X M I H Z B Y 3 Problem 2: Football Pool Problem Bayern M¨unchen : 1. FC Kaiserslautern 0 Homburg : Hannover 96 1 1860 M¨unchen : FC Karlsruhe 2 1. FC K¨oln : Hertha BSC 0 Wolfsburg : Stuttgart 0 Bremen : Unterhaching 2 Frankfurt : Rostock 0 Bielefeld : Duisburg 1 Feiburg : Schalke 04 2 4 Problem 3: In-Memory Database Systems Database Corrupted! User memory malloc malloc malloc malloc Code memory 5 Problem 4: Bulk Data Distribution 6 What do they have in common? They all can be solved using (algorithmic) coding theory 7 Basic Idea of Coding Adding redundancy to be able to correct! MESSAGE MESSAGE MESSAGE MASSAGE Objectives: • Add as little redundancy as possible; • correct as many errors as possible. 8 Codes A code of block-length n over the alphabet GF (q) is a set of vectors in GF (q)n. If the set forms a vector space over GF (q), then the code is called linear. [n, k]q-code 9 Encoding Problem ? ? ? source ? ? ? Codewords Efficiency! 10 Linear Codes: Generator Matrix Generator matrix source codeword k n k x n O(n2) 11 Linear Codes: Parity Check Matrix 0 Parity check matrix codeword O(n2) after O(n3) preprocessing. 12 The Decoding Problem: Maximum Likelihood Decoding abcd 7→ abcd | abcd | abcd Received: a x c d | a b z d | abcd.
    [Show full text]
  • Vector Quantization and Signal Compression the Kluwer International Series in Engineering and Computer Science
    VECTOR QUANTIZATION AND SIGNAL COMPRESSION THE KLUWER INTERNATIONAL SERIES IN ENGINEERING AND COMPUTER SCIENCE COMMUNICATIONS AND INFORMATION THEORY Consulting Editor: Robert Gallager Other books in the series: Digital Communication. Edward A. Lee, David G. Messerschmitt ISBN: 0-89838-274-2 An Introduction to Cryptolog)'. Henk c.A. van Tilborg ISBN: 0-89838-271-8 Finite Fields for Computer Scientists and Engineers. Robert J. McEliece ISBN: 0-89838-191-6 An Introduction to Error Correcting Codes With Applications. Scott A. Vanstone and Paul C. van Oorschot ISBN: 0-7923-9017-2 Source Coding Theory.. Robert M. Gray ISBN: 0-7923-9048-2 Switching and TraffIC Theory for Integrated BroadbandNetworks. Joseph Y. Hui ISBN: 0-7923-9061-X Advances in Speech Coding, Bishnu Atal, Vladimir Cuperman and Allen Gersho ISBN: 0-7923-9091-1 Coding: An Algorithmic Approach, John B. Anderson and Seshadri Mohan ISBN: 0-7923-9210-8 Third Generation Wireless Information Networks, edited by Sanjiv Nanda and David J. Goodman ISBN: 0-7923-9128-3 VECTOR QUANTIZATION AND SIGNAL COMPRESSION by Allen Gersho University of California, Santa Barbara Robert M. Gray Stanford University ..... SPRINGER SCIENCE+BUSINESS" MEDIA, LLC Library of Congress Cataloging-in-Publication Data Gersho, A1len. Vector quantization and signal compression / by Allen Gersho, Robert M. Gray. p. cm. -- (K1uwer international series in engineering and computer science ; SECS 159) Includes bibliographical references and index. ISBN 978-1-4613-6612-6 ISBN 978-1-4615-3626-0 (eBook) DOI 10.1007/978-1-4615-3626-0 1. Signal processing--Digital techniques. 2. Data compression (Telecommunication) 3. Coding theory. 1. Gray, Robert M., 1943- .
    [Show full text]
  • Voting Rules As Error-Correcting Codes ∗ Ariel D
    Artificial Intelligence 231 (2016) 1–16 Contents lists available at ScienceDirect Artificial Intelligence www.elsevier.com/locate/artint ✩ Voting rules as error-correcting codes ∗ Ariel D. Procaccia, Nisarg Shah , Yair Zick Computer Science Department, Carnegie Mellon University, Pittsburgh, PA 15213, USA a r t i c l e i n f o a b s t r a c t Article history: We present the first model of optimal voting under adversarial noise. From this viewpoint, Received 21 January 2015 voting rules are seen as error-correcting codes: their goal is to correct errors in the input Received in revised form 13 October 2015 rankings and recover a ranking that is close to the ground truth. We derive worst-case Accepted 20 October 2015 bounds on the relation between the average accuracy of the input votes, and the accuracy Available online 27 October 2015 of the output ranking. Empirical results from real data show that our approach produces Keywords: significantly more accurate rankings than alternative approaches. © Social choice 2015 Elsevier B.V. All rights reserved. Voting Ground truth Adversarial noise Error-correcting codes 1. Introduction Social choice theory develops and analyzes methods for aggregating the opinions of individuals into a collective decision. The prevalent approach is motivated by situations in which opinions are subjective, such as political elections, and focuses on the design of voting rules that satisfy normative properties [1]. An alternative approach, which was proposed by the marquis de Condorcet in the 18th Century, had confounded schol- ars for centuries (due to Condorcet’s ambiguous writing) until it was finally elucidated by Young [34].
    [Show full text]
  • Adinkras for Mathematicians
    TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 366, Number 6, June 2014, Pages 3325–3355 S 0002-9947(2014)06031-5 Article electronically published on February 6, 2014 ADINKRAS FOR MATHEMATICIANS YAN X. ZHANG Abstract. Adinkras are graphical tools created to study representations of supersymmetry algebras. Besides having inherent interest for physicists, the study of adinkras has already shown non-trivial connections with coding the- ory and Clifford algebras. Furthermore, adinkras offer many easy-to-state and accessible mathematical problems of algebraic, combinatorial, and computa- tional nature. We survey these topics for a mathematical audience, make new connections to other areas (homological algebra and poset theory), and solve some of these said problems, including the enumeration of all hyper- cube adinkras up through dimension 5 and the enumeration of odd dashings of adinkras for any dimension. 1. Introduction In a series of papers starting with [11], different subsets of the “DFGHILM collaboration” (Doran, Faux, Gates, H¨ubsch, Iga, Landweber, Miller) have built and extended the machinery of adinkras. Following the ubiquitous spirit of visual diagrams in physics, adinkras are combinatorial objects that encode information about the representation theory of supersymmetry algebras. Adinkras have many intricate links with other fields such as graph theory, Clifford theory, and coding theory. Each of these connections provide many problems that can be compactly communicated to a (non-specialist) mathematician. This paper is a humble attempt to bridge the language gap and generate communication. In short, adinkras are chromotopologies (a class of edge-colored bipartite graphs) with two additional structures, a dashing condition on the edges and a ranking condition on the vertices.
    [Show full text]
  • Group Algebra and Coding Theory
    Group algebras and Coding Theory Marinˆes Guerreiro ∗† September 7, 2015 Abstract Group algebras have been used in the context of Coding Theory since the beginning of the latter, but not in its full power. The work of Ferraz and Polcino Milies entitled Idempotents in group algebras and minimal abelian codes (Finite Fields and their Applications, 13, (2007) 382-393) gave origin to many thesis and papers linking these two subjects. In these works, the techniques of group algebras are mainly brought into play for the computing of the idempotents that generate the minimal codes and the minimum weight of such codes. In this paper I summarize the main results of the work done by doctorate students and research partners of Polcino Milies and Ferraz. Keywords: idempotents, group algebra, coding theory. Dedicated to Prof. C´esar Polcino Milies on the occasion of his seventieth birthday. arXiv:1412.8042v2 [math.RA] 4 Sep 2015 1 Introduction The origins of Information Theory and Error Correcting Codes Theory are in the papers by Shannon [64] and Hamming [35], where they settled the theoretical foundations for such theories. ∗M. Guerreiro is with Departamento de Matem´atica, Universidade Federal de Vi¸cosa, CEP 36570-000 - Vi¸cosa-MG (Brasil), supported by CAPES, PROCAD 915/2010 and FAPEMIG, APQ CEX 00438-08 and PCE-00151-12(Brazil). E-mail: [email protected]. †This work was presented at Groups, Rings and Group Rings 2014 on the occasion of the 70th Birthday of C´esar Polcino Milies, Ubatuba-SP (Brasil) 21-26 July, 2014. 1 For a non empty finite set A, called alphabet, a code C of length n is n simply a proper subset of A and an n-tuple (a0, a1,...,an−1) ∈ C is called a word of the code C.
    [Show full text]
  • Andrew Viterbi
    Andrew Viterbi Interview conducted by Joel West, PhD December 15, 2006 Interview conducted by Joel West, PhD on December 15, 2006 Andrew Viterbi Dr. Andrew J. Viterbi, Ph.D. serves as President of the Viterbi Group LLC and Co- founded it in 2000. Dr. Viterbi co-founded Continuous Computing Corp. and served as its Chief Technology Officer from July 1985 to July 1996. From July 1983 to April 1985, he served as the Senior Vice President and Chief Scientist of M/A-COM Inc. In July 1985, he co-founded QUALCOMM Inc., where Dr. Viterbi served as the Vice Chairman until 2000 and as the Chief Technical Officer until 1996. Under his leadership, QUALCOMM received international recognition for innovative technology in the areas of digital wireless communication systems and products based on Code Division Multiple Access (CDMA) technologies. From October 1968 to April 1985, he held various Executive positions at LINKABIT (M/A-COM LINKABIT after August 1980) and served as the President of the M/A-COM LINKABIT. In 1968, Dr. Viterbi Co-founded LINKABIT Corp., where he served as an Executive Vice President and later as the President in the early 1980's. Dr. Viterbi served as an Advisor at Avalon Ventures. He served as the Vice-Chairman of Continuous Computing Corp. since July 1985. During most of his period of service with LINKABIT, Dr. Viterbi served as the Vice-Chairman and a Director. He has been a Director of Link_A_Media Devices Corporation since August 2010. He serves as a Director of Continuous Computing Corp., Motorola Mobility Holdings, Inc., QUALCOMM Flarion Technologies, Inc., The International Engineering Consortium and Samsung Semiconductor Israel R&D Center Ltd.
    [Show full text]