California State University, Northridge an Error

Total Page:16

File Type:pdf, Size:1020Kb

California State University, Northridge an Error CALIFORNIA STATE UNIVERSITY, NORTHRIDGE AN ERROR DETECTING AND CORRECTING SYSTEM FOR MAGNETIC STORAGE DISKS A project submitted in partial satisfaction of the requirements for the degree of Master of Science in Engineering by David Allan Kieselbach June, 1981 The project of David Allan Kieselbach is approved: Nagi M. Committee Chairman California State University, Northridge ii ACKNOWLEDGMENTS I wish to thank Professor Nagi El Naga who helped in numerous ways by suggesting, reviewing and criticizing the entire project. I also wish to thank my employer, Hughes Aircraft Company, who sponsored my Master of Science studies under the Hughes Fellowship Program. In addition the company prov~ded the services of the Technical Typing Center to aid in the preparation of the project manuscript. Specifically I offer my sincere gratitude to Sharon Scott and Aiko Ogata who are responsible for the excellent quality of typing and art work of the figures. For assistance in proof reading and remaining a friend throughout the long ordeal my special thanks to Christine Wacker. Finally to my parents, Dulcie and Henry who have always given me their affection and encouragement I owe my utmost appreciation and debt of gratitude. iii TABLE OF CONTENTS CHAPTER I. INTRODUCTION Page 1.1 Introduction . 1 1.2 Objectives •••••.• 4 1.3 Project Outline 5 CHAPTER II. EDCC CODES 2.1 Generator Matrix 11 2.1.1 Systematic Generator Matrix •. 12 2.2 Parity Check Matrix 16 2.3 Cyclic Codes .• . 19 2.4 Analytic Methods of Code Construction . 21 2.4.1 Hamming Codes • • 21 2.4.2 Fire Codes 21 2.4.3 Burton Codes- 24 2.4.4 BCH Codes • • . 25 CHAPTER III. ENCODING 3.1 The Mathematics of Encoding a Cyclic Code 26 3.2 Encoding Via the Generator Matrix 32 3.3 Encoding Implementation for Cyclic Codes . 36 3.3.1 Encoding With an (n-k) Stage Shift Register • 38 3.3.1.1 Multiplier Circuit Encoder • • . 38 3.3.1.2 Divider Circuit Encoder (Unmodified) 38 3.3.1.3 Divider Circuit En~oder (Modified) • 40 3.3.2 Encoding With a k Stage Shift Register 43 3.4 Shortened Cyclic Codes • 49 iv TABLE OF CONTENTS (Continued) CHAPTER IV. DECODING Page 4.1 Error Detection 51 4.2 Decoding Procedures 55 4.2.1 Table Look-up Decoding 56 4.2.2 Meggitt Decoding Technique • • • • 57 4.2.3 Error Trapping • • . • 61 4.2.4 Trial and Error Decoding . • . 64 4.2.5 Majority Logic Decoding ••...•... 66 4.2.6 Algebraic Procedures . • . 68 4.3 Examples of Practical Burst Error Correcting Decoders • . 68 4.4 Decoding Shortened Codes • 76 4.5 Encoder/Decoders • 78 CHAPTER V. INTERLEAVING 5.1 Symbol Interleaving 82 5 .1.1 Interleaving Via Code Expansion . • 82 5 .1.2 Interleaving Via Manipulation of Code Vectors .. • • . • • . • 83 5.1.2.1 I-Encoder Interleaving • . • 87 5.1.2.2 Array Manipulation Interleaving 87 5.2 Block Interleaving •.•..••••.••.•• 90 CHAPTER VI. EDCC SYSTEM DESIGN 6.1 Introduction • • • • • 92 6.2 Code and Interleaving Degree Selection • 96 6.2.1 Determination of G(X) for the EDC • 96 6.2.2 Determination of tne Interleaving Degree (EDCC) • . • • • • . 97 6.2.3 Determination of G(X) for the EDCC 105 6.3 Encoding Process .•. 107 6.3.1 EDC Encoding 107 6.3.2 EDCC Encoding . • • . 107 6.3.3 Combined EDC and EDCC Encoding 110 6.4 Decoding and Error Correction 114 6.4.1 EDCC Decoding .••••• 114 6.4.1.1 Premultiplication Polynomial 117 v TABLE OF CONTENTS (Continued) CHAPTER VI (Continued)' Page 6.4.2 Error Correction Algorithm 118 6.4.3 EDC Decoding 120 6.4.4 Combined EDCC and EDC Decoding 121 6.5 Interleaving . 127 6.6 System Control 132 CHAPTER VII. EDCC SYSTEM SIMULATION 7.1 Description of the Simulation Program 144 7. 1 .1 Program Structure • . • . 146 7.1.1.1 Program Subroutines 147 7.2 Execution of the Simulation Program 149 7.3 Results of the Simulation Program 150 CHAPTER VIII. PERFORMANCE MEASUREMENTS 8.1 Undetectability 152 8.2 Mistakability 156 8.3 Unreliability 159 8.4 Evaluation of the EDC System Parameters 159 8.5 Conclusions 162 REFERENCES 164 APPENDIX A - Flowcharts of Simulation Program and Subroutines • 165 APPENDIX B - Source Code for EDCC Simulation Program 178 APPENDIX C - Results from EDCC Simulation Program . • . 186 vi LIST OF FIGURES CHAPTER I Page 1.1 Block diagram of a general data communication or storage system 3 CHAPTER II 2.1 Block diagram of a one-way data communication system employing an error correcting code • . 7 2.2 Block diagram of a two-way data communication system using error detection and retransmission • 7 2.3 Classes of codes 10 2.4 Encoding circuit for a convolutional code 10 2.5 Matrix constructed of code vectors 14 2.6 Systematic code format 14 CHAPTER III 6 5 3.1 A circuit for dividing by x + x + x4 + x3 + 1 (internal XOR) . 39 6 5 4 3 3.2 A circuit for dividing by x + x + x + x + 1 (external XOR) . 39 3 3.3 Unmodified encoder, G(X) = X +X+ 1 .•. 40 3.4 General circuit for a modified (n-k) stage encoder • • • . • . 41 3.5 Modified encoder for a (15,9) code 44 3.6 General circuit for a k stage encoder • 48 3.7 k stage encoding circuit for (7,4) cyclic code 48 3.8 An (n,k) code shortened by S bits ...• 50 vii LIST OF FIGURES (Continued) CHAPTER IV Page 4.1 The impact of the error polynomial 12 13 14 E(X) = X + X + X on a (15,9) code • • • • • • • 54 4.2 Process of shifting the syndrome until I the degree of the error polynomial is less than that of G(X) 54 I 4.3 Look-up table decoder for an (n,k) code . 58 4.4 A Meggitt decoder for an (n,k) cyclic code 60 4.5 Error trapping decoder for a burst EDCC . 63 4.6 Format of end around burst of length t 65 4. 7 (a) Unmodified error trapping decoder for a (15,9) code .•.••..••.... 70 4.7(b) Corrector/buffer circuit for the unmodified decoder of figure 4.7(a) .••• 71 4.8 Modified error trapping decoder for a (15,9) code • . • . • . • • • • .••••.•• 72 4.9 Modified error trapping decoder for a (15,9) code shortened by 2 bits 79 4.10 Modified error trapping encoder/decoder for a general (n,k) code • • . • • • . 80 CHAPTER V 5.1 Symbol interleaving the (15, 9) code to degree 5 . 86 5.2 Interleaving via I encoders . • . 86 5.3 Interleaving via array manipulation 88 5.4 A symbol interleaved data stream, I = 2 88 CHAPTER VI -I 6.1 Block diagram of a disk system 94 6.2 Error detecting and correcting system • 95 6.3 EDC format (n,k) = (380,360) 98 6.4 EDCC combined data field for code length 98 6.5 Relationship between code rate and burst correction capability . 103 viii LIST OF FIGURES (Continued) CHAPTER VI (Continued) Page 6.6 Relationship between code rate and interleaving degree • • • 104 6.7 Format for system base EDCC . • 106 6.8 Irreducible polynomials of determination of G (X) • • • • • • • .. • • • • • • • • • 106 6.9 Feedback pattern for G(X) = 1 + x2 + x5 + x9 + x11 + x14 • 108 3 20 6.10 EDC generation circuit, G(X) = 1 +X +X . • • . • 109 6.11 EDCC generation circuit • 111 6.12 Combined encoding circuit • • 112 6.13 Flow chart of encoding process • 113 6.14 EDCC decoding and error correction circuit 116 6.15 Format of the shortened Fire Code . 117 6.16 Fortran program calculation of T(X) for the EDCC • 119 6.17 Subdivisions of the syndrome register • . 118 6.18 EDC decoding circuit . 122 6.19 Combined decoding and error correction circuit 123 6.20 Flowchart of decoding process • • • • • • 125 6.21 The affect of interleaving a (33,19) EDCC to degree I = 20 • • • . • • • • • • • . 128 6.22 Interleaving circuit block diagram 129 6.23 Interleaving circuit 131 6.24 Alternative vertical address generator 133 6.25 EDC encoder/decoder circuit • • • ••• 135 6.26 EDCC encoder/decoder circuit • 136 6.27 Block diagram of the EDCC system 138 6.28 Timing diagram for the WRITE cycle 140 6.29 Timing diagram for the READ cycle • 141 ix LIST OF FIGURES (Continued) CHAPTER VII Page 7.1 EDCC computer simulation block diagram • 143 . I CHAPTER VIII 8.1 Graphic representation of an error detection system . • • . • . .. • . • . 153 8.2 Graphic representation of an EDCC system • • 156 8.3 Code domain of the EDCC system designed with dual codes • . • . • • • • . • • 160 X LIST OF TABLES CHAPTER II Page 2.1 Burst error correcting Fire Codes with b < 10 . • • . • • 23 2.2 A comparison of code construction techniques 25 CHAPTER III 3.1 A (7,4) cyclic code generated by 3 G(X) = X + X + 1 . • . 27 3.2 A (7,4) systematic cyclic code generated 3 by G(X) = x + X + 1 . • • • • . 31 3.3 Internal operation of a modified encoder 45 3.4 Comparison of the encoding techniques • • . 45 CHAPTER IV 4.1 Syndrome generation of an error free received word • • • • • . • • • • • 73 4.2 Syndrome generation of a corrupted received word • • . • • . • • • . 74 4.3 Unmodified decoding versus modified decoding 75 CHAPTER VI 6.1 Possible code lengths and interleave degrees with 380 data bits • • . • . • • 99 6.2 Code selection process for a sector of 380 bits and b > 100 • . • . 102 xi ABSTRACT AN ERROR DETECTING AND CORRECTING SYSTEM FOR MAGNETIC STORAGE DISKS by David Allan Kieselbach Master of Science in Engineering This project presents an error detecting and correcting system for a magnetic disk storage device.
Recommended publications
  • Coset Codes. II. Binary Lattices and Related Codes
    1152 IEEE TRANSACTIONSON INFORMATIONTHEORY, VOL. 34, NO. 5, SEPTEMBER 1988 Coset Codes-Part 11: Binary Lattices and Related Codes G. DAVID FORNEY, JR., FELLOW, IEEE Invited Paper Abstract -The family of Barnes-Wall lattices (including D4 and E,) of increases by a factor of 21/2 (1.5 dB) for each doubling of lengths N = 2“ and their principal sublattices, which are useful in con- dimension. structing coset codes, are generated by iteration of a simple construction What may be obscured by the length of this paper is called the “squaring construction.” The closely related Reed-Muller codes are generated by the same construction. The principal properties of these that the construction of these lattices is extremely simple. codes and lattices, including distances, dimensions, partitions, generator The only building blocks needed are the set Z of ordinary matrices, and duality properties, are consequences of the general proper- integers, with its infinite chain of two-way partitions ties of iterated squaring constructions, which also exhibit the interrelation- 2/22/4Z/. , and an elementary construction that we ships between codes and lattices of different lengths. An extension called call the “squaring construction,” which produces chains of the “cubing construction” generates good codes and lattices of lengths 2N-tuples with certain guaranteed distance properties from N = 3.2”, including the Golay code and Leech lattice, with the use of special bases for 8-space. Another related construction generates the chains of N-tuples. Iteration of this construction produces Nordstrom-Robinson code and an analogous 16-dimensional nonlattice the entire family of lattices, determines their minimum packing.
    [Show full text]
  • Decoding Algorithms of Reed-Solomon Code
    Master’s Thesis Computer Science Thesis no: MCS-2011-26 October 2011 Decoding algorithms of Reed-Solomon code Szymon Czynszak School of Computing Blekinge Institute of Technology SE – 371 79 Karlskrona Sweden This thesis is submitted to the School of Computing at Blekinge Institute o f Technology in partial fulfillment of the requirements for the degree of Master of Science in Computer Science. The thesis is equivalent to 20 weeks of full time studies. Contact Information: Author(s): Szymon Czynszak E-mail: [email protected] University advisor(s): Mr Janusz Biernat, prof. PWR, dr hab. in ż. Politechnika Wrocławska E-mail: [email protected] Mr Martin Boldt, dr Blekinge Institute of Technology E-mail: [email protected] Internet : www.bth.se/com School of Computing Phone : +46 455 38 50 00 Blekinge Institute of Technology Fax : +46 455 38 50 57 SE – 371 79 Karlskrona Sweden ii Abstract Reed-Solomon code is nowadays broadly used in many fields of data trans- mission. Using of error correction codes is divided into two main operations: information coding before sending information into communication channel and decoding received information at the other side. There are vast of decod- ing algorithms of Reed-Solomon codes, which have specific features. There is needed knowledge of features of algorithms to choose correct algorithm which satisfies requirements of system. There are evaluated cyclic decoding algo- rithm, Peterson-Gorenstein-Zierler algorithm, Berlekamp-Massey algorithm, Sugiyama algorithm with erasures and without erasures and Guruswami- Sudan algorithm. There was done implementation of algorithms in software and in hardware.
    [Show full text]
  • Finite Fields: an Introduction. Part
    Finite fields: An introduction. Part II. Vlad Gheorghiu Department of Physics Carnegie Mellon University Pittsburgh, PA 15213, U.S.A. August 7, 2008 Vlad Gheorghiu (CMU) Finite fields: An introduction. Part II. August 7, 2008 1 / 18 Outline 1 Brief review of Part I (Jul 2, 2008) 2 Construction of finite fields Polynomials over finite fields n Explicit construction of the finite field Fq, with q = p . Examples 3 Classical coding theory Decoding methods The Coset-Leader Algorithm Examples Vlad Gheorghiu (CMU) Finite fields: An introduction. Part II. August 7, 2008 2 / 18 Examples The structure of finite fields 2 Classical codes over finite fields Linear codes: basic properties Encoding methods Hamming distance as a metric Brief review of Part I (Jul 2, 2008) Brief review of Part I 1 Finite fields Definitions Vlad Gheorghiu (CMU) Finite fields: An introduction. Part II. August 7, 2008 3 / 18 The structure of finite fields 2 Classical codes over finite fields Linear codes: basic properties Encoding methods Hamming distance as a metric Brief review of Part I (Jul 2, 2008) Brief review of Part I 1 Finite fields Definitions Examples Vlad Gheorghiu (CMU) Finite fields: An introduction. Part II. August 7, 2008 3 / 18 2 Classical codes over finite fields Linear codes: basic properties Encoding methods Hamming distance as a metric Brief review of Part I (Jul 2, 2008) Brief review of Part I 1 Finite fields Definitions Examples The structure of finite fields Vlad Gheorghiu (CMU) Finite fields: An introduction. Part II. August 7, 2008 3 / 18 Encoding methods Hamming distance as a metric Brief review of Part I (Jul 2, 2008) Brief review of Part I 1 Finite fields Definitions Examples The structure of finite fields 2 Classical codes over finite fields Linear codes: basic properties Vlad Gheorghiu (CMU) Finite fields: An introduction.
    [Show full text]
  • Efficient Representation of Binary Nonlinear Codes
    Des. Codes Cryptogr. DOI 10.1007/s10623-014-0028-4 Efficient representation of binary nonlinear codes: constructions and minimum distance computation Mercè Villanueva · Fanxuan Zeng · Jaume Pujol Received: 16 December 2013 / Revised: 22 November 2014 / Accepted: 24 November 2014 © Springer Science+Business Media New York 2014 Abstract A binary nonlinear code can be represented as a union of cosets of a binary linear subcode. In this paper, the complexity of some algorithms to obtain this representation is analyzed. Moreover, some properties and constructions of new codes from given ones in terms of this representation are described. Algorithms to compute the minimum distance of binary nonlinear codes, based on known algorithms for linear codes, are also established, along with an algorithm to decode such codes. All results are written in such a way that they can be easily transformed into algorithms, and the performance of these algorithms is evaluated. Keywords Nonlinear code · Kernel · Minimum distance · Minimum weight · Decoding · Algorithms Mathematics Subject Classification 94B60 · 94B25 · 94B35 1 Introduction Z Zn Let 2 be the ring of integers modulo 2 and let 2 be the set of all binary vectors of length n. ( ,v) ,v ∈ Zn The (Hamming) distance d u between two vectors u 2 is the number of coordinates v w ( ) ∈ Zn w ( ) = ( , ) in which u and differ. The (Hamming) weight t u of u 2 is t u d u 0 ,where ( , , ) Zn 0 is the all-zero vector of length n.An n M d binary code C is a subset of 2 with M vectors and minimum Hamming distance d.
    [Show full text]
  • Part II Linear Codes
    Part II Linear codes CHAPTER 2: Linear codes ABSTRACT Most of the important codes are special types of so-called linear codes. Linear codes are of very large importance because they have very concise description, very nice properties, very easy encoding and, in principle, easy to describe decoding. prof. Jozef Gruska IV054 2. Linear codes 2/39 Linear codes Linear codes are special sets of words of the length n over an alphabet f0; ::; q − 1g, n where q is a power of prime. Since now on sets of words Fq will be considered as vector spaces V (n; q) of vectors of length n with elements from the set f0; ::; q − 1g and arithmetical operations will be taken modulo q. Definition A subset C ⊆ V (n; q) is a linear code if 1 u + v 2 C for all u; v 2 C 2 au 2 C for all u 2 C; a 2 GF (q) Example Codes C1; C2; C3 introduced in Lecture 1 are linear codes. Lemma A subset C ⊆ V (n; q) is a linear code if one of the following conditions is satisfied 1 C is a subspace of V (n; q) 2 sum of any two codewords from C is in C (for the case q = 2) If C is a k-dimensional subspace of V (n; q), then C is called[ n; k]-code. It has qk codewords. If minimal distance of C is d, then it is called[ n; k; d] code. Linear codes are also called "group codes". prof. Jozef Gruska IV054 2.
    [Show full text]
  • Coding Theory Via Groebner Bases
    Coding Theory via Groebner Bases Vom Promotionsausschuss der Technischen Universitat¨ Hamburg-Harburg zur Erlangung des akademischen Grades Doktorin der Naturwissenschaften genehmigte Dissertation von Mehwish Saleemi aus Islamabad 2012 1. Gutachter: Prof. Dr. Karl-Heinz Zimmermann Institute fur¨ Rechnertechnologie, Technische Universitat¨ Hamburg-Harburg 2. Gutachter: Prof. Dr. Rudolf Scharlau Fakultat¨ fur¨ Mathematik, Technische Universitat¨ Dortmund Tag der mundlichen¨ Prufung:¨ 14.02.2012 Vorsitzender des Prufungsausschusses:¨ Prof. Dr. Dieter Gollmann Institut fur¨ Sicherheit in verteilten Anwendungen, Technische Universitat¨ Hamburg- Harburg Abstract Coding theory plays an important role in efficient transmission of data over noisy communication channels. It consists of two steps; the first step is to encode the data to reduce its sensitivity to noise during transmission, and the second step is to decode the received data by detecting and correcting the noise induced errors. In this thesis an algebraic approach is used to develop efficient encoding and decod- ing algorithms for a very commonly used class of linear codes, the Reed-Muller codes and the Golay codes. To develop the approach first the algebraic structure of linear codes is explored. For this, the reduced Groebner basis for a class of ideals in commutative poly- nomial rings is constructed. The extension of these ideals to a residue class ring enabled us to find the parameters of the corresponding codes. It is found that the corresponding codes contains the primitive Reed-Muller codes. The added advan- tage of this approach is that, once these Groebner bases are constructed a standard procedure can be used to develop encoding and decoding processes. A binomial ideal, defined as a sum of toric ideal and a prime ideal over some arbitrary field, is explored.
    [Show full text]
  • New Decoding Methods for LDPC Codes on Error and Error-Erasure Channels
    New Decoding Methods for LDPC Codes on Error and Error-Erasure Channels MASTER OF SCIENCE THESIS by İlke ALTIN August 19, 2010 Committee Members Supervisor : Dr. ir. J.H. Weber Readers : Prof. dr. ir. I. Niemegeers Dr. ir. H. Wang Wireless and Mobile Communications Group Faculty of Electrical Engineering, Mathematics and Computer Science Delft University of Technology Delft, The Netherlands ii İlke ALTIN Master of Science Thesis Abstract New Decoding Methods for LDPC Codes on Error and Error-Erasure Channels For low-end devices with limited battery or computational power, low complexity decoders are beneficial. In this research we have searched for low complexity decoder alternatives for error and error-erasure channels. We have especially focused on low complexity error erasure decoders, which is a topic that has not been studied by many researchers. The separation of erasures from errors idea [1] seemed profitable to design a new error erasure decoder, so we have also worked on this idea. However, the methods that are described in that paper are not realizable for practical values of code length and number of erasures. Thus, a new separation method; the rank completer is proposed, which is realizable. In the part of the research that is related to error decoding, we have proposed a modification to reliability ratio based bit flipping algorithm [2], which improves the BER performance with very small additional complexity. In the part that is related to error erasure decoding, we have given a new guessing algorithm that performs better than some known guessing algorithms, and a new error erasure decoder that uses the rank completer idea.
    [Show full text]
  • Viterbi Decoding of Convolutional Codes
    MIT 6.02 DRAFT Lecture Notes Last update: February 29, 2012 Comments, questions or bug reports? Please contact hari at mit.edu CHAPTER 8 Viterbi Decoding of Convolutional Codes This chapter describes an elegant and efficient method to decode convolutional codes, whose construction and encoding we described in the previous chapter. This decoding method avoids explicitly enumerating the 2N possible combinations of N-bit parity bit sequences. This method was invented by Andrew Viterbi ’57 and bears his name. 8.1 The Problem At the receiver, we have a sequence of voltage samples corresponding to the parity bits that the transmitter has sent. For simplicity, and without loss of generality, we will assume that the receiver picks a suitable sample for the bit, or averages the set of samples corre- sponding to a bit, digitizes that value to a “0” or “1” by comparing to the threshold voltage (the demapping step), and propagates that bit decision to the decoder. Thus, we have a received bit sequence, which for a convolutionally-coded stream cor- responds to the stream of parity bits. If we decode this received bit sequence with no other information from the receiver’s sampling and demapper, then the decoding pro- cess is termed hard-decision decoding (“hard decoding”). If, instead (or in addition), the decoder is given the stream of voltage samples and uses that “analog” information (in digitized form, using an analog-to-digital conversion) in decoding the data, we term the process soft-decision decoding (“soft decoding”). The Viterbi decoder can be used in either case.
    [Show full text]
  • High Throughput Error Correction in Information Reconciliation For
    www.nature.com/scientificreports OPEN High throughput error correction in information reconciliation for semiconductor superlattice secure key distribution Jianguo Xie1,3, Han Wu2,3, Chao Xia1, Peng Ding2, Helun Song2, Liwei Xu1 & Xiaoming Chen1* Semiconductor superlattice secure key distribution (SSL-SKD) has been experimentally demonstrated to be a novel scheme to generate and agree on the identical key in unconditional security just by public channel. The error correction in the information reconciliation procedure is introduced to eliminate the inevitable diferences of analog systems in SSL-SKD. Nevertheless, the error correction has been proved to be the performance bottleneck of information reconciliation for high computational complexity. Hence, it determines the fnal secure key throughput of SSL-SKD. In this paper, diferent frequently-used error correction codes, including BCH codes, LDPC codes, and Polar codes, are optimized separately to raise the performance, making them usable in practice. Firstly, we perform multi-threading to support multi-codeword decoding for BCH codes and Polar codes and updated value calculation for LDPC codes. Additionally, we construct lookup tables to reduce redundant calculations, such as logarithmic table and antilogarithmic table for fnite feld computation. Our experimental results reveal that our proposed optimization methods can signifcantly promote the efciency of SSL-SKD, and three error correction codes can reach the throughput of Mbps and provide a minimum secure key rate of 99%. Semiconductor superlattice secure key distribution (SSL-SKD) is a new secure key distribution technique based on chaos synchronization in superlattice PUF pairs 1 driven by a synchronizing digital signal. SSL-SKD only uses the public channel with all electronic structures to create and provide secure key data for cryptography in unconditional security2.
    [Show full text]
  • On the Optimality of the Hamming Metric for Decoding Block Codes Over Binary Additive Noise Channels
    On the Optimality of the Hamming Metric for Decoding Block Codes over Binary Additive Noise Channels by Ghady Azar A thesis submitted to the Department of Mathematics and Statistics in conformity with the requirements for the degree of Master of Applied Science Queen’s University Kingston, Ontario, Canada July 2013 Copyright c Ghady Azar, 2013 Abstract Most of the basic concepts of algebraic coding theory are derived for the mem- oryless binary symmetric channel. These concepts do not necessarily hold for time-varying channels or for channels with memory. However, errors in real-life channels seem to occur in bursts rather than independently, suggesting that these channels exhibit some statistical dependence or memory. Nonetheless, the same algebraic codes are still commonly used in current communication systems that employ interleaving to spread channel error bursts over the set of received code- words to make the channel appear memoryless to the block decoder. This method suffers from immediate shortcomings as it fails to exploit the channel’s memory while adding delay to the system. We study optimal maximum likelihood block decoding of binary codes sent over several binary additive channels with infinite and finite memory. We derive con- ditions on general binary codes and channels parameters under which maximum likelihood and minimum distance decoding are equivalent. The channels consid- ered in this work are the infinite and finite memory Polya contagion channels [1], the queue-based channel [29], and the Gilbert-Elliott channel [9, 12]. We also present results on the optimality of classical perfect and quasi-perfect codes when used over the aforementioned channels under maximum likelihood decoding.
    [Show full text]
  • Capacity-Achieving Guessing Random Additive Noise Decoding
    Capacity-Achieving Guessing Random Additive Noise Decoding The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation Duffy, Ken R. et al. “Capacity-Achieving Guessing Random Additive Noise Decoding.” IEEE Transactions on Information Theory, 65, 7 (July 2019): 4023 - 4040 © 2019 The Author(s) As Published 10.1109/TIT.2019.2896110 Publisher Institute of Electrical and Electronics Engineers (IEEE) Version Author's final manuscript Citable link https://hdl.handle.net/1721.1/129702 Terms of Use Creative Commons Attribution-Noncommercial-Share Alike Detailed Terms http://creativecommons.org/licenses/by-nc-sa/4.0/ 1 Capacity-achieving Guessing Random Additive Noise Decoding (GRAND) Ken R. Duffy∗, Jiange Li† and Muriel Medard´ † ∗Hamilton Institute, Maynooth University, Ireland. E-mail: [email protected]. †Research Laboratory of Electronics, Massachusetts Institute of Technology, Cambridge, MA 02139, U. S. A. E-mail: [email protected], [email protected]. Abstract We introduce a new algorithm for realizing Maximum Likelihood (ML) decoding in discrete channels with or without memory. In it, the receiver rank orders noise sequences from most likely to least likely. Subtracting noise from the received signal in that order, the first instance that results in a member of the code-book is the ML decoding. We name this algorithm GRAND for Guessing Random Additive Noise Decoding. We establish that GRAND is capacity-achieving when used with random code-books. For rates below capacity we identify error exponents, and for rates beyond capacity we identify success exponents. We determine the scheme’s complexity in terms of the number of computations the receiver performs.
    [Show full text]
  • AI Coding: Learning to Construct Error Correction Codes
    IEEE TRANSACTIONS ON COMMUNICATIONS, VOL. XX, NO. X, OCTOBER 2019 1 AI Coding: Learning to Construct Error Correction Codes Lingchen Huang, Member, IEEE, Huazi Zhang, Member, IEEE, Rong Li, Member, IEEE, Yiqun Ge, Member, IEEE, Jun Wang, Member, IEEE, Abstract—In this paper, we investigate an artificial- intelligence (AI) driven approach to design error correction Code codes (ECC). Classic error-correction code design based upon Performance coding-theoretic principles typically strives to optimize some performance-related code property such as minimum Hamming distance, decoding threshold, or subchannel reliability ordering. In contrast, AI-driven approaches, such as reinforcement learn- Coding AI ing (RL) and genetic algorithms, rely primarily on optimization methods to learn the parameters of an optimal code within a cer- Theory Techniques tain code family. We employ a constructor-evaluator framework, in which the code constructor can be realized by various AI algorithms and the code evaluator provides code performance Code metric measurements. The code constructor keeps improving the code construction to maximize code performance that is Construction evaluated by the code evaluator. As examples, we focus on RL and genetic algorithms to construct linear block codes and polar codes. The results show that comparable code performance can Fig. 1: Error correction code design logic be achieved with respect to the existing codes. It is noteworthy that our method can provide superior performances to classic constructions in certain cases (e.g., list decoding for polar codes). improve its code performance. Equivalently, given a target er- ror rate, we optimize code design to maximize the achievable Index Terms—Machine learning, Code construction, Artificial code rate, i.e.
    [Show full text]