Notes 1: Introduction, Linear Codes January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Hamming Weight Attacks on Cryptographic Hardware – Breaking Masking Defense?
Hamming Weight Attacks on Cryptographic Hardware { Breaking Masking Defense? Marcin Gomu lkiewicz1 and Miros law Kuty lowski12 1 Cryptology Centre, Poznan´ University 2 Institute of Mathematics, Wroc law University of Technology, ul. Wybrzeze_ Wyspianskiego´ 27 50-370 Wroc law, Poland Abstract. It is believed that masking is an effective countermeasure against power analysis attacks: before a certain operation involving a key is performed in a cryptographic chip, the input to this operation is com- bined with a random value. This has to prevent leaking information since the input to the operation is random. We show that this belief might be wrong. We present a Hamming weight attack on an addition operation. It works with random inputs to the addition circuit, hence masking even helps in the case when we cannot control the plaintext. It can be applied to any round of the encryption. Even with moderate accuracy of measuring power consumption it de- termines explicitly subkey bits. The attack combines the classical power analysis (over Hamming weight) with the strategy of the saturation at- tack performed using a random sample. We conclude that implementing addition in cryptographic devices must be done very carefully as it might leak secret keys used for encryption. In particular, the simple key schedule of certain algorithms (such as IDEA and Twofish) combined with the usage of addition might be a serious danger. Keywords: cryptographic hardware, side channel cryptanalysis, Ham- ming weight, power analysis 1 Introduction Symmetric encryption algorithms are often used to protect data stored in inse- cure locations such as hard disks, archive copies, and so on. -
A [49 16 6 ]-Linear Code As Product of the [7 4 2 ] Code Due to Aunu and the Hamming [7 4 3 ] Code
General Letters in Mathematics, Vol. 4, No.3 , June 2018, pp.114 -119 Available online at http:// www.refaad.com https://doi.org/10.31559/glm2018.4.3.4 A [49 16 6 ]-Linear Code as Product Of The [7 4 2 ] Code Due to Aunu and the Hamming [7 4 3 ] Code CHUN, Pamson Bentse1, IBRAHIM Alhaji Aminu2, and MAGAMI, Mohe'd Sani3 1 Department Of Mathematics, Plateau State University Bokkos, Jos Nigeria [email protected] 2 Department Of Mathematics, Usmanu Danfodiyo University Sokoto, Sokoto Nigeria [email protected] 3 Department Of Mathematics, Usmanu Danfodiyo University Sokoto, Sokoto Nigeria [email protected] Abstract: The enumeration of the construction due to "Audu and Aminu"(AUNU) Permutation patterns, of a [ 7 4 2 ]- linear code which is an extended code of the [ 6 4 1 ] code and is in one-one correspondence with the known [ 7 4 3 ] - Hamming code has been reported by the Authors. The [ 7 4 2 ] linear code, so constructed was combined with the known Hamming [ 7 4 3 ] code using the ( u|u+v)-construction method to obtain a new hybrid and more practical single [14 8 3 ] error- correcting code. In this paper, we provide an improvement by obtaining a much more practical and applicable double error correcting code whose extended version is a triple error correcting code, by combining the same codes as in [1]. Our goal is achieved through using the product code construction approach with the aid of some proven theorems. Keywords: Cayley tables, AUNU Scheme, Hamming codes, Generator matrix,Tensor product 2010 MSC No: 97H60, 94A05 and 94B05 1. -
Coding Theory: Linear Error-Correcting Codes Coding Theory Basic Definitions Error Detection and Correction Finite Fields Anna Dovzhik
Coding Theory: Linear Error- Correcting Codes Anna Dovzhik Outline Coding Theory: Linear Error-Correcting Codes Coding Theory Basic Definitions Error Detection and Correction Finite Fields Anna Dovzhik Linear Codes Hamming Codes Finite Fields Revisited BCH Codes April 23, 2014 Reed-Solomon Codes Conclusion Coding Theory: Linear Error- Correcting 1 Coding Theory Codes Basic Definitions Anna Dovzhik Error Detection and Correction Outline Coding Theory 2 Finite Fields Basic Definitions Error Detection and Correction 3 Finite Fields Linear Codes Linear Codes Hamming Codes Hamming Codes Finite Fields Finite Fields Revisited Revisited BCH Codes BCH Codes Reed-Solomon Codes Reed-Solomon Codes Conclusion 4 Conclusion Definition A q-ary word w = w1w2w3 ::: wn is a vector where wi 2 A. Definition A q-ary block code is a set C over an alphabet A, where each element, or codeword, is a q-ary word of length n. Basic Definitions Coding Theory: Linear Error- Correcting Codes Anna Dovzhik Definition If A = a ; a ;:::; a , then A is a code alphabet of size q. Outline 1 2 q Coding Theory Basic Definitions Error Detection and Correction Finite Fields Linear Codes Hamming Codes Finite Fields Revisited BCH Codes Reed-Solomon Codes Conclusion Definition A q-ary block code is a set C over an alphabet A, where each element, or codeword, is a q-ary word of length n. Basic Definitions Coding Theory: Linear Error- Correcting Codes Anna Dovzhik Definition If A = a ; a ;:::; a , then A is a code alphabet of size q. Outline 1 2 q Coding Theory Basic Definitions Error Detection Definition and Correction Finite Fields A q-ary word w = w1w2w3 ::: wn is a vector where wi 2 A. -
Superregular Matrices and Applications to Convolutional Codes
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Repositório Institucional da Universidade de Aveiro Superregular matrices and applications to convolutional codes P. J. Almeidaa, D. Nappa, R.Pinto∗,a aCIDMA - Center for Research and Development in Mathematics and Applications, Department of Mathematics, University of Aveiro, Aveiro, Portugal. Abstract The main results of this paper are twofold: the first one is a matrix theoretical result. We say that a matrix is superregular if all of its minors that are not trivially zero are nonzero. Given a a × b, a ≥ b, superregular matrix over a field, we show that if all of its rows are nonzero then any linear combination of its columns, with nonzero coefficients, has at least a − b + 1 nonzero entries. Secondly, we make use of this result to construct convolutional codes that attain the maximum possible distance for some fixed parameters of the code, namely, the rate and the Forney indices. These results answer some open questions on distances and constructions of convolutional codes posted in the literature [6, 9]. Key words: convolutional code, Forney indices, optimal code, superregular matrix 2000MSC: 94B10, 15B33, 15B05 1. Introduction Several notions of superregular matrices (or totally positive) have appeared in different areas of mathematics and engineering having in common the specification of some properties regarding their minors [2, 3, 5, 11, 14]. In the context of coding theory these matrices have entries in a finite field F and are important because they can be used to generate linear codes with good distance properties. -
Reed-Solomon Code and Its Application 1 Equivalence
IERG 6120 Coding Theory for Storage Systems Lecture 5 - 27/09/2016 Reed-Solomon Code and Its Application Lecturer: Kenneth Shum Scribe: Xishi Wang 1 Equivalence of Codes Two linear codes are said to be equivalent if one of them can be obtained from the other by means of a sequence of transformations of the following types: (i) a permutation of the positions of the code; (ii) multiplication of symbols in a fixed position by a non-zero scalar in F . Note that these transformations can be applied to all code symbols. 2 Reed-Solomon Code In RS code each symbol in the codewords is the evaluation of a polynomial at one point α, namely, 2 1 3 6 α 7 6 2 7 6 α 7 f(α) = c0 c1 c2 ··· ck−1 6 7 : 6 . 7 4 . 5 αk−1 The whole codeword is given by n such evaluations at distinct points α1; ··· ; αn, 2 1 1 ··· 1 3 6 α1 α2 ··· αn 7 6 2 2 2 7 6 α1 α2 ··· αn 7 T f(α1) f(α2) ··· f(αn) = c0 c1 c2 ··· ck−1 6 7 = c G: 6 . .. 7 4 . 5 k−1 k−1 k−1 α1 α2 ··· αn G is the generator matrix of RSq(n; k; d) code. The order of writing down the field elements α1 to αn is not important as far as the code structure is concerned, as any permutation of the αi's give an equivalent code. i j=1;:::;n The generator matrix G is a Vandermonde matrix, which is of the form [aj]i=0;:::;k−1. -
Type I Codes Over GF(4)
Type I Codes over GF(4) Hyun Kwang Kim¤ San 31, Hyoja Dong Department of Mathematics Pohang University of Science and Technology Pohang, 790-784, Korea e-mail: [email protected] Dae Kyu Kim School of Electronics & Information Engineering Chonbuk National University Chonju, Chonbuk 561-756, Korea e-mail: [email protected] Jon-Lark Kimy Department of Mathematics University of Louisville Louisville, KY 40292, USA e-mail: [email protected] Abstract It was shown by Gaborit el al. [10] that a Euclidean self-dual code over GF (4) with the property that there is a codeword whose Lee weight ´ 2 (mod 4) is of interest because of its connection to a binary singly-even self-dual code. Such a self-dual code over GF (4) is called Type I. The purpose of this paper is to classify all Type I codes of lengths up to 10 and extremal Type I codes of length 12, and to construct many new extremal Type I codes over GF (4) of ¤The present study was supported by Com2MaC-KOSEF, POSTECH BSRI research fund, and grant No. R01-2006-000-11176-0 from the Basic Research Program of the Korea Science & Engineering Foundation. ycorresponding author, supported in part by a Project Completion Grant from the University of Louisville. 1 lengths from 14 to 22 and 34. As a byproduct, we construct a new extremal singly-even self-dual binary [36; 18; 8] code, and a new ex- tremal singly-even self-dual binary [68; 34; 12] code with a previously unknown weight enumerator W2 for ¯ = 95 and γ = 1. -
Hamming Code - Wikipedia, the Free Encyclopedia Hamming Code Hamming Code
Hamming code - Wikipedia, the free encyclopedia Hamming code Hamming code From Wikipedia, the free encyclopedia In telecommunication, a Hamming code is a linear error-correcting code named after its inventor, Richard Hamming. Hamming codes can detect and correct single-bit errors, and can detect (but not correct) double-bit errors. In contrast, the simple parity code cannot detect errors where two bits are transposed, nor can it correct the errors it can find. Contents [hide] • 1 History • 2 Codes predating Hamming ♦ 2.1 Parity ♦ 2.2 Two-out-of-five code ♦ 2.3 Repetition • 3 Hamming codes • 4 Example using the (11,7) Hamming code • 5 Hamming code (7,4) ♦ 5.1 Hamming matrices ♦ 5.2 Channel coding ♦ 5.3 Parity check ♦ 5.4 Error correction • 6 Hamming codes with additional parity • 7 See also • 8 References • 9 External links History Hamming worked at Bell Labs in the 1940s on the Bell Model V computer, an electromechanical relay-based machine with cycle times in seconds. Input was fed in on punch cards, which would invariably have read errors. During weekdays, special code would find errors and flash lights so the operators could correct the problem. During after-hours periods and on weekends, when there were no operators, the machine simply moved on to the next job. Hamming worked on weekends, and grew increasingly frustrated with having to restart his programs from scratch due to the unreliability of the card reader. Over the next few years he worked on the problem of error-correction, developing an increasingly powerful array of algorithms. -
Reed-Solomon Error Correction
R&D White Paper WHP 031 July 2002 Reed-Solomon error correction C.K.P. Clarke Research & Development BRITISH BROADCASTING CORPORATION BBC Research & Development White Paper WHP 031 Reed-Solomon Error Correction C. K. P. Clarke Abstract Reed-Solomon error correction has several applications in broadcasting, in particular forming part of the specification for the ETSI digital terrestrial television standard, known as DVB-T. Hardware implementations of coders and decoders for Reed-Solomon error correction are complicated and require some knowledge of the theory of Galois fields on which they are based. This note describes the underlying mathematics and the algorithms used for coding and decoding, with particular emphasis on their realisation in logic circuits. Worked examples are provided to illustrate the processes involved. Key words: digital television, error-correcting codes, DVB-T, hardware implementation, Galois field arithmetic © BBC 2002. All rights reserved. BBC Research & Development White Paper WHP 031 Reed-Solomon Error Correction C. K. P. Clarke Contents 1 Introduction ................................................................................................................................1 2 Background Theory....................................................................................................................2 2.1 Classification of Reed-Solomon codes ...................................................................................2 2.2 Galois fields............................................................................................................................3 -
Linear Block Codes
Linear Block codes (n, k) Block codes k – information bits n - encoded bits Block Coder Encoding operation n-digit codeword made up of k-information digits and (n-k) redundant parity check digits. The rate or efficiency for this code is k/n. 푘 푁푢푚푏푒푟 표푓 푛푓표푟푚푎푡표푛 푏푡푠 퐶표푑푒 푒푓푓푐푒푛푐푦 푟 = = 푛 푇표푡푎푙 푛푢푚푏푒푟 표푓 푏푡푠 푛 푐표푑푒푤표푟푑 Note: unlike source coding, in which data is compressed, here redundancy is deliberately added, to achieve error detection. SYSTEMATIC BLOCK CODES A systematic block code consists of vectors whose 1st k elements (or last k-elements) are identical to the message bits, the remaining (n-k) elements being check bits. A code vector then takes the form: X = (m0, m1, m2,……mk-1, c0, c1, c2,…..cn-k) Or X = (c0, c1, c2,…..cn-k, m0, m1, m2,……mk-1) Systematic code: information digits are explicitly transmitted together with the parity check bits. For the code to be systematic, the k-information bits must be transmitted contiguously as a block, with the parity check bits making up the code word as another contiguous block. Information bits Parity bits A systematic linear block code will have a generator matrix of the form: G = [P | Ik] Systematic codewords are sometimes written so that the message bits occupy the left-hand portion of the codeword and the parity bits occupy the right-hand portion. Parity check matrix (H) Will enable us to decode the received vectors. For each (kxn) generator matrix G, there exists an (n-k)xn matrix H, such that rows of G are orthogonal to rows of H i.e., GHT = 0, where HT is the transpose of H. -
16.1 Digital “Modes”
Contents 16.1 Digital “Modes” 16.5 Networking Modes 16.1.1 Symbols, Baud, Bits and Bandwidth 16.5.1 OSI Networking Model 16.1.2 Error Detection and Correction 16.5.2 Connected and Connectionless 16.1.3 Data Representations Protocols 16.1.4 Compression Techniques 16.5.3 The Terminal Node Controller (TNC) 16.1.5 Compression vs. Encryption 16.5.4 PACTOR-I 16.2 Unstructured Digital Modes 16.5.5 PACTOR-II 16.2.1 Radioteletype (RTTY) 16.5.6 PACTOR-III 16.2.2 PSK31 16.5.7 G-TOR 16.2.3 MFSK16 16.5.8 CLOVER-II 16.2.4 DominoEX 16.5.9 CLOVER-2000 16.2.5 THROB 16.5.10 WINMOR 16.2.6 MT63 16.5.11 Packet Radio 16.2.7 Olivia 16.5.12 APRS 16.3 Fuzzy Modes 16.5.13 Winlink 2000 16.3.1 Facsimile (fax) 16.5.14 D-STAR 16.3.2 Slow-Scan TV (SSTV) 16.5.15 P25 16.3.3 Hellschreiber, Feld-Hell or Hell 16.6 Digital Mode Table 16.4 Structured Digital Modes 16.7 Glossary 16.4.1 FSK441 16.8 References and Bibliography 16.4.2 JT6M 16.4.3 JT65 16.4.4 WSPR 16.4.5 HF Digital Voice 16.4.6 ALE Chapter 16 — CD-ROM Content Supplemental Files • Table of digital mode characteristics (section 16.6) • ASCII and ITA2 code tables • Varicode tables for PSK31, MFSK16 and DominoEX • Tips for using FreeDV HF digital voice software by Mel Whitten, KØPFX Chapter 16 Digital Modes There is a broad array of digital modes to service various needs with more coming. -
Short Low-Rate Non-Binary Turbo Codes Gianluigi Liva, Bal´Azs Matuz, Enrico Paolini, Marco Chiani
Short Low-Rate Non-Binary Turbo Codes Gianluigi Liva, Bal´azs Matuz, Enrico Paolini, Marco Chiani Abstract—A serial concatenation of an outer non-binary turbo decoding thresholds lie within 0.5 dB from the Shannon code with different inner binary codes is introduced and an- limit in the coherent case, for a wide range of coding rates 1 alyzed. The turbo code is based on memory- time-variant (1/3 ≤ R ≤ 1/96). Remarkably, a similar result is achieved recursive convolutional codes over high order fields. The resulting codes possess low rates and capacity-approaching performance, in the noncoherent detection framework as well. thus representing an appealing solution for spread spectrum We then focus on the specific case where the inner code is communications. The performance of the scheme is investigated either an order-q Hadamard code or a first order length-q Reed- on the additive white Gaussian noise channel with coherent and Muller (RM) code, due to their simple fast Hadamard trans- noncoherent detection via density evolution analysis. The pro- form (FHT)-based decoding algorithms. The proposed scheme posed codes compare favorably w.r.t. other low rate constructions in terms of complexity/performance trade-off. Low error floors can be thus seen either as (i) a serial concatenation of an outer and performances close to the sphere packing bound are achieved Fq-based turbo code with an inner Hadamard/RM code with down to small block sizes (k = 192 information bits). antipodal signalling, or (ii) as a coded modulation Fq-based turbo/LDPC code with q-ary (bi-) orthogonal modulation. -
Linear Block Codes Vector Spaces
Linear Block Codes Vector Spaces Let GF(q) be fixed, normally q=2 • The set Vn consisting of all n-dimensional vectors (a n-1, a n-2,•••a0), ai ∈ GF(q), 0 ≤i ≤n-1 forms an n-dimensional vector space. • In Vn, we can also define the inner product of any two vectors a = (a n-1, a n-2,•••a0) and b = (b n-1, b n-2,•••b0) n−1 i ab= ∑ abi i i=0 • If a•b = 0, then a and b are said to be orthogonal. Linear Block codes • Let C ⊂ Vn be a block code consisting of M codewords. • C is said to be linear if a linear combination of two codewords C1 and C2, a 1C1+a 2C2, is still a codeword, that is, C forms a subspace of Vn. a 1, ∈ a2 GF(q) • The dimension of the linear block code (a subspace of Vn) is denoted as k. M=qk. Dual codes Let C ⊂Vn be a linear code with dimension K. ┴ The dual code C of C consists of all vectors in Vn that are orthogonal to every vector in C. The dual code C┴ is also a linear code; its dimension is equal to n-k. Generator Matrix Assume q=2 Let C be a binary linear code with dimension K. Let {g1,g2,…gk} be a basis for C, where gi = (g i(n-1) , g i(n-2) , … gi0 ), 1 ≤i≤K. Then any codeword Ci in C can be represented uniquely as a linear combination of {g1,g2,…gk}.