RPIT - RPIT - RPIT - RPIT - PREPRINT -- PREPRINT -- PREPRINT --PREPRINT -- arXiv:cs/0610074v2 [cs.IT] 12 Oct 2006 eea edSlmncds n octntn hmwt an with them concatenating code. block and interl inner by codes, obtained Reed–Solomon are Base which several investigated, collaborativ analyzed. are concaten designs codes, work, and Reed–Solomon code the interleaved considered this of in framework multi-seque is the In errors by synthesis codes bursts. the Reed–Solomon shift-register in interleaved that of occur provided decoding distance, signal minimum received patterns the error of half correction approaches, the decoding allow schemes algebraic met interleaved decoding collaborative a algebraic by codewords classical However, these codewo by Usually, independently several codes. decoded interleaving Reed–Solomon stor by ordinary data generated of and are transmission, They data systems. processing, data numerous rtv eoig hf-eitrsnhss utpeseq multiple co synthesis, codes, shift-register concatenated decoding, codes, orative IRS heterogeneous codes, IRS etaserdwtotntc,atrwihti eso wil version this which publication after possible notice, have for without IEEE transferred paper the to be Zvenigorod, submitted this in been has 2006 in Wor work (ACCT-10) lea Theory o Theory presented Workshop Information International on Coding 10th results ITSOC Combinatorial the is IEEE and the Zealand, 2005 Sidorenko New the of Rotorua, R. on Some V. presented respectively. been RU. 867/15 RAS under BO IITP, Germany, and (DFG), 867/14, Forschungsgemeinschaft Deutsche by ros h eoe o h ne oewl sal create an usually ou Kiayias, the Bleichenbacher, will for codes, decoder IRS code the decoding of inner For input codes. the the at errors for burst codes correlated decoder even IRS ran since the independent consider designs, statistically to errors, code inducing suggested concatenated models channel is in simultaneously. for codes it scheme outer [10], as interleaved and also the [7], erro correlated [3], of if In words effective most all are affect codes bur error IRS [9], where since applications occur, in in Reed–Solo considered considered mainly Interleaved are are publications. codes other codes and Parvaresh IRS [11], as Moreover, [10], well [8]. as Vardy [7], Bleichenbacher, and Høholdt [3], [6 and [5], Thommesen, Shokrollahi [2], and Justesen, Minder [1], Brown, l [4], Yung Garg and authors Kiayias, and different by Lee, investigated Krachkovsky, are They publications. Solomon h oko ldmrR ioek n er cmd ssuppor is Schmidt Georg and Sidorenko R. Vladimir of work The ne Terms Index Abstract nrcn er,cd ein with designs code years, recent In edSlmnCdsadCnaeae Code Concatenated and Codes Reed–Solomon eateto eeomnctosadApidInformation Applied and Telecommunications of Department nelae edSlmncdsaeapidin applied are codes Reed–Solomon Interleaved — IS oe eetetpco eea scientific several of topic the were codes (IRS) nelae edSlmncds homogeneous codes, Reed–Solomon Interleaved — { georg.schmidt,vladimir.sidorenko,martin.bossert er Schmidt, Georg olbrtv eoigo Interleaved of Decoding Collaborative .I I. NTRODUCTION tdn ebr IEEE Member, Student atnBossert Martin nelae Reed– interleaved esuperseded. be l oyih may Copyright . leri and Algebraic n uences usa This Russia. rjc BO project Designs so in kshop beyond efrom ve eaving hods. on d llab- dom mon such ated age also nce rds ike sts ter ted eirMme,IEEE Member, Senior re ldmrR Sidorenko, R. Vladimir , rs ], d e agrta aftemnmmdsac fteReed–Solomon the of distance minimum the half than larger rgnosIScdsmyb neetn,eg hnue as used when e.g. Het interesting, of [12]. be codes in outer may and investigated codes been codes), IRS have IRS erogeneous properties heterogeneous their them of calling some (without conside [3] been in first have constructions Heterogeneous codes. o nqeylctn pto allo up This of locating words. several uniquely instead the for codewords in Reed–Solomon independently all them locating loca in which jointly a applied, errors Hence, be position. the can same strategy the decoding at collaborative symbol erroneous an have may a.Mr rcsl ecnie R oe ossigof consisting codes IRS consider codes we Reed–Solomon precisely More way. algorithm. decoding list powerful w decoding interpolation, quite IRS polynomial a multivariate yields consider of [8] context the Vardy in and the Parvaresh, on based approach. algorithm an propose [4] Yung utdb us ro,ie,a ro hc fet comple a the affects which arrangement, error the an of i.e., error, column burst a by rupted the decoding to collaborati comparison by achieved in be can decoding decoding which gain, allow the the bounds on of These th estimating probability. bounds failure derive analyze decoding we and To error decoder, this synthesis. based of is shift-register behavior which multi-sequence codes, IRS on heterogeneous and code. IRS homogeneous heterogeneous a decoding of problem transforme is the code into Reed–Solomon o low-rate problem single the the a where half [15], decoding beyond in code described like Reed–Solomon distance single minimum [14]. a Zinoviev of code by IRS decoding heterogeneous the and of [13] application Zyablov interesting Another and Blokh by described ahcode each dimensions oe neednl pt aftemnmmdsac ya by distance minimum the half to standard up independently codes arxlk eitdi i.1 ecl h e fmatrices of set If the code. call IRS We an 1. way Fig. this in in obtainable depicted like matrix a r ifrn,w altersligIScode IRS resulting the call we different, are A ( otpeiu ulctoscnie nyhmgnosIRS homogeneous only consider publications previous Most nti ae,w osdrIScdsi ahrgeneral rather a in codes IRS consider we paper, this In foeclm ftearneetsonb i.1i cor- is 1 Fig. by shown arrangement the of column one If both of decoding collaborative for method a propose We l ) ecl h code the call we , one iiu Distance Minimum Bounded A K hoy nvriyo l,Germany Ulm, of University Theory, ( ℓ ) (1) eeaie octntdcodes concatenated generalized , ℓ K , 1 = (2) l , . . . , A K , . . . , ebr IEEE Member, homogeneous (1) } t , n rag hmrwws into row-wise them arrange and , @uni-ulm.de A ros nmn ae vnif even cases many in errors, (2) ( l ) etk oeodfrom codeword a take We . l , . . . , edSlmncodewords Reed–Solomon R oe ftecodes the if code, IRS BD decoder. (BMD) A and , A ( l (1) ) Welch–Berlekamp flength of l heterogeneous = Reed–Solomon nrdcdand introduced A (2) = N · · · hich and is s t red for ws tes . ve = te is d e 1 - f l 2
N Definition 1 (Discrete Fourier Transform (DFT) over Fq) n 1 information symbols redundancy symbols Let p(x) = p0 + p1x + + pn 1x − be a polynomial of PSfrag replacements ··· − K(1) degree deg (p(x)) n 1 with coefficients from Fq. Further, ≤ − (2) let α Fq be some element of order n. Then, the polynomial K ∈ K(3) n 1 l P (x)= F (p(x)) = P0 + P1x + + Pn 1x − ··· − i K(l) whose coefficients are calculated by Pi = p(α ) is called the Discrete Fourier Transform of p(x) over the field Fq. burst error The inverse of the Discrete Fourier Transform pursuant to Fig. 1. Interleaved Reed–Solomon code. Definition 1 can be calculated in the well-known way as follows: Let P (x) be a polynomial of deg (P (x)) n 1, and denote by ≤ − code with the largest dimension. 1 n 1 Algebraic decoding of a single Reed–Solomon codeword p(x)= F − (P (x)) = p0 + p1x + + pn 1x − ··· − can efficiently be performed by the Berlekamp–Massey al- the Inverse Discrete Fourier Transform. Then, the coefficients gorithm [16], [17], which is based on a single-sequence of are calculated by 1 i . shift-register synthesis approach. It is mentioned in [2] and pi p(x) pi = n− P (α− ) In the terminology of the Fourier Transform, we formally [3], that decoding of interleaved Reed–Solomon codes can be performed on the basis of a multi-sequence shift-register call p(x) time domain polynomial and P (x) the frequency synthesis algorithm. Such an algorithm is described by Feng domain polynomial or the spectrum of p(x). We now describe Reed–Solomon codes based on the Discrete Fourier Trans- and Tzeng in [18] and [19]. For homogeneous IRS codes, this algorithm provides an effective method for collaborative form. The following definition does not describe the class of decoding. However, it is shown in [20] and [21] that the Reed–Solomon codes in its most general way. However, with Feng–Tzeng algorithm does not always work correctly for regard to a concise notation, it is adequate for our purposes: sequences of varying length. Thus, it cannot be applied for Definition 2 (Reed–Solomon (RS) code) the heterogeneous case, since the different dimensions of the Let codes (1), (2),..., (l) result in syndrome sequences of A A A K 1 different lengths. Hence, we use the multi-sequence shift- − i A(x) = Aix , Ai Fq register synthesis algorithm proposed in [20] to decode both, { } ∈ ( i=0 ) homogeneous and heterogeneous IRS codes. The complexity X of this shift-register based approach is similar to the complex- be the set of all polynomials of deg (A(x)) < K with coeffi- F F ity of applying the Berlekamp–Massey algorithm to all words cients Ai from q, and let α q be some element of order N < q. Then, a Reed–Solomon∈ code = (q; N,K,D) of the interleaved Reed–Solomon code independently. A RS In the following sections, we describe the basic principles of length N, dimension K, and minimum Hamming distance D = N K +1 can be defined as the set of polynomials behind collaborative IRS decoding. Based on this, we derive − the maximum correction radius. Furthermore, we discuss how 1 , a(x)= F − (A(x)) A(x) A(x) . the code properties and the decoding performance is influenced A | ∈{ } by the IRS code design, i.e., by the choice of the parameters The codewords of are represented by the polynomials a(x)= (1) (2) (l) A n 1 of , ,..., . Moreover, we derive bounds on the a0 + a1x + + an 1x − , or alternatively by the n-tuples A A A ··· − failure and error probability, which allow us to estimate the a = (a0,a1,...,an 1). performance gain, which is achievable by collaborative decod- − ing. On the basis of these results, we investigate concatenated Reed–Solomon codes fulfill the Singleton Bound with equality, code designs with outer IRS codes and inner block codes, and are therefore Maximum Distance Separable (MDS). This and derive bounds on the overall decoding performance. These means that from any set of K correct symbols, a Reed– considerations are supplemented by Monte Carlo simulations, Solomon codeword can uniquely be reconstructed. to assess the quality of our bounds. An interleaved Reed–Solomon code is now obtained by taking l Reed–Solomon codes according to Definition 2 and grouping them row-wise into a matrix: II.INTERLEAVED REED–SOLOMON CODES As already mentioned, the codewords of an IRS code are Definition 3 (Interleaved Reed–Solomon (IRS) code) matrices whose rows are the codewords of the Reed–Solomon Let (ℓ) = q; N,K(ℓ),D(ℓ) , ℓ = 1,...,l, be l Reed– codes (1), (2),..., (l). Before we formally define IRS SolomonA codesRS of length N according to Definition 2. Then, an codes,A we brieflyA introduceA classical Reed–Solomon codes, interleaved Reed–Solomon code is the set of matrices basically to establish the notation we use in the following. a(1) Reed–Solomon codes can conveniently be defined in the a(2) frequency domain, i.e., by the Fourier Transform of their code- A , . , a(ℓ) (ℓ), ℓ =1,...,l . words. For this purpose, we consider the following definition . ∈ A of the Discrete Fourier Transform over finite fields: a(l) PREPRINT -- PREPRINT -- PREPRINT -- PREPRINT -- 3
(1) (2) If all l Reed–Solomon codes are equivalent, i.e., = = of the erroneous symbols. Therefore, Λ(x)=Λ0 +Λ1x + (l) A A t = = , the IRS code is called homogeneous. +Λtx is called error locator polynomial. An error locator ···Otherwise,A we sayA that the IRS code is heterogeneous. ···polynomial was first applied by Peterson [22] for decoding BCH codes. Instead of considering an IRS codeword as matrix l N Since the roots of the error locator polynomial are not A with elements from the extension field F , we× can also q modified by multiplying Λ(x) by a constant factor, and since consider it as a row vector a with elements from the extension Λ0 = 0, Λ(x) can always be normalized w.l.o.g. in such field F l . In this extension field representation, an IRS code q a way6 that Λ = 1. Equation (1) forms a linear system of l k(ℓ) 0 is a code of length N, cardinality ℓ=1 q , and minimum N equations. In this system, t equations only depend on the (ℓ) F distance min1 ℓ l N K +1 over ql . Hence, for the ≤ ≤ − M = N K known syndrome coefficients S0,...,SM 1 and homogeneous case, i.e., if (ℓ) =Q (q; N,K,D), ℓ = − − A RS the unknown coefficients Λ1,..., Λt. With these t equations, 1,...,l, we obtain a code of length N, dimension K, and we write the matrix equation minimum distance D = N K +1 over F l . This means − q that homogeneous IRS codes fulfill the Singleton Bound with S0 S1 ... St 1 Λt St − − equality and are therefore MDS. However, since the minimum S1 S2 ... St Λt 1 St+1 distance of a heterogeneous IRS code is determined by the . . . .− = − . , . . . . . weakest Reed–Solomon code, i.e., the Reed–Solomon code SM t 1 SM t ... SM 2 Λ1 SM 1 with the smallest minimum distance, heterogeneous IRS codes − − − − − − are generally not MDS. S Λ T (2) |which is a linear{z system of N } | K{z t} equations| {z and} t III. COLLABORATIVE DECODING OF INTERLEAVED − − EED OLOMON ODES unknowns. Hence, (2) cannot have a unique solution, if t > R –S C N K N K −2 and we are never able to correct more than −2 Before we explain the concept of collaborative interleaved N K ⌊ ⌋ errors. If t −2 , solving (2) yields a unique error locator Reed–Solomon decoding, we consider the conventional case of polynomial,≤⌊ and hence⌋ also the locations of the erroneous a single Reed–Solomon code = (q; N,K,D). Decoding A RS symbols. Basically, (2) can be solved by standard methods is usually performed by an algebraic Bounded Minimum Dis- from linear algebra. This approach, also known as Peterson tance (BMD) decoder, which corrects all errors with weights Algorithm, is described in [23]. However, today the Peterson up to half the minimum code distance. Algorithm is of minor practical relevance since there exist algorithms which utilize the structure of (2) to solve the system A. BMD Decoding up to Half the Minimum Distance of equations in a more efficient way. Assume that a codeword a(x) is transmitted over To understand how the structure of (2) can be exploited, we ∈ A a noisy channel, which adds an error polynomial e(x) = observe that we are able to state (2) in the form of the linear N 1 e0 + e1x + + eN 1x − over Fq, so that we observe recursion the word ··· − at the output of the channel. If r(x)= a(x)+ e(x) t e(x) 0, the Discrete Fourier Transform R(x)= F (r(x)) = (ℓ) ≡ Si = Λj Si j , i = t,...,N K 1 (3) R + R x + + R xN 1 will be a polynomial of degree − − − − 0 1 N 1 − j=1 smaller than···K due− to Definition 2, i.e., the coefficients X RK ,...,RN 1 will be zero. However, if e(x) / , we obtain of length t. In this way, the problem of calculating Λ(x) is − ∈ A deg (R(x)) K, which means that some of the coefficients transformed to the problem of finding the smallest integer t and ≥ RK ,...,RN 1 are non-zero. These coefficients only depend the connection weights Λ1,..., Λt for recursively generating − F N K 1 on the transformed error E(x) = (e(x)) = E0 + E1x + the syndrome sequence = Si i=0− − . This problem is N 1 S { } + EN 1x − . Hence, we can consider the coefficients equivalent to the problem of determining the shortest possible ··· − RK ,...,RN 1 as syndrome coefficients and denote them by linear feedback shift-register, which is capable of generating − Sj = RK j = EK j , j =0,...,N K 1. the syndrome sequence . This single-sequence shift-register + + − − To correct t errors, the standard approach for algebraic synthesis problem is solvedS very efficiently by the Berlekamp– Reed–Solomon decoding is to define the polynomial λ(x) = Massey algorithm [16], [17]. N 1 λ0 +λ1x+ +λN 1x − , such that it has a zero coefficient When Λ(x) is determined, and hence the erroneous posi- ··· − λj = 0, whenever the corresponding coefficient ej of the tions are known, the most difficult part of decoding is accom- error polynomial e(x) is non-zero. For all error free positions, plished. If the error locations are known, the error values can i.e., for all positions for which ej = 0, the corresponding always be uniquely determined, as long as t N K. Error ≤ − coefficient λj is defined to be non-zero. Consequently, λj ej = evaluation can be performed using several standard techniques · 0 holds for all positions j =0,...,N 1. Due to the properties like Recursive Extension [24] or the Forney algorithm [25]. of the Discrete Fourier Transform, this− relation is transformed into Λ(x) E(x) 0 mod xN 1 . (1) B. Collaborative Error Location · ≡ − The spectrum Λ(x) is a polynomial of degree t, whose roots Now, we consider an interleaved Reed–Solomon code pur- j1 j2 jt α− , α− ,...,α− correspond to the locations j1, j2,...,jt suant to Definition 3. We assume that we observe a received PREPRINT -- PREPRINT -- PREPRINT -- PREPRINT -- 4
word of length t. All l linear recursions use the same connection (1) (1) (1) r a e weights Λ1,..., Λt to combine the syndrome coefficients of r(2) a(2) e(2) the l received words. Hence, the error locator polynomial R = . = . + . , . . . Λ(x) can be calculated by finding the smallest integer t and (l) (l) (l) the connection weights Λ1,..., Λt for recursively generating r a e (ℓ) (ℓ) N K 1 all l different syndrome sequences = Si i=0− − . A E This is equivalent to synthesizing theS shortest linear{ } feedback where the ℓth row of R consists| {z of} the| coefficients{z } of the re- shift-register capable of generating the l syndrome sequences ceived word r(ℓ)(x)= a(ℓ)(x)+e(ℓ)(x). Each error polynomial (1),..., (l). For a homogeneous Reed–Solomon code, i.e., S S e(ℓ)(x) has t(ℓ) t non-zero coefficients located at the indices if all l sequences have the same length, this multi-sequence ≤ i1,i2,...,it(ℓ) j1, j2,...,jt . Hence, R = A + E shift-register problem can be solved by the Feng-Tzeng al- can{ be written} as ⊆ sum { of a codeword} A A and an error gorithm [18], [19]. However, for a heterogeneous IRS code, matrix E with exactly t non-zero columns.∈ For each row in the sequences (1),..., (l) may be of different length. It is S S R, we are able to calculate a syndrome sequence (ℓ) = demonstrated in [20] and [21] that the Feng–Tzeng algorithm (ℓ) N K(ℓ) 1 S (ℓ) does not always yield the shortest shift-register for varying Si i=0− − . However, since we allow the codes to{ have} different dimensions K(ℓ), the calculated syndromeA length sequences. Hence, we apply the varying length shift- sequences may be of different lengths. Nevertheless, as long register synthesis algorithm proposed in [20] to calculate as t
Once a t-valid error locator polynomial Λ(x) is successfully D. Joint Error and Erasure Correction calculated for an IRS code, error evaluation can be performed In [11], an algorithm for joint error and erasure decoding independently for all l Reed–Solomon codewords by the of IRS codes is proposed. Like the algorithm presented in standard techniques also used for classical BMD decoding. [4] for decoding errors only, the algorithm from [11] uses a Using Algorithm 1 for calculating Λ(x) yields the same Welch-Berlekamp approach, which is based on solving a linear computational complexity as locating the errors in the l code- system of equations. We briefly explain how Algorithm 2 can words independently by the Berlekamp–Massey algorithm. be modified to allow joint error and erasure correction based Consequently, we are able to increase the error correction on shift-register synthesis. radius above half the minimum distance of the Reed–Solomon Technically, erasures occur if the receiver is not able to code with the smallest distance, without increasing the decod- detect any meaningful symbol at the output of the channel. ing complexity. This means that the decoder does not know the value, but the Altogether, our collaborative decoding strategy for IRS position of an erased symbol. Formally, we define an erasure codes consists of four steps. First, we calculate l syndromes as follows: Let ξ denote an erased symbol, and define it to be (1),..., (l), next, we synthesize an error locator polynomial F F S S a special symbol ξ / q. Moreover, let the addition of α q Λ(x), then we check whether this polynomial is t-valid, and and ξ be defined by∈α + ξ = ξ + α = ξ. Now, assume that∈ we whether t is smaller than some maximum error correcting observe a received word radius tmax, which will be specified later. If we have a t- valid Λ-polynomial, we calculate values of the errors e(ℓ)(x) R = A + E + E , (ℓ) independently for all received words r (x), and obtain esti- where the rows of the matrix mates ǫ(1) aˆ(ℓ)(x)= r(ℓ)(x) e(ℓ)(x) − ǫ(2) E for ℓ = 1,...,l. If Λ(x) is not t-valid, we get a decoding = . . failure. The complete decoding algorithm is summarized by ǫ(l) Algorithm 2. (ℓ) (ℓ) are the coefficients of the erasure polynomials ǫ (x)= ǫ0 + Algorithm 2: Collaborative IRS Decoder (ℓ) (ℓ) n 1 ǫ1 x + + ǫn 1x − over 0, ξ , ℓ = 1,...,l. Each non- r(1) zero element··· in −E indicates an{ erased} symbol. In contrast to R . input: received word = . errors, the erased symbols are detectable in the received word r(l) ! R, which means that they do not have to be located, only use DFT to calculate syndromes (1),..., (l) their values have to be determined. For this reason, we do not synthesize t, Λ(x) by AlgorithmS 1 S require the erasures to occur in column bursts. if t tmax and Λ(x) is t-valid then To perform joint error and erasure decoding for IRS codes, for≤ each ℓ from 1 to l do we proceed like described in [26] and define an erasure locator evaluate errors, and calculate e(ℓ) polynomial (ℓ) (ℓ) (ℓ) calculate aˆ = r e (ℓ) (ℓ) i(ℓ) i(ℓ) i − Ψ (x) , 1 α 1 x 1 α 2 x 1 α ϑℓ x = else − − ··· − (ℓ) (ℓ) (ℓ) decoding failure =1+Ψ x +Ψ x2 + +Ψ xϑℓ 1 2 ··· ϑℓ aˆ(1) for each of the rows ℓ = 1,...,l in our IRS code. The . A (ℓ) (ℓ) (ℓ) output: A = . or decoding failure i i iϑ . roots α− 1 , α− 2 ,...,α− ℓ indicate the positions of the (l) ! ∈ aˆ erasures in the erasure polynomial ǫ(ℓ)(x), i.e., the coefficients b (ℓ) (ℓ) (ℓ) of ǫ(ℓ)(x) at the positions i ,i ,...,i are equal to ξ, and 1 2 ϑℓ Generally, depending on the errors added by the channel, all other coefficients are zero. Now we consider the received Algorithm 2 may yield three different results: polynomial r(ℓ)(x) whose coefficients are the ℓth row of R. (ℓ) (ℓ) 1) The algorithm may obtain a correct result, i.e., a(ℓ)(x) From r (x) we create the polynomial r¯ (x) by replacing aˆ(ℓ)(x) ℓ =1,...,l. ≡ all erasure elements by an arbitrary field element. Then, we ¯(ℓ) F (ℓ) ˘(ℓ) ¯(ℓ) 2) The algorithm∀ may obtain an erroneous result, i.e., ℓ calculate R (x)= r¯ (x) and R (x)= R (x)Ψ(x). (ℓ) (ℓ) (ℓ) (ℓ) ∃ ∈ It is explained in [26] that R˘(ℓ)(x) = R˘ + R˘ x + + [1,...,l]: a (x) aˆ (x). 0 1 6≡ ˘(ℓ) n 1 (···ℓ) 3) The algorithm may not yield a result at all, i.e., it may Rn 1x − is the spectrum of a modified received word r˘ (x) yield a decoding failure. with− errors at the same positions as r(ℓ)(x). Moreover, the last (ℓ) ˘(ℓ) All three events occur with a certain probability. In the N K ϑℓ coefficients of R (x) are only influenced by the errors,− so− that we are able to obtain the modified syndromes following, we denote the probability for a correct decision by (ℓ) ˘(ℓ) ˘(ℓ) N K 1 ϑℓ ˘(ℓ) ˘(ℓ) = S − − − , where S = R (ℓ) , i = Pc, the probability for an erroneous decision by Pe, and the i i=0 i K +ϑℓ+i S { }(ℓ) probability for a decoding failure by Pf . Hence, the probability 0,...,N K ϑℓ 1, ℓ = 1,...,l. This means that − − − Pw for obtaining a wrong decoding result is calculated by each erasure symbol in the row ℓ reduces the length of the (ℓ) Pw =1 Pc = Pe + Pf . usable syndrome ˘ by one. In this way, we obtain l (varying PREPRINT− -- PREPRINT -- PREPRINTS -- PREPRINT -- 6
length) syndrome sequences, which we feed into Algorithm 1 Proof: Assume that Algorithm 2 is applied to the error to calculate a pair t, Λ(x). Once, a t-valid polynomial Λ(x) matrix e(1) is calculated and t tmax, we proceed like described in (2) [26] to obtain the error≤ and erasure values independently for e E = . . each row, using the error and erasure locator polynomial . (ℓ) (ℓ) Ω (x)=Ψ (x)Λ(x). e(l) In some sense, error and erasure decoding of an IRS code is equivalent to error only decoding of a heterogeneous IRS In its fist step, Algorithm 2 computes the Discrete Fourier code, since in both cases we process varying length syndrome Transform for every row of E: sequences. For this reason, we do not further consider the error (1) (1) (1) (1) F e E0 E1 ... EN 1 and erasure case explicitly, but keep in mind that it is closely F e(2) (2) (2) (2)− E0 E1 ... EN 1 related to the case of heterogeneous IRS codes. F (E)= = − . . . . . . . . (l) (l) (l) (l) F e E E ... E IV. ERROR CORRECTING RADIUS 0 1 N 1 − The last N K(ℓ ) elements of the ℓth row coincide with the If the number of errors t is smaller than half the mini- elements of− the syndrome sequence (ℓ). Hence, we formally mum distance of the Reed–Solomon code with the smallest define the syndrome matrix S redundancy, we are always able to correct all errors by solving (1) (1) (1) (3) separately for all l Reed–Solomon words using standard ξ ξ EK(1) EK(1) +1 ... EN 1 techniques like the Berlekamp–Massey algorithm. Certainly, it (2) (2) (2) (2)− ξ EK(2) EK(2) +1 EK(2) +2 ... EN 1 is important that Algorithm 2 is also capable of decoding all S(E)= − . . . errors in this case. To prove that Algorithm 2 actually has this . . . property, we first consider the following lemma: (l) (l) ξξ ξ EK(l) ... EN 1 − (1) (1) (1) ξ ξ S S ... S Fn 0 1 N K(1) 1 Lemma 1 Let v, and w be two vectors in q , such that the (2) (2) (2) (2)− − ξ S0 S1 S2 ... SN K(2) 1 Hamming weights of v and w satisfy wt (v) < q 1 or = − − − F . . . wt (w) < q 1. Then, there always exists an element β q, . . . − ∈ β =0, such that the support of v + βw satisfies ξξ ξS(l) ... S(l) 6 0 N 1 − supp (v + βw) = supp (v) supp (w) . (6) by replacing the first K(ℓ) elements of the ℓth row by erasure ∪ symbols, i.e., by symbols ξ / Fq, which fulfill α+ξ = ξ+α = ∈ ξ, and α ξ = ξ α = ξ for an arbitrary element α Fq. In other words,· S·(E) is an l N matrix containing∈ the l Proof: Since wt (v) < q 1, or wt (w) < q 1 the cardi- syndrome sequences aligned to× the right, and prepended by nality of the intersection =− supp (v) supp (w−) of the sup- D ∩ erasure symbols. From the linearity of the Discrete Fourier ports of v = (v0, v1,...,vN 1) and w = (w0, w1,...,wN 1) Transform we know that is at most = min wt (v−) , wt (w) < q 1. To obtain− (6), |D| { } − F F F vi + βwi =0 has to be satisfied for all i . In other words, E = (LE)= L (E) , 6 ∈ D vi β = i . (7) and consequently alsoe S(E)= LS(E). 6 −wi ∀ ∈ D Since (5) is a linear recursion, we know (see e.g. [27]) that Equation (7) can always be satisfied choosing a suitable β, if the rows of S(E) satisfye (5) with some pair t, Λ(x), then since < q 1, and therefore there exists at least one the rows of LS(E) also satisfy (5) with the same pair t, 1 β =0,|D| which fulfills− (7). Λ(x). Since L is a non-singular matrix, E = L− E, and we 6 Now, we consider a received word R = A + E, where A conclude that S(E) satisfies (5) with the pair t, Λ(x), if and is a codeword of an IRS code and E is an error matrix. Since only if S(E) satisfies (5) with the same pair t, Λ(ex). Thus, the syndromes (1),..., (l) calculated by Algorithm 2 only if Algorithm 2 has a unique solution t, Λ(x) for S(E), then depend on E andS not onSA, we can assume w.l.o.g. the case Algorithme 2 has the same unique solution for S(E), and vice A = 0 to analyze Algorithm 2. In other words, we are able versa. to apply Algorithm 2 directly to the error matrix E instead of e Theorem 1 (Guaranteed Correcting Radius) Consider an the received word R. interleaved Reed–Solomon code pursuant to Definition 3. Assume that this code is corrupted by an error matrix E with Lemma 2 Assume that E is an l N error matrix and L is × t non-zero columns. Then, Algorithm 2 always yields a unique a non-singular l l matrix over Fq. Moreover, let E = LE × and correct solution, i.e., all t column errors can be corrected, be another error matrix, whose rows are linear combinations as long as t satisfies of the rows of E. Then, Algorithm 2 applied to Eeyields a unique pair t, Λ(x), if and only if Algorithm 2 applied to E N K t tg = − , (8) yields the same unique pair t, Λ(x). ≤ $ 2 % PREPRINT -- PREPRINT --e PREPRINT --b PREPRINT -- 7
where dimension. This gives rise to the following theorem on the K = max K(ℓ) maximum error correcting radius: 1 ℓ l ≤ ≤ n o is the maximum dimensionb among the l Reed–Solomon codes Theorem 2 (Maximum Correcting Radius) Consider an (1), (2),..., (ℓ). interleaved Reed–Solomon code pursuant to Definition 3. A A A Assume that a word of this code is corrupted by t column Proof: Assume that a codeword of the interleaved Reed– errors. Then, Algorithm 2 may only find a unique and correct Solomon code is corrupted by an error matrix E with t non- solution, if t satisfies zero columns. We distinguish two different cases: (1) (1) l 1) If wt e = supp e = t, then the Berlekamp– t tmax = min N K¯ ,N K , (10) (1) ≤ l +1 − − Massey algorithm applied to the sequence yields a pair S t, Λ(x) which is the shortest length solution of (3), provided where b l that t satisfies (8). Since we consider column errors, the pair 1 K¯ = K(ℓ) t, Λ(x) is also a solution of (3) for the remaining sequences l (2),..., (l). In other words, t, Λ(x) is the shortest length ℓ=1 S S X solution of (5), which is also obtained by applying Algorithm1 is the average dimension of the l Reed–Solomon codes. to the l sequences (1), (2),..., (l). Consequently, Algo- rithm 2 yields the correctS S solution.S Proof: To find a unique and correct solution, the linear system of equations (4) with t unknowns and the set of linear e(1) e(1) 2) Now, we consider the case wt = supp < recursions (5) of length t must have a unique solution, which t, i.e., the case that the column errors only affect t(1) < t means that rank (Sl) = t. This is only possible, if (9) is symbols in e(1). Since satisfied, since otherwise the matrix Sl has less than t rows. N K Hence, in order to be able to locate the errors, t must not be t − < q 1 larger than the first argument in (10). Even if the locations of ≤ 2 − $ % the errors are known, the error values can only be calculated b is always satisfied due to Definition 2, we are able to recur- if t is not larger than the number of redundancy symbols in sively apply Lemma 1 to obtain coefficients , every Reed–Solomon code. Thus, if t>N K, it is not l β1,β2,...,βl − β1 =1, such that the linear combination possible to evaluate the errors in at least one Reed–Solomon codeword. Consequently, t must also not be largerb than the l second argument in (10). e˜(1) = β e(ℓ) ℓ The Theorems 1 and 2 characterize the collaborative de- ℓ X=1 coding approach described by the linear system of equations is an error vector with t non-zero coefficients, i.e., we have (4) or equivalently by the set of linear recursions (5). If the (1) supp e˜ = t. Hence, we define the non-singular matrix number of column errors satisfy (8), i.e., if the number of errors is smaller than half the minimum distance of the Reed– 1 β ... β 2 l Solomon code with the largest dimension, (4) and hence also 0 1 ... 0 (5) have a unique solution, and we are always able to correctly L = . . . . , . . .. . reconstruct the transmitted codeword A. If the number of 0 0 ... 1 errors lies between the guaranteed error correcting radius specified by (8) and the maximum correcting radius specified and use it to calculate the error matrix E = LE, which has t by (10), (4) and (5) may still have a unique solution, which non-zero error symbols in its first row. Thus, we know from enables us to correct errors beyond half the minimum distance case 1) that Algorithm 1 synthesizes thee correct pair t, Λ(x) of the component codes. In this sense, our collaborative for E, and that Algorithm 2 consequently yields the correct decoding approach yields a Bounded Distance (BD) decoder, solution for the matrix E. Due to Lemma 2, we know that whose maximum error correction radius lies beyond half the Algorithme 2 also yields the correct solution for the error matrix minimum distance. However, as soon as the number of column E, which proves Theoreme 1 for the case 2). errors exceeds the limit specified by (8), we are not able any more to guarantee that (4) and hence also (5) has a unique Beyond the guaranteed error correcting radius we may be solution. In fact, the solutions of (4) and (5) will be ambiguous able to locate the errors, as long as the number of unknowns with some probability Pf > 0, since rank (Sl) < t for some in (4) is not larger than the number of equations, i.e., as long error patterns. as The possibility to obtain an ambiguous solution is inherent l l 1 to Bounded Distance decoders which correct errors within a t N K(ℓ) (9) radius beyond half the minimum distance, since some correct- ≤ l +1 − l ℓ ! X=1 ing spheres are inevitably overlapping in this case. Basically, is satisfied. Moreover, we are only able to evaluate the errors, three strategies are conceivable to cope with this problem: if the number of errors does not exceed the number of 1) The decoder selects on single codeword out of all valid redundancy symbols in the component code with the largest decoding results PREPRINT -- PREPRINT -- PREPRINT -- PREPRINT -- 8
2) The decoder returns a list of all valid decoding results over all component codes. Though, if l is small enough, the that is, a list of all codewords which lie inside a sphere redundancy may be distributed unequally, as long as we only with radius tmax around the received word y. care about the maximum error correcting radius, and not about 3) If the decoding result is not unique, i.e., if there exists the minimum distance of the IRS code. However, the minimum more than one valid decoding result, the decoder yields distance is only optimal in the homogeneous case. This is a decoding failure explained by the fact that homogeneous IRS codes are MDS, Since finding all valid solutions is algebraically difficult with while the minimum distance of heterogeneous IRS codes our approach and usually requires a high computational effort, is determined by the Reed–Solomon code with the largest the decoders we consider here apply the third strategy: when- dimension. Nevertheless, heterogeneous IRS codes can be ever (4) does not have a unique solution, Algorithm 2 may interesting e.g. in the context of generalized concatenated code yield a decoding failure (we formalize this statement later in constructions introduced and described by Blokh and Zyablov Lemma 5). [13] and Zinoviev [14], for decoding single Reed–Solomon codes beyond half the minimum distance like described in V. DISTRIBUTIONOFTHE REDUNDANCY [15], and for several other applications. Definition 3 does not restrict the choice of the dimensions of VI. COLLABORATIVE DECODING PERFORMANCE the l Reed–Solomon codes used to construct a heterogeneous IRS code. However, to ensure a good performance under To evaluate the performance of our decoding strategy, we collaborative decoding, we should apply some restrictions on would like to estimate the error probability Pe, and the failure the dimensions K(1),...,K(l). probability Pf , which together result in the probability Pw From (10) we observe that the maximum error correction for obtaining a wrong decoding result. We are particularly radius depends on the average redundancy N K¯ and the interested in the failure probability Pf (t) in the range tg