Error Correction Via Linear Programming

Error Correction Via Linear Programming

Error Correction via Linear Programming Emmanuel Candes Mark Rudelson Terence Tao Applied and Computational Department of Mathematics Department of Mathematics Mathematics, Caltech, University of Missouri, University of California, Pasadena, CA 91125, USA Columbia, MO 65203, USA Los Angeles, CA 90095, USA [email protected] [email protected] [email protected] Roman Vershynin Department of Mathematics University of California, Davis, CA 9595616, USA [email protected] Abstract ear codes, sparse solutions to underdetermined systems, `1- minimization, linear programming, restricted orthonormal- Suppose we wish to transmit a vector f ∈ Rn reliably. A ity, Gaussian random matrices. frequently discussed approach consists in encoding f with an m by n coding matrix A. Assume now that a fraction 1 Introduction of the entries of Af are corrupted in a completely arbitrary fashion by an error e. We do not know which entries are affected nor do we know how they are affected. Is it possi- 1.1 The error correction problem ble to recover f exactly from the corrupted m-dimensional vector y0 = Af + e? This paper considers the model problem of recovering This paper proves that under suitable conditions on the an input vector f ∈ Rn from corrupted measurements coding matrix A, the input f is the unique solution to the y0 = Af + e. Here, A is an m by n matrix (we will as- P sume throughout the paper that m > n), and e ∈ m is `1-minimization problem (kxk`1 := i |xi|) R an unknown vector of errors. We will assume that at most 0 ˜ min ky − Afk`1 r entries are corrupted, thus at most r entries of e are non- ˜ n f∈R zero, but apart from this restriction e will be arbitrary. The provided that the fraction of corrupted entries is not too problem we consider is whether it is possible to recover f large, i.e. does not exceed some strictly positive constant exactly from the data y. And if so, how? ρ∗ ( numerical values for ρ∗ are actually given). In other In its abstract form, our problem is of course equivalent words, f can be recovered exactly by solving a simple to the classical error correcting problem which arises in cod- convex optimization problem; in fact, a linear program. ing theory as we may think of A as a linear code; a linear We report on numerical experiments suggesting that `1- code is a given collection of codewords which are vectors m minimization is amazingly effective; f is recovered exactly a1, . , an ∈ R —the columns of the matrix A. Given a even in situations where a very significant fraction of the vector f ∈ Rn (the “plaintext”) we can then generate a vec- output is corrupted. tor Af in Rm (the “ciphertext”); if A has full rank, then one In the case when the measurement matrix A is Gaus- can clearly recover the plaintext f from the ciphertext Af. sian, the problem is equivalent to that of counting low- But now we suppose that the ciphertext Af is corrupted by dimensional facets of a convex polytope, and in particular an arbitrary vector e ∈ Rm with at most r non-zero entries of a random section of the unit cube. In this case we can so that the corrupted ciphertext is of the form Af + e. The strengthen the results somewhat by using a geometric func- question is then: given the coding matrix A and Af +e, can tional analysis approach. one recover f exactly? Let us say that the linear code A is a (m, n, r)-error cor- Keywords. Linear codes, decoding of (random) lin- recting code if one can recover f from Af + e whenever e has at most r non-zero coefficients. As is well-known, 1.2 Solution via `1-minimization if the number r of corrupted entries is too large, then of course we have no hope of having a (m, n, r)-error correct- To recover f accurately from corrupted data y0 = Af + ing code. For instance, if n + 2r > m then elementary e, we consider solving the following `1-minimization (or linear algebra shows that there exist plaintexts f, f 0 ∈ Rn Basis Pursuit) problem and errors e, e0 ∈ Rm with at most r non-zero coefficients 0 0 0 ˜ each such that Af + e = Af + e , and so one cannot dis- (P1) min ky − Afk`1 (1.2) ˜ n tinguish f from f 0 in this case. In particular, if the fraction f∈R ρ := r of corrupted entries exceeds 1/2 then an (m, n, r)- m or equivalently error correcting code is impossible regardless of how large one makes m with respect to n. 0 0 (P ) min ky − yk` 1 n 1 This situation raises an important question: for which y˜∈R fraction ρ of the corrupted entries is accurate decoding pos- Thus the minimizer y? to (P 0) is the metric projection of y0 sible with practical algorithms? That is, with algorithms 1 onto the vector space Y with respect to the ` norm. This whose complexity is at most polynomial in the length m of 1 is a convex program which can be classically reformulated the codewords? as a linear program. Indeed, the `1-minimization problem Setting y := Af, and letting Y ⊂ m be the image R (P1) is equivalent to of A, we can rephrase the problem more geometrically as follows: how to reconstruct a vector y in an n-dimensional m m 0 m X 0 ˜ subspace Y of R from a vector y ∈ R that differs from min ti, −t ≤ y − Af ≤ t, y in at most r coordinates? i=1 If the matrix A is chosen randomly, for instance by the where the optimization variables are t ∈ Rm and f˜ ∈ n Gaussian ensemble, then it is easy to show that with prob- R (as is standard, the generalized vector inequality x ≤ y ability one that all the plaintexts can be distinguished in an means that x ≤ y for every coordinate i). Hence, (P ) information-theoretic sense as soon as n + 2r ≤ m. How- i i 1 is an LP with inequality constraints and can be solved ef- ever, this result provides no algorithm for recovering the ficiently using standard or specialized optimization algo- plaintext f from the corrupted ciphertext Af +e, other than rithms, see [4], [5]. brute force search, which has exponential complexity in m. The main claim of this paper is that for suitable coding Based on analogy with discrete (e.g. finite field) analogues matrices A, the solution f ? to our linear program is actually of this problem, to obtain a polynomial-time recovery algo- exact; f ? = f! (Equivalently, y? = y.) rithm it is more reasonable to expect as a necessary condi- tion the Gilbert-Varshamov bound Y Y n/m ≥ 1 − H(Cr/n) (1.1) u which is fundamental in coding theory (see [34]); here y y’ y=u y’ H(x) is the entropy function, and C, c, c1, etc. will be used to denote various positive absolute constants. This heuris- tic can be made rigorous if one requires a certain stability property for the recovery algorithm; see Section 6. (MLS) (BP) One can instead consider a mean square approach, based on the minimization problem The potential of Basis Pursuit for exact re- construction is illustrated by the following 0 ˜ (P2) min ky − Afk`2 heuristics, essentially due to [18]. The minimizer u ˜ n f∈R 0 to (P2) is the contact point where the smallest Euclidean ball centered at y0 meets the subspace Y . That contact or equivalently point is in general different from y. The situation is much 0 0 0 better in (P1): typically the solution coincides with y. (P2) min ky − y˜k`2 0 y˜∈Y The minimizer u to (P1) is the contact point where the smallest octahedron centered at y0 (the ball with respect but the minimizer f ? (resp. y?) may be arbitrarily far away to the 1-norm) meets Y . Because the vector y − y0 lies from the plaintext f (resp. y) since we have no size control in a low-dimensional coordinate subspace, the octahedron on the error e. has a wedge at y. Thus, many subspaces Y through y will 0 00 miss the octahedron of radius y − y (as opposed to the This convex program (P1 ) is easily seen to be equivalent to 0 0 Euclidean ball). This forces the solution u to (P1), which (P1) or (P1), and may be recast as an LP. is the contact point of the octahedron, to coincide with y. We now state the first main result of this paper, which we The idea of using the 1-norm instead of the 2-norm for prove in Section 2. better data recovery has been explored since mid-seventies n in various applied areas, in particular geophysics and statis- Theorem 1.1. Let f ∈ R , let A be an m × n matrix, let 0 tics (early history can be found in [47]). With the subse- e have at most r non-zero entries, let y = Af + e, and let quent development of fast interior point methods in Linear B be a matrix whose kernel equals the range of A. Suppose we also have the condition Programming, (P1) turned into an effectively solvable prob- lem, and was put forward more recently by Donoho and his δ + 3δ < 2, (1.6) collaborators, triggering massive experimental and theoret- 3r 4r ical work [5, 7, 8, 10, 14, 15, 17–21, 23, 27, 33, 45–47]. and let e ∈ Rm have at most r entries non-zero.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    14 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us