
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 4, APRIL 2006 1289 Compressed Sensing David L. Donoho, Member, IEEE Abstract—Suppose is an unknown vector in (a digital the important information about the signals/images—in effect, image or signal); we plan to measure general linear functionals not acquiring that part of the data that would eventually just be of and then reconstruct. If is known to be compressible “thrown away” by lossy compression. Moreover, the protocols by transform coding with a known transform, and we recon- struct via the nonlinear procedure defined here, the number of are nonadaptive and parallelizable; they do not require knowl- measurements can be dramatically smaller than the size . edge of the signal/image to be acquired in advance—other than Thus, certain natural classes of images with pixels need only knowledge that the data will be compressible—and do not at- S P a @ I R log @ AA nonadaptive nonpixel samples for tempt any “understanding” of the underlying object to guide faithful recovery, as opposed to the usual pixel samples. an active or adaptive sensing strategy. The measurements made More specifically, suppose has a sparse representation in some orthonormal basis (e.g., wavelet, Fourier) or tight frame in the compressed sensing protocol are holographic—thus, not (e.g., curvelet, Gabor)—so the coefficients belong to an ball simple pixel samples—and must be processed nonlinearly. for H I. The most important coefficients in that In specific applications, this principle might enable dra- I P I expansion allow reconstruction with P error @ A.Itis matically reduced measurement time, dramatically reduced a @ log@ AA possible to design nonadaptive measurements sampling rates, or reduced use of analog-to-digital converter allowing reconstruction with accuracy comparable to that attain- able with direct knowledge of the most important coefficients. resources. Moreover, a good approximation to those important coeffi- cients is extracted from the measurements by solving a linear A. Transform Compression Background program—Basis Pursuit in signal processing. The nonadaptive Our treatment is abstract and general, but depends on one spe- measurements have the character of “random” linear combi- cific assumption which is known to hold in many settings of nations of basis/frame elements. Our results use the notions of optimal recovery, of -widths, and information-based complexity. signal and image processing: the principle of transform sparsity. We estimate the Gel’fand -widths of balls in high-dimensional We suppose that the object of interest is a vector , which Euclidean space in the case H I, and give a criterion can be a signal or image with samples or pixels, and that there identifying near-optimal subspaces for Gel’fand -widths. We is an orthonormal basis for which can show that “most” subspaces are near-optimal, and show that be, for example, an orthonormal wavelet basis, a Fourier basis, convex optimization (Basis Pursuit) is a near-optimal way to extract information derived from these near-optimal subspaces. or a local Fourier basis, depending on the application. (As ex- plained later, the extension to tight frames such as curvelet or Index Terms—Adaptive sampling, almost-spherical sections of Banach spaces, Basis Pursuit, eigenvalues of random matrices, Gabor frames comes for free.) The object has transform coeffi- Gel’fand -widths, information-based complexity, integrated cients , and these are assumed sparse in the sense sensing and processing, minimum I-norm decomposition, op- that, for some and for some timal recovery, Quotient-of-a-Subspace theorem, sparse solution of linear equations. (I.1) I. INTRODUCTION Such constraints are actually obeyed on natural classes of sig- S our modern technology-driven civilization acquires and nals and images; this is the primary reason for the success of A exploits ever-increasing amounts of data, “everyone” now standard compression tools based on transform coding [1]. To knows that most of the data we acquire “can be thrown away” fix ideas, we mention two simple examples of constraint. with almost no perceptual loss—witness the broad success of • Bounded Variation model for images. Here image bright- lossy compression formats for sounds, images, and specialized ness is viewed as an underlying function on the technical data. The phenomenon of ubiquitous compressibility unit square , which obeys (essentially) raises very natural questions: why go to so much effort to ac- quire all the data when most of what we get will be thrown away? Can we not just directly measure the part that will not end up being thrown away? In this paper, we design compressed data acquisition proto- The digital data of interest consist of pixel sam- cols which perform as if it were possible to directly acquire just ples of produced by averaging over pixels. We take a wavelet point of view; the data are seen as a su- perposition of contributions from various scales. Let Manuscript received September 18, 2004; revised December 15, 2005. The author is with the Department of Statistics, Stanford University, Stanford, denote the component of the data at scale , and let CA 94305 USA (e-mail: [email protected]). denote the orthonormal basis of wavelets at scale , con- Communicated by A. Høst-Madsen, Associate Editor for Detection and Estimation. taining elements. The corresponding coefficients Digital Object Identifier 10.1109/TIT.2006.871582 obey . 0018-9448/$20.00 © 2006 IEEE 1290 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 4, APRIL 2006 • Bump Algebra model for spectra. Here a spectrum (e.g., use “OR/IBC” as a generic label for work taking place in those mass spectrum or magnetic resonance spectrum) is fields, admittedly being less than encyclopedic about various modeled as digital samples of an underlying scholarly contributions. function on the real line which is a superposition of We have a class of possible objects of interest, and are so-called spectral lines of varying positions, amplitudes, interested in designing an information operator and linewidths. Formally that samples pieces of information about , and an algorithm that offers an approximate reconstruction of . Here the information operator takes the form Here the parameters are line locations, are ampli- tudes/polarities, and are linewidths, and represents a where the are sampling kernels, not necessarily sampling lineshape, for example the Gaussian, although other pro- pixels or other simple features of the signal; however, they are files could be considered. We assume the constraint where nonadaptive, i.e., fixed independently of . The algorithm is , which in applications represents an en- an unspecified, possibly nonlinear reconstruction operator. ergy or total mass constraint. Again we take a wavelet We are interested in the error of reconstruction and in the viewpoint, this time specifically using smooth wavelets. behavior of optimal information and optimal algorithms. Hence, The data can be represented as a superposition of con- we consider the minimax error as a standard of comparison tributions from various scales. Let denote the com- ponent of the spectrum at scale , and let denote the orthonormal basis of wavelets at scale , containing elements. The corresponding coefficients again obey So here, all possible methods of nonadaptive linear sampling are , [2]. allowed, and all possible methods of reconstruction are allowed. While in these two examples, the constraint appeared, other In our application, the class of objects of interest is the set constraints with can appear naturally as well; of objects where obeys (I.1) for a given see below. For some readers, the use of norms with and . Denote then may seem initially strange; it is now well understood that the norms with such small are natural mathematical measures of sparsity [3], [4]. As decreases below , more and more sparsity Our goal is to evaluate and to have practical is being required. Also, from this viewpoint, an constraint schemes which come close to attaining it. based on requires no sparsity at all. Note that in each of these examples, we also allowed for sep- C. Four Surprises arating the object of interest into subbands, each one of which obeys an constraint. In practice, in the following we stick Here is the main quantitative phenomenon of interest for this with the view that the object of interest is a coefficient vector paper. obeying the constraint (I.1), which may mean, from an applica- Theorem 1: Let be a sequence of problem sizes with tion viewpoint, that our methods correspond to treating various , , and , , . Then for subbands separately, as in these examples. there is so that The key implication of the constraint is sparsity of the transform coefficients. Indeed, we have trivially that, if de- (I.3) notes the vector with everything except the largest coeffi- cients set to We find this surprising in four ways. First, compare (I.3) with (I.2) (I.2). We see that the forms are similar, under the calibration . In words: the quality of approximation to for , with a constant depending only on which could be gotten by using the biggest transform coef- . Thus, for example, to approximate with error ,we ficients can be gotten by using the pieces of need to keep only the biggest terms in . nonadaptive information provided by . The surprise is that we would not know in advance which transform coefficients are B. Optimal Recovery/Information-Based Complexity likely to be the important ones in this approximation, yet the Background optimal information operator
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages18 Page
-
File Size-