Acta Numerica http://journals.cambridge.org/ANU
Additional services for Acta Numerica:
Email alerts: Click here Subscriptions: Click here Commercial reprints: Click here Terms of use : Click here
On firstorder algorithms for l 1/nuclear norm minimization
Yurii Nesterov and Arkadi Nemirovski
Acta Numerica / Volume 22 / May 2013, pp 509 575 DOI: 10.1017/S096249291300007X, Published online: 02 April 2013
Link to this article: http://journals.cambridge.org/abstract_S096249291300007X
How to cite this article: Yurii Nesterov and Arkadi Nemirovski (2013). On firstorder algorithms for l 1/nuclear norm minimization. Acta Numerica, 22, pp 509575 doi:10.1017/S096249291300007X
Request Permissions : Click here
Downloaded from http://journals.cambridge.org/ANU, IP address: 130.207.93.116 on 30 Apr 2013 Acta Numerica (2013), pp. 509–575 c Cambridge University Press, 2013 doi:10.1017/S096249291300007X Printed in the United Kingdom
On first-order algorithms for 1/nuclear norm minimization
Yurii Nesterov∗ Center for Operations Research and Econometrics, 34 voie du Roman Pays, 1348, Louvain-la-Neuve, Belgium E-mail: [email protected]
Arkadi Nemirovski† School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332, USA E-mail: [email protected]
In the past decade, problems related to l1/nuclear norm minimization have attracted much attention in the signal processing, machine learning and opti- mization communities. In this paper, devoted to 1/nuclear norm minimiza- tion as ‘optimization beasts’, we give a detailed description of two attractive first-order optimization techniques for solving problems of this type. The first one, aimed primarily at lasso-type problems, comprises fast gradient meth- ods applied to composite minimization formulations. The second approach, aimed at Dantzig-selector-type problems, utilizes saddle-point first-order algo- rithms and reformulation of the problem of interest as a generalized bilinear saddle-point problem. For both approaches, we give complete and detailed complexity analyses and discuss the application domains.
CONTENTS 1 Introduction 510 2 Composite minimization and lasso problems 518 3 Saddle-point algorithms for Dantzig selector problems 531 Appendix: Proofs 551 References 570
∗ Partly supported by Direction de la Recherche Scientifique, Communaut´eFran¸caise de Belgique, via grant Action de Recherche Concert´e ARC 04/09-315, and by the Laboratory of Structural Methods of Data Analysis in Predictive Modelling via the RF government grant 11.G34.31.0073c. † Supported in part by NSF via grants DMS-0914785 and CMMI 1232623, by BSF via grant 2008302, and by ONR via grant N000140811104. 510 Y. Nesterov and A. Nemirovski
1. Introduction 1.1. Motivation The goal of this paper is to review recent progress in applications of first- order optimization methods to problems of 1/nuclear norm minimization. For us, the problems of primary interest are the lasso-type problems, 2 Opt = min λ x + Ax − b 2 , (1.1a) x∈E and the Dantzig-selector-type problems, Opt = min x : Ax − b p ≤ δ . (1.1b) x∈E Here E is a Euclidean space, · is a given norm on E (often different from the Euclidean norm), x → Ax is a linear mapping from E to some Rm,and b ∈ Rm is a given vector. In (1.1a), λ>0 is a given penalty parameter.In 1 (1.1b), δ ≥ 0, and p is either 2 (‘ 2-fit’), or ∞ (‘uniform fit’). We will be mainly (but not exclusively) interested in the following cases:
• 1 minimization,whereE = Rn and · = · 1, • nuclear norm minimization,whereE = Rµ×ν is the space of µ × ν matrices and · = · nuc is the nuclear norm, the sum of the singular values of a matrix.
The main source of 1 minimization problems of the form (1.1a,b) is sparsity-oriented signal processing, where the goal is to recover an unknown ‘signal’ x ∈ Rn from its noisy observations y = Bx + ξ,whereξ is the observation error. The signal is known to be sparse, i.e., it has at most s n non-zero entries. The most interesting case here is compressive sensing, where the observations are ‘deficient’ (m =dimy