A Unified Successive Pseudo-Convex Approximation Framework

Total Page:16

File Type:pdf, Size:1020Kb

A Unified Successive Pseudo-Convex Approximation Framework 1 A Unified Successive Pseudo-Convex Approximation Framework Yang Yang and Marius Pesavento Abstract—In this paper, we propose a successive pseudo-convex because problem (2) can be rewritten into a problem with the approximation algorithm to efficiently compute stationary points form of (1) by the help of auxiliary variables: for a large class of possibly nonconvex optimization problems. The stationary points are obtained by solving a sequence of minimize f(x) + y successively refined approximate problems, each of which is x;y (3) much easier to solve than the original problem. To achieve subject to x 2 X ; g(x) ≤ y: convergence, the approximate problem only needs to exhibit a weak form of convexity, namely, pseudo-convexity. We show that We do not assume that f(x) is convex, so (1) is in general a the proposed framework not only includes as special cases a nonconvex optimization problem. The focus of this paper is on number of existing methods, for example, the gradient method the development of efficient iterative algorithms for computing and the Jacobi algorithm, but also leads to new algorithms which the stationary points of problem (1). The optimization problem enjoy easier implementation and faster convergence speed. We (1) represents general class of optimization problems with a also propose a novel line search method for nondifferentiable optimization problems, which is carried out over a properly vast number of diverse applications. Consider for example constructed differentiable function with the benefit of a simpli- the sum-rate maximization in the MIMO multiple access fied implementation as compared to state-of-the-art line search channel (MAC) [1], the broadcast channel (BC) [2, 3] and the techniques that directly operate on the original nondifferentiable interference channel (IC) [4, 5, 6, 7, 8, 9], where f(x) is the objective function. The advantages of the proposed algorithm sum-rate function of multiple users (to be maximized) while are shown, both theoretically and numerically, by several ex- ample applications, namely, MIMO broadcast channel capacity the set X characterizes the users’ power constraints. In the computation, energy efficiency maximization in massive MIMO context of the MIMO IC, (1) is a nonconvex problem and NP- systems and LASSO in sparse signal recovery. hard [5]. As another example, consider portfolio optimization Index Terms—Energy efficiency, exact line search, LASSO, in which f(x) represents the expected return of the portfolio massive MIMO, MIMO broadcast channel, nonconvex optimiza- (to be maximized) and the set X characterizes the trading tion, nondifferentiable optimization, successive convex approxi- constraints [10]. Furthermore, in sparse (l1-regularized) linear mation. regression, f(x) denotes the least square function and g(x) is the sparsity regularization function [11, 12]. Commonly used iterative algorithms belong to the class of I. INTRODUCTION descent direction methods such as the conditional gradient In this paper, we propose an iterative algorithm to solve the method and the gradient projection method for the differen- following general optimization problem: tiable problem (1) [13] and the proximal gradient method for the nondifferentiable problem (2) [14, 15], which often suffer minimize f(x) from slow convergence. To speed up the convergence, the x (1) block coordinate descent (BCD) method that uses the notion of subject to x 2 X ; the nonlinear best-response has been widely studied [13, Sec. where X ⊆ Rn is a closed and convex set, and f(x): Rn ! 2.7]. In particular, this method is applicable if the constraint R is a proper and differentiable function with a continuous set of (1) has a Cartesian product structure X = X1 ×:::×XK gradient. We assume that problem (1) has a solution. such that Problem (1) also includes some class of nondifferentiable minimize f(x1;:::; xK ) x=(x )K optimization problems, if the nondifferentiable function g(x) k k=1 (4) arXiv:1506.04972v2 [math.OC] 7 Apr 2016 is convex: subject to xk 2 Xk; k = 1; : : : ; K: minimize f(x) + g(x) The BCD method is an iterative algorithm: in each iteration, x t+1 (2) only one variable is updated by its best-response xk = subject to x 2 X ; t+1 t+1 t t arg minxk2Xk f(x1 ;:::; xk−1; xk; xk+1;:::; xK ) (i.e., the point that minimizes f(x) with respect to (w.r.t.) the variable Y. Yang is with Intel Deutschland GmbH, Germany (email: xk only while the remaining variables are fixed to their values [email protected]). M. Pesavento is with Communication Systems Group, Darmstadt University of the preceding iteration) and the variables are updated se- of Technology, Germany (email: [email protected]). quentially. This method and its variants have been successfully The authors acknowledge the financial support of the Seventh Framework adopted to many practical problems [1, 6, 7, 10, 16]. Programme for Research of the European Commission under grant number ADEL-619647 and the EXPRESS project within the DFG priority program When the number of variables is large, the convergence CoSIP (DFG-SPP 1798). speed of the BCD method may be slow due to the sequential 2 nature of the update. A parallel variable update based on the 2) The stepsizes can be determined based on the problem best-response seems attractive as a mean to speed up the structure, typically resulting in faster convergence than in cases updating procedure, however, the convergence of a parallel where constant stepsizes [2, 10, 18] and decreasing stepsizes best-response algorithm is only guaranteed under rather re- [8, 19] are used. For example, a constant stepsize can be used strictive conditions, c.f. the diagonal dominance condition on when f(x) is given as the difference of two convex functions the objective function f(x1;:::; xK ) [17], which is not only as in DC programming [21]. When the objective function difficult to satisfy but also hard to verify. If f(x1;:::; xK ) is nondifferentiable, we propose a new exact/successive line is convex, the parallel algorithms converge if the stepsize search method that is carried out over a properly constructed is inversely proportional to the number of block variables differentiable function. Thus it is much easier to implement K. This choice of stepsize, however, tends to be overly than state-of-the-art techniques that operate on the original conservative in systems with a large number of block variables nondifferentiable objective function directly. and inevitably slows down the convergence [2, 10, 18]. In the proposed algorithm, the exact/successive line search A recent progress in parallel algorithms has been made in is used to determine the stepsize and it can be implemented in [8, 9, 19, 20], in which it was shown that the stationary point a centralized controller, whose existence presence is justified of (1) can be found by solving a sequence of successively for particular applications, e.g., the base station in the MIMO refined approximate problems of the original problem (1), and BC, and the portfolio manager in multi-portfolio optimization convergence to a stationary point is established if, among other [10]. We remark that also in applications in which centralized conditions, the approximate function (the objective function controller are not admitted, however, the line search procedure of the approximate problem) and stepsizes are properly se- does not necessarily imply an increased signaling burden when lected. The parallel algorithms proposed in [8, 9, 19, 20] are it is implemented in a distributed manner among different essentially descent direction methods. A description on how distributed processors. For example, in the LASSO problem to construct the approximate problem such that the convexity studied in Sec. IV-C, the stepsize based on the exact line of the original problem is preserved as much as possible is search can be computed in closed-form and it does not incur also contained in [8, 9, 19, 20] to achieve faster convergence any additional signaling as in predetermined stepsizes, e.g., than standard descent directions methods such as classical decreasing stepsizes and constant stepsizes. Besides, even conditional gradient method and gradient projection method. in cases where the line search procedure induces additional Despite its novelty, the parallel algorithms proposed in signaling, the burden is often fully amortized by the significant [8, 9, 19, 20] suffer from two limitations. Firstly, the approx- increase in the convergence rate. imate function must be strongly convex, and this is usually The rest of the paper is organized as follows. In Sec. guaranteed by artificially adding a quadratic regularization II we introduce the mathematical background. The novel term to the original objective function f(x), which how- iterative method is proposed and its convergence is analyzed ever may destroy the desirable characteristic structure of the in Sec. III; its connection to several existing descent direction original problem that could otherwise be exploited, e.g., to algorithms is presented there. In Sec. IV, several applications obtain computationally efficient closed-form solutions of the are considered: the sum rate maximization problem of MIMO approximate problems [6]. Secondly, the algorithms require BC, the energy efficiency maximization of a massive MIMO the use of a decreasing stepsize. On the one hand, a slow system to illustrate the advantage of the proposed approximate decay of the stepsize is preferable to make notable progress function, and the LASSO problem to illustrate the advantage and to achieve satisfactory convergence speed; on the other of the proposed stepsize. The paper is finally concluded in hand, theoretical convergence is guaranteed only when the Sec. V. stepsize decays fast enough. In practice, it is a difficult task Notation: We use x, x and X to denote a scalar, vector on its own to find a decay rate for the stepsize that provides and matrix, respectively.
Recommended publications
  • Lagrangian Duality in Convex Optimization
    Lagrangian Duality in Convex Optimization LI, Xing A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Philosophy in Mathematics The Chinese University of Hong Kong July 2009 ^'^'^xLIBRARr SYSTEf^N^J Thesis/Assessment Committee Professor LEUNG, Chi Wai (Chair) Professor NG, Kung Fu (Thesis Supervisor) Professor LUK Hing Sun (Committee Member) Professor HUANG, Li Ren (External Examiner) Abstract of thesis entitled: Lagrangian Duality in Convex Optimization Submitted by: LI, Xing for the degree of Master of Philosophy in Mathematics at the Chinese University of Hong Kong in July, 2009. Abstract In convex optimization, a strong duality theory which states that the optimal values of the primal problem and the Lagrangian dual problem are equal and the dual problem attains its maximum plays an important role. With easily visualized physical meanings, the strong duality theories have also been wildly used in the study of economical phenomenon and operation research. In this thesis, we reserve the first chapter for an introduction on the basic principles and some preliminary results on convex analysis ; in the second chapter, we will study various regularity conditions that could lead to strong duality, from the classical ones to recent developments; in the third chapter, we will present stable Lagrangian duality results for cone-convex optimization problems under continuous linear perturbations of the objective function . In the last chapter, Lagrange multiplier conditions without constraint qualifications will be discussed. 摘要 拉格朗日對偶理論主要探討原問題與拉格朗日對偶問題的 最優值之間“零對偶間隙”成立的條件以及對偶問題存在 最優解的條件,其在解決凸規劃問題中扮演著重要角色, 並在經濟學運籌學領域有著廣泛的應用。 本文將系統地介紹凸錐規劃中的拉格朗日對偶理論,包括 基本規範條件,閉凸錐規範條件等,亦會涉及無規範條件 的序列拉格朗日乘子。 ACKNOWLEDGMENTS I wish to express my gratitude to my supervisor Professor Kung Fu Ng and also to Professor Li Ren Huang, Professor Chi Wai Leung, and Professor Hing Sun Luk for their guidance and valuable suggestions.
    [Show full text]
  • Arxiv:2011.09194V1 [Math.OC]
    Noname manuscript No. (will be inserted by the editor) Lagrangian duality for nonconvex optimization problems with abstract convex functions Ewa M. Bednarczuk · Monika Syga Received: date / Accepted: date Abstract We investigate Lagrangian duality for nonconvex optimization prob- lems. To this aim we use the Φ-convexity theory and minimax theorem for Φ-convex functions. We provide conditions for zero duality gap and strong duality. Among the classes of functions, to which our duality results can be applied, are prox-bounded functions, DC functions, weakly convex functions and paraconvex functions. Keywords Abstract convexity · Minimax theorem · Lagrangian duality · Nonconvex optimization · Zero duality gap · Weak duality · Strong duality · Prox-regular functions · Paraconvex and weakly convex functions 1 Introduction Lagrangian and conjugate dualities have far reaching consequences for solution methods and theory in convex optimization in finite and infinite dimensional spaces. For recent state-of the-art of the topic of convex conjugate duality we refer the reader to the monograph by Radu Bot¸[5]. There exist numerous attempts to construct pairs of dual problems in non- convex optimization e.g., for DC functions [19], [34], for composite functions [8], DC and composite functions [30], [31] and for prox-bounded functions [15]. In the present paper we investigate Lagrange duality for general optimiza- tion problems within the framework of abstract convexity, namely, within the theory of Φ-convexity. The class Φ-convex functions encompasses convex l.s.c. Ewa M. Bednarczuk Systems Research Institute, Polish Academy of Sciences, Newelska 6, 01–447 Warsaw Warsaw University of Technology, Faculty of Mathematics and Information Science, ul.
    [Show full text]
  • Log-Concave Duality in Estimation and Control Arxiv:1607.02522V1
    Log-Concave Duality in Estimation and Control Robert Bassett,˚ Michael Casey,: and Roger J-B Wets; Department of Mathematics, Univ. California Davis Abstract In this paper we generalize the estimation-control duality that exists in the linear- quadratic-Gaussian setting. We extend this duality to maximum a posteriori estimation of the system's state, where the measurement and dynamical system noise are inde- pendent log-concave random variables. More generally, we show that a problem which induces a convex penalty on noise terms will have a dual control problem. We provide conditions for strong duality to hold, and then prove relaxed conditions for the piece- wise linear-quadratic case. The results have applications in estimation problems with nonsmooth densities, such as log-concave maximum likelihood densities. We conclude with an example reconstructing optimal estimates from solutions to the dual control problem, which has implications for sharing solution methods between the two types of problems. 1 Introduction We consider the problem of estimating the state of a noisy dynamical system based only on noisy measurements of the system. In this paper, we assume linear dynamics, so that the progression of state variables is Xt`1 “FtpXtq ` Wt`1; t “ 0; :::; T ´ 1 (1) X0 “W0 (2) arXiv:1607.02522v1 [math.OC] 8 Jul 2016 Xt is the state variable{a random vector indexed by a discrete time-step t which ranges from 0 to some final time T . All of the results in this paper still hold in the case that the dimension nx of Xt is time-dependent, but for notational convenience we will assume that Xt P R for t “ 0; :::; T .
    [Show full text]
  • Lecture 11: October 8 11.1 Primal and Dual Problems
    10-725/36-725: Convex Optimization Fall 2015 Lecture 11: October 8 Lecturer: Ryan Tibshirani Scribes: Tian Tong Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications. They may be distributed outside this class only with the permission of the Instructor. 11.1 Primal and dual problems 11.1.1 Lagrangian Consider a general optimization problem (called as primal problem) min f(x) (11.1) x subject to hi(x) ≤ 0; i = 1; ··· ; m `j(x) = 0; j = 1; ··· ; r: We define its Lagrangian as m r X X L(x; u; v) = f(x) + uihi(x) + vj`j(x): i=1 j=1 Lagrange multipliers u 2 Rm; v 2 Rr. Lemma 11.1 At each feasible x, f(x) = supu≥0;v L(x; u; v), and the supremum is taken iff u ≥ 0 satisfying uihi(x) = 0; i = 1; ··· ; m. Pm Proof: At each feasible x, we have hi(x) ≤ 0 and `(x) = 0, thus L(x; u; v) = f(x) + i=1 uihi(x) + Pr j=1 vj`j(x) ≤ f(x). The last inequality becomes equality iff uihi(x) = 0; i = 1; ··· ; m. Proposition 11.2 The optimal value of the primal problem, named as f ?, satisfies: f ? = inf sup L(x; u; v): x u≥0;v ? Proof: First considering feasible x (marked as x 2 C), we have f = infx2C f(x) = infx2C supu≥0;v L(x; u; v). Second considering non-feasible x, since supu≥0;v L(x; u; v) = 1 for any x2 = C, infx=2C supu≥0;v L(x; u; v) = ? 1.
    [Show full text]
  • Epsilon Solutions and Duality in Vector Optimization
    NOT FOR QUOTATION WITHOUT THE PERMISSION OF THE AUTHOR EPSILON SOLUTIONS AND DUALITY IN VECTOR OPTIMIZATION Istvdn VdLyi May 1987 WP-87-43 Working Papers are interim reports on work of the International Institute for Applied Systems Analysis and have received only limited review. Views or opinions expressed herein do not necessarily represent those of the Institute or of its National Member Organizations. INTERNATIONAL INSTITUTE FOR APPLIED SYSTEMS ANALYSIS 2361 Laxenburg, Austria PREFACE This paper is a continuation of the author's previous investigations in the theory of epsilon-solutions in convex vector optimization and serves as a theoretical background for the research of SDS in the field of multicriteria optimization. With the stress laid on duality theory, the results presented here give some insight into the problems arising when exact solutions have to be substituted by approximate ones. Just like in the scalar case, the available computational techniques fre- quently lead to such a situation in multicriteria optimization. Alexander B. Kurzhanski Chairman Systems and Decision Sciences Area CONTENTS 1. Introduction 2. Epsilon Optimal Elements 3. Perturbation Map and Duality 4. Conical Supports 5. Conclusion 6. References EPSILON SOLUTIONS AND DUALITY IN VECTOR OF'TIMIZATION Istv&n V&lyi The study of epsilon solutions in vector optimization problems was started in 1979 by S. S. Kutateladze [I]. These types of solutions are interesting because of their relation to nondifferentiable optimization and the vector valued extensions of Ekeland's variational principle as considered by P. Loridan [2] and I. Vdlyi [3], but computational aspects are perhaps even more important. In practical situations, namely, we often stop the calculations at values that we consider sufficiently close to the optimal solution, or use algorithms that result in some approximates of the Pareto set.
    [Show full text]
  • CONJUGATE DUALITY for GENERALIZED CONVEX OPTIMIZATION PROBLEMS Anulekha Dhara and Aparna Mehra
    JOURNAL OF INDUSTRIAL AND Website: http://AIMsciences.org MANAGEMENT OPTIMIZATION Volume 3, Number 3, August 2007 pp. 415–427 CONJUGATE DUALITY FOR GENERALIZED CONVEX OPTIMIZATION PROBLEMS Anulekha Dhara and Aparna Mehra Department of Mathematics Indian Institute of Technology Delhi Hauz Khas, New Delhi-110016, India (Communicated by M. Mathirajan) Abstract. Equivalence between a constrained scalar optimization problem and its three conjugate dual models is established for the class of general- ized C-subconvex functions. Applying these equivalent relations, optimality conditions in terms of conjugate functions are obtained for the constrained multiobjective optimization problem. 1. Introduction and preliminaries. Two basic approaches have emerged in lit- erature for studying nonlinear programming duality namely the Lagrangian ap- proach and the Conjugate Function approach. The latter approach is based on the perturbation theory which in turn helps to study stability of the optimization prob- lem with respect to variations in the data parameters. That is, the optimal solution of the problem can be considered as a function of the perturbed parameters and the effect of the changes in the parameters on the objective value of the problem can be investigated. In this context, conjugate duality helps to analyze the geometric and the economic aspects of the problem. It is because of this reason that since its early inception conjugate duality has been extensively studied by several authors for convex programming problems (see, [1, 6] and references therein). The class of convex functions possess rich geometrical properties that enable us to examine the conjugation theory. However, in non convex case, many of these properties fail to hold thus making it difficult to develop conjugate duality for such classes of functions.
    [Show full text]
  • On Strong and Total Lagrange Duality for Convex Optimization Problems
    On strong and total Lagrange duality for convex optimization problems Radu Ioan Bot¸ ∗ Sorin-Mihai Grad y Gert Wanka z Abstract. We give some necessary and sufficient conditions which completely characterize the strong and total Lagrange duality, respectively, for convex op- timization problems in separated locally convex spaces. We also prove similar statements for the problems obtained by perturbing the objective functions of the primal problems by arbitrary linear functionals. In the particular case when we deal with convex optimization problems having infinitely many convex in- equalities as constraints, the conditions we work with turn into the so-called Farkas-Minkowski and locally Farkas-Minkowski conditions for systems of convex inequalities, recently used in the literature. Moreover, we show that our new ∗Faculty of Mathematics, Chemnitz University of Technology, D-09107 Chemnitz, Germany, e-mail: [email protected]. yFaculty of Mathematics, Chemnitz University of Technology, D-09107 Chemnitz, Germany, e-mail: [email protected]. zFaculty of Mathematics, Chemnitz University of Technology, D-09107 Chemnitz, Germany, e-mail: [email protected] 1 results extend some existing ones in the literature. Keywords. Conjugate functions, Lagrange dual problem, basic constraint qualification, (locally) Farkas-Minkowski condition, stable strong duality AMS subject classification (2000). 49N15, 90C25, 90C34 1 Introduction Consider a convex optimization problem (P ) inf f(x), x2U; g(x)∈−C where X and Y are separated locally convex vector spaces, U is a non-empty closed convex subset of X, C is a non-empty closed convex cone in Y , f : X ! R is a proper convex lower semicontinuous function and g : X ! Y • is a proper C-convex function.
    [Show full text]
  • Conjugate Duality and Optimization CBMS-NSF REGIONAL CONFERENCE SERIES in APPLIED MATHEMATICS
    Conjugate Duality and Optimization CBMS-NSF REGIONAL CONFERENCE SERIES IN APPLIED MATHEMATICS A series of lectures on topics of current research interest in applied mathematics under the direction of the Conference Board of the Mathematical Sciences, supported by the National Science Foundation and published by SIAM. GARRETT BIRKHOFF, The Numerical Solution of Elliptic Equations D. V. LINDLEY, Bayesian Statistics, A Review R. S. VARGA, Functional Analysis and Approximation Theory in Numerical Analysis R. R. BAHADUR, Some Limit Theorems in Statistics PATRICK BILLINGSLEY, Weak Convergence of Measures: Applications in Probability J. L. LIONS, Some Aspects of the Optimal Control of Distributed Parameter Systems ROGER PENROSE, Techniques of Differential Topology in Relativity HERMAN CHERNOFF, Sequential Analysis and Optimal Design J. DURBIN, Distribution Theory for Tests Based on the Sample Distribution Function SOL I. RUBINOW, Mathematical Problems in the Biological Sciences P. D. LAX, Hyperbolic Systems of Conservation Laws and the Mathematical Theory of Shock Waves I. J. SCHOENBERG, Cardinal Spline Interpolation IVAN SINGER, The Theory of Best Approximation and Functional Analysis WERNER C. RHEINBOLDT, Methods of Solving Systems of Nonlinear Equations HANS F. WEINBERGER, Variational Methods for Eigenvalue Approximation R. TYRRELL ROCKAFELLAR, Conjugate Duality and Optimization SIR JAMES LIGHTHILL, Mathematical Biofluiddynamics GERARD SALTON, Theory of Indexing CATHLEEN S. MORAWETZ, Notes on Time Decay and Scattering for Some Hyperbolic Problems F. HOPPENSTEADT, Mathematical Theories of Populations: Demographics, Genetics and Epidemics RICHARD ASKEY, Orthogonal Polynomials and Special Functions L. E. PAYNE, Improperly Posed Problems in Partial Differential Equations S. ROSEN, Lectures on the Measurement and Evaluation of the Performance of Computing Systems HERBERT B.
    [Show full text]
  • Lec 13 - Math Basis for Rate Distortion Optimization
    ECE 5578 Multimedia Communication Lec 13 - Math Basis for Rate Distortion Optimization Zhu Li Dept of CSEE, UMKC Office: FH560E, Email: [email protected], Ph: x 2346. http://l.web.umkc.edu/lizhu slides created with WPS Office Linux and EqualX LaTex equation editor ECE 5578 Multimedia Communciation, 2018 p.1 Outline Recap of HEVC system R-D Optimization Basis Summary ECE 5578 Multimedia Communciation, 2018 p.2 HEVC Coding Structure Slide Credit: Vivienne Sze & Madhukar Budagavi, ISCAS 2014 Tutorial Quad Tree Decomposition: Ref: G. Schuster, PhD Thesis, 1996: Optimal Allocation of Bits Among Motion, Segmentation and Residual ECE 5578 Multimedia Communciation, 2018 p.3 HEVC Transforms Transform + Quant: ECE 5578 Multimedia Communciation, 2018 p.4 HEVC Intra-Prediction Intra-Prediction Modes ECE 5578 Multimedia Communciation, 2018 p.5 Intra-Predicted Basis As if it is a 1-non zero coefficient transform… Ref: J. Laniema and W.-J. Han, “Intra Picture Prediction in HEVC”, Chapter in, Springer- Velag Book on High Efficiency Video Coding (HEVC): Algorithms and Architectures, Springer, 2014. Ed. V. Sze et. Al. ECE 5578 Multimedia Communciation, 2018 p.6 HEVC Entropy Coding Binary Arithmetic Coding: ECE 5578 Multimedia Communciation, 2018 p.7 Parallel Processing Tools: Slice/Tile Credit: Vivienne Sze & Madhukar Budagavi, ISCAS 2014 Tutorial ECE 5578 Multimedia Communciation, 2018 p.8 HEVC Resources Main Spec: . http://www.itu.int/ITU-T/recommendaBons/rec.aspx?rec=11885 T-CSVT Special Issue: . 2012: Combined Issue on HEVC Standard and Research: http://ieeexplore.ieee.org/xpl/tocresult.jsp?isnumber=6403920 . 2016: Special Issue on HEVC Extensions and Efficient HEVC Implementations: http://ieeexplore.ieee.org/xpl/tocresult.jsp?isnumber=7372356 Springer Book: .
    [Show full text]
  • Conjugate Duality
    LECTURE 5: CONJUGATE DUALITY 1. Primal problem and its conjugate dual 2. Duality theory and optimality conditions 3. Relation to other type of dual problems 4. Linear conic programming problems Motivation of conjugate duality Min f(x) (over R) = Max g(y) (over R) and h(y) = - g(y) • f(x) + h(y) = f(x) – g(y) can be viewed as “duality gap” • Would like to have (i) weak duality 0 (ii) strong duality 0 = f(x*) + h(y*) Where is the duality information • Recall the fundamental result of Fenchel’s conjugate inequality • Need a structure such that in general and at optimal solutions 0 = <x*, y*> = f(x*) + h(y*) Concept of dual cone • Let X be a cone in • Define its dual cone • Properties: (i) Y is a cone. (ii) Y is a convex set. (iii) Y is a closed set. Observations Conjugate (Geometric) duality Dual side information • Conjugate dual function • Dual cone Properties: 1. Y is a cone in 2. Y is closed and convex 3. both are closed and convex. Conjugate (Geometric) dual problem Observations Conjugate duality theory Conjugate duality theory Proof Conjugate duality theory Example – Standard form LP Conjugate dual problem Dual LP Example – Karmarkar form LP Example – Karmarkar form LP Example – Karmarkar form LP • Conjugate dual problem becomes which is an unconstrained convex programming problem. Illustration Example – Posynomial programming • Nonconvex programming problem • Transformation Posynomial programming • Primal problem: Conjugate dual problem: Dual Posynomial Programming h(y) , supx2En [< x; y > −f (x)] < +1 Pn Pn xj Let g(x) = j=1 xj yj − log j=1 cj e ∗ xj @g cj e = 0 ) yj = x∗ @xj Pn j j=1 cj e Pn ) yj > 0 and j=1 yj = 1: ) Ω is closed.
    [Show full text]
  • Fenchel Duality Theory and a Primal-Dual Algorithm on Riemannian Manifolds
    Fenchel Duality Theory and a Primal-Dual Algorithm on Riemannian Manifolds Ronny Bergmann∗ Roland Herzog∗ Maurício Silva Louzeiro∗ Daniel Tenbrinck† José Vidal-Núñez‡ This paper introduces a new notion of a Fenchel conjugate, which generalizes the classical Fenchel conjugation to functions dened on Riemannian manifolds. We investigate its properties, e.g., the Fenchel–Young inequality and the characterization of the convex subdierential using the analogue of the Fenchel–Moreau Theorem. These properties of the Fenchel conjugate are employed to derive a Riemannian primal-dual optimization algorithm, and to prove its convergence for the case of Hadamard manifolds under appropriate assumptions. Numerical results illustrate the performance of the algorithm, which competes with the recently derived Douglas–Rachford algorithm on manifolds of nonpositive curvature. Furthermore, we show numerically that our novel algorithm even converges on manifolds of positive curvature. Keywords. convex analysis, Fenchel conjugate function, Riemannian manifold, Hadamard manifold, primal-dual algorithm, Chambolle–Pock algorithm, total variation AMS subject classications (MSC2010). 49N15, 49M29, 90C26, 49Q99 1 Introduction Convex analysis plays an important role in optimization, and an elaborate theory on convex analysis and conjugate duality is available on locally convex vector spaces. Among the vast references on this topic, we mention Bauschke, Combettes, 2011 for convex analysis and monotone operator techniques, Ekeland, ∗Technische Universität Chemnitz, Faculty of Mathematics, 09107 Chemnitz, Germany ([email protected] arXiv:1908.02022v4 [math.NA] 9 Oct 2020 chemnitz.de, https://www.tu-chemnitz.de/mathematik/part dgl/people/bergmann, ORCID 0000-0001-8342- 7218, [email protected], https://www.tu-chemnitz.de/mathematik/part dgl/people/herzog, ORCID 0000-0003-2164-6575, [email protected], https://www.tu-chemnitz.de/ mathematik/part dgl/people/louzeiro, ORCID 0000-0002-4755-3505).
    [Show full text]
  • Duality for Almost Convex Optimization Problems Via the Perturbation Approach
    Duality for almost convex optimization problems via the perturbation approach Radu Ioan Bot¸ ∗ G´abor Kassay † Gert Wanka ‡ Abstract. We deal with duality for almost convex finite dimensional op- timization problems by means of the classical perturbation approach. To this aim some standard results from the convex analysis are extended to the case of almost convex sets and functions. The duality for some classes of primal-dual problems is derived as a special case of the general approach. The sufficient regularity conditions we need for guaranteeing strong duality are proved to be similar to the ones in the convex case. Keywords. almost convex sets, almost convex functions, conjugate func- tions, duality, perturbation approach AMS subject classification (2000). 26A51, 42A50, 49N15 1 Introduction Dealing with duality for a given optimization problem is one of the main features in mathematical programming and convex analysis both from theoretical and practical point of view. There is a well developed theory of duality for convex optimization problems in finite dimensional spaces, as one can read for instance in [15]. Distinct dual problems have been investigated by using the so-called perturbation approach in [16]. This is based on the theory of conjugate functions and describes how a dual problem can be assigned to a primal one ([5]). Generalized convex functions are those non-convex functions which possess at least one valuable property of convex functions. The growing interest in them during the last decades comes with no surprise since they are often more suitable than convex functions to describe practical problems originated from economics, management science, engineering, etc.
    [Show full text]