Convex Analysis

Total Page:16

File Type:pdf, Size:1020Kb

Convex Analysis Convex Analysis Taught by Professor Adrian Lewis Notes by Mateo D´ıaz [email protected] Cornell University Spring, 2018 Last updated May 18, 2018. The latest version is online here. Notes template by David Mehrle. Contents 1 Convex Sets .................................................. 4 2 Convex functions............................................... 6 2.1 Convexity and calculus......................................... 7 2.1.1 Gradients ............................................ 7 2.1.2 Recognizing smooth convexity................................ 8 2.2 Continuity of convex functions.................................... 8 2.3 Subgradients............................................... 9 3 Duality..................................................... 10 3.1 Fenchel Duality............................................. 10 3.2 Conic programs............................................. 12 3.3 Convex calculus............................................. 14 3.4 Lagrangian duality........................................... 15 4 Algorithms................................................... 19 4.1 Augmented Lagrangian Method ................................... 19 4.2 Proximal Point method and fixed points............................... 20 4.3 Smooth Minimization.......................................... 25 4.4 Splitting Problems and Alternating projections........................... 26 4.5 Proximal Gradient ........................................... 28 4.6 Douglas-Rachford............................................ 29 4.6.1 Consensus optimization.................................... 30 4.7 ADMM: Alternating Directions Method of Multipliers....................... 31 4.8 Splitting ................................................. 34 4.9 Projected subgradient method..................................... 34 4.10 Accelerated Proximal Gradient .................................... 36 5 Variational analysis.............................................. 38 5.1 Nonconvex calculus........................................... 38 5.2 Subgradients and nonconvexity.................................... 40 5.3 Inverse problems ............................................ 43 5.4 Linearizing sets............................................. 47 5.5 Optimality conditions ......................................... 49 1 Contents by Lecture Lecture 01 on 30 January 2018.......................................... 4 Lecture 02 on February 1 2018.......................................... 5 Lecture 03 on February 6 2018.......................................... 8 Lecture 04 on February 8 2018.......................................... 9 Lecture 05 on February 13 2018 ......................................... 10 Lecture 06 on February 15 2018 ......................................... 12 Lecture 07 on February 22 2018 ......................................... 15 Lecture 08 on February 27 2018 ......................................... 18 Lecture 09 on March 1 2018 ........................................... 20 Lecture 10 on March 6 2018 ........................................... 21 Lecture 11 on March 8 2018 ........................................... 24 Lecture 12 on March 13 2018........................................... 26 Lecture 13 on March 15 2018........................................... 29 Lecture 14 on March 20 2018........................................... 31 Lecture 15 on March 22 2018........................................... 33 Lecture 16 on April 10 2018 ........................................... 34 Lecture 17 on April 12 2018 ........................................... 36 Lecture 18 on April 17 2018 ........................................... 38 Lecture 19 on April 19 2018 ........................................... 40 Lecture 20 on April 24 2018 ........................................... 42 Lecture 21 on April 26 2018 ........................................... 43 Lecture 22 on May 1 2018............................................. 45 Lecture 23 on May 3 2018............................................. 46 Lecture 24 on May 8 2018............................................. 48 2 Disclousure These notes are not official. I do not accept any responsibility or liability for the accuracy, content, complete- ness, correctness, or reliability of the information contained on these notes. Please feel free to email me if you find any mistakes or typos. 3 Lecture 01: Convex Sets 30 January 2018 1 Convex Sets We will mainly consider a real Euclidean space E (finite-dimensional Hilbert space). We denote the inner product as h·, ·i. Example 1. Few examples: • The standard space. Rn, hx, yi = x>y. • Symmetric matrices. Sn := fsymmetric n × n matricesg, hX, Yi = tr(X>Y). Definition 1. A set C ⊆ Rn is convex if for all x, y 2 E, λ 2 [0, 1] we have λx + (1 - λ)y 2 C. (a) Nonconvex set (b) Convex set Example 2. Here are two simple examples: • Closed Halfspaces. H := fx j a>x ≤ βg for some fixed β 2 R and a nonzero a 2 E∗. • Unit ball. Define kxk2 = hx, xi, then B = fx j kxk ≤ 1g. Proposition 1. Arbitrary intersections of convex sets are convex. Proof. Exercise. Example 3. Consider: • Polyhedrons. They are intersections of halfspaces. n n > • PSD Matrices. S+ = fX 2 S j y Xy ≥ 0g, this is convex because 0 ≤ y>Xy = tr(ytXy) = tr(Xyy>) = hX, yy>i, n therefore each constraint defines a halfspaces and thus S+ is an infinite intersection of halfspaces. Proposition 2. Closures of convex sets are convex. Proof. Exercise. But, how about interiors? Lemma 1 (Accessibility). For a convex set S if x 2 int S, y 2 cl S, then λx + (1 - λ)y 2 int S 80 ≤ λ < 1. 4 Lecture 02: Convex Sets February 1 2018 Proof. Assume first that y 2 S. Then there exists δ > 0 such that x + δB ⊂ S. By convexity, for all λ 2 (0, 1) λ(x + δB) + (1 - λ)y ⊆ S = λ(x + (1 - λ)y) + λδB 2 S which implies that λ(x + (1 - λ)y) 2 int S. ) Now suppose that y 2 cl S, then there exists a sequence yk 2 S such that yk y. By the first argument Å (1 - λ) ã λ(x + (1 - λ)y) = λ(x + (1 - λ)y ) + (1 - λ)(y - y ) = λ x + (y - y ) +(1 - λ)y . k k λ ! k k zk Notice that when k is big enough zk lies in the interior of S.| Thus by{z our first argument} this convex combination lies in the interior of S. Corollary 1. The interior of convex sets is convex and assuming that int S 6= Æ, then cl int S = cl S. Question 1. For convex S, int cl S = int S? Theorem 1 (Best approximation). Any nonempty closed convex S has a unique shortest vector x¯, i.e. kx¯k ≤ kxk for all x 2 S. Moreover is characterized by the following property hx¯, x - x¯i ≥ 0. (1.1) Proof. Existence. Choose any x^ 2 S, consider S1 = S \ kx^kB is nonempty and compact. Thus the continuous function k · k has a minimizer on that set, call it x¯. For any x 2 S, either kxk > kx^k ≥ kx¯k or kxk ≤ kx^k and thus kxk ≥ kx^k. Characterization. We know for x 2 S and t 2 [0, 1] such that kx^k2 ≤ kx¯ + t(x - x¯)k2, then expanding 0 ≤ tk(x - x¯)k2 + 2hx¯, (x - x¯)i. By letting t 0 we get the result. Uniqueness. Suppose that there exist another point x0 2 S which also satisfies (1.1). Thus ! 0 0 0 hx¯, x - x¯i ≥ 0 and hx ,x ¯ - x i ≥ 0. Then if we add the two inequalities we get kx¯ - x0k ≤ 0 sox ¯ = x0. Theorem 2 (Basic separation). If S is closed and convex and y 2/ S then there exists a halfspaces H such that S ⊆ H and y 2/ H. Proof. Without loss of generality take y = 0 and use previous result. Theorem 3 (Supporting hyperplane (Hann-Banach)). Suppose that S is convex and int S 6= Æ and let y 2/ int S, then there exist a halfspace H containing S with y 2 bd Proof. Without loss of generality assume that 0 2 int S. For each r 2 N+, define the vector Å 1ã z = 1 + y. k r By the Accessibility Lemma then zk 2/ cl S, by basic separation there is a separating hyperplane given by ak 6= 0 so that har, zki ≥ har, xi 8x 2 S. Without loss of generality assume that kakk = 1, then by compactness some subsequence converges to a. Thus, ha, yi ≥ ha, xi 8x 2 S. 5 Lecture 02: Convex functions February 1 2018 2 Convex functions There is a very natural way to extend our results, of the previous class, from sets to functions. The kind of trick to do that is to allow your function to have infinite values, namely ± . We will think about functions of the form f : E R := [- , ]. 1 Definition 2. Let f : E R := [- , ] be a function, then its epigraph is defined as ! 1 1 f := f(x r) E R j f(x) rg ! 1 1epi , 2 × ≤ Definition 3. We say that a function f is convex if epi f is convex. Figure 2: Epigraph of the function f = (x2 - 1)2 In the other direction, we can define Definition 4. Given a set C 2 E, define the indicator function as 0 if x 2 C, δ (x) := C otherwise. Then it is possible to show that C ⊂ E is convex if, and only if, δ is convex. 1 C Definition 5. We say that a function f : S R is convex if f + δS is convex. Equivalently, S is convex and f(λx + (1 - λ)y) λf(x) + (1 - λ)f(y) x y S λ [0 1] ≤ ! 8 , 2 , 2 , . (2.1) Definition 6. We say that a function f : S R is strictly convex if inequality (2.1) holds strictly. Example 4. In particular, norms are strictly convex, k · k , k · k , k · k , ::: . ! 1 2 Definition 7. We say that x 2 E is a minimizer for f : E1 R if f(x) f(x) x E ≥ ! 8 2 . Proposition 3. Minimizers of strictly convex functions are unique. Proof. Easy to check. Definition 8. We say that x 2 E is a local minimizer for f : E R if there exist a neighbor U ⊂ E of x such that f(x) f(x) x U ≥ 8! 2 . 6 Lecture 02: Convexity and calculus February 1 2018 Proposition 4. For convex f, local minimizers are minimizers. Proof. First, we need an easy 1-dimensional result for convex functions. g(t) Claim 1. Suppose that g : R+ R convex with g(0) = 0. Then t 7 t is a nondecreasing function. Proof. Exercise. ! ! Suppose that x is a local minimizer. Now, given any x 2 E we would like to show that f(x) ≥ f(x). Define g(t) = f(x + t(x - x)) - f(x), it is easy to see that function is convex because f is. Since x is a local minimizer then for small enough t g(t) g(1) 0 ≤ ≤ = f(x)- f(x), t 1 where the second inequality follows by the claim. 2.1 Convexity and calculus Convexity is a one-dimensional idea, so if we understand how it behaves on the reals then we are in good shape for everything else.
Recommended publications
  • On Choquet Theorem for Random Upper Semicontinuous Functions
    CORE Metadata, citation and similar papers at core.ac.uk Provided by Elsevier - Publisher Connector International Journal of Approximate Reasoning 46 (2007) 3–16 www.elsevier.com/locate/ijar On Choquet theorem for random upper semicontinuous functions Hung T. Nguyen a, Yangeng Wang b,1, Guo Wei c,* a Department of Mathematical Sciences, New Mexico State University, Las Cruces, NM 88003, USA b Department of Mathematics, Northwest University, Xian, Shaanxi 710069, PR China c Department of Mathematics and Computer Science, University of North Carolina at Pembroke, Pembroke, NC 28372, USA Received 16 May 2006; received in revised form 17 October 2006; accepted 1 December 2006 Available online 29 December 2006 Abstract Motivated by the problem of modeling of coarse data in statistics, we investigate in this paper topologies of the space of upper semicontinuous functions with values in the unit interval [0,1] to define rigorously the concept of random fuzzy closed sets. Unlike the case of the extended real line by Salinetti and Wets, we obtain a topological embedding of random fuzzy closed sets into the space of closed sets of a product space, from which Choquet theorem can be investigated rigorously. Pre- cise topological and metric considerations as well as results of probability measures and special prop- erties of capacity functionals for random u.s.c. functions are also given. Ó 2006 Elsevier Inc. All rights reserved. Keywords: Choquet theorem; Random sets; Upper semicontinuous functions 1. Introduction Random sets are mathematical models for coarse data in statistics (see e.g. [12]). A the- ory of random closed sets on Hausdorff, locally compact and second countable (i.e., a * Corresponding author.
    [Show full text]
  • 1 Overview 2 Elements of Convex Analysis
    AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 2 | Wednesday, January 27th 1 Overview In our previous lecture we discussed several applications of optimization, introduced basic terminol- ogy, and proved Weierstass' theorem (circa 1830) which gives a sufficient condition for existence of an optimal solution to an optimization problem. Today we will discuss basic definitions and prop- erties of convex sets and convex functions. We will then prove the separating hyperplane theorem and see an application of separating hyperplanes in machine learning. We'll conclude by describing the seminal perceptron algorithm (1957) which is designed to find separating hyperplanes. 2 Elements of Convex Analysis We will primarily consider optimization problems over convex sets { sets for which any two points are connected by a line. We illustrate some convex and non-convex sets in Figure 1. Definition. A set S is called a convex set if any two points in S contain their line, i.e. for any x1; x2 2 S we have that λx1 + (1 − λ)x2 2 S for any λ 2 [0; 1]. In the previous lecture we saw linear regression an example of an optimization problem, and men- tioned that the RSS function has a convex shape. We can now define this concept formally. n Definition. For a convex set S ⊆ R , we say that a function f : S ! R is: • convex on S if for any two points x1; x2 2 S and any λ 2 [0; 1] we have that: f (λx1 + (1 − λ)x2) ≤ λf(x1) + 1 − λ f(x2): • strictly convex on S if for any two points x1; x2 2 S and any λ 2 [0; 1] we have that: f (λx1 + (1 − λ)x2) < λf(x1) + (1 − λ) f(x2): We illustrate a convex function in Figure 2.
    [Show full text]
  • CORE View Metadata, Citation and Similar Papers at Core.Ac.Uk
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Bulgarian Digital Mathematics Library at IMI-BAS Serdica Math. J. 27 (2001), 203-218 FIRST ORDER CHARACTERIZATIONS OF PSEUDOCONVEX FUNCTIONS Vsevolod Ivanov Ivanov Communicated by A. L. Dontchev Abstract. First order characterizations of pseudoconvex functions are investigated in terms of generalized directional derivatives. A connection with the invexity is analysed. Well-known first order characterizations of the solution sets of pseudolinear programs are generalized to the case of pseudoconvex programs. The concepts of pseudoconvexity and invexity do not depend on a single definition of the generalized directional derivative. 1. Introduction. Three characterizations of pseudoconvex functions are considered in this paper. The first is new. It is well-known that each pseudo- convex function is invex. Then the following question arises: what is the type of 2000 Mathematics Subject Classification: 26B25, 90C26, 26E15. Key words: Generalized convexity, nonsmooth function, generalized directional derivative, pseudoconvex function, quasiconvex function, invex function, nonsmooth optimization, solution sets, pseudomonotone generalized directional derivative. 204 Vsevolod Ivanov Ivanov the function η from the definition of invexity, when the invex function is pseudo- convex. This question is considered in Section 3, and a first order necessary and sufficient condition for pseudoconvexity of a function is given there. It is shown that the class of strongly pseudoconvex functions, considered by Weir [25], coin- cides with pseudoconvex ones. The main result of Section 3 is applied to characterize the solution set of a nonlinear programming problem in Section 4. The base results of Jeyakumar and Yang in the paper [13] are generalized there to the case, when the function is pseudoconvex.
    [Show full text]
  • Applications of Convex Analysis Within Mathematics
    Noname manuscript No. (will be inserted by the editor) Applications of Convex Analysis within Mathematics Francisco J. Arag´onArtacho · Jonathan M. Borwein · Victoria Mart´ın-M´arquez · Liangjin Yao July 19, 2013 Abstract In this paper, we study convex analysis and its theoretical applications. We first apply important tools of convex analysis to Optimization and to Analysis. We then show various deep applications of convex analysis and especially infimal convolution in Monotone Operator Theory. Among other things, we recapture the Minty surjectivity theorem in Hilbert space, and present a new proof of the sum theorem in reflexive spaces. More technically, we also discuss autoconjugate representers for maximally monotone operators. Finally, we consider various other applications in mathematical analysis. Keywords Adjoint · Asplund averaging · autoconjugate representer · Banach limit · Chebyshev set · convex functions · Fenchel duality · Fenchel conjugate · Fitzpatrick function · Hahn{Banach extension theorem · infimal convolution · linear relation · Minty surjectivity theorem · maximally monotone operator · monotone operator · Moreau's decomposition · Moreau envelope · Moreau's max formula · Moreau{Rockafellar duality · normal cone operator · renorming · resolvent · Sandwich theorem · subdifferential operator · sum theorem · Yosida approximation Mathematics Subject Classification (2000) Primary 47N10 · 90C25; Secondary 47H05 · 47A06 · 47B65 Francisco J. Arag´onArtacho Centre for Computer Assisted Research Mathematics and its Applications (CARMA),
    [Show full text]
  • Convex) Level Sets Integration
    Journal of Optimization Theory and Applications manuscript No. (will be inserted by the editor) (Convex) Level Sets Integration Jean-Pierre Crouzeix · Andrew Eberhard · Daniel Ralph Dedicated to Vladimir Demjanov Received: date / Accepted: date Abstract The paper addresses the problem of recovering a pseudoconvex function from the normal cones to its level sets that we call the convex level sets integration problem. An important application is the revealed preference problem. Our main result can be described as integrating a maximally cycli- cally pseudoconvex multivalued map that sends vectors or \bundles" of a Eu- clidean space to convex sets in that space. That is, we are seeking a pseudo convex (real) function such that the normal cone at each boundary point of each of its lower level sets contains the set value of the multivalued map at the same point. This raises the question of uniqueness of that function up to rescaling. Even after normalising the function long an orienting direction, we give a counterexample to its uniqueness. We are, however, able to show uniqueness under a condition motivated by the classical theory of ordinary differential equations. Keywords Convexity and Pseudoconvexity · Monotonicity and Pseudomono- tonicity · Maximality · Revealed Preferences. Mathematics Subject Classification (2000) 26B25 · 91B42 · 91B16 Jean-Pierre Crouzeix, Corresponding Author, LIMOS, Campus Scientifique des C´ezeaux,Universit´eBlaise Pascal, 63170 Aubi`ere,France, E-mail: [email protected] Andrew Eberhard School of Mathematical & Geospatial Sciences, RMIT University, Melbourne, VIC., Aus- tralia, E-mail: [email protected] Daniel Ralph University of Cambridge, Judge Business School, UK, E-mail: [email protected] 2 Jean-Pierre Crouzeix et al.
    [Show full text]
  • Ce Document Est Le Fruit D'un Long Travail Approuvé Par Le Jury De Soutenance Et Mis À Disposition De L'ensemble De La Communauté Universitaire Élargie
    AVERTISSEMENT Ce document est le fruit d'un long travail approuvé par le jury de soutenance et mis à disposition de l'ensemble de la communauté universitaire élargie. Il est soumis à la propriété intellectuelle de l'auteur. Ceci implique une obligation de citation et de référencement lors de l’utilisation de ce document. D'autre part, toute contrefaçon, plagiat, reproduction illicite encourt une poursuite pénale. Contact : [email protected] LIENS Code de la Propriété Intellectuelle. articles L 122. 4 Code de la Propriété Intellectuelle. articles L 335.2- L 335.10 http://www.cfcopies.com/V2/leg/leg_droi.php http://www.culture.gouv.fr/culture/infos-pratiques/droits/protection.htm THESE` pr´esent´eepar NGUYEN Manh Cuong en vue de l'obtention du grade de DOCTEUR DE L'UNIVERSITE´ DE LORRAINE (arr^et´eminist´eriel du 7 Ao^ut 2006) Sp´ecialit´e: INFORMATIQUE LA PROGRAMMATION DC ET DCA POUR CERTAINES CLASSES DE PROBLEMES` EN APPRENTISSAGE ET FOUILLE DE DONEES.´ Soutenue le 19 mai 2014 devant le jury compos´ede Rapporteur BENNANI Youn`es Professeur, Universit´eParis 13 Rapporteur RAKOTOMAMONJY Alain Professeur, Universit´ede Rouen Examinateur GUERMEUR Yann Directeur de recherche, LORIA-Nancy Examinateur HEIN Matthias Professeur, Universit´eSaarland Examinateur PHAM DINH Tao Professeur ´em´erite, INSA-Rouen Directrice de th`ese LE THI Hoai An Professeur, Universit´ede Lorraine Co-encadrant CONAN-GUEZ Brieuc MCF, Universit´ede Lorraine These` prepar´ ee´ a` l'Universite´ de Lorraine, Metz, France au sein de laboratoire LITA Remerciements Cette th`ese a ´et´epr´epar´ee au sein du Laboratoire d'Informatique Th´eorique et Appliqu´ee (LITA) de l'Universit´ede Lorraine - France, sous la co-direction du Madame le Professeur LE THI Hoai An, Directrice du laboratoire Laboratoire d'Informatique Th´eorique et Ap- pliqu´ee(LITA), Universit´ede Lorraine et Ma^ıtre de conf´erence CONAN-GUEZ Brieuc, Institut Universitaire de Technologie (IUT), Universit´ede Lorraine, Metz.
    [Show full text]
  • Lipschitz Continuity of Convex Functions
    Lipschitz Continuity of Convex Functions Bao Tran Nguyen∗ Pham Duy Khanh† November 13, 2019 Abstract We provide some necessary and sufficient conditions for a proper lower semi- continuous convex function, defined on a real Banach space, to be locally or globally Lipschitz continuous. Our criteria rely on the existence of a bounded selection of the subdifferential mapping and the intersections of the subdifferential mapping and the normal cone operator to the domain of the given function. Moreover, we also point out that the Lipschitz continuity of the given function on an open and bounded (not necessarily convex) set can be characterized via the existence of a bounded selection of the subdifferential mapping on the boundary of the given set and as a consequence it is equivalent to the local Lipschitz continuity at every point on the boundary of that set. Our results are applied to extend a Lipschitz and convex function to the whole space and to study the Lipschitz continuity of its Moreau envelope functions. Keywords Convex function, Lipschitz continuity, Calmness, Subdifferential, Normal cone, Moreau envelope function. Mathematics Subject Classification (2010) 26A16, 46N10, 52A41 1 Introduction arXiv:1911.04886v1 [math.FA] 12 Nov 2019 Lipschitz continuous and convex functions play a significant role in convex and nonsmooth analysis. It is well-known that if the domain of a proper lower semicontinuous convex function defined on a real Banach space has a nonempty interior then the function is continuous over the interior of its domain [3, Proposition 2.111] and as a consequence, it is subdifferentiable (its subdifferential is a nonempty set) and locally Lipschitz continuous at every point in the interior of its domain [3, Proposition 2.107].
    [Show full text]
  • On the Ekeland Variational Principle with Applications and Detours
    Lectures on The Ekeland Variational Principle with Applications and Detours By D. G. De Figueiredo Tata Institute of Fundamental Research, Bombay 1989 Author D. G. De Figueiredo Departmento de Mathematica Universidade de Brasilia 70.910 – Brasilia-DF BRAZIL c Tata Institute of Fundamental Research, 1989 ISBN 3-540- 51179-2-Springer-Verlag, Berlin, Heidelberg. New York. Tokyo ISBN 0-387- 51179-2-Springer-Verlag, New York. Heidelberg. Berlin. Tokyo No part of this book may be reproduced in any form by print, microfilm or any other means with- out written permission from the Tata Institute of Fundamental Research, Colaba, Bombay 400 005 Printed by INSDOC Regional Centre, Indian Institute of Science Campus, Bangalore 560012 and published by H. Goetze, Springer-Verlag, Heidelberg, West Germany PRINTED IN INDIA Preface Since its appearance in 1972 the variational principle of Ekeland has found many applications in different fields in Analysis. The best refer- ences for those are by Ekeland himself: his survey article [23] and his book with J.-P. Aubin [2]. Not all material presented here appears in those places. Some are scattered around and there lies my motivation in writing these notes. Since they are intended to students I included a lot of related material. Those are the detours. A chapter on Nemyt- skii mappings may sound strange. However I believe it is useful, since their properties so often used are seldom proved. We always say to the students: go and look in Krasnoselskii or Vainberg! I think some of the proofs presented here are more straightforward. There are two chapters on applications to PDE.
    [Show full text]
  • NOTE on the CONTINUITY of M-CONVEX and L-CONVEX FUNCTIONS in CONTINUOUS VARIABLES 1. Introduction Two Kinds of Convexity Concept
    Journal of the Operations Research Society of Japan 2008, Vol. 51, No. 4, 265-273 NOTE ON THE CONTINUITY OF M-CONVEX AND L-CONVEX FUNCTIONS IN CONTINUOUS VARIABLES Kazuo Murota Akiyoshi Shioura University of Tokyo Tohoku University (Received March 27, 2008) Abstract M-convex and L-convex functions in continuous variables constitute subclasses of convex func- tions with nice combinatorial properties. In this note we give proofs of the fundamental facts that closed proper M-convex and L-convex functions are continuous on their effective domains. Keywords: Combinatorial optimization, convex function, continuity, submodular func- tion, matroid. 1. Introduction Two kinds of convexity concepts, called M-convexity and L-convexity, play primary roles in the theory of discrete convex analysis [6]. They are originally introduced for functions in integer variables by Murota [4, 5], and then for functions in continuous variables by Murota{ Shioura [8, 10]. M-convex and L-convex functions in continuous variables constitute subclasses of convex functions with additional combinatorial properties such as submodularity and diagonal dom- inance (see, e.g., [6{11]). Fundamental properties of M-convex and L-convex functions are investigated in [9], such as equivalent axioms, subgradients, directional derivatives, etc. Con- jugacy relationship between M-convex and L-convex functions under the Legendre-Fenchel transformation is shown in [10]. Subclasses of M-convex and L-convex functions are investi- gated in [8] (polyhedral M-convex and L-convex functions) and in [11] (quadratic M-convex and L-convex functions). As variants of M-convex and L-convex functions, the concepts of M\-convex and L\-convex functions are also introduced by Murota{Shioura [8, 10], where \M\" and \L\" should be read \M-natural" and \L-natural," respectively.
    [Show full text]
  • Optimization Over Nonnegative and Convex Polynomials with and Without Semidefinite Programming
    OPTIMIZATION OVER NONNEGATIVE AND CONVEX POLYNOMIALS WITH AND WITHOUT SEMIDEFINITE PROGRAMMING GEORGINA HALL ADISSERTATION PRESENTED TO THE FACULTY OF PRINCETON UNIVERSITY IN CANDIDACY FOR THE DEGREE OF DOCTOR OF PHILOSOPHY RECOMMENDED FOR ACCEPTANCE BY THE DEPARTMENT OF OPERATIONS RESEARCH AND FINANCIAL ENGINEERING ADVISER:PROFESSOR AMIR ALI AHMADI JUNE 2018 c Copyright by Georgina Hall, 2018. All rights reserved. Abstract The problem of optimizing over the cone of nonnegative polynomials is a fundamental problem in computational mathematics, with applications to polynomial optimization, con- trol, machine learning, game theory, and combinatorics, among others. A number of break- through papers in the early 2000s showed that this problem, long thought to be out of reach, could be tackled by using sum of squares programming. This technique however has proved to be expensive for large-scale problems, as it involves solving large semidefinite programs (SDPs). In the first part of this thesis, we present two methods for approximately solving large- scale sum of squares programs that dispense altogether with semidefinite programming and only involve solving a sequence of linear or second order cone programs generated in an adaptive fashion. We then focus on the problem of finding tight lower bounds on polyno- mial optimization problems (POPs), a fundamental task in this area that is most commonly handled through the use of SDP-based sum of squares hierarchies (e.g., due to Lasserre and Parrilo). In contrast to previous approaches, we provide the first theoretical framework for constructing converging hierarchies of lower bounds on POPs whose computation simply requires the ability to multiply certain fixed polynomials together and to check nonnegativ- ity of the coefficients of their product.
    [Show full text]
  • Sum of Squares and Polynomial Convexity
    CONFIDENTIAL. Limited circulation. For review only. Sum of Squares and Polynomial Convexity Amir Ali Ahmadi and Pablo A. Parrilo Abstract— The notion of sos-convexity has recently been semidefiniteness of the Hessian matrix is replaced with proposed as a tractable sufficient condition for convexity of the existence of an appropriately defined sum of squares polynomials based on sum of squares decomposition. A multi- decomposition. As we will briefly review in this paper, by variate polynomial p(x) = p(x1; : : : ; xn) is said to be sos-convex if its Hessian H(x) can be factored as H(x) = M T (x) M (x) drawing some appealing connections between real algebra with a possibly nonsquare polynomial matrix M(x). It turns and numerical optimization, the latter problem can be re- out that one can reduce the problem of deciding sos-convexity duced to the feasibility of a semidefinite program. of a polynomial to the feasibility of a semidefinite program, Despite the relative recency of the concept of sos- which can be checked efficiently. Motivated by this computa- convexity, it has already appeared in a number of theo- tional tractability, it has been speculated whether every convex polynomial must necessarily be sos-convex. In this paper, we retical and practical settings. In [6], Helton and Nie use answer this question in the negative by presenting an explicit sos-convexity to give sufficient conditions for semidefinite example of a trivariate homogeneous polynomial of degree eight representability of semialgebraic sets. In [7], Lasserre uses that is convex but not sos-convex. sos-convexity to extend Jensen’s inequality in convex anal- ysis to linear functionals that are not necessarily probability I.
    [Show full text]
  • EE364 Review Session 2
    EE364 Review EE364 Review Session 2 Convex sets and convex functions • Examples from chapter 3 • Installing and using CVX • 1 Operations that preserve convexity Convex sets Convex functions Intersection Nonnegative weighted sum Affine transformation Composition with an affine function Perspective transformation Perspective transformation Pointwise maximum and supremum Minimization Convex sets and convex functions are related via the epigraph. • Composition rules are extremely important. • EE364 Review Session 2 2 Simple composition rules Let h : R R and g : Rn R. Let f(x) = h(g(x)). Then: ! ! f is convex if h is convex and nondecreasing, and g is convex • f is convex if h is convex and nonincreasing, and g is concave • f is concave if h is concave and nondecreasing, and g is concave • f is concave if h is concave and nonincreasing, and g is convex • EE364 Review Session 2 3 Ex. 3.6 Functions and epigraphs. When is the epigraph of a function a halfspace? When is the epigraph of a function a convex cone? When is the epigraph of a function a polyhedron? Solution: If the function is affine, positively homogeneous (f(αx) = αf(x) for α 0), and piecewise-affine, respectively. ≥ Ex. 3.19 Nonnegative weighted sums and integrals. r 1. Show that f(x) = i=1 αix[i] is a convex function of x, where α1 α2 αPr 0, and x[i] denotes the ith largest component of ≥ ≥ · · · ≥ ≥ k n x. (You can use the fact that f(x) = i=1 x[i] is convex on R .) P 2. Let T (x; !) denote the trigonometric polynomial T (x; !) = x + x cos ! + x cos 2! + + x cos(n 1)!: 1 2 3 · · · n − EE364 Review Session 2 4 Show that the function 2π f(x) = log T (x; !) d! − Z0 is convex on x Rn T (x; !) > 0; 0 ! 2π .
    [Show full text]