Lecture 18 Subgradients

Total Page:16

File Type:pdf, Size:1020Kb

Lecture 18 Subgradients Lecture 18 Subgradients November 3, 2008 Lecture 18 Outline • Existence of Subgradients • Subdifferential Properties • Optimality Conditions Convex Optimization 1 Lecture 18 Convex-Constrained Non-differentiable Minimization minimize f(x) subject to x ∈ C • Characteristics: • The function f : Rn → R is convex and possibly non-differentiable • The set C ⊆ Rn is nonempty closed and convex • The optimal value f ∗ is finite • Our focus here is non-differentiability Convex Optimization 2 Lecture 18 Definition of Subgradient and Subdifferential Def. A vector s ∈ Rn is a subgradient of f at xˆ ∈ dom f when f(x) ≥ f(ˆx) + sT (x − xˆ) for all x ∈ dom f Def. A subdifferential of f at xˆ ∈ dom f is the set of all subgradients s of f at xˆ ∈ dom f • The subdifferential of f at xˆ is denoted by ∂f(ˆx) • When f is differentible at xˆ, we have ∂f(ˆx) = {∇f(ˆx)} (the subdif- ferential is a singleton) • Examples sign(x) for x 6= 0 f(x) = |x|, ∂f(0) = [−1, 1] for x = 0 x2 + 2|x| − 3 for |x| > 1 f(x) = 0 for |x| ≤ 1 Convex Optimization 3 Lecture 18 Subgradients and Epigraph • Let s be a subgradient of f at xˆ: f(x) ≥ f(ˆx) + sT (x − xˆ) for all x ∈ dom f • The subgradient inequality is equivalent to −sT xˆ + f(ˆx) ≤ −sT x + f(x) for all x ∈ dom f • Let f(x) > −∞ for all x ∈ Rn. Then n epi f = {(x, w) | f(x) ≤ w, x ∈ R } Thus, −sT xˆ + f(ˆx) ≤ −sT x + w for all (x, w) ∈ epi f, equivalent to T T " −s # " xˆ # " −s # " x # ≤ for all (x, w) ∈ epi f 1 f(ˆx) 1 w Therefore, the hyperplane n o H = (x, γ) ∈ Rn+1 | (−s, 1)T (x, γ) = (−s, 1)T (ˆx, f(ˆx)) supports epi f at the vector (ˆx, f(ˆx)) Convex Optimization 4 Lecture 18 Subdifferential Set Properties Theorem 1 A subdifferential set ∂f(ˆx) is convex and closed Proof H7. Theorem 2 (Existence) Let f be convex with a nonempty dom f. Then: (a) For x ∈ relint(dom f), we have ∂f(x) 6= ∅. (b) ∂f(x) 6= ∅ is nonempty and bounded if and only if x ∈ int(dom f). Implications • The subdifferential ∂f(ˆx) is nonempty compact convex set for every xˆ in the interior of dom f. • When dom f = Rn, ∂f(x) is nonempty compact convex set for all x Convex Optimization 5 Lecture 18 Partial Proof: If xˆ ∈ int(dom f), then ∂f(ˆx) is nonempty and bounded. • ∂f(ˆx) Nonempty. Let xˆ be in the interior of dom f. The vector (ˆx, f(ˆx)) does not belong to the interior of epi f. The epigraph epi f is convex and by the Supporting Hyperlane Theorem, there is a vector (d, β) ∈ Rn+1, (d, β) 6= 0 such that dT xˆ + βf(ˆx) ≤ dT x + βw for all (x, w) ∈ epi f We have epi f = {(x, w) | f(x) ≤ w, x ∈ dom f}. Hence, dT xˆ + βf(ˆx) ≤ dT x + βw for all x ∈ dom f, f(x) ≤ w We must have β ≥ 0. We cannot have β = 0 (it would imply d = 0). Dividing by β, we see that −d/β is a subgradient of f at xˆ Convex Optimization 6 Lecture 18 • ∂f(ˆx) Bounded. By the subgradient inequality, we have f(x) ≥ f(ˆx) + sT (x − xˆ) for all x ∈ dom f Suppose that the subdifferential ∂f(ˆx) is unbounded. Let sk be a sequence of subgradients in ∂f(ˆx) with kskk → ∞. Since xˆ lies in the interior of domain, there exists a δ > 0 such that xˆ + δy ∈ dom f for any y ∈ n. Letting x =x ˆ + δ sk for any k, we R kskk have ! sk f xˆ + δ ≥ f(ˆx) + δkskk for all k kskk As k → ∞, we have f xˆ + δ sk − f(ˆx) → ∞. kskk However, this relation contradicts the continuity of f at xˆ. [Recall, a convex function is continuous over the interior of its domain.] √ Example Consider f(x) = − x with dom f = {x | x ≥ 0}. We have ∂f(0) = ∅. Note that 0 is not in the interior of the domain of f Convex Optimization 7 Lecture 18 Boundedness of the Subdifferential Sets Theorem 2 Let f : Rn → R be convex and let X be a bounded set. Then, the set ∪x∈X∂f(x) is bounded. Proof Assume that there is an unbounded sequence of subgradients sk, i.e., lim kskk = ∞, k→∞ where sk ∈ ∂f(xk) for some xk ∈ X. The sequence {xk} is bounded, so n it has a convergent subsequence, say {xk}K converging to some x ∈ R . sk Consider dk = for k ∈ K. This is a bounded sequence, and it has a kskk Convex Optimization 8 Lecture 18 0 convergent subsequence, say {dk}K0 with K ⊆ K. Let d be the limit of {dk}K0. Since sk ∈ ∂f(xk), we have for each k, T f(xk + dk) ≥ f(xk) + sk dk = f(xk) + kskk. By letting k → ∞ with k ∈ K0, we see that lim sup[f(xk + dk) − f(xk)] ≥ lim sup kskk. k→∞ k→∞ k∈K0 k∈K0 By continuity of f, lim sup[f(xk + dk) − f(xk)] = f(x + d) − f(x), k→∞ k∈K0 hence finite, implying that kskk is bounded. This is a contradition ({sk} was assumed to be unbounded). Convex Optimization 9 Lecture 18 Continuity of the Subdifferential n Theorem 3 Let f : R → R be convex and let {xk} converge to some n x ∈ R . Let sk ∈ ∂f(xk) for all k. Then, the sequence {sk} is bounded and every of its limit points is a subgradient of f at x. Proof H7. Convex Optimization 10 Lecture 18 Subdifferential and Directional Derivatives Definition The directional derivative f 0(x; d) of f at x along direction d is the following limiting value f(x + αd) − f(x) f 0(x; d) = lim . α→0 α f(x+αd)−f(x) • When f is convex, the ratio α is nondecreasing function of α > 0, and as α decreases to zero, the ratio converges to some value or decreases to −∞. (HW) Theorem 4 Let x ∈ int(dom f). Then, the directional derivative f 0(x; d) is finite for all d ∈ Rn. In particular, we have f 0(x; d) = max sT d. s∈∂f(x) Convex Optimization 11 Lecture 18 Proof When x ∈ int(dom f), the subdifferential ∂f(x) is nonempty and compact. Using the subgradient defining relation, we can see that f 0(x; d) ≥ sT d for all s ∈ ∂f(x). Therefore, f 0(x; d) ≥ max sT d. s∈∂f(x) To show that actually equality holds, we rely on Separating Hyperplane Theorem. Define C1 = {(z, w) | z ∈ dom f, f(z) < w}, 0 C2 = {(y, v) | y = x + αd, v = f(x) + αf (x; d), α ≥ 0}. These sets are nonempty, convex, and disjoint (HW). By the Separating Convex Optimization 12 Lecture 18 Hyperplane Theorem, there exists a nonzero vector (a, β) ∈ Rn+1 such that aT (x + αd) + β(f(x) + αf 0(x; d)) ≤ aT z + βw, (1) for all α ≥ 0, z ∈ dom f, and f(z) < w. We must have β ≥ 0 - why? We cannot have β = 0 -why? Thus, β > 0 and we can divide by β the relation in (1), and obtain with ˜a = a/β, ˜aT (x + αd) + f(x) + αf 0(x; d) ≤ ˜aT z + w, (2) for all α ≥ 0, z ∈ dom f, and f(z) < w. Choosing α = 0 and letting w ↓ f(z), we see ˜aT x + f(x) ≤ ˜aT z + f(z), implying that f(x) − ˜aT (z − x) ≤ f(z) for all z ∈ dom f. Therefore −˜a ∈ ∂f(x). Convex Optimization 13 Lecture 18 Letting z = x, w ↓ f(z) and α = 1 in (2), we obtain ˜aT (x + d) + f(x) + f 0(x; d) ≤ ˜aT x + f(x), 0 T 0 T implying f (x; d) ≤ −˜a d. In view of f (x; d) ≥ maxs∈∂f(x) s d, it follows that f 0(x; d) = max sT d, s∈∂f(x) where the maximum is attained at the “constructed” subgradient −˜a. Convex Optimization 14 Lecture 18 Optimality Conditions: Unconstrained Case Unconstrained optimization minimize f(x) Assumption • The function f is convex (non-differentiable) and proper [f proper means f(x) > −∞ for all x and dom f 6= ∅] Theorem Under this assumption, a vector x∗ minimizes f over Rn if and only if 0 ∈ ∂f(x∗) • The result is a generalization of ∇f(x∗) = 0 • Proof x∗ is optimal if and only if f(x) ≥ f(x∗) for all x, or equivalently ∗ T ∗ n f(x) ≥ f(x ) + 0 (x − x ) for all x ∈ R Thus, x∗ is optimal if and only if 0 ∈ ∂f(x∗) Convex Optimization 15 Lecture 18 Examples • The function f(x) = |x| sign(x) for x 6= 0 ∂f(0) = [−1, 1] for x = 0 The minimum is at x∗ = 0, and evidently 0 ∈ ∂f(0) • The function f(x) = kxk x for x 6= 0 ∂f(x) = kxk {s | ksk ≤ 1} for x = 0 Again, the minimum is at x∗ = 0 and 0 ∈ ∂f(0) Convex Optimization 16 Lecture 18 • The function f(x) = max{x2 + 2x − 3, x2 − 2x − 3, 4} 2 x − 2x − 3 for x < −1 f(x) = 4 for x ∈ [−1, 1] 2 x + 2x − 3 for x > 1 2 2 1 x − for x > [−4, 0] for x = −1 ∂f(x) = 0 for x ∈ (−1, 1) [0, 4] for x = 1 2x + 2 for x > 1 • The optimal set is X∗ = [−1, 1] • For every x∗ ∈ X∗, we have 0 ∈ ∂f(x∗) Convex Optimization 17 Lecture 18 Optimality Conditions: Constrained Case Constrained optimization minimize f(x) subject to x ∈ C Assumption • The function f is convex (non-differentiable) and proper • The set C is nonempty closed and convex Theorem Under this assumption, a vector x∗ ∈ C minimizes f over the set C if and only if there exists a subgradient d ∈ ∂f(x∗) such that dT (x − x∗) ≥ 0 for all x ∈ C • The result is a generalization of ∇f(x∗)T (x − x∗) ≥ 0 for x ∈ C Convex Optimization 18.
Recommended publications
  • On Choquet Theorem for Random Upper Semicontinuous Functions
    CORE Metadata, citation and similar papers at core.ac.uk Provided by Elsevier - Publisher Connector International Journal of Approximate Reasoning 46 (2007) 3–16 www.elsevier.com/locate/ijar On Choquet theorem for random upper semicontinuous functions Hung T. Nguyen a, Yangeng Wang b,1, Guo Wei c,* a Department of Mathematical Sciences, New Mexico State University, Las Cruces, NM 88003, USA b Department of Mathematics, Northwest University, Xian, Shaanxi 710069, PR China c Department of Mathematics and Computer Science, University of North Carolina at Pembroke, Pembroke, NC 28372, USA Received 16 May 2006; received in revised form 17 October 2006; accepted 1 December 2006 Available online 29 December 2006 Abstract Motivated by the problem of modeling of coarse data in statistics, we investigate in this paper topologies of the space of upper semicontinuous functions with values in the unit interval [0,1] to define rigorously the concept of random fuzzy closed sets. Unlike the case of the extended real line by Salinetti and Wets, we obtain a topological embedding of random fuzzy closed sets into the space of closed sets of a product space, from which Choquet theorem can be investigated rigorously. Pre- cise topological and metric considerations as well as results of probability measures and special prop- erties of capacity functionals for random u.s.c. functions are also given. Ó 2006 Elsevier Inc. All rights reserved. Keywords: Choquet theorem; Random sets; Upper semicontinuous functions 1. Introduction Random sets are mathematical models for coarse data in statistics (see e.g. [12]). A the- ory of random closed sets on Hausdorff, locally compact and second countable (i.e., a * Corresponding author.
    [Show full text]
  • EE364 Review Session 2
    EE364 Review EE364 Review Session 2 Convex sets and convex functions • Examples from chapter 3 • Installing and using CVX • 1 Operations that preserve convexity Convex sets Convex functions Intersection Nonnegative weighted sum Affine transformation Composition with an affine function Perspective transformation Perspective transformation Pointwise maximum and supremum Minimization Convex sets and convex functions are related via the epigraph. • Composition rules are extremely important. • EE364 Review Session 2 2 Simple composition rules Let h : R R and g : Rn R. Let f(x) = h(g(x)). Then: ! ! f is convex if h is convex and nondecreasing, and g is convex • f is convex if h is convex and nonincreasing, and g is concave • f is concave if h is concave and nondecreasing, and g is concave • f is concave if h is concave and nonincreasing, and g is convex • EE364 Review Session 2 3 Ex. 3.6 Functions and epigraphs. When is the epigraph of a function a halfspace? When is the epigraph of a function a convex cone? When is the epigraph of a function a polyhedron? Solution: If the function is affine, positively homogeneous (f(αx) = αf(x) for α 0), and piecewise-affine, respectively. ≥ Ex. 3.19 Nonnegative weighted sums and integrals. r 1. Show that f(x) = i=1 αix[i] is a convex function of x, where α1 α2 αPr 0, and x[i] denotes the ith largest component of ≥ ≥ · · · ≥ ≥ k n x. (You can use the fact that f(x) = i=1 x[i] is convex on R .) P 2. Let T (x; !) denote the trigonometric polynomial T (x; !) = x + x cos ! + x cos 2! + + x cos(n 1)!: 1 2 3 · · · n − EE364 Review Session 2 4 Show that the function 2π f(x) = log T (x; !) d! − Z0 is convex on x Rn T (x; !) > 0; 0 ! 2π .
    [Show full text]
  • NP-Hardness of Deciding Convexity of Quartic Polynomials and Related Problems
    NP-hardness of Deciding Convexity of Quartic Polynomials and Related Problems Amir Ali Ahmadi, Alex Olshevsky, Pablo A. Parrilo, and John N. Tsitsiklis ∗y Abstract We show that unless P=NP, there exists no polynomial time (or even pseudo-polynomial time) algorithm that can decide whether a multivariate polynomial of degree four (or higher even degree) is globally convex. This solves a problem that has been open since 1992 when N. Z. Shor asked for the complexity of deciding convexity for quartic polynomials. We also prove that deciding strict convexity, strong convexity, quasiconvexity, and pseudoconvexity of polynomials of even degree four or higher is strongly NP-hard. By contrast, we show that quasiconvexity and pseudoconvexity of odd degree polynomials can be decided in polynomial time. 1 Introduction The role of convexity in modern day mathematical programming has proven to be remarkably fundamental, to the point that tractability of an optimization problem is nowadays assessed, more often than not, by whether or not the problem benefits from some sort of underlying convexity. In the famous words of Rockafellar [39]: \In fact the great watershed in optimization isn't between linearity and nonlinearity, but convexity and nonconvexity." But how easy is it to distinguish between convexity and nonconvexity? Can we decide in an efficient manner if a given optimization problem is convex? A class of optimization problems that allow for a rigorous study of this question from a com- putational complexity viewpoint is the class of polynomial optimization problems. These are op- timization problems where the objective is given by a polynomial function and the feasible set is described by polynomial inequalities.
    [Show full text]
  • Arxiv:2103.15804V4 [Math.AT] 25 May 2021
    Decorated Merge Trees for Persistent Topology Justin Curry, Haibin Hang, Washington Mio, Tom Needham, and Osman Berat Okutan Abstract. This paper introduces decorated merge trees (DMTs) as a novel invariant for persistent spaces. DMTs combine both π0 and Hn information into a single data structure that distinguishes filtrations that merge trees and persistent homology cannot distinguish alone. Three variants on DMTs, which empha- size category theory, representation theory and persistence barcodes, respectively, offer different advantages in terms of theory and computation. Two notions of distance|an interleaving distance and bottleneck distance|for DMTs are defined and a hierarchy of stability results that both refine and generalize existing stability results is proved here. To overcome some of the computational complexity inherent in these dis- tances, we provide a novel use of Gromov-Wasserstein couplings to compute optimal merge tree alignments for a combinatorial version of our interleaving distance which can be tractably estimated. We introduce computational frameworks for generating, visualizing and comparing decorated merge trees derived from synthetic and real data. Example applications include comparison of point clouds, interpretation of persis- tent homology of sliding window embeddings of time series, visualization of topological features in segmented brain tumor images and topology-driven graph alignment. Contents 1. Introduction 1 2. Decorated Merge Trees Three Different Ways3 3. Continuity and Stability of Decorated Merge Trees 10 4. Representations of Tree Posets and Lift Decorations 16 5. Computing Interleaving Distances 20 6. Algorithmic Details and Examples 24 7. Discussion 29 References 30 Appendix A. Interval Topology 34 Appendix B. Comparison and Existence of Merge Trees 35 Appendix C.
    [Show full text]
  • Linear Algebra Review
    CSE 203B: Convex Optimization Week 4 Discuss Session 1 Contents • Convex functions (Ref. Chap.3) • Review: definition, first order condition, second order condition, operations that preserve convexity • Epigraph • Conjugate function • Dual norm 2 Review of Convex Function • Definition ( often simplified by restricting to a line) • First order condition: a global underestimator • Second order condition • Operations that preserve convexity • Nonnegative weighted sum • Composition with affine function • Pointwise maximum and supremum • Composition • Minimization See Chap 3.2, try to prove why the convexity • Perspective is preserved with those operations. 3 Epigraph • 훼-sublevel set of 푓: 푅푛 → 푅 퐶훼 = 푥 ∈ 푑표푚 푓 푓 푥 ≤ 훼} sublevel sets of a convex function are convex for any value of 훼. • Epigraph of 푓: 푅푛 → 푅 is defined as 퐞퐩퐢 푓 = (푥, 푡) 푥 ∈ 푑표푚 푓, 푓 푥 ≤ 푡} ⊆ 푅푛+1 • A function is convex iff its epigraph is a convex set. 4 Relation between convex sets and convex functions • A function is convex iff its epigraph is a convex set. • Consider a convex function 푓 and 푥, 푦 ∈ 푑표푚 푓 푡 ≥ 푓 푦 ≥ 푓 푥 + 훻푓 푥 푇(푦 − 푥) epi 풇 First order condition for convexity • The hyperplane supports epi 풇 at (푥, 푓 푥 ), for any 푡 푥 푦, 푡 ∈ 퐞퐩퐢 푓 ⇒ 훻푓 푥 푇 푦 − 푥 + 푓 푥 − 푡 ≤ 0 훻푓 푥 푇 푦 푥 ⇒ − ≤ 0 −1 푡 푓 푥 Supporting hyperplane, derived from first order condition 5 Pointwise Supremum • If for each 푦 ∈ 푈, 푓(푥, 푦): 푅푛 → 푅 is convex in 푥, then function 푔(푥) = sup 푓(푥, 푦) 푦∈푈 is convex in 푥.
    [Show full text]
  • Lecture 3: Convex Sets and Functions 3.1 Convex Sets
    EE 227A: Convex Optimization and Applications January 24, 2012 Lecture 3: Convex Sets and Functions Lecturer: Laurent El Ghaoui Reading assignment: Chapters 2 (except x2.6) and sections 3.1, 3.2, 3.3 of BV. You can also look at section 3.1 of the web textbook. 3.1 Convex Sets Definition. A subset C of Rn is convex if and only if it contains the line segment between any two points in it: 8 x1; x2 2 C; 8 θ1 ≥ 0; θ2 ≥ 0; θ1 + θ2 = 1 : θ1x1 + θ2x2 2 C: Some important special cases of convex sets are the following. • The set is said to be an affine subspace if it contains the entire line passing through any two points. This corresponds to the condition above, with θ1; θ2 arbitrary. Subspaces and affine subspaces are convex. • The set is said to be a convex cone if the condition above holds, but with the restriction θ1 + θ2 = 1 removed. Examples. • The convex hull of a set of points fx1; : : : ; xmg is defined as ( m m ) X m X Co(x1; : : : ; xm) := λixi : λ 2 R+ ; λi = 1 ; i=1 i=1 and is convex. The conic hull: ( m ) X m λixi : λ 2 R+ i=1 is a convex cone. • For a 2 Rn, and b 2 R, the hyperplane H = fx : aT x = bg is affine. The half-space fx : aT x ≤ bg is convex. 3-1 EE 227A Lecture 3 | January 24, 2012 Sp'12 n×n n • For a square, non-singular matrix R 2 R , and xc 2 R , the ellipsoid fxc + Ru : kuk2 ≤ 1g is convex.
    [Show full text]
  • Convex Optimization Scai Lab Study Group
    Convex Optimization ScAi Lab Study Group Zhiping (Patricia) Xiao University of California, Los Angeles Winter 2020 Outline 2 Introduction: Convex Optimization Convexity Convex Functions' Properties Definition of Convex Optimization Convex Optimization General Strategy Learning Algorithms Convergence Analysis Examples Source of Materials 3 Textbook: I Convex Optimization and Intro to Linear Algebra by Prof. Boyd and Prof. Vandenberghe Course Materials: I ECE236B, ECE236C offered by Prof. Vandenberghe I CS260 Lecture 12 offered by Prof. Quanquan Gu Notes: I My previous ECE236B notes and ECE236C final report. I My previous CS260 Cheat Sheet. Related Papers: I Accelerated methods for nonconvex optimization I Lipschitz regularity of deep neural networks: analysis and efficient estimation Introduction: Convex Optimization q Notations 5 I iff: if and only if I R+ = fx 2 R j x ≥ 0g I R++ = fx 2 R j x > 0g I int K: interior of set K, not its boundary. I Generalized inequalities (textbook 2.4), based on a proper cone K (convex, closed, solid, pointed | if x 2 K and −x 2 K then x = 0): I x K y () y − x 2 K I x ≺K y () y − x 2 int K n n T I Positive semidefinite matrix X 2 S+, 8y 2 R , y Xy ≥ 0 () X 0. Convexity of Set 6 Set C is convex iff the line segment between any two points in C lies in C, i.e. 8x1; x2 2 C and 8θ 2 [0; 1], we have: θx1 + (1 − θ)x2 2 C Both convex and nonconvex sets have convex hull, which is defined as: k k X X conv C = f θixi j xi 2 C; θi ≥ 0; i = 1; 2; : : : ; k; θi = 1g i=1 i=1 Convexity of Set (Examples from Textbook) 7 Figure: Left: convex, middle & right: nonconvex.
    [Show full text]
  • Convex Sets and Functions (Part III)
    Convex Sets and Functions (part III) Prof. Dan A. Simovici UMB 1 / 14 Outline 1 The epigraph and hypograph of functions 2 Constructing Convex Functions 2 / 14 The epigraph and hypograph of functions Definition Let (S; O) be a topological space and let f : S −! R^ be a function. Its epigraph is the set epi(f ) = f(x; y) 2 S × R j f (x) 6 yg: The hypograph of f is the set hyp(f ) = f(x; y) 2 S × R j y 6 f (x)g: 3 / 14 The epigraph and hypograph of functions 2 The epigraph of a function f : R −! R is the dotted area in R located above the graph of the function f ; the hypograph of f is the dotted area below the graph. y y x x (a) (b) 4 / 14 The epigraph and hypograph of functions Note that the intersection epi(f ) \ hyp(f ) = f(x; y) 2 S × R j y = f (x)g is the graph of the function f . If f (x) = 1, then (x; 1) 62 epi(f ). Thus, for the function f1 defined by f1(x) = 1 for x 2 S we have epi(f1) = ;. 5 / 14 The epigraph and hypograph of functions Theorem Let f : S −! R be a function defined on the convex subset S of a real linear space L. Then, f is convex on S if and only if its epigraph is a convex subset of S × R; f is concave if and only if its hypograph is a convex subset of S × R.
    [Show full text]
  • Convex Analysis
    Convex Analysis Taught by Professor Adrian Lewis Notes by Mateo D´ıaz [email protected] Cornell University Spring, 2018 Last updated May 18, 2018. The latest version is online here. Notes template by David Mehrle. Contents 1 Convex Sets .................................................. 4 2 Convex functions............................................... 6 2.1 Convexity and calculus......................................... 7 2.1.1 Gradients ............................................ 7 2.1.2 Recognizing smooth convexity................................ 8 2.2 Continuity of convex functions.................................... 8 2.3 Subgradients............................................... 9 3 Duality..................................................... 10 3.1 Fenchel Duality............................................. 10 3.2 Conic programs............................................. 12 3.3 Convex calculus............................................. 14 3.4 Lagrangian duality........................................... 15 4 Algorithms................................................... 19 4.1 Augmented Lagrangian Method ................................... 19 4.2 Proximal Point method and fixed points............................... 20 4.3 Smooth Minimization.......................................... 25 4.4 Splitting Problems and Alternating projections........................... 26 4.5 Proximal Gradient ........................................... 28 4.6 Douglas-Rachford............................................ 29 4.6.1 Consensus optimization...................................
    [Show full text]
  • On a Conic Approach to Convex Analysis 1 Introduction
    On a Conic Approach to Convex Analysis Jan Brinkhuis February 2008 Econometric Institute Report EI 2008-05 Abstract The aim of this paper is to make an attempt to justify the main results from Convex Analysis by one elegant tool, the conification method, which consists of three steps: conify, work with convex cones, deconify. It is based on the fact that the standard operations and facts (‘the calculi’) are much simpler for special convex sets, convex cones. By considering suitable classes of convex cones, we get the standard operations and facts for all situations in the complete generality that is required. The main advantages of this conification method are that the standard operations— linear image, inverse linear image, closure, the duality operator, the binary operations and the inf- operator—are defined on all objects of each class of convex objects—convex sets, convex functions, convex cones and sublinear functions—and that moreover the standard facts—such as the duality theorem—hold forall closed convex objects. This requires that the analysis is carried out in the context of convex objects over cosmic space, the space that is obtained from ordinary space by adding a horizon, representing the directions of ordinary space. 1 Introduction Reduction to conic calculus. This paper presents some progress on joint work with V.M.Tikhomirov that was published in [7]. It is an attempt to develop a general and universal tool for justifying the main results of Convex Analysis. The idea is as follows. Many standard operations and facts are much simpler and more complete for special convex objects, convex cones.
    [Show full text]
  • Generalized Convexity in Mathematical Programming
    BULL. AUSTRAL. MATH. SOC. 90C25, 90C30 VOL. 27 (1983), 185-202. GENERALIZED CONVEXITY IN MATHEMATICAL PROGRAMMING B. MOND The role of convexity in the development of mathematical programming is reviewed. Many recent generalizations of convexity and their applications to optimization theory are discussed. 1. Introduction As is well known, convexity plays a key role in mathematical programming. However, functions that are, in some sense, close to, but not quite, convex often have many of the important properties of convex functions. This, plus the fact that many practical problems involve functions that are 'almost1 convex has led to many generalizations of the notion of convex functions. Here we survey some of these extensions of convexity and indicate briefly the role they play in optimization theory. Most of the results in Sections 2 to 5 can be found in the books by Avriel [7], Mangasarian [76] and Martos [77] while the results in Section 6 on cone-convexity can be found in the book by Craven [JS]. In view of this, no additional references or original sources will be given for results in these sections except for the new dual for non-linear programs in Section 5 that does not appear in any of the above-mentioned books. More detailed references will be given in the later sections. Proofs will only be given where they are extremely brief and not generally available in the literature. It should be noted that most proofs Received 7 October 1982. An invited lecture delivered to the Annual Congress of the South African Mathematical Society, October 1982.
    [Show full text]
  • GEOMETRIC APPROACH to CONVEX SUBDIFFERENTIAL CALCULUS October 10, 2018
    GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, 2018 BORIS S. MORDUKHOVICH1 and NGUYEN MAU NAM2 Dedicated to Franco Giannessi and Diethard Pallaschke with great respect Abstract. In this paper we develop a geometric approach to convex subdifferential calculus in finite dimensions with employing some ideas of modern variational analysis. This approach allows us to obtain natural and rather easy proofs of basic results of convex subdifferential calculus in full generality and also derive new results of convex analysis concerning optimal value/marginal func- tions, normals to inverse images of sets under set-valued mappings, calculus rules for coderivatives of single-valued and set-valued mappings, and calculating coderivatives of solution maps to param- eterized generalized equations governed by set-valued mappings with convex graphs. Key words. convex analysis, generalized differentiation, geometric approach, convex separation, normal cone, subdifferential, coderivative, calculus rules, maximum function, optimal value function AMS subject classifications. 49J52, 49J53, 90C31 1 Introduction The notion of subdifferential (collection of subgradients) for nondifferentiable convex func- tions was independently introduced and developed by Moreau [15] and Rockafellar [19] who were both influenced by Fenchel [6]. Since then this notion has become one of the most central concepts of convex analysis and its various applications, including first of all convex optimization. The underlying difference between the standard derivative of a differentiable function and the subdifferential of a convex function at a given point is that the subdif- ferential is a set (of subgradients) which reduces to a singleton (gradient) if the function is differentiable. Due to the set-valuedness of the subdifferential, deriving calculus rules for it is a significantly more involved task in comparison with classical differential calculus.
    [Show full text]