On the Ekeland Variational Principle with Applications and Detours
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Boosting Algorithms for Maximizing the Soft Margin
Boosting Algorithms for Maximizing the Soft Margin Manfred K. Warmuth∗ Karen Glocer Gunnar Ratsch¨ Dept. of Engineering Dept. of Engineering Friedrich Miescher Laboratory University of California University of California Max Planck Society Santa Cruz, CA, U.S.A. Santa Cruz, CA, U.S.A. T¨ubingen, Germany Abstract We present a novel boosting algorithm, called SoftBoost, designed for sets of bi- nary labeled examples that are not necessarily separable by convex combinations of base hypotheses. Our algorithm achieves robustness by capping the distribu- tions on the examples. Our update of the distribution is motivated by minimizing a relative entropy subject to the capping constraints and constraints on the edges of the obtained base hypotheses. The capping constraints imply a soft margin in the dual optimization problem. Our algorithm produces a convex combination of hypotheses whose soft margin is within δ of its maximum. We employ relative en- ln N tropy projection methods to prove an O( δ2 ) iteration bound for our algorithm, where N is number of examples. We compare our algorithm with other approaches including LPBoost, Brown- Boost, and SmoothBoost. We show that there exist cases where the numberof iter- ations required by LPBoost grows linearly in N instead of the logarithmic growth for SoftBoost. In simulation studies we show that our algorithm converges about as fast as LPBoost, faster than BrownBoost, and much faster than SmoothBoost. In a benchmark comparison we illustrate the competitiveness of our approach. 1 Introduction Boosting methods have been used with great success in many applications like OCR, text classifi- cation, natural language processing, drug discovery, and computational biology [13]. -
Computational Geometry – Problem Session Convex Hull & Line Segment Intersection
Computational Geometry { Problem Session Convex Hull & Line Segment Intersection LEHRSTUHL FUR¨ ALGORITHMIK · INSTITUT FUR¨ THEORETISCHE INFORMATIK · FAKULTAT¨ FUR¨ INFORMATIK Guido Bruckner¨ 04.05.2018 Guido Bruckner¨ · Computational Geometry { Problem Session Modus Operandi To register for the oral exam we expect you to present an original solution for at least one problem in the exercise session. • this is about working together • don't worry if your idea doesn't work! Guido Bruckner¨ · Computational Geometry { Problem Session Outline Convex Hull Line Segment Intersection Guido Bruckner¨ · Computational Geometry { Problem Session Definition of Convex Hull Def: A region S ⊆ R2 is called convex, when for two points p; q 2 S then line pq 2 S. The convex hull CH(S) of S is the smallest convex region containing S. Guido Bruckner¨ · Computational Geometry { Problem Session Definition of Convex Hull Def: A region S ⊆ R2 is called convex, when for two points p; q 2 S then line pq 2 S. The convex hull CH(S) of S is the smallest convex region containing S. Guido Bruckner¨ · Computational Geometry { Problem Session Definition of Convex Hull Def: A region S ⊆ R2 is called convex, when for two points p; q 2 S then line pq 2 S. The convex hull CH(S) of S is the smallest convex region containing S. In physics: Guido Bruckner¨ · Computational Geometry { Problem Session Definition of Convex Hull Def: A region S ⊆ R2 is called convex, when for two points p; q 2 S then line pq 2 S. The convex hull CH(S) of S is the smallest convex region containing S. -
On Quasi Norm Attaining Operators Between Banach Spaces
ON QUASI NORM ATTAINING OPERATORS BETWEEN BANACH SPACES GEUNSU CHOI, YUN SUNG CHOI, MINGU JUNG, AND MIGUEL MART´IN Abstract. We provide a characterization of the Radon-Nikod´ymproperty in terms of the denseness of bounded linear operators which attain their norm in a weak sense, which complement the one given by Bourgain and Huff in the 1970's. To this end, we introduce the following notion: an operator T : X ÝÑ Y between the Banach spaces X and Y is quasi norm attaining if there is a sequence pxnq of norm one elements in X such that pT xnq converges to some u P Y with }u}“}T }. Norm attaining operators in the usual (or strong) sense (i.e. operators for which there is a point in the unit ball where the norm of its image equals the norm of the operator) and also compact operators satisfy this definition. We prove that strong Radon-Nikod´ymoperators can be approximated by quasi norm attaining operators, a result which does not hold for norm attaining operators in the strong sense. This shows that this new notion of quasi norm attainment allows to characterize the Radon-Nikod´ymproperty in terms of denseness of quasi norm attaining operators for both domain and range spaces, completing thus a characterization by Bourgain and Huff in terms of norm attaining operators which is only valid for domain spaces and it is actually false for range spaces (due to a celebrated example by Gowers of 1990). A number of other related results are also included in the paper: we give some positive results on the denseness of norm attaining Lipschitz maps, norm attaining multilinear maps and norm attaining polynomials, characterize both finite dimensionality and reflexivity in terms of quasi norm attaining operators, discuss conditions to obtain that quasi norm attaining operators are actually norm attaining, study the relationship with the norm attainment of the adjoint operator and, finally, present some stability results. -
On the Dual Formulation of Boosting Algorithms
On the Dual Formulation of Boosting Algorithms Chunhua Shen, and Hanxi Li Abstract—We study boosting algorithms from a new perspective. We show that the Lagrange dual problems of ℓ1 norm regularized AdaBoost, LogitBoost and soft-margin LPBoost with generalized hinge loss are all entropy maximization problems. By looking at the dual problems of these boosting algorithms, we show that the success of boosting algorithms can be understood in terms of maintaining a better margin distribution by maximizing margins and at the same time controlling the margin variance. We also theoretically prove that, approximately, ℓ1 norm regularized AdaBoost maximizes the average margin, instead of the minimum margin. The duality formulation also enables us to develop column generation based optimization algorithms, which are totally corrective. We show that they exhibit almost identical classification results to that of standard stage-wise additive boosting algorithms but with much faster convergence rates. Therefore fewer weak classifiers are needed to build the ensemble using our proposed optimization technique. Index Terms—AdaBoost, LogitBoost, LPBoost, Lagrange duality, linear programming, entropy maximization. ✦ 1 INTRODUCTION that the hard-margin LPBoost does not perform well in OOSTING has attracted a lot of research interests since most cases although it usually produces larger minimum B the first practical boosting algorithm, AdaBoost, was margins. More often LPBoost has worse generalization introduced by Freund and Schapire [1]. The machine learn- performance. In other words, a higher minimum margin ing community has spent much effort on understanding would not necessarily imply a lower test error. Breiman [11] how the algorithm works [2], [3], [4]. -
1 Overview 2 Elements of Convex Analysis
AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 2 | Wednesday, January 27th 1 Overview In our previous lecture we discussed several applications of optimization, introduced basic terminol- ogy, and proved Weierstass' theorem (circa 1830) which gives a sufficient condition for existence of an optimal solution to an optimization problem. Today we will discuss basic definitions and prop- erties of convex sets and convex functions. We will then prove the separating hyperplane theorem and see an application of separating hyperplanes in machine learning. We'll conclude by describing the seminal perceptron algorithm (1957) which is designed to find separating hyperplanes. 2 Elements of Convex Analysis We will primarily consider optimization problems over convex sets { sets for which any two points are connected by a line. We illustrate some convex and non-convex sets in Figure 1. Definition. A set S is called a convex set if any two points in S contain their line, i.e. for any x1; x2 2 S we have that λx1 + (1 − λ)x2 2 S for any λ 2 [0; 1]. In the previous lecture we saw linear regression an example of an optimization problem, and men- tioned that the RSS function has a convex shape. We can now define this concept formally. n Definition. For a convex set S ⊆ R , we say that a function f : S ! R is: • convex on S if for any two points x1; x2 2 S and any λ 2 [0; 1] we have that: f (λx1 + (1 − λ)x2) ≤ λf(x1) + 1 − λ f(x2): • strictly convex on S if for any two points x1; x2 2 S and any λ 2 [0; 1] we have that: f (λx1 + (1 − λ)x2) < λf(x1) + (1 − λ) f(x2): We illustrate a convex function in Figure 2. -
CORE View Metadata, Citation and Similar Papers at Core.Ac.Uk
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Bulgarian Digital Mathematics Library at IMI-BAS Serdica Math. J. 27 (2001), 203-218 FIRST ORDER CHARACTERIZATIONS OF PSEUDOCONVEX FUNCTIONS Vsevolod Ivanov Ivanov Communicated by A. L. Dontchev Abstract. First order characterizations of pseudoconvex functions are investigated in terms of generalized directional derivatives. A connection with the invexity is analysed. Well-known first order characterizations of the solution sets of pseudolinear programs are generalized to the case of pseudoconvex programs. The concepts of pseudoconvexity and invexity do not depend on a single definition of the generalized directional derivative. 1. Introduction. Three characterizations of pseudoconvex functions are considered in this paper. The first is new. It is well-known that each pseudo- convex function is invex. Then the following question arises: what is the type of 2000 Mathematics Subject Classification: 26B25, 90C26, 26E15. Key words: Generalized convexity, nonsmooth function, generalized directional derivative, pseudoconvex function, quasiconvex function, invex function, nonsmooth optimization, solution sets, pseudomonotone generalized directional derivative. 204 Vsevolod Ivanov Ivanov the function η from the definition of invexity, when the invex function is pseudo- convex. This question is considered in Section 3, and a first order necessary and sufficient condition for pseudoconvexity of a function is given there. It is shown that the class of strongly pseudoconvex functions, considered by Weir [25], coin- cides with pseudoconvex ones. The main result of Section 3 is applied to characterize the solution set of a nonlinear programming problem in Section 4. The base results of Jeyakumar and Yang in the paper [13] are generalized there to the case, when the function is pseudoconvex. -
POLARS and DUAL CONES 1. Convex Sets
POLARS AND DUAL CONES VERA ROSHCHINA Abstract. The goal of this note is to remind the basic definitions of convex sets and their polars. For more details see the classic references [1, 2] and [3] for polytopes. 1. Convex sets and convex hulls A convex set has no indentations, i.e. it contains all line segments connecting points of this set. Definition 1 (Convex set). A set C ⊆ Rn is called convex if for any x; y 2 C the point λx+(1−λ)y also belongs to C for any λ 2 [0; 1]. The sets shown in Fig. 1 are convex: the first set has a smooth boundary, the second set is a Figure 1. Convex sets convex polygon. Singletons and lines are also convex sets, and linear subspaces are convex. Some nonconvex sets are shown in Fig. 2. Figure 2. Nonconvex sets A line segment connecting two points x; y 2 Rn is denoted by [x; y], so we have [x; y] = f(1 − α)x + αy; α 2 [0; 1]g: Likewise, we can consider open line segments (x; y) = f(1 − α)x + αy; α 2 (0; 1)g: and half open segments such as [x; y) and (x; y]. Clearly any line segment is a convex set. The notion of a line segment connecting two points can be generalised to an arbitrary finite set n n by means of a convex combination. Given a finite set fx1; x2; : : : ; xpg ⊂ R , a point x 2 R is the convex combination of x1; x2; : : : ; xp if it can be represented as p p X X x = αixi; αi ≥ 0 8i = 1; : : : ; p; αi = 1: i=1 i=1 Given an arbitrary set S ⊆ Rn, we can consider all possible finite convex combinations of points from this set. -
José M. Vega-Guzmán
Jos´eM. Vega-Guzm´an Curriculum Vitae Contact Department of Mathematics Tel: (409) 880-8792 Information College of Arts and Sciences Fax: (409) 880-8794 Lamar University email: [email protected] 200-C Lucas Building URL: https://sites.google.com/site/profjmvg/ Beaumont, TX 77710 Education PhD (Applied Mathematics) 2008{2013 • Arizona State University, Tempe, AZ • Co-Advisors: Sergei K. Suslov and Carlos Castillo-Chavez • Dissertation Title: Solution Methods for Certain Evolution Equations Master in Natural Sciences (Mathematics) 2006{2008 • Arizona State University, Tempe, AZ B.S. (MathEd) 2000{2005 • Universidad de Puerto Rico, Cayey, PR • Advisor: Maria A. Avi~n´o Research Interests Nonlinear Wave Propagation. Quantum Mechanics. Integrability of Nonlinear Evolution Equations. Applications of ODE's and PDE's in the Natural Sciences. Mathematical Modeling. Publications 2017 Vega{Guzm´an,J.M.; Alqahtani, R.; Zhou, Q.; Mahmood, M.F.; Moshokoa, S.P.; Ullah, M.Z.; Biswas, A.; Belic, M. Optical solitons for Lakshmanan-Porsezian-Daniel model with spatio-temporal dispersion using the method of undetermined coefficients. Optik - International Journal for Light and Electron Optics. 144, 115{123. 2017 Wang, G.; Kara, A.H.; Vega{Guzm´an,J.M.; Biswas, A. Group analysis, nonlinear self-adjointness, conser- vation laws and soliton solutions for the mKdV systems. Nonlinear Analysis: Modeling and Control. 22(3), 334{346. 2017 Vega{Guzm´an,J.M.; Ullah, M.Z.; Asma, M.; Zhou, Q.; Biswas, A. Dispersive solitons in magneto-optic waveguides. Superlattices and Microstructures. 103, 161{170. 2017 Vega{Guzm´an,J.M.; Mahmood, M.F.; Zhou, Q.; Triki, H.; Arnous, A.H.; Biswas, A.; Moshokoa, S.P.; Belic, M. -
Applications of Convex Analysis Within Mathematics
Noname manuscript No. (will be inserted by the editor) Applications of Convex Analysis within Mathematics Francisco J. Arag´onArtacho · Jonathan M. Borwein · Victoria Mart´ın-M´arquez · Liangjin Yao July 19, 2013 Abstract In this paper, we study convex analysis and its theoretical applications. We first apply important tools of convex analysis to Optimization and to Analysis. We then show various deep applications of convex analysis and especially infimal convolution in Monotone Operator Theory. Among other things, we recapture the Minty surjectivity theorem in Hilbert space, and present a new proof of the sum theorem in reflexive spaces. More technically, we also discuss autoconjugate representers for maximally monotone operators. Finally, we consider various other applications in mathematical analysis. Keywords Adjoint · Asplund averaging · autoconjugate representer · Banach limit · Chebyshev set · convex functions · Fenchel duality · Fenchel conjugate · Fitzpatrick function · Hahn{Banach extension theorem · infimal convolution · linear relation · Minty surjectivity theorem · maximally monotone operator · monotone operator · Moreau's decomposition · Moreau envelope · Moreau's max formula · Moreau{Rockafellar duality · normal cone operator · renorming · resolvent · Sandwich theorem · subdifferential operator · sum theorem · Yosida approximation Mathematics Subject Classification (2000) Primary 47N10 · 90C25; Secondary 47H05 · 47A06 · 47B65 Francisco J. Arag´onArtacho Centre for Computer Assisted Research Mathematics and its Applications (CARMA), -
Convex) Level Sets Integration
Journal of Optimization Theory and Applications manuscript No. (will be inserted by the editor) (Convex) Level Sets Integration Jean-Pierre Crouzeix · Andrew Eberhard · Daniel Ralph Dedicated to Vladimir Demjanov Received: date / Accepted: date Abstract The paper addresses the problem of recovering a pseudoconvex function from the normal cones to its level sets that we call the convex level sets integration problem. An important application is the revealed preference problem. Our main result can be described as integrating a maximally cycli- cally pseudoconvex multivalued map that sends vectors or \bundles" of a Eu- clidean space to convex sets in that space. That is, we are seeking a pseudo convex (real) function such that the normal cone at each boundary point of each of its lower level sets contains the set value of the multivalued map at the same point. This raises the question of uniqueness of that function up to rescaling. Even after normalising the function long an orienting direction, we give a counterexample to its uniqueness. We are, however, able to show uniqueness under a condition motivated by the classical theory of ordinary differential equations. Keywords Convexity and Pseudoconvexity · Monotonicity and Pseudomono- tonicity · Maximality · Revealed Preferences. Mathematics Subject Classification (2000) 26B25 · 91B42 · 91B16 Jean-Pierre Crouzeix, Corresponding Author, LIMOS, Campus Scientifique des C´ezeaux,Universit´eBlaise Pascal, 63170 Aubi`ere,France, E-mail: [email protected] Andrew Eberhard School of Mathematical & Geospatial Sciences, RMIT University, Melbourne, VIC., Aus- tralia, E-mail: [email protected] Daniel Ralph University of Cambridge, Judge Business School, UK, E-mail: [email protected] 2 Jean-Pierre Crouzeix et al. -
Applications of LP Duality
Lecture 6 Duality of LPs and Applications∗ Last lecture we introduced duality of linear programs. We saw how to form duals, and proved both the weak and strong duality theorems. In this lecture we will see a few more theoretical results and then begin discussion of applications of duality. 6.1 More Duality Results 6.1.1 A Quick Review Last time we saw that if the primal (P) is max c>x s:t: Ax ≤ b then the dual (D) is min b>y s:t: A>y = c y ≥ 0: This is just one form of the primal and dual and we saw that the transformation from one to the other is completely mechanical. The duality theorem tells us that if (P) and (D) are a primal-dual pair then we have one of the three possibilities 1. Both (P) and (D) are infeasible. 2. One is infeasible and the other is unbounded. 3. Both are feasible and if x∗ and y∗ are optimal solutions to (P) and (D) respectively, then c>x∗ = b>y∗. *Lecturer: Anupam Gupta. Scribe: Deepak Bal. 1 LECTURE 6. DUALITY OF LPS AND APPLICATIONS 2 6.1.2 A Comment about Complexity Note that the duality theorem (and equivalently, the Farkas Lemma) puts several problems related to LP feasibility and solvability in NP \ co-NP. E.g., Consider the question of whether the equational form LP Ax = b; x ≥ 0 is feasible. If the program is feasible, we may efficiently verify this by checking that a \certificate” point satisfies the equations. By taking this point to be a vertex and appealing to Hwk1 (Problem 4), we see that we may represent this certificate point in size polynomial in the size of the input. -
Introduction to the Modern Calculus of Variations
MA4G6 Lecture Notes Introduction to the Modern Calculus of Variations Filip Rindler Spring Term 2015 Filip Rindler Mathematics Institute University of Warwick Coventry CV4 7AL United Kingdom [email protected] http://www.warwick.ac.uk/filiprindler Copyright ©2015 Filip Rindler. Version 1.1. Preface These lecture notes, written for the MA4G6 Calculus of Variations course at the University of Warwick, intend to give a modern introduction to the Calculus of Variations. I have tried to cover different aspects of the field and to explain how they fit into the “big picture”. This is not an encyclopedic work; many important results are omitted and sometimes I only present a special case of a more general theorem. I have, however, tried to strike a balance between a pure introduction and a text that can be used for later revision of forgotten material. The presentation is based around a few principles: • The presentation is quite “modern” in that I use several techniques which are perhaps not usually found in an introductory text or that have only recently been developed. • For most results, I try to use “reasonable” assumptions, not necessarily minimal ones. • When presented with a choice of how to prove a result, I have usually preferred the (in my opinion) most conceptually clear approach over more “elementary” ones. For example, I use Young measures in many instances, even though this comes at the expense of a higher initial burden of abstract theory. • Wherever possible, I first present an abstract result for general functionals defined on Banach spaces to illustrate the general structure of a certain result.