CS168: the Modern Algorithmic Toolbox Lecture #18: Linear and Convex Programming, with Applications to Sparse Recovery
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Lecture 3 1 Geometry of Linear Programs
ORIE 6300 Mathematical Programming I September 2, 2014 Lecture 3 Lecturer: David P. Williamson Scribe: Divya Singhvi Last time we discussed how to take dual of an LP in two different ways. Today we will talk about the geometry of linear programs. 1 Geometry of Linear Programs First we need some definitions. Definition 1 A set S ⊆ <n is convex if 8x; y 2 S, λx + (1 − λ)y 2 S, 8λ 2 [0; 1]. Figure 1: Examples of convex and non convex sets Given a set of inequalities we define the feasible region as P = fx 2 <n : Ax ≤ bg. We say that P is a polyhedron. Which points on this figure can have the optimal value? Our intuition from last time is that Figure 2: Example of a polyhedron. \Circled" corners are feasible and \squared" are non feasible optimal solutions to linear programming problems occur at \corners" of the feasible region. What we'd like to do now is to consider formal definitions of the \corners" of the feasible region. 3-1 One idea is that a point in the polyhedron is a corner if there is some objective function that is minimized there uniquely. Definition 2 x 2 P is a vertex of P if 9c 2 <n with cT x < cT y; 8y 6= x; y 2 P . Another idea is that a point x 2 P is a corner if there are no small perturbations of x that are in P . Definition 3 Let P be a convex set in <n. Then x 2 P is an extreme point of P if x cannot be written as λy + (1 − λ)z for y; z 2 P , y; z 6= x, 0 ≤ λ ≤ 1. -
4. Convex Optimization Problems
Convex Optimization — Boyd & Vandenberghe 4. Convex optimization problems optimization problem in standard form • convex optimization problems • quasiconvex optimization • linear optimization • quadratic optimization • geometric programming • generalized inequality constraints • semidefinite programming • vector optimization • 4–1 Optimization problem in standard form minimize f0(x) subject to f (x) 0, i =1,...,m i ≤ hi(x)=0, i =1,...,p x Rn is the optimization variable • ∈ f : Rn R is the objective or cost function • 0 → f : Rn R, i =1,...,m, are the inequality constraint functions • i → h : Rn R are the equality constraint functions • i → optimal value: p⋆ = inf f (x) f (x) 0, i =1,...,m, h (x)=0, i =1,...,p { 0 | i ≤ i } p⋆ = if problem is infeasible (no x satisfies the constraints) • ∞ p⋆ = if problem is unbounded below • −∞ Convex optimization problems 4–2 Optimal and locally optimal points x is feasible if x dom f and it satisfies the constraints ∈ 0 ⋆ a feasible x is optimal if f0(x)= p ; Xopt is the set of optimal points x is locally optimal if there is an R> 0 such that x is optimal for minimize (over z) f0(z) subject to fi(z) 0, i =1,...,m, hi(z)=0, i =1,...,p z x≤ R k − k2 ≤ examples (with n =1, m = p =0) f (x)=1/x, dom f = R : p⋆ =0, no optimal point • 0 0 ++ f (x)= log x, dom f = R : p⋆ = • 0 − 0 ++ −∞ f (x)= x log x, dom f = R : p⋆ = 1/e, x =1/e is optimal • 0 0 ++ − f (x)= x3 3x, p⋆ = , local optimum at x =1 • 0 − −∞ Convex optimization problems 4–3 Implicit constraints the standard form optimization problem has an implicit -
Metaheuristics1
METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming. -
Extreme Points and Basic Solutions
EXTREME POINTS AND BASIC SOLUTIONS: In Linear Programming, the feasible region in Rn is defined by P := {x ∈ Rn | Ax = b, x ≥ 0}. The set P , as we have seen, is a convex subset of Rn. It is called a convex polytope. The term convex polyhedron refers to convex polytope which is bounded. Polytopes in two dimensions are often called polygons. Recall that the vertices of a convex polytope are what we called extreme points of that set. Recall that extreme points of a convex set are those which cannot be represented as a proper convex combination of two other (distinct) points of the convex set. It may, or may not be the case that a convex set has any extreme points as shown by the example in R2 of the strip S := {(x, y) ∈ R2 | 0 ≤ x ≤ 1, y ∈ R}. On the other hand, the square defined by the inequalities |x| ≤ 1, |y| ≤ 1 has exactly four extreme points, while the unit disk described by the ineqality x2 + y2 ≤ 1 has infinitely many. These examples raise the question of finding conditions under which a convex set has extreme points. The answer in general vector spaces is answered by one of the “big theorems” called the Krein-Milman Theorem. However, as we will see presently, our study of the linear programming problem actually answers this question for convex polytopes without needing to call on that major result. The algebraic characterization of the vertices of the feasible polytope confirms the obser- vation that we made by following the steps of the Simplex Algorithm in our introductory example. -
The Feasible Region for Consecutive Patterns of Permutations Is a Cycle Polytope
Séminaire Lotharingien de Combinatoire XX (2020) Proceedings of the 32nd Conference on Formal Power Article #YY, 12 pp. Series and Algebraic Combinatorics (Ramat Gan) The feasible region for consecutive patterns of permutations is a cycle polytope Jacopo Borga∗1, and Raul Penaguiaoy 1 1Department of Mathematics, University of Zurich, Switzerland Abstract. We study proportions of consecutive occurrences of permutations of a given size. Specifically, the feasible limits of such proportions on large permutations form a region, called feasible region. We show that this feasible region is a polytope, more precisely the cycle polytope of a specific graph called overlap graph. This allows us to compute the dimension, vertices and faces of the polytope. Finally, we prove that the limits of classical occurrences and consecutive occurrences are independent, in some sense made precise in the extended abstract. As a consequence, the scaling limit of a sequence of permutations induces no constraints on the local limit and vice versa. Keywords: permutation patterns, cycle polytopes, overlap graphs. This is a shorter version of the preprint [11] that is currently submitted to a journal. Many proofs and details omitted here can be found in [11]. 1 Introduction 1.1 Motivations Despite not presenting any probabilistic result here, we give some motivations that come from the study of random permutations. This is a classical topic at the interface of combinatorics and discrete probability theory. There are two main approaches to it: the first concerns the study of statistics on permutations, and the second, more recent, looks arXiv:2003.12661v2 [math.CO] 30 Jun 2020 for the limits of permutations themselves. -
Linear Programming
Stanford University | CS261: Optimization Handout 5 Luca Trevisan January 18, 2011 Lecture 5 In which we introduce linear programming. 1 Linear Programming A linear program is an optimization problem in which we have a collection of variables, which can take real values, and we want to find an assignment of values to the variables that satisfies a given collection of linear inequalities and that maximizes or minimizes a given linear function. (The term programming in linear programming, is not used as in computer program- ming, but as in, e.g., tv programming, to mean planning.) For example, the following is a linear program. maximize x1 + x2 subject to x + 2x ≤ 1 1 2 (1) 2x1 + x2 ≤ 1 x1 ≥ 0 x2 ≥ 0 The linear function that we want to optimize (x1 + x2 in the above example) is called the objective function.A feasible solution is an assignment of values to the variables that satisfies the inequalities. The value that the objective function gives 1 to an assignment is called the cost of the assignment. For example, x1 := 3 and 1 2 x2 := 3 is a feasible solution, of cost 3 . Note that if x1; x2 are values that satisfy the inequalities, then, by summing the first two inequalities, we see that 3x1 + 3x2 ≤ 2 that is, 1 2 x + x ≤ 1 2 3 2 1 1 and so no feasible solution has cost higher than 3 , so the solution x1 := 3 , x2 := 3 is optimal. As we will see in the next lecture, this trick of summing inequalities to verify the optimality of a solution is part of the very general theory of duality of linear programming. -
IEOR 269, Spring 2010 Integer Programming and Combinatorial Optimization
IEOR 269, Spring 2010 Integer Programming and Combinatorial Optimization Professor Dorit S. Hochbaum Contents 1 Introduction 1 2 Formulation of some ILP 2 2.1 0-1 knapsack problem . 2 2.2 Assignment problem . 2 3 Non-linear Objective functions 4 3.1 Production problem with set-up costs . 4 3.2 Piecewise linear cost function . 5 3.3 Piecewise linear convex cost function . 6 3.4 Disjunctive constraints . 7 4 Some famous combinatorial problems 7 4.1 Max clique problem . 7 4.2 SAT (satisfiability) . 7 4.3 Vertex cover problem . 7 5 General optimization 8 6 Neighborhood 8 6.1 Exact neighborhood . 8 7 Complexity of algorithms 9 7.1 Finding the maximum element . 9 7.2 0-1 knapsack . 9 7.3 Linear systems . 10 7.4 Linear Programming . 11 8 Some interesting IP formulations 12 8.1 The fixed cost plant location problem . 12 8.2 Minimum/maximum spanning tree (MST) . 12 9 The Minimum Spanning Tree (MST) Problem 13 i IEOR269 notes, Prof. Hochbaum, 2010 ii 10 General Matching Problem 14 10.1 Maximum Matching Problem in Bipartite Graphs . 14 10.2 Maximum Matching Problem in Non-Bipartite Graphs . 15 10.3 Constraint Matrix Analysis for Matching Problems . 16 11 Traveling Salesperson Problem (TSP) 17 11.1 IP Formulation for TSP . 17 12 Discussion of LP-Formulation for MST 18 13 Branch-and-Bound 20 13.1 The Branch-and-Bound technique . 20 13.2 Other Branch-and-Bound techniques . 22 14 Basic graph definitions 23 15 Complexity analysis 24 15.1 Measuring quality of an algorithm . -
Math 112 Review for Exam Ii (Ws 12 -18)
MATH 112 REVIEW FOR EXAM II (WS 12 -18) I. Derivative Rules • There will be a page or so of derivatives on the exam. You should know how to apply all the derivative rules. (WS 12 and 13) II. Functions of One Variable • Be able to find local optima, and to distinguish between local and global optima. • Be able to find the global maximum and minimum of a function y = f(x) on the interval from x = a to x = b, using the fact that optima may only occur where f(x) has a horizontal tangent line and at the endpoints of the interval. Step 1: Compute the derivative f’(x). Step 2: Find all critical points (values of x at which f’(x) = 0.) Step 3: Plug all the values of x from Step 2 that are in the interval from a to b and the endpoints of the interval into the function f(x). Step 4: Sketch a rough graph of f(x) and pick off the global max and min. • Understand the following application: Maximizing TR(q) starting with a demand curve. (WS 15) • Understand how to use the Second Derivative Test. (WS 16) If a is a critical point for f(x) (that is, f’(a) = 0), and the second derivative is: f ’’(a) > 0, then f(x) has a local min at x = a. f ’’(a) < 0, then f(x) has a local max at x = a. f ’’(a) = 0, then the test tells you nothing. IMPORTANT! For the Second Derivative Test to work, you must have f’(a) = 0 to start with! For example, if f ’’(a) > 0 but f ’(a) ≠ 0, then the graph of f(x) is concave up at x = a but f(x) does not have a local min there. -
The Feasible Region for Consecutive Patterns of Permutations Is a Cycle Polytope
Algebraic Combinatorics Draft The feasible region for consecutive patterns of permutations is a cycle polytope Jacopo Borga & Raul Penaguiao Abstract. We study proportions of consecutive occurrences of permutations of a given size. Specifically, the limit of such proportions on large permutations forms a region, called feasible region. We show that this feasible region is a polytope, more precisely the cycle polytope of a specific graph called overlap graph. This allows us to compute the dimension, vertices and faces of the polytope, and to determine the equations that define it. Finally we prove that the limit of classical occurrences and consecutive occurrences are in some sense independent. As a consequence, the scaling limit of a sequence of permutations induces no constraints on the local limit and vice versa. (1,0,0,0,0,0) 1 1 (0; 2 ; 2 ; 0; 0; 0) 1 1 1 1 (0; 0; 2 ; 2 ; 0; 0) (0; 2 ; 0; 2 ; 0; 0) 1 1 (0; 0; 0; 2 ; 2 ; 0) arXiv:1910.02233v2 [math.CO] 30 Jun 2020 Figure 1. The four-dimensional polytope P3 given by the six pat- terns of size three (see Eq. (2) for a precise definition). We highlight in light-blue one of the six three-dimensional facets of P3. This facet is a pyramid with square base. The polytope itself is a four-dimensional pyramid, whose base is the highlighted facet. This paper has been prepared using ALCO author class on 1st July 2020. Keywords. Permutation patterns, cycle polytopes, overlap graphs. Acknowledgements. This work was completed with the support of the SNF grants number 200021- 172536 and 200020-172515. -
Mathematical Modelling and Applications of Particle Swarm Optimization
Master’s Thesis Mathematical Modelling and Simulation Thesis no: 2010:8 Mathematical Modelling and Applications of Particle Swarm Optimization by Satyobroto Talukder Submitted to the School of Engineering at Blekinge Institute of Technology In partial fulfillment of the requirements for the degree of Master of Science February 2011 Contact Information: Author: Satyobroto Talukder E-mail: [email protected] University advisor: Prof. Elisabeth Rakus-Andersson Department of Mathematics and Science, BTH E-mail: [email protected] Phone: +46455385408 Co-supervisor: Efraim Laksman, BTH E-mail: [email protected] Phone: +46455385684 School of Engineering Internet : www.bth.se/com Blekinge Institute of Technology Phone : +46 455 38 50 00 SE – 371 79 Karlskrona Fax : +46 455 38 50 57 Sweden ii ABSTRACT Optimization is a mathematical technique that concerns the finding of maxima or minima of functions in some feasible region. There is no business or industry which is not involved in solving optimization problems. A variety of optimization techniques compete for the best solution. Particle Swarm Optimization (PSO) is a relatively new, modern, and powerful method of optimization that has been empirically shown to perform well on many of these optimization problems. It is widely used to find the global optimum solution in a complex search space. This thesis aims at providing a review and discussion of the most established results on PSO algorithm as well as exposing the most active research topics that can give initiative for future work and help the practitioner improve better result with little effort. This paper introduces a theoretical idea and detailed explanation of the PSO algorithm, the advantages and disadvantages, the effects and judicious selection of the various parameters. -
Introduction to Integer Programming – Integer Programming Models
15.053/8 March 14, 2013 Introduction to Integer Programming – Integer programming models 1 Quotes of the Day “Somebody who thinks logically is a nice contrast to the real world.” -- The Law of Thumb “Take some more tea,” the March Hare said to Alice, very earnestly. “I’ve had nothing yet,” Alice replied in an offended tone, “so I can’t take more.” “You mean you can’t take less,” said the Hatter. “It’s very easy to take more than nothing.” -- Lewis Carroll in Alice in Wonderland 2 Combinatorial optimization problems INPUT: A description of the data for an instance of the problem FEASIBLE SOLUTIONS: there is a way of determining from the input whether a given solution x’ (assignment of values to decision variables) is feasible. Typically in combinatorial optimization problems there is a finite number of possible solutions. OBJECTIVE FUNCTION: For each feasible solution x’ there is an associated objective f(x’). Minimization problem. Find a feasible solution x* that minimizes f( ) among all feasible solutions. 3 Example 1: Traveling Salesman Problem INPUT: a set N of n points in the plane FEASIBLE SOLUTION: a tour that passes through each point exactly once. OBJECTIVE: minimize the length of the tour. 4 Example 2: Balanced Partition INPUT: A set of positive integers a1, …, an FEASIBLE SOLUTION: a partition of {1, 2, … n} into two disjoint sets S and T. – S ∩ T = ∅, S∪T = {1, … , n} OBJECTIVE : minimize | ∑i∈S ai - ∑i∈T ai | Example: 7, 10, 13, 17, 20, 22 These numbers sum to 89 The best split is {10, 13, 22} and {7, 17, 20}. -
Integer Optimization Methods for Machine Learning by Allison an Chang Sc.B
Integer Optimization Methods for Machine Learning by Allison An Chang Sc.B. Applied Mathematics, Brown University (2007) Submitted to the Sloan School of Management in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Operations Research at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 2012 c Massachusetts Institute of Technology 2012. All rights reserved. Author............................................. ................. Sloan School of Management May 18, 2012 Certified by......................................... ................. Dimitris Bertsimas Boeing Leaders for Global Operations Professor Co-Director, Operations Research Center Thesis Supervisor Certified by......................................... ................. Cynthia Rudin Assistant Professor of Statistics Thesis Supervisor Accepted by......................................... ................ Patrick Jaillet Dugald C. Jackson Professor Department of Electrical Engineering and Computer Science Co-Director, Operations Research Center 2 Integer Optimization Methods for Machine Learning by Allison An Chang Submitted to the Sloan School of Management on May 18, 2012, in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Operations Research Abstract In this thesis, we propose new mixed integer optimization (MIO) methods to ad- dress problems in machine learning. The first part develops methods for supervised bipartite ranking, which arises in prioritization tasks in diverse domains such as infor- mation retrieval, recommender systems, natural language processing, bioinformatics, and preventative maintenance. The primary advantage of using MIO for ranking is that it allows for direct optimization of ranking quality measures, as opposed to current state-of-the-art algorithms that use heuristic loss functions. We demonstrate using a number of datasets that our approach can outperform other ranking methods. The second part of the thesis focuses on reverse-engineering ranking models.