Combinatorial Optimization Top Ten List Discrete Mathematics 2000 As Selected by William R

Total Page:16

File Type:pdf, Size:1020Kb

Combinatorial Optimization Top Ten List Discrete Mathematics 2000 As Selected by William R Combinatorial Optimization Top Ten List Discrete Mathematics 2000 as selected by William R. Pulleyblank T. J. Watson Research Center IBM Corporation Yorktown Heights, NY Euler's Theorem 1736 Theorem: A graph has a Euler tour if and only if there are zero or two nodes of odd degree L. Euler, Solutio Problematis ad Geometriam Situs Pertinentis, Commentarii Academiae Scientiarum Imperalis Petropolitanae 8 (1736) 128-140. " As far as the problem of the seven bridges of Koenigsberg is concerned, it can be solved by making an exhaustive list of all possible routes, and then finding whether or not any route satisfies the conditions of the problem. Because of the number of possibilities, this method of solution would be too difficult and laborious, and in other problems with more bridges, it would be impossible..." " So whatever arrangement may be proposed, one can easily determine whether or not a journey can be made, crossing each bridge once, by the following rules: If there are more than two areas to which an odd number of bridges lead, then such a journey is impossible. If, however, the number of bridges is odd for exactly two areas, then the journey is possible if it starts in either of these areas. If, finally, there are no areas to which an odd number of bridges leads, then the required journey can be accomplished starting from any area. " Algorithm first published 137 years later C. Hierholzer, Ueber die Moeglichkeit, einen Linienzug ohne Wiederholung und ohne Unterbrechnung zu umfahren, Mathematische Annalen 6 (1873), 30-32. - start at an odd degree node; - walk until you are stuck * what's left has no odd degree nodes - solve the remaining problem and splice in the answer. see N.L. Biggs, E.K.Lloyd, R.J.Wilson, Graph Theory 1736-1936, Oxford University Press, 1976. Chinese Postman problem Guan Meigu, Graphic Programming using odd and even points, Chinese Math. 1 (1962) 273-277 The streets were given lengths, and any (even) number of odd degree intersections was permitted. The problem now became: Find the shortest tour traversing each street at least once. Solved polynomially in J. Edmonds and E.L. Johnson, Matching, Euler Tours and the Chinese Postman, Mathematical Programming 5 (1973) 88-124. Many extensions - one in particular - P.D. Seymour, The matroids with the max-flow min-cut property, J. Combin. Theory Ser. B 23 (1977) 189-222. Herbert J. Ryser, Combinatorial Mathematics, The Mathematical Association of America, 1963. " Combinatorial mathematics cuts across the many subdivisions of mathematics, and this makes a formal definition difficult. ... Two general types of problems appear throughout the literature. In the first the existence of the prescribed configuration is in doubt, and the study attempts to settle this issue. These we call existence problems. In the second the existence of the configuration is known, and the study attempts to determine the number of configurations or the classification of these configurations according to types. These we call enumeration problems. " March 1971 -- Second Louisiana Conference on Combinatorics, Graph Theory and Computing, LSU, Baton Rouge, Ryser gave an invited address and included a third type of problem: How do you efficiently find or construct the configurations? Max-flow Min-cut Theorem 1956 Given a directed graph G = (V, E),with finite arc capacities ( uj : j? E), and designated source and sink nodes s and z, the maximum value of an s - z flow is the minimum capacity of an s - z cut. L.R. Ford Jr. and D.R. Fulkerson, Maximal flow through a network, Canadian J. of Math. 8 (1956) 399-404. A. Kotzig, "Suvislost' a Pravidelina Suvislost Konecnych Grafor", Bratislava: Vysoka Skola Economicka (1956). P. Elias, A. Feinstein and C.E. Shannon, A note on the maximum flow through a network, IRE Transactions on Information Theory IT2 (1956) 117-119. Equivalent Theorems: Menger's Theorem (1927) (edge or node disjoint paths) Dilworth's Theorem (1950) (antichains in partially ordered sets) P. Hall's Theorem (1935) (Systems of Discrete Representatives) Koenig's Theorem (1931) (maximum matching and minimum cover in a bipartite graph) Hoffman's Circulation Theorem (1960) Gale's Supply Demand Flow (1957) ... These were all linked by total unimodularity - Hoffman and Kruskal (1956) Write down a "natural" linear programming formulation, then there will be an integral optimal solution ... and some extensions ... Flow augmenting paths Dinit (1970) and Edmonds-Karp (1972) - shortest flow augmenting paths give polynomial performance Preflow-push algorithms Generalized flows Multicommodity flows ... Seymour's Theorem (1980): [A good characterization of total unimodularity] - a matrix is totally unimodular if and only if it can be composed from network matrices, their transposes, and two exceptions. P.D. Seymour, Decomposition of regular matroids, J. of Combin. Theory Ser. B 28 (1980) 305-359. Nonbipartite matching algorithm and polyhedral characterization 1965. Jack Edmonds, Paths, trees and flowers, Canadian J. of Maths. 17 (1965) 449-467. There exists a polynomially bounded algorithm which finds a maximum matching in a nonbipartite graph. Generalized the Hungarian method of Kuhn 1955 and Munkres 1957 for bipartite graphs Alternating paths not enough You shrink the odd sets! Paths, trees and flowers (Edmonds 1965) - 2. Digression ... I am claiming, as a mathematical result, the existence of a good algorithm for finding a maximum cardinality matching in a graph. ... an algorithm whose difficulty grows only algebraically with the size of the graph... ...For practical purposes the difference between algebraic and exponential order is often more crucial than the difference between finite and non-finite. also, Cobham (1965) Foundation of Polyhedral Combinatorics The convex hull of the incidence vectors x of the matchings of a nonbipartite graph is defined by xj m 0 for all j ? E, ?@ 5 (xj : j (v)) [ 1 for all v ?V, ? 5 (xj : j E(S)) [ (|S| -1)/2 for all odd cardinality S <V. J. Edmonds, Maximum matching and a polyhedron with 0-1 vertices, J. Res. Nat. Bureau of Standards B 69 (1965) 125-130. Extensions: bidirected capacitated weighted matchings - includes network flows, edge covers and natchings parity constraints - includes T-joins, cuts and postman problems stable sets in claw free graphs path systems Matroid Intersection Theorem 1970 Matroids had been introduced by Hassler Whitney in 1935 as an abstraction of linear independence. A family F of (so called independent) subsets of a set E forms a matroid if 1. all subsets of an independent set are independent; 2. for any A < E, all maximal independent subsets of A have the same cardinality - denoted by r(A). (If 1. is satisfied, then F is called an independence system.) Examples - edge sets of forests in a graph, linearly independent sets of columns of a matrix, systems of distinct representatives of a set family, ... Matroid Optimization Problem: Let each element e c E have a weight we. Find an independent set I for which 5 (we: ec I) is maximized. Theorem (Rado 1957, Edmonds 1970): An independence system is a matroid if and only if the greedy algorithm solves the Matroid Optimization Problem, for any vector w of weights. Theorem (Edmonds 1970): The incidence vectors x of the independent sets of a matroid are the vertices of the polyhedron defined by xe m 0 for all ec E, 5(xe: e? A) [ r(A) for all A ` E. Matroid Intersection Theorem (Edmonds(1970): Let M1 = (E, F1) and M2 = (E, F2) be matroids on a set E. Then maximum |I| = minimum {r1(S)+r2(E\S)}. c 3 ` I F1 F2 S E Weighted Intersection Theorem (Edmonds 1970): Let M1 = (E, F1) and M2 = (E, F2) be matroids on a set E. Let each element e c E have weight we. Then x is the incidence vector of a maximum weight set independent in both matroids if and only if x is a basic optimal solution to the linear program maximize wx subject to xe m 0 for all ec E, 5(xe: e? A) [ r1(A) for all A ` E, 5(xe: e? A) [ r2(A) for all A ` E. Consequences: Koenig's Theorem for maximum matchings in bipartite graphs; Edmonds' Matroid Partition Theorem Perfect's Theorem on Systems of Common Representatives Extensions: Polymatroids and polymatroid intersection Submodular and supermodular functions Unification with nonbipartite matching: Matroid parity problem, or matroid matching - cardinality case for linear matroids (Lovasz 1980) Cook's Theorem 1971 If there exists a polynomially bounded algorithm for satisfiability, then if a class of decision problems has a polynomial length certificate for a "YES" answer, then there is a polynomial time algorithm to solve any decision problem in the class. S.A. Cook, The complexity of theorem-proving procedures, Proc. 3rd Ann. ACM Symp. on Theory of Computing, Association for Computing Machinery, New York (1971). NP - problems having short "YES" certificates - polynomially solvable on nondeterministic Turing machines P - problems solvable in polynomial time Cook's Theorem: SATISFIABILITY is NP-complete - a polynomial algorithm for SATISFIABILITY implies P = NP is P = NP ? R. M. Karp 1972 Reducibility among combinatorial problems in R.E. Miller and J.W. Thatcher (eds.), Complexity of Computer Computations, Plenum Press, New York, 85-103. Showed that many well known problems were NP-complete - Hamiltonian circuit; large independent set of nodes; scheduling problems; 3-d matching ... M.R. Garey and D. S. Johnson, Computers and Intractability: A Guide to the Theory of NP-Completeness, Freeman, San Francisco, 1979 * More than 300 NP-complete problems Conventional Wisdom: P CNP. Garey and Johnson - 12 open problems: 1. Graph Isomorphism open 2. Subgraph homeomorphism for a fixed graph P 3. Graph genus NP-complete 4. Chordal graph completion NP-complete 5. Chromatic Index NP-complete 6. Spanning tree parity P 7.
Recommended publications
  • A Lagrangian Decomposition Approach Combined with Metaheuristics for the Knapsack Constrained Maximum Spanning Tree Problem
    MASTERARBEIT A Lagrangian Decomposition Approach Combined with Metaheuristics for the Knapsack Constrained Maximum Spanning Tree Problem ausgeführt am Institut für Computergraphik und Algorithmen der Technischen Universität Wien unter der Anleitung von Univ.-Prof. Dipl.-Ing. Dr.techn. Günther Raidl und Univ.-Ass. Dipl.-Ing. Dr.techn. Jakob Puchinger durch Sandro Pirkwieser, Bakk.techn. Matrikelnummer 0116200 Simmeringer Hauptstraße 50/30, A-1110 Wien Datum Unterschrift Abstract This master’s thesis deals with solving the Knapsack Constrained Maximum Spanning Tree (KCMST) problem, which is a so far less addressed NP-hard combinatorial optimization problem belonging to the area of network design. Thereby sought is a spanning tree whose profit is maximal, but at the same time its total weight must not exceed a specified value. For this purpose a Lagrangian decomposition approach, which is a special variant of La- grangian relaxation, is applied to derive upper bounds. In the course of the application the problem is split up in two subproblems, which are likewise to be maximized but easier to solve on its own. The subgradient optimization method as well as the Volume Algorithm are deployed to solve the Lagrangian dual problem. To derive according lower bounds, i.e. feasible solutions, a simple Lagrangian heuristic is applied which is strengthened by a problem specific local search. Furthermore an Evolutionary Algorithm is presented which uses a suitable encoding for the solutions and appropriate operators, whereas the latter are able to utilize heuristics based on defined edge-profits. It is shown that simple edge-profits, derived straightforward from the data given by an instance, are of no benefit.
    [Show full text]
  • Chapter 8 Stochastic Gradient / Subgradient Methods
    Chapter 8 Stochastic gradient / subgradient methods Contents (class version) 8.0 Introduction........................................ 8.2 8.1 The subgradient method................................. 8.5 Subgradients and subdifferentials................................. 8.5 Properties of subdifferentials.................................... 8.7 Convergence of the subgradient method.............................. 8.10 8.2 Example: Hinge loss with 1-norm regularizer for binary classifier design...... 8.17 8.3 Incremental (sub)gradient method............................ 8.19 Incremental (sub)gradient method................................. 8.21 8.4 Stochastic gradient (SG) method............................. 8.23 SG update.............................................. 8.23 Stochastic gradient algorithm: convergence analysis....................... 8.26 Variance reduction: overview................................... 8.33 Momentum............................................. 8.35 Adaptive step-sizes......................................... 8.37 8.5 Example: X-ray CT reconstruction........................... 8.41 8.1 © J. Fessler, April 12, 2020, 17:55 (class version) 8.2 8.6 Summary.......................................... 8.50 8.0 Introduction This chapter describes two families of algorithms: • subgradient methods • stochastic gradient methods aka stochastic gradient descent methods Often we turn to these methods as a “last resort,” for applications where none of the methods discussed previously are suitable. Many machine learning applications,
    [Show full text]
  • Parity Systems and the Delta-Matroid Intersection Problem
    Parity Systems and the Delta-Matroid Intersection Problem Andr´eBouchet ∗ and Bill Jackson † Submitted: February 16, 1998; Accepted: September 3, 1999. Abstract We consider the problem of determining when two delta-matroids on the same ground-set have a common base. Our approach is to adapt the theory of matchings in 2-polymatroids developed by Lov´asz to a new abstract system, which we call a parity system. Examples of parity systems may be obtained by combining either, two delta- matroids, or two orthogonal 2-polymatroids, on the same ground-sets. We show that many of the results of Lov´aszconcerning ‘double flowers’ and ‘projections’ carry over to parity systems. 1 Introduction: the delta-matroid intersec- tion problem A delta-matroid is a pair (V, ) with a finite set V and a nonempty collection of subsets of V , called theBfeasible sets or bases, satisfying the following axiom:B ∗D´epartement d’informatique, Universit´edu Maine, 72017 Le Mans Cedex, France. [email protected] †Department of Mathematical and Computing Sciences, Goldsmiths’ College, London SE14 6NW, England. [email protected] 1 the electronic journal of combinatorics 7 (2000), #R14 2 1.1 For B1 and B2 in and v1 in B1∆B2, there is v2 in B1∆B2 such that B B1∆ v1, v2 belongs to . { } B Here P ∆Q = (P Q) (Q P ) is the symmetric difference of two subsets P and Q of V . If X\ is a∪ subset\ of V and if we set ∆X = B∆X : B , then we note that (V, ∆X) is a new delta-matroid.B The{ transformation∈ B} (V, ) (V, ∆X) is calledB a twisting.
    [Show full text]
  • Oasics-SOSA-2019-14.Pdf (0.4
    Simple Greedy 2-Approximation Algorithm for the Maximum Genus of a Graph Michal Kotrbčík Department of Computer Science, Comenius University, 842 48 Bratislava, Slovakia [email protected] Martin Škoviera1 Department of Computer Science, Comenius University, 842 48 Bratislava, Slovakia [email protected] Abstract The maximum genus γM (G) of a graph G is the largest genus of an orientable surface into which G has a cellular embedding. Combinatorially, it coincides with the maximum number of disjoint pairs of adjacent edges of G whose removal results in a connected spanning subgraph of G. In this paper we describe a greedy 2-approximation algorithm for maximum genus by proving that removing pairs of adjacent edges from G arbitrarily while retaining connectedness leads to at least γM (G)/2 pairs of edges removed. As a consequence of our approach we also obtain a 2-approximate counterpart of Xuong’s combinatorial characterisation of maximum genus. 2012 ACM Subject Classification Theory of computation → Design and analysis of algorithms → Graph algorithms analysis, Mathematics of computing → Graph algorithms, Mathematics of computing → Graphs and surfaces Keywords and phrases maximum genus, embedding, graph, greedy algorithm Digital Object Identifier 10.4230/OASIcs.SOSA.2019.14 Acknowledgements The authors would like to thank Rastislav Královič and Jana Višňovská for reading preliminary versions of this paper and making useful suggestions. 1 Introduction One of the paradigms in topological graph theory is the study of all surface embeddings of a given graph. The maximum genus γM (G) parameter of a graph G is then the maximum integer g such that G has a cellular embedding in the orientable surface of genus g.A result of Duke [12] implies that a graph G has a cellular embedding in the orientable surface of genus g if and only if γ(G) ≤ g ≤ γM (G) where γ(G) denotes the (minimum) genus of G.
    [Show full text]
  • A Tight Extremal Bound on the Lovász Cactus Number in Planar Graphs
    A Tight Extremal Bound on the Lovász Cactus Number in Planar Graphs Parinya Chalermsook Aalto University, Espoo, Finland parinya.chalermsook@aalto.fi Andreas Schmid Max Planck Institute for Informatics, Saarbrücken, Germany [email protected] Sumedha Uniyal Aalto University, Espoo, Finland sumedha.uniyal@aalto.fi Abstract A cactus graph is a graph in which any two cycles are edge-disjoint. We present a constructive proof of the fact that any plane graph G contains a cactus subgraph C where C contains at least 1 a 6 fraction of the triangular faces of G. We also show that this ratio cannot be improved by showing a tight lower bound. Together with an algorithm for linear matroid parity, our bound implies two approximation algorithms for computing “dense planar structures” inside any graph: (i) 1 A 6 approximation algorithm for, given any graph G, finding a planar subgraph with a maximum 1 number of triangular faces; this improves upon the previous 11 -approximation; (ii) An alternate (and 4 arguably more illustrative) proof of the 9 approximation algorithm for finding a planar subgraph with a maximum number of edges. Our bound is obtained by analyzing a natural local search strategy and heavily exploiting the exchange arguments. Therefore, this suggests the power of local search in handling problems of this kind. 2012 ACM Subject Classification Mathematics of computing → Graph theory Keywords and phrases Graph Drawing, Matroid Matching, Maximum Planar Subgraph, Local Search Algorithms Digital Object Identifier 10.4230/LIPIcs.STACS.2019.19 Related Version Full Version: https://arxiv.org/abs/1804.03485. Funding Parinya Chalermsook: Part of this work was done while PC and AS were visiting the Simons Institute for the Theory of Computing.
    [Show full text]
  • Subgradient Method
    Subgradient Method Ryan Tibshirani Convex Optimization 10-725/36-725 Last last time: gradient descent Consider the problem min f(x) x n for f convex and differentiable, dom(f) = R . Gradient descent: (0) n choose initial x 2 R , repeat (k) (k−1) (k−1) x = x − tk · rf(x ); k = 1; 2; 3;::: Step sizes tk chosen to be fixed and small, or by backtracking line search If rf Lipschitz, gradient descent has convergence rate O(1/) Downsides: • Requires f differentiable this lecture • Can be slow to converge next lecture 2 Subgradient method n Now consider f convex, with dom(f) = R , but not necessarily differentiable Subgradient method: like gradient descent, but replacing gradients with subgradients. I.e., initialize x(0), repeat (k) (k−1) (k−1) x = x − tk · g ; k = 1; 2; 3;::: where g(k−1) 2 @f(x(k−1)), any subgradient of f at x(k−1) Subgradient method is not necessarily a descent method, so we (k) (0) (k) keep track of best iterate xbest among x ; : : : x so far, i.e., f(x(k) ) = min f(x(i)) best i=0;:::k 3 Outline Today: • How to choose step sizes • Convergence analysis • Intersection of sets • Stochastic subgradient method 4 Step size choices • Fixed step sizes: tk = t all k = 1; 2; 3;::: • Diminishing step sizes: choose to meet conditions 1 1 X 2 X tk < 1; tk = 1; k=1 k=1 i.e., square summable but not summable Important that step sizes go to zero, but not too fast Other options too, but important difference to gradient descent: all step sizes options are pre-specified, not adaptively computed 5 Convergence analysis n Assume that f convex, dom(f) = R , and also that f is Lipschitz continuous with constant G > 0, i.e., jf(x) − f(y)j ≤ Gkx − yk2 for all x; y Theorem: For a fixed step size t, subgradient method satisfies lim f(x(k) ) ≤ f ? + G2t=2 k!1 best Theorem: For diminishing step sizes, subgradient method sat- isfies lim f(x(k) ) = f ? k!1 best 6 Basic bound, and convergence rate (0) ? Letting R = kx − x k2, after k steps, we have the basic bound R2 + G2 Pk t2 f(x(k) ) − f(x?) ≤ i=1 i best Pk 2 i=1 ti Previous theorems follow from this.
    [Show full text]
  • An Efficient Solution Methodology for Mixed-Integer Programming Problems Arising in Power Systems" (2016)
    University of Connecticut OpenCommons@UConn Doctoral Dissertations University of Connecticut Graduate School 12-15-2016 An Efficient Solution Methodology for Mixed- Integer Programming Problems Arising in Power Systems Mikhail Bragin University of Connecticut - Storrs, [email protected] Follow this and additional works at: https://opencommons.uconn.edu/dissertations Recommended Citation Bragin, Mikhail, "An Efficient Solution Methodology for Mixed-Integer Programming Problems Arising in Power Systems" (2016). Doctoral Dissertations. 1318. https://opencommons.uconn.edu/dissertations/1318 An Efficient Solution Methodology for Mixed-Integer Programming Problems Arising in Power Systems Mikhail Bragin, PhD University of Connecticut, 2016 For many important mixed-integer programming (MIP) problems, the goal is to obtain near- optimal solutions with quantifiable quality in a computationally efficient manner (within, e.g., 5, 10 or 20 minutes). A traditional method to solve such problems has been Lagrangian relaxation, but the method suffers from zigzagging of multipliers and slow convergence. When solving mixed-integer linear programming (MILP) problems, the recently adopted branch-and-cut may also suffer from slow convergence because when the convex hull of the problems has complicated facial structures, facet- defining cuts are typically difficult to obtain, and the method relies mostly on time-consuming branching operations. In this thesis, the novel Surrogate Lagrangian Relaxation method is developed and its convergence is proved to the optimal multipliers, without the knowledge of the optimal dual value and without fully optimizing the relaxed problem. Moreover, for practical implementations a stepsizing formula, that guarantees convergence without requiring the optimal dual value, has been constructively developed. The key idea is to select stepsizes in a way that distances between Lagrange multipliers at consecutive iterations decrease, and as a result, Lagrange multipliers converge to a unique limit.
    [Show full text]
  • Subgradient Methods
    Subgradient Methods Stephen Boyd and Almir Mutapcic Notes for EE364b, Stanford University, Winter 2006-07 April 13, 2008 Contents 1 Introduction 2 2 Basic subgradient method 2 2.1 Negativesubgradientupdate. .... 2 2.2 Stepsizerules................................... 3 2.3 Convergenceresults.............................. .. 4 3 Convergence proof 4 3.1 Assumptions.................................... 4 3.2 Somebasicinequalities . ... 5 3.3 Aboundonthesuboptimalitybound . ... 7 3.4 Astoppingcriterion.............................. .. 8 3.5 Numericalexample ................................ 8 4 Alternating projections 9 4.1 Optimal step size choice when f ⋆ isknown................... 9 4.2 Finding a point in the intersection of convex sets . ......... 11 4.3 Solving convex inequalities . ..... 14 4.4 Positive semidefinite matrix completion . ........ 15 5 Projected subgradient method 16 5.1 Numericalexample ................................ 18 6 Projected subgradient for dual problem 18 6.1 Numericalexample ................................ 20 7 Subgradient method for constrained optimization 21 7.1 Numericalexample ................................ 24 8 Speeding up subgradient methods 24 1 1 Introduction The subgradient method is a very simple algorithm for minimizing a nondifferentiable convex function. The method looks very much like the ordinary gradient method for differentiable functions, but with several notable exceptions: The subgradient method applies directly to nondifferentiable f. • The step lengths are not chosen via a line search, as in the ordinary gradient method. • In the most common cases, the step lengths are fixed ahead of time. Unlike the ordinary gradient method, the subgradient method is not a descent method; • the function value can (and often does) increase. The subgradient method is readily extended to handle problems with constraints. Subgradient methods can be much slower than interior-point methods (or Newton’s method in the unconstrained case).
    [Show full text]
  • 1 Base Polytopes
    6.883 Learning with Combinatorial Structure Note for Lecture 8 Authors: Hongyi Zhang A bit of history in combinatorial optimization. Last time we talked about the Lovász extension, which plays an important role in the optimization of submodular functions. Actually, despite the name Lovász, Jack Edmonds is another important figure who made huge contribution to this concept, and submodular optimization in general. In DISCML 20111, there was a brief introduction to Jack Edmonds’ contributions to mathematics and computer science by Jeff Bilmes. 1 Base polytopes 1.1 Examples of base polytopes Recall that in the last lecture we defined the base polytope BF of a submodular function, and showed that there is a greedy algorithm to solve the optimization problem max y>x y2BF and the solution is closely related to Lovász extension. Now we shall look at several other examples where interesting concepts turn out to be the base polytopes of some submodu- lar (or supermodular) functions. 1.1.1 Probability simplex Let the ground set be V such that jVj = n. Define F (S) = minfjSj; 1g;S ⊆ V, in last lecture we already saw that F (S) is submodular. It is easy to check that BF is a probability sim- plex. In fact, by definition ( ) X X BF = y : yi = F (V) and yi ≤ F (S) 8S ⊆ V i2V i2S ( n ) X = y : yi = 1 and yi ≥ 0; 8i i=1 n which is a standard (n−1)-simplex in R . 1http://las.ethz.ch/discml/discml11.html NIPS Workshop on Discrete Optimization in Ma- chine Learning 2011 1 1.1.2 Permutahedron PjSj Let the ground set be V such that jVj = n.
    [Show full text]
  • Algebraic Algorithms in Combinatorial Optimization
    Algebraic Algorithms in Combinatorial Optimization CHEUNG, Ho Yee A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree of Master of Philosophy in Computer Science and Engineering The Chinese University of Hong Kong September 2011 Thesis/Assessment Committee Professor Shengyu Zhang (Chair) Professor Lap Chi Lau (Thesis Supervisor) Professor Andrej Bogdanov (Committee Member) Professor Satoru Iwata (External Examiner) Abstract In this thesis we extend the recent algebraic approach to design fast algorithms for two problems in combinatorial optimization. First we study the linear matroid parity problem, a common generalization of graph matching and linear matroid intersection, that has applications in various areas. We show that Harvey's algo- rithm for linear matroid intersection can be easily generalized to linear matroid parity. This gives an algorithm that is faster and simpler than previous known al- gorithms. For some graph problems that can be reduced to linear matroid parity, we again show that Harvey's algorithm for graph matching can be generalized to these problems to give faster algorithms. While linear matroid parity and some of its applications are challenging generalizations, our results show that the al- gebraic algorithmic framework can be adapted nicely to give faster and simpler algorithms in more general settings. Then we study the all pairs edge connectivity problem for directed graphs, where we would like to compute minimum s-t cut value between all pairs of vertices. Using a combinatorial approach it is not known how to solve this problem faster than computing the minimum s-t cut value for each pair of vertices separately.
    [Show full text]
  • Matroid Matching: the Power of Local Search
    Matroid Matching: the Power of Local Search [Extended Abstract] Jon Lee Maxim Sviridenko Jan Vondrák IBM Research IBM Research IBM Research Yorktown Heights, NY Yorktown Heights, NY San Jose, CA [email protected] [email protected] [email protected] ABSTRACT can be easily transformed into an NP-completeness proof We consider the classical matroid matching problem. Un- for a concrete class of matroids (see [46]). An important weighted matroid matching for linear matroids was solved result of Lov´asz is that (unweighted) matroid matching can by Lov´asz, and the problem is known to be intractable for be solved in polynomial time for linear matroids (see [35]). general matroids. We present a PTAS for unweighted ma- There have been several attempts to generalize Lov´asz' re- troid matching for general matroids. In contrast, we show sult to the weighted case. Polynomial-time algorithms are that natural LP relaxations have an Ω(n) integrality gap known for some special cases (see [49]), but for general linear and moreover, Ω(n) rounds of the Sherali-Adams hierarchy matroids there is only a pseudopolynomial-time randomized are necessary to bring the gap down to a constant. exact algorithm (see [8, 40]). More generally, for any fixed k ≥ 2 and > 0, we obtain a In this paper, we revisit the matroid matching problem (k=2 + )-approximation for matroid matching in k-uniform for general matroids. Our main result is that while LP- hypergraphs, also known as the matroid k-parity problem. based approaches including the Sherali-Adams hierarchy fail As a consequence, we obtain a (k=2 + )-approximation for to provide any meaningful approximation, a simple local- the problem of finding the maximum-cardinality set in the search algorithm gives a PTAS (in the unweighted case).
    [Show full text]
  • On the Links Between Probabilistic Graphical Models and Submodular Optimisation Senanayak Sesh Kumar Karri
    On the Links between Probabilistic Graphical Models and Submodular Optimisation Senanayak Sesh Kumar Karri To cite this version: Senanayak Sesh Kumar Karri. On the Links between Probabilistic Graphical Models and Submodular Optimisation. Machine Learning [cs.LG]. Université Paris sciences et lettres, 2016. English. NNT : 2016PSLEE047. tel-01753810 HAL Id: tel-01753810 https://tel.archives-ouvertes.fr/tel-01753810 Submitted on 29 Mar 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THESE` DE DOCTORAT de l’Universite´ de recherche Paris Sciences Lettres PSL Research University Prepar´ ee´ a` l’Ecole´ normale superieure´ On the Links between Probabilistic Graphical Models and Submodular Optimisation Liens entre modeles` graphiques probabilistes et optimisation sous-modulaire Ecole´ doctorale n◦386 ECOLE´ DOCTORALE DE SCIENCES MATHEMATIQUES´ DE PARIS CENTRE Specialit´ e´ INFORMATIQUE COMPOSITION DU JURY : M Andreas Krause ETH Zurich, Rapporteur M Nikos Komodakis ENPC Paris, Rapporteur M Francis Bach Inria Paris, Directeur de these` Soutenue par Senanayak Sesh Kumar KARRI le 27.09.2016 M Josef Sivic ENS Paris, Membre du Jury Dirigee´ par Francis BACH M Antonin Chambolle CMAP EP Paris, Membre du Jury M Guillaume Obozinski ENPC Paris, Membre du Jury RESEARCH UNIVERSITY PARIS ÉCOLENORMALE SUPÉRIEURE What is the purpose of life? Proof and Conjecture, ....
    [Show full text]