IEOR 269, Spring 2010 Integer Programming and Combinatorial Optimization

Total Page:16

File Type:pdf, Size:1020Kb

IEOR 269, Spring 2010 Integer Programming and Combinatorial Optimization IEOR 269, Spring 2010 Integer Programming and Combinatorial Optimization Professor Dorit S. Hochbaum Contents 1 Introduction 1 2 Formulation of some ILP 2 2.1 0-1 knapsack problem . 2 2.2 Assignment problem . 2 3 Non-linear Objective functions 4 3.1 Production problem with set-up costs . 4 3.2 Piecewise linear cost function . 5 3.3 Piecewise linear convex cost function . 6 3.4 Disjunctive constraints . 7 4 Some famous combinatorial problems 7 4.1 Max clique problem . 7 4.2 SAT (satisfiability) . 7 4.3 Vertex cover problem . 7 5 General optimization 8 6 Neighborhood 8 6.1 Exact neighborhood . 8 7 Complexity of algorithms 9 7.1 Finding the maximum element . 9 7.2 0-1 knapsack . 9 7.3 Linear systems . 10 7.4 Linear Programming . 11 8 Some interesting IP formulations 12 8.1 The fixed cost plant location problem . 12 8.2 Minimum/maximum spanning tree (MST) . 12 9 The Minimum Spanning Tree (MST) Problem 13 i IEOR269 notes, Prof. Hochbaum, 2010 ii 10 General Matching Problem 14 10.1 Maximum Matching Problem in Bipartite Graphs . 14 10.2 Maximum Matching Problem in Non-Bipartite Graphs . 15 10.3 Constraint Matrix Analysis for Matching Problems . 16 11 Traveling Salesperson Problem (TSP) 17 11.1 IP Formulation for TSP . 17 12 Discussion of LP-Formulation for MST 18 13 Branch-and-Bound 20 13.1 The Branch-and-Bound technique . 20 13.2 Other Branch-and-Bound techniques . 22 14 Basic graph definitions 23 15 Complexity analysis 24 15.1 Measuring quality of an algorithm . 24 15.1.1 Examples . 24 15.2 Growth of functions . 26 15.3 Definitions for asymptotic comparisons of functions . 26 15.4 Properties of asymptotic notation . 26 15.5 Caveats of complexity analysis . 27 16 Complexity classes and NP-completeness 28 16.1 Search vs. Decision . 28 16.2 The class NP ........................................ 29 17 Addendum on Branch and Bound 30 18 Complexity classes and NP-completeness 31 18.1 Search vs. Decision . 31 18.2 The class NP ........................................ 32 18.2.1 Some Problems in NP ............................... 33 18.3 The class co-NP ...................................... 34 18.3.1 Some Problems in co-NP ............................. 34 18.4 NPand co-NP ........................................ 34 18.5 NP-completeness and reductions . 35 18.5.1 Reducibility . 35 18.5.2 NP-Completeness . 36 19 The Chinese Checkerboard Problem and a First Look at Cutting Planes 40 19.1 Problem Setup . 40 19.2 The First Integer Programming Formulation . 40 19.3 An Improved ILP Formulation . 40 20 Cutting Planes 43 20.1 Chinese checkers . 43 20.2 Branch and Cut . 44 IEOR269 notes, Prof. Hochbaum, 2010 iii 21 The geometry of Integer Programs and Linear Programs 45 21.1 Cutting planes for the Knapsack problem . 45 21.2 Cutting plane approach for the TSP . 46 22 Gomory cuts 46 23 Generating a Feasible Solution for TSP 47 24 Diophantine Equations 48 24.1 Hermite Normal Form . 49 25 Optimization with linear set of equality constraints 51 26 Balas's additive algorithm 51 27 Held and Karp's algorithm for TSP 54 28 Lagrangian Relaxation 55 29 Lagrangian Relaxation 59 30 Complexity of Nonlinear Optimization 61 30.1 Input for a Polynomial Function . 61 30.2 Table Look-Up . 61 30.3 Examples of Non-linear Optimization Problems . 62 30.4 Impossibility of strongly polynomial algorithms for nonlinear (non-quadratic) opti- mization . 64 31 The General Network Flow Problem Setup 65 32 Shortest Path Problem 68 33 Maximum Flow Problem 69 33.1 Setup . 69 33.2 Algorithms . 71 33.2.1 Ford-Fulkerson algorithm . 72 33.2.2 Capacity scaling algorithm . 74 33.3 Maximum-flow versus Minimum Cost Network Flow . 75 33.4 Formulation . 75 34 Minimum Cut Problem 77 34.1 Minimum s-t Cut Problem . 77 34.2 Formulation . 78 35 Selection Problem 79 36 A Production/Distribution Network: MCNF Formulation 81 36.1 Problem Description . 81 36.2 Formulation as a Minimum Cost Network Flow Problem . 83 36.3 The Optimal Solution to the Chairs Problem . 84 IEOR269 notes, Prof. Hochbaum, 2010 iv 37 Transhipment Problem 87 38 Transportation Problem 87 38.1 Production/Inventory Problem as Transportation Problem . 88 39 Assignment Problem 90 40 Maximum Flow Problem 90 40.1 A Package Delivery Problem . 90 41 Shortest Path Problem 91 41.1 An Equipment Replacement Problem . 92 42 Maximum Weight Matching 93 42.1 An Agent Scheduling Problem with Reassignment . 93 43 MCNF Hierarchy 96 44 The maximum/minimum closure problem 96 44.1 A practical example: open-pit mining . 96 44.2 The maximum closure problem . 98 45 Integer programs with two variables per inequality 101 45.1 Monotone IP2 . 101 45.2 Non{monotone IP2 . 103 46 Vertex cover problem 104 46.1 Vertex cover on bipartite graphs . 105 46.2 Vertex cover on general graphs . 105 47 The convex cost closure problem 107 47.1 The threshold theorem . 108 47.2 Naive algorithm for solving (ccc) . 111 47.3 Solving (ccc) in polynomial time using binary search . 111 47.4 Solving (ccc) using parametric minimum cut . 111 48 The s-excess problem 113 48.1 The convex s-excess problem . 114 48.2 Threshold theorem for linear edge weights . 114 48.3 Variants / special cases . 115 49 Forest Clearing 116 50 Producing memory chips (VLSI layout) 117 51 Independent set problem 117 51.1 Independent Set v.s. Vertex Cover . 118 51.2 Independent set on bipartite graphs . 118 IEOR269 notes, Prof. Hochbaum, 2010 v 52 Maximum Density Subgraph 119 52.1 Linearizing ratio problems . 119 52.2 Solving the maximum density subgraph problem . 119 53 Parametric cut/flow problem 120 53.1 The parametric cut/flow problem with convex function (tentative title) . 121 54 Average k-cut problem 122 55 Image segmentation problem 122 55.1 Solving the normalized cut variant problem . 123 55.2 Solving the λ-question with a minimum cut procedure . 125 56 Duality of Max-Flow and MCNF Problems 127 56.1 Duality of Max-Flow Problem: Minimum Cut . 127 56.2 Duality of MCNF problem . 127 57 Variant of Normalized Cut Problem 128 57.1 Problem Setup . 129 57.2 Review of Maximum Closure Problem . 130 57.3 Review of Maximum s-Excess Problem . 131 57.4 Relationship between s-excess and maximum closure problems . 132 57.5 Solution Approach for Variant of the Normalized Cut problem . 132 58 Markov Random Fields 133 59 Examples of 2 vs. 3 in Combinatorial Optimization 134 59.1 Edge Packing vs. Vertex Packing . 134 59.2 Chinese Postman Problem vs. Traveling Salesman Problem . 135 59.3 2SAT vs. 3SAT . ..
Recommended publications
  • 4. Convex Optimization Problems
    Convex Optimization — Boyd & Vandenberghe 4. Convex optimization problems optimization problem in standard form • convex optimization problems • quasiconvex optimization • linear optimization • quadratic optimization • geometric programming • generalized inequality constraints • semidefinite programming • vector optimization • 4–1 Optimization problem in standard form minimize f0(x) subject to f (x) 0, i =1,...,m i ≤ hi(x)=0, i =1,...,p x Rn is the optimization variable • ∈ f : Rn R is the objective or cost function • 0 → f : Rn R, i =1,...,m, are the inequality constraint functions • i → h : Rn R are the equality constraint functions • i → optimal value: p⋆ = inf f (x) f (x) 0, i =1,...,m, h (x)=0, i =1,...,p { 0 | i ≤ i } p⋆ = if problem is infeasible (no x satisfies the constraints) • ∞ p⋆ = if problem is unbounded below • −∞ Convex optimization problems 4–2 Optimal and locally optimal points x is feasible if x dom f and it satisfies the constraints ∈ 0 ⋆ a feasible x is optimal if f0(x)= p ; Xopt is the set of optimal points x is locally optimal if there is an R> 0 such that x is optimal for minimize (over z) f0(z) subject to fi(z) 0, i =1,...,m, hi(z)=0, i =1,...,p z x≤ R k − k2 ≤ examples (with n =1, m = p =0) f (x)=1/x, dom f = R : p⋆ =0, no optimal point • 0 0 ++ f (x)= log x, dom f = R : p⋆ = • 0 − 0 ++ −∞ f (x)= x log x, dom f = R : p⋆ = 1/e, x =1/e is optimal • 0 0 ++ − f (x)= x3 3x, p⋆ = , local optimum at x =1 • 0 − −∞ Convex optimization problems 4–3 Implicit constraints the standard form optimization problem has an implicit
    [Show full text]
  • Metaheuristics1
    METAHEURISTICS1 Kenneth Sörensen University of Antwerp, Belgium Fred Glover University of Colorado and OptTek Systems, Inc., USA 1 Definition A metaheuristic is a high-level problem-independent algorithmic framework that provides a set of guidelines or strategies to develop heuristic optimization algorithms (Sörensen and Glover, To appear). Notable examples of metaheuristics include genetic/evolutionary algorithms, tabu search, simulated annealing, and ant colony optimization, although many more exist. A problem-specific implementation of a heuristic optimization algorithm according to the guidelines expressed in a metaheuristic framework is also referred to as a metaheuristic. The term was coined by Glover (1986) and combines the Greek prefix meta- (metá, beyond in the sense of high-level) with heuristic (from the Greek heuriskein or euriskein, to search). Metaheuristic algorithms, i.e., optimization methods designed according to the strategies laid out in a metaheuristic framework, are — as the name suggests — always heuristic in nature. This fact distinguishes them from exact methods, that do come with a proof that the optimal solution will be found in a finite (although often prohibitively large) amount of time. Metaheuristics are therefore developed specifically to find a solution that is “good enough” in a computing time that is “small enough”. As a result, they are not subject to combinatorial explosion – the phenomenon where the computing time required to find the optimal solution of NP- hard problems increases as an exponential function of the problem size. Metaheuristics have been demonstrated by the scientific community to be a viable, and often superior, alternative to more traditional (exact) methods of mixed- integer optimization such as branch and bound and dynamic programming.
    [Show full text]
  • A Branch-And-Price Approach with Milp Formulation to Modularity Density Maximization on Graphs
    A BRANCH-AND-PRICE APPROACH WITH MILP FORMULATION TO MODULARITY DENSITY MAXIMIZATION ON GRAPHS KEISUKE SATO Signalling and Transport Information Technology Division, Railway Technical Research Institute. 2-8-38 Hikari-cho, Kokubunji-shi, Tokyo 185-8540, Japan YOICHI IZUNAGA Information Systems Research Division, The Institute of Behavioral Sciences. 2-9 Ichigayahonmura-cho, Shinjyuku-ku, Tokyo 162-0845, Japan Abstract. For clustering of an undirected graph, this paper presents an exact algorithm for the maximization of modularity density, a more complicated criterion to overcome drawbacks of the well-known modularity. The problem can be interpreted as the set-partitioning problem, which reminds us of its integer linear programming (ILP) formulation. We provide a branch-and-price framework for solving this ILP, or column generation combined with branch-and-bound. Above all, we formulate the column gen- eration subproblem to be solved repeatedly as a simpler mixed integer linear programming (MILP) problem. Acceleration tech- niques called the set-packing relaxation and the multiple-cutting- planes-at-a-time combined with the MILP formulation enable us to optimize the modularity density for famous test instances in- cluding ones with over 100 vertices in around four minutes by a PC. Our solution method is deterministic and the computation time is not affected by any stochastic behavior. For one of them, column generation at the root node of the branch-and-bound tree arXiv:1705.02961v3 [cs.SI] 27 Jun 2017 provides a fractional upper bound solution and our algorithm finds an integral optimal solution after branching. E-mail addresses: (Keisuke Sato) [email protected], (Yoichi Izunaga) [email protected].
    [Show full text]
  • Independence Number, Vertex (Edge) ୋ Cover Number and the Least Eigenvalue of a Graph
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Linear Algebra and its Applications 433 (2010) 790–795 Contents lists available at ScienceDirect Linear Algebra and its Applications journal homepage: www.elsevier.com/locate/laa The vertex (edge) independence number, vertex (edge) ୋ cover number and the least eigenvalue of a graph ∗ Ying-Ying Tan a, Yi-Zheng Fan b, a Department of Mathematics and Physics, Anhui University of Architecture, Hefei 230601, PR China b School of Mathematical Sciences, Anhui University, Hefei 230039, PR China ARTICLE INFO ABSTRACT Article history: In this paper we characterize the unique graph whose least eigen- Received 4 October 2009 value attains the minimum among all graphs of a fixed order and Accepted 4 April 2010 a given vertex (edge) independence number or vertex (edge) cover Available online 8 May 2010 number, and get some bounds for the vertex (edge) independence Submitted by X. Zhan number, vertex (edge) cover number of a graph in terms of the least eigenvalue of the graph. © 2010 Elsevier Inc. All rights reserved. AMS classification: 05C50 15A18 Keywords: Graph Adjacency matrix Vertex (edge) independence number Vertex (edge) cover number Least eigenvalue 1. Introduction Let G = (V,E) be a simple graph of order n with vertex set V = V(G) ={v1,v2, ...,vn} and edge set E = E(G). The adjacency matrix of G is defined to be a (0, 1)-matrix A(G) =[aij], where aij = 1ifvi is adjacent to vj and aij = 0 otherwise. The eigenvalues of the graph G are referred to the eigenvalues of A(G), which are arranged as λ1(G) λ2(G) ··· λn(G).
    [Show full text]
  • Linear Programming Notes X: Integer Programming
    Linear Programming Notes X: Integer Programming 1 Introduction By now you are familiar with the standard linear programming problem. The assumption that choice variables are infinitely divisible (can be any real number) is unrealistic in many settings. When we asked how many chairs and tables should the profit-maximizing carpenter make, it did not make sense to come up with an answer like “three and one half chairs.” Maybe the carpenter is talented enough to make half a chair (using half the resources needed to make the entire chair), but probably she wouldn’t be able to sell half a chair for half the price of a whole chair. So, sometimes it makes sense to add to a problem the additional constraint that some (or all) of the variables must take on integer values. This leads to the basic formulation. Given c = (c1, . , cn), b = (b1, . , bm), A a matrix with m rows and n columns (and entry aij in row i and column j), and I a subset of {1, . , n}, find x = (x1, . , xn) max c · x subject to Ax ≤ b, x ≥ 0, xj is an integer whenever j ∈ I. (1) What is new? The set I and the constraint that xj is an integer when j ∈ I. Everything else is like a standard linear programming problem. I is the set of components of x that must take on integer values. If I is empty, then the integer programming problem is a linear programming problem. If I is not empty but does not include all of {1, . , n}, then sometimes the problem is called a mixed integer programming problem.
    [Show full text]
  • 3.1 Matchings and Factors: Matchings and Covers
    1 3.1 Matchings and Factors: Matchings and Covers This copyrighted material is taken from Introduction to Graph Theory, 2nd Ed., by Doug West; and is not for further distribution beyond this course. These slides will be stored in a limited-access location on an IIT server and are not for distribution or use beyond Math 454/553. 2 Matchings 3.1.1 Definition A matching in a graph G is a set of non-loop edges with no shared endpoints. The vertices incident to the edges of a matching M are saturated by M (M-saturated); the others are unsaturated (M-unsaturated). A perfect matching in a graph is a matching that saturates every vertex. perfect matching M-unsaturated M-saturated M Contains copyrighted material from Introduction to Graph Theory by Doug West, 2nd Ed. Not for distribution beyond IIT’s Math 454/553. 3 Perfect Matchings in Complete Bipartite Graphs a 1 The perfect matchings in a complete b 2 X,Y-bigraph with |X|=|Y| exactly c 3 correspond to the bijections d 4 f: X -> Y e 5 Therefore Kn,n has n! perfect f 6 matchings. g 7 Kn,n The complete graph Kn has a perfect matching iff… Contains copyrighted material from Introduction to Graph Theory by Doug West, 2nd Ed. Not for distribution beyond IIT’s Math 454/553. 4 Perfect Matchings in Complete Graphs The complete graph Kn has a perfect matching iff n is even. So instead of Kn consider K2n. We count the perfect matchings in K2n by: (1) Selecting a vertex v (e.g., with the highest label) one choice u v (2) Selecting a vertex u to match to v K2n-2 2n-1 choices (3) Selecting a perfect matching on the rest of the vertices.
    [Show full text]
  • Catalogue of Graph Polynomials
    Catalogue of graph polynomials J.A. Makowsky April 6, 2011 Contents 1 Graph polynomials 3 1.1 Comparinggraphpolynomials. ........ 3 1.1.1 Distinctivepower.............................. .... 3 1.1.2 Substitutioninstances . ...... 4 1.1.3 Uniformalgebraicreductions . ....... 4 1.1.4 Substitutioninstances . ...... 4 1.1.5 Substitutioninstances . ...... 4 1.1.6 Substitutioninstances . ...... 4 1.2 Definabilityofgraphpolynomials . ......... 4 1.2.1 Staticdefinitions ............................... ... 4 1.2.2 Dynamicdefinitions .............................. 4 1.2.3 SOL-definablepolynomials ............................ 4 1.2.4 Generalizedchromaticpolynomials . ......... 4 2 A catalogue of graph polynomials 5 2.1 Polynomialsfromthezoo . ...... 5 2.1.1 Chromaticpolynomial . .... 5 2.1.2 Chromatic symmetric function . ...... 6 2.1.3 Adjointpolynomials . .... 6 2.1.4 TheTuttepolynomial . ... 7 2.1.5 Strong Tutte symmetric function . ....... 7 2.1.6 Tutte-Grothendieck invariants . ........ 8 2.1.7 Aweightedgraphpolynomial . ..... 8 2.1.8 Chainpolynomial ............................... 9 2.1.9 Characteristicpolynomial . ....... 10 2.1.10 Matchingpolynomial. ..... 11 2.1.11 Theindependentsetpolynomial . ....... 12 2.1.12 Thecliquepolynomial . ..... 14 2.1.13 Thevertex-coverpolynomial . ...... 14 2.1.14 Theedge-coverpolynomial . ..... 15 2.1.15 TheMartinpolynomial . 16 2.1.16 Interlacepolynomial . ...... 17 2.1.17 Thecoverpolynomial . 19 2.1.18 Gopolynomial ................................. 21 2.1.19 Stabilitypolynomial . ...... 23 1 2.1.20 Strong U-polynomial...............................
    [Show full text]
  • Integer Linear Programming
    Introduction Linear Programming Integer Programming Integer Linear Programming Subhas C. Nandy ([email protected]) Advanced Computing and Microelectronics Unit Indian Statistical Institute Kolkata 700108, India. Introduction Linear Programming Integer Programming Organization 1 Introduction 2 Linear Programming 3 Integer Programming Introduction Linear Programming Integer Programming Linear Programming A technique for optimizing a linear objective function, subject to a set of linear equality and linear inequality constraints. Mathematically, maximize c1x1 + c2x2 + ::: + cnxn Subject to: a11x1 + a12x2 + ::: + a1nxn b1 ≤ a21x1 + a22x2 + ::: + a2nxn b2 : ≤ : am1x1 + am2x2 + ::: + amnxn bm ≤ xi 0 for all i = 1; 2;:::; n. ≥ Introduction Linear Programming Integer Programming Linear Programming In matrix notation, maximize C T X Subject to: AX B X ≤0 where C is a n≥ 1 vector |- cost vector, A is a m× n matrix |- coefficient matrix, B is a m × 1 vector |- requirement vector, and X is an n× 1 vector of unknowns. × He developed it during World War II as a way to plan expenditures and returns so as to reduce costs to the army and increase losses incurred by the enemy. The method was kept secret until 1947 when George B. Dantzig published the simplex method and John von Neumann developed the theory of duality as a linear optimization solution. Dantzig's original example was to find the best assignment of 70 people to 70 jobs subject to constraints. The computing power required to test all the permutations to select the best assignment is vast. However, the theory behind linear programming drastically reduces the number of feasible solutions that must be checked for optimality.
    [Show full text]
  • A Branch-And-Bound Algorithm for Zero-One Mixed Integer
    A BRANCH-AND-BOUND ALGORITHM FOR ZERO- ONE MIXED INTEGER PROGRAMMING PROBLEMS Ronald E. Davis Stanford University, Stanford, California David A. Kendrick University of Texas, Austin, Texas and Martin Weitzman Yale University, New Haven, Connecticut (Received August 7, 1969) This paper presents the results of experimentation on the development of an efficient branch-and-bound algorithm for the solution of zero-one linear mixed integer programming problems. An implicit enumeration is em- ployed using bounds that are obtained from the fractional variables in the associated linear programming problem. The principal mathematical result used in obtaining these bounds is the piecewise linear convexity of the criterion function with respect to changes of a single variable in the interval [0, 11. A comparison with the computational experience obtained with several other algorithms on a number of problems is included. MANY IMPORTANT practical problems of optimization in manage- ment, economics, and engineering can be posed as so-called 'zero- one mixed integer problems,' i.e., as linear programming problems in which a subset of the variables is constrained to take on only the values zero or one. When indivisibilities, economies of scale, or combinatoric constraints are present, formulation in the mixed-integer mode seems natural. Such problems arise frequently in the contexts of industrial scheduling, investment planning, and regional location, but they are by no means limited to these areas. Unfortunately, at the present time the performance of most compre- hensive algorithms on this class of problems has been disappointing. This study was undertaken in hopes of devising a more satisfactory approach. In this effort we have drawn on the computational experience of, and the concepts employed in, the LAND AND DoIGE161 Healy,[13] and DRIEBEEKt' I algorithms.
    [Show full text]
  • Matroid Partitioning Algorithm Described in the Paper Here with Ray’S Interest in Doing Ev- Erything Possible by Using Network flow Methods
    Chapter 7 Matroid Partition Jack Edmonds Introduction by Jack Edmonds This article, “Matroid Partition”, which first appeared in the book edited by George Dantzig and Pete Veinott, is important to me for many reasons: First for per- sonal memories of my mentors, Alan J. Goldman, George Dantzig, and Al Tucker. Second, for memories of close friends, as well as mentors, Al Lehman, Ray Fulker- son, and Alan Hoffman. Third, for memories of Pete Veinott, who, many years after he invited and published the present paper, became a closest friend. And, finally, for memories of how my mixed-blessing obsession with good characterizations and good algorithms developed. Alan Goldman was my boss at the National Bureau of Standards in Washington, D.C., now the National Institutes of Science and Technology, in the suburbs. He meticulously vetted all of my math including this paper, and I would not have been a math researcher at all if he had not encouraged it when I was a university drop-out trying to support a baby and stay-at-home teenage wife. His mentor at Princeton, Al Tucker, through him of course, invited me with my child and wife to be one of the three junior participants in a 1963 Summer of Combinatorics at the Rand Corporation in California, across the road from Muscle Beach. The Bureau chiefs would not approve this so I quit my job at the Bureau so that I could attend. At the end of the summer Alan hired me back with a big raise. Dantzig was and still is the only historically towering person I have known.
    [Show full text]
  • New Approximation Algorithms for Minimum Weighted Edge Cover
    New Approximation Algorithms for Minimum Weighted Edge Cover S M Ferdous∗ Alex Potheny Arif Khanz Abstract The K-Nearest Neighbor graph is used to spar- We describe two new 3=2-approximation algorithms and a sify data sets, which is an important step in graph-based new 2-approximation algorithm for the minimum weight semi-supervised machine learning. Here one has a few edge cover problem in graphs. We show that one of the 3=2-approximation algorithms, the Dual Cover algorithm, labeled items, many unlabeled items, and a measure of computes the lowest weight edge cover relative to previously similarity between pairs of items; we are required to known algorithms as well as the new algorithms reported label the remaining items. A popular approach for clas- here. The Dual Cover algorithm can also be implemented to be faster than the other 3=2-approximation algorithms on sification is to generate a similarity graph between the serial computers. Many of these algorithms can be extended items to represent both the labeled and unlabeled data, to solve the b-Edge Cover problem as well. We show the and then to use a label propagation algorithm to classify relation of these algorithms to the K-Nearest Neighbor graph construction in semi-supervised learning and other the unlabeled items [23]. In this approach one builds applications. a complete graph out of the dataset and then sparsi- fies this graph by computing a K-Nearest Neighbor 1 Introduction graph [22]. This sparsification leads to efficient al- An Edge Cover in a graph is a subgraph such that gorithms, but also helps remove noise which can af- every vertex has at least one edge incident on it in fect label propagation [11].
    [Show full text]
  • Edge Guarding Plane Graphs
    Edge Guarding Plane Graphs Master Thesis of Paul Jungeblut At the Department of Informatics Institute of Theoretical Informatics Reviewers: Prof. Dr. Dorothea Wagner Prof. Dr. Peter Sanders Advisor: Dr. Torsten Ueckerdt Time Period: 26.04.2019 – 25.10.2019 KIT – The Research University in the Helmholtz Association www.kit.edu Statement of Authorship I hereby declare that this document has been composed by myself and describes my own work, unless otherwise acknowledged in the text. I also declare that I have read the Satzung zur Sicherung guter wissenschaftlicher Praxis am Karlsruher Institut für Technologie (KIT). Paul Jungeblut, Karlsruhe, 25.10.2019 iii Abstract Let G = (V; E) be a plane graph. We say that a face f of G is guarded by an edge vw 2 E if at least one vertex from fv; wg is on the boundary of f. For a planar graph class G the function ΓG : N ! N maps n to the minimal number of edges needed to guard all faces of any n-vertex graph in G. This thesis contributes new bounds on ΓG for several graph classes, in particular Γ Γ Γ on 4;stacked for stacked triangulations, on for quadrangulations and on sp for series parallel graphs. Specifically we show that • b(2n - 4)=7c 6 Γ4;stacked(n) 6 b2n=7c, b(n - 2)=4c Γ (n) bn=3c • 6 6 and • b(n - 2)=3c 6 Γsp(n) 6 bn=3c. Note that the bounds for stacked triangulations and series parallel graphs are tight (up to a small constant). For quadrangulations we identify the non-trivial subclass 2 Γ (n) bn=4c of -degenerate quadrangulations for which we further prove ;2-deg 6 matching the lower bound.
    [Show full text]