The Adaptive Complexity of Submodular Optimization

Total Page:16

File Type:pdf, Size:1020Kb

The Adaptive Complexity of Submodular Optimization The Adaptive Complexity of Submodular Optimization The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Balkanski, Eric. 2019. The Adaptive Complexity of Submodular Optimization. Doctoral dissertation, Harvard University, Graduate School of Arts & Sciences. Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:42029793 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA The Adaptive Complexity of Submodular Optimization a dissertation presented by Eric Balkanski to The School of Engineering and Applied Sciences in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the subject of Computer Science Harvard University Cambridge, Massachusetts May 2019 © 2019 Eric Balkanski All rights reserved. Dissertation advisor: Yaron Singer Eric Balkanski The Adaptive Complexity of Submodular Optimization Abstract In this thesis, we develop a new optimization technique that leads to exponentially faster algorithms for solving submodular optimization problems. For the canonical problem of maximizing a non-decreasing submodular function under a cardinality constraint, it is well known that the celebrated greedy algorithm which iteratively adds elements whose marginal contribution is largest achieves a 1 1/e approximation, which ≠ is optimal. The optimal approximation guarantee of the greedy algorithm comes at a price of high adaptivity. The adaptivity of an algorithm is the number of sequential rounds it makes when polynomially-many function evaluations can be executed in parallel in each round. Since submodular optimization is regularly applied on very large datasets, adaptivity is crucial as algorithms with low adaptivity enable dramatic speedups in parallel computing time. Submodular optimization has been studied for well over forty years now, and somewhat surprisingly, there was no known constant-factor approximation algorithm for submodular maximization whose adaptivity is sublinear in the size of the ground set n. Our main contribution is a novel optimization technique called adaptive sampling which leads to constant factor approximation algorithms for submodular maximization in only logarithmically many adaptive rounds. This is an exponential speedup in the parallel runtime for submodular maximization compared to previous constant factor approximation algorithms. Furthermore, we show that no algorithm can achieve a constant factor approximation in o˜(log n) rounds. Thus, the adaptive complexity of submodular maximization, i.e., the minimum number of rounds r such that there exists an r-adaptive algorithm which achieves a constant factor approximation, is logarithmic up to lower order terms. iii Contents 1 Introduction 1 1.1 Adaptive Sampling: An Exponential Speedup . 4 1.2 ResultsOverview ................................. 5 1.2.1 From Predictions to Decisions: Optimization from Samples . 6 1.2.2 Faster Parallel Algorithms Through Adaptive Sampling . 8 1.3 Preliminaries . 12 1.4 Discussion About Adaptivity . 14 1.4.1 Adaptivity in Other Areas . 14 1.4.2 Related Models of Parallel Computation . 15 1.4.3 Applications of Adaptivity . 17 2 Non-Adaptive Optimization 20 2.1 From Predictions to Decisions: Optimization from Samples . 20 2.1.1 The Optimization from Samples Model . 23 2.1.2 Optimization from Samples is Equivalent to Non-Adaptivity . 25 2.1.3 Overview of Results . 27 2.2 Optimization from Samples Algorithms . 29 2.2.1 Curvature . 30 2.2.2 Learning to Influence . 35 iv 2.2.3 General Submodular Maximization . 54 2.3 The Limitations of Optimization from Samples . 62 2.3.1 A Framework for Hardness of Optimization from Samples . 62 2.3.2 Submodular Maximization . 65 2.3.3 Maximum Coverage . 67 2.3.4 Curvature . 84 2.4 References and Acknowledgments . 91 3 Adaptive Optimization 92 3.1 Adaptivity . 93 3.1.1 The Adaptive Complexity Model . 93 3.1.2 The Adaptivity Landscape for Submodular Optimization . 94 3.1.3 Main Result . 95 3.1.4 Adaptive Sampling: a Coupling of Learning and Optimization . 96 3.1.5 Overview of Results . 96 3.2 Adaptive Algorithms . 99 3.2.1 An Algorithm with Logarithmic Adaptivity . 99 3.2.2 Experiments . 117 3.2.3 The Optimal Approximation . 123 3.2.4 Non-monotone Functions . 141 3.2.5 Matroid Constraints . 166 3.3 Adaptivity Lower Bound . 205 3.4 References and Acknowledgments . 214 v Acknowledgments First and foremost, I would like to thank my advisor Yaron Singer. Yaron has taught me everything about research, from long-term research vision to choosing the right font for a presentation. I have been extremely fortunate to have an advisor who has always believed in me, cares so much about my success, and tirelessly mentored and advised me through every stage of my PhD. My family has been a vital source of support through the years. I am grateful to my parents, Cecile and Yves, for all the encouragement and freedom to explore every project and idea I had since a very young age. Thank you to my siblings Sophie and Jefffor all the good times laughing together and the endless singing in the car during vacations, I cannot wait for our next family vacation. Thanks also to my girlfriend Meghan for the emotional support and dealing with me during paper deadlines. I am very grateful for my two internships at Google Research NYC, which have played an important role in my PhD. During these two summers, I met and worked with wonderful people, broadened my research horizon, and explored new directions. Thank you in particular to my hosts and collaborators Umar Syed, Sergei Vassilvitskii, Balu Sivan, Renato Paes Leme, and Vahab Mirrokni. I have also had the chance to work with incredible collaborators from whom I have learned a lot and who made significant contributions to this thesis: Aviad Rubinstein, Jason Hartline, Andreas Krause, Baharan Mirzasoleiman, Nicole Immorlica, Amir Globerson, Nir Rosenfeld, and Adam Breuer. I am also grateful for the numerous discussions about the content of this thesis with Thibaut Horel, whose breadth of knowledge has been very helpful. I am fortunate to have spent my PhD working in a fun and warm environment. Thank you to my officemates through the years, Emma Heikensten, Siri Isaksson, Dimitris Kalimeris, Gal Kaplun, Sharon Qian, Greg Stoddard, Bo Waggoner and Ming Yin, I will miss the vi great atmosphere in MD115. The EconCS group and Maxwell Dworkin have been amazingly friendly and supportive places to grow academically, thanks in particular to David Parkes, Yiling Chen, Jean Pouget-Abadie, Hongyao Ma, Chara Podimata, Jarek Blasiok, Debmalya Mandal, Preetum Nakkiran, Lior Seeman, Goran Radanovic, and Yang Liu, and to my research committee, Yaron Singer, David Parkes, Sasha Rush, and Michael Mitzenmacher. I am fortunate for the countless and diverse opportunities I had during my undergraduate studies at Carnegie Mellon which helped me grow. These opportunities lead me to valuable research, teaching, leadership, social, and athletic experiences. Thank you in particular to my undergraduate research advisor, Ariel Procaccia, who was the first person to tell me he believed I could become a professor, and my academic advisor John Mackey for all the opportunities. This thesis was supported in part by a Google PhD Fellowship and a Smith Family Graduate Science and Engineering Fellowship. vii Chapter 1 Introduction The field of optimization has recently expanded to new application domains, where complex decision-making tasks are challenging existing frameworks and techniques. A main difficulty is that the scale of these tasks is growing at a vertiginous rate. Innovative frameworks and techniques are needed to capture modern application domains as well as address the challenges posed by large scale computation. We begin by discussing three application domains where optimization has recently played an important role. The first domain is genomics. Recent developments in computational biology that allow processing large amounts of genomic data have been a catalyst for progress towards understanding the human genome. One approach that helps to understand and process massive gene datasets is to cluster gene sequences. Given gene clusters, a small representative subset of genes, in other words a summary of the entire dataset, can be obtained by choosing one gene sequence from each cluster. The problem of clustering gene sequences is an example of a large scale optimization problem. A different domain is recommender systems. Twenty years ago, a Friday movie night would require browsing through multiple aisles at Blockbuster to find the right movie. Today, streaming services use recommender systems to suggest a personalized collection of movies to 1 a user. Recommender systems are not limited to movies, but are also used for music and online retail. Given the preferences of different users, optimization techniques are used to find diverse and personalized collections of movies, songs, or products. The third domain, ride-sharing services, did not even exist a decade ago. Ride-sharing services have revolutionized transportation by allowing passengers to request a ride to their desired destination in a few seconds using their smartphone. These companies face novel and complex optimization tasks. One example is driver dispatch, which is the problem of assigning drivers to different locations, in order to match the demand from riders. Gene sequences, user ratings for music and movies, rider-driver locations are examples of large datasets that we wish to harness using optimization. Novel algorithmic techniques and frameworks that are adapted for these new domains are needed. A standard approach to these optimization problems is to use an algorithmic technique called the greedy approach. Informally, a greedy algorithm takes small local steps towards building a global solution to a problem. For example, for the driver dispatch problem in New York City, a greedy algorithm might first dispatch a first driver to bustling midtown Manhattan.
Recommended publications
  • Even Factors, Jump Systems, and Discrete Convexity
    Even Factors, Jump Systems, and Discrete Convexity Yusuke KOBAYASHI¤ Kenjiro TAKAZAWAy May, 2007 Abstract A jump system, which is a set of integer lattice points with an exchange property, is an extended concept of a matroid. Some combinatorial structures such as the degree sequences of the matchings in an undirected graph are known to form a jump system. On the other hand, the maximum even factor problem is a generalization of the maximum matching problem into digraphs. When the given digraph has a certain property called odd- cycle-symmetry, this problem is polynomially solvable. The main result of this paper is that the degree sequences of all even factors in a digraph form a jump system if and only if the digraph is odd-cycle-symmetric. Furthermore, as a gen- eralization, we show that the weighted even factors induce M-convex (M-concave) functions on jump systems. These results suggest that even factors are a natural generalization of matchings and the assumption of odd-cycle-symmetry of digraphs is essential. 1 Introduction In the study of combinatorial optimization, extensions of matroids are introduced as abstract con- cepts including many combinatorial objects. A number of optimization problems on matroidal structures can be solved in polynomial time. One of the extensions of matroids is a jump system of Bouchet and Cunningham [2]. A jump system is a set of integer lattice points with an exchange property (to be described in Section 2.1); see also [18, 22]. It is a generalization of a matroid [4], a delta-matroid [1, 3, 8], and a base polyhedron of an integral polymatroid (or a submodular sys- tem) [14].
    [Show full text]
  • Invitation to Matroid Theory
    Invitation to Matroid Theory Gregory Henselman-Petrusek January 21, 2021 Abstract This text is a companion to the short course Invitation to matroid theory taught in the Univer- sity of Oxford Centre for TDA in January 2021. It was first written for algebraic topologists, but should be suitable for all audiences; no background in matroid theory is assumed! 1 How to use this text Matroid theory is a beautiful, powerful, and accessible. Many important subjects can be understood and even researched with just a handful of concepts and definitions from matroid theory. Even so, getting use out of matroids often depends on knowing the right keywords for an internet search, and finding the right keyword depends on two things: (i) a precise knowledge of the core terms/definitions of matroid theory, and (ii) a birds eye view of the network of relationships that interconnect some of the main ideas in the field. These are the focus of this text; it falls far short of a comprehensive overview, but provides a strong start to (i) and enough of (ii) to build an interesting (and fun!) network of ideas including some of the most important concepts in matroid theory. The proofs of most results can be found in standard textbooks (e.g. Oxley, Matroid theory). Exercises are chosen to maximize the fraction intellectual reward time required to solve Therefore, if an exercise takes more than a few minutes to solve, it is fine to look up answers in a textbook! Struggling is not prerequisite to learning, at this level. 2 References The following make no attempt at completeness.
    [Show full text]
  • Matroid Theory Release 9.4
    Sage 9.4 Reference Manual: Matroid Theory Release 9.4 The Sage Development Team Aug 24, 2021 CONTENTS 1 Basics 1 2 Built-in families and individual matroids 77 3 Concrete implementations 97 4 Abstract matroid classes 149 5 Advanced functionality 161 6 Internals 173 7 Indices and Tables 197 Python Module Index 199 Index 201 i ii CHAPTER ONE BASICS 1.1 Matroid construction 1.1.1 Theory Matroids are combinatorial structures that capture the abstract properties of (linear/algebraic/...) dependence. For- mally, a matroid is a pair M = (E; I) of a finite set E, the groundset, and a collection of subsets I, the independent sets, subject to the following axioms: • I contains the empty set • If X is a set in I, then each subset of X is in I • If two subsets X, Y are in I, and jXj > jY j, then there exists x 2 X − Y such that Y + fxg is in I. See the Wikipedia article on matroids for more theory and examples. Matroids can be obtained from many types of mathematical structures, and Sage supports a number of them. There are two main entry points to Sage’s matroid functionality. The object matroids. contains a number of con- structors for well-known matroids. The function Matroid() allows you to define your own matroids from a variety of sources. We briefly introduce both below; follow the links for more comprehensive documentation. Each matroid object in Sage comes with a number of built-in operations. An overview can be found in the documen- tation of the abstract matroid class.
    [Show full text]
  • Matroid Matching: the Power of Local Search
    Matroid Matching: the Power of Local Search [Extended Abstract] Jon Lee Maxim Sviridenko Jan Vondrák IBM Research IBM Research IBM Research Yorktown Heights, NY Yorktown Heights, NY San Jose, CA [email protected] [email protected] [email protected] ABSTRACT can be easily transformed into an NP-completeness proof We consider the classical matroid matching problem. Un- for a concrete class of matroids (see [46]). An important weighted matroid matching for linear matroids was solved result of Lov´asz is that (unweighted) matroid matching can by Lov´asz, and the problem is known to be intractable for be solved in polynomial time for linear matroids (see [35]). general matroids. We present a PTAS for unweighted ma- There have been several attempts to generalize Lov´asz' re- troid matching for general matroids. In contrast, we show sult to the weighted case. Polynomial-time algorithms are that natural LP relaxations have an Ω(n) integrality gap known for some special cases (see [49]), but for general linear and moreover, Ω(n) rounds of the Sherali-Adams hierarchy matroids there is only a pseudopolynomial-time randomized are necessary to bring the gap down to a constant. exact algorithm (see [8, 40]). More generally, for any fixed k ≥ 2 and > 0, we obtain a In this paper, we revisit the matroid matching problem (k=2 + )-approximation for matroid matching in k-uniform for general matroids. Our main result is that while LP- hypergraphs, also known as the matroid k-parity problem. based approaches including the Sherali-Adams hierarchy fail As a consequence, we obtain a (k=2 + )-approximation for to provide any meaningful approximation, a simple local- the problem of finding the maximum-cardinality set in the search algorithm gives a PTAS (in the unweighted case).
    [Show full text]
  • Small Space Stream Summary for Matroid Center Sagar Kale EPFL, Lausanne, Switzerland Sagar.Kale@Epfl.Ch
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Dagstuhl Research Online Publication Server Small Space Stream Summary for Matroid Center Sagar Kale EPFL, Lausanne, Switzerland sagar.kale@epfl.ch Abstract In the matroid center problem, which generalizes the k-center problem, we need to pick a set of centers that is an independent set of a matroid with rank r. We study this problem in streaming, where elements of the ground set arrive in the stream. We first show that any randomized one-pass streaming algorithm that computes a better than ∆-approximation for partition-matroid center must use Ω(r2) bits of space, where ∆ is the aspect ratio of the metric and can be arbitrarily large. This shows a quadratic separation between matroid center and k-center, for which the Doubling algorithm [7] gives an 8-approximation using O(k)-space and one pass. To complement this, we give a one-pass algorithm for matroid center that stores at most O(r2 log(1/ε)/ε) points (viz., stream summary) among which a (7 + ε)-approximate solution exists, which can be found by brute force, or a (17 + ε)-approximation can be found with an efficient algorithm. If we are allowed a second pass, we can compute a (3 + ε)-approximation efficiently. We also consider the problem of matroid center with z outliers and give a one-pass algorithm that outputs a set of O((r2 + rz) log(1/ε)/ε) points that contains a (15 + ε)-approximate solution. Our techniques extend to knapsack center and knapsack center with z outliers in a straightforward way, and we get algorithms that use space linear in the size of a largest feasible set (as opposed to quadratic space for matroid center).
    [Show full text]
  • Quasipolynomial Representation of Transversal Matroids with Applications in Parameterized Complexity
    Quasipolynomial Representation of Transversal Matroids with Applications in Parameterized Complexity Daniel Lokshtanov1, Pranabendu Misra2, Fahad Panolan3, Saket Saurabh4, and Meirav Zehavi5 1 University of Bergen, Bergen, Norway [email protected] 2 The Institute of Mathematical Sciences, HBNI, Chennai, India [email protected] 3 University of Bergen, Bergen, Norway [email protected] 4 University of Bergen, Bergen, Norway and The Institute of Mathematical Sciences, HBNI, Chennai, India [email protected] 5 Ben-Gurion University, Beersheba, Israel [email protected] Abstract Deterministic polynomial-time computation of a representation of a transversal matroid is a longstanding open problem. We present a deterministic computation of a so-called union rep- resentation of a transversal matroid in time quasipolynomial in the rank of the matroid. More precisely, we output a collection of linear matroids such that a set is independent in the trans- versal matroid if and only if it is independent in at least one of them. Our proof directly implies that if one is interested in preserving independent sets of size at most r, for a given r ∈ N, but does not care whether larger independent sets are preserved, then a union representation can be computed deterministically in time quasipolynomial in r. This consequence is of independent interest, and sheds light on the power of union representation. Our main result also has applications in Parameterized Complexity. First, it yields a fast computation of representative sets, and due to our relaxation in the context of r, this computation also extends to (standard) truncations. In turn, this computation enables to efficiently solve various problems, such as subcases of subgraph isomorphism, motif search and packing problems, in the presence of color lists.
    [Show full text]
  • An Optimal Approximation for Submodular Maximization Under A
    An Optimal Approximation for Submodular Maximization undera Matroid Constraint in the Adaptive Complexity Model Eric Balkanski Aviad Rubinstein Yaron Singer Harvard University Stanford University Harvard University [email protected] [email protected] [email protected] Abstract In this paper we study submodular maximization under a matroid constraint in the adaptive complexity model. This model was recently introduced in the context of submodular optimiza- tion in [BS18a] to quantify the information theoretic complexity of black-box optimization in a parallel computation model. Informally, the adaptivity of an algorithm is the number of sequential rounds it makes when each round can execute polynomially-many function evalu- ations in parallel. Since submodular optimization is regularly applied on large datasets we seek algorithms with low adaptivity to enable speedups via parallelization. Consequently, a recent line of work has been devoted to designing constant factor approximation algorithms for maximizing submodular functions under various constraints in the adaptive complexity model [BS18a, BS18b, BBS18, BRS19, EN19, FMZ19, CQ19, ENV18, FMZ18]. Despite the burst in work on submodular maximization in the adaptive complexity model, the fundamental problem of maximizing a monotone submodular function under a matroid constraint has remained elusive. In particular, all known techniques fail for this problem and there are no known constant factor approximation algorithms whose adaptivity is sublinear in the rank of the matroid k or in the worst case sublinear in the size of the ground set n. In this paper we present an approximation algorithm for the problem of maximizing a mono- tone submodular function under a matroid constraint in the adaptive complexity model.
    [Show full text]
  • Even Circuits in Oriented Matroids
    EVEN CIRCUITS IN ORIENTED MATROIDS KARL HEUER, RAPHAEL STEINER, AND SEBASTIAN WIEDERRECHT Abstract. In this paper we generalise the even directed cycle problem, which asks whether a given digraph contains a directed cycle of even length, to orientations of regular matroids. We define non-even oriented matroids generalising non-even digraphs, which played a central role in resolving the computational complexity of the even dicycle problem. Then we show that the problem of detecting an even directed circuit in a regular matroid is polynomially equivalent to the recognition of non-even oriented matroids. Our main result is a precise characterisation of the class of non-even oriented bond matroids in terms of forbidden minors, which complements an existing characterisation of non-even oriented graphic matroids by Seymour and Thomassen. §1. Introduction Deciding whether a given digraph contains a directed cycle, briefly dicycle, of even length is a fundamental problem for digraphs and often referred to as the even dicycle problem. The computational complexity of this problem was unknown for a long time and several polynomial time equivalent problems have been found [11–13,19]. The question about the computational complexity was resolved by Robertson, Seymour and Thomas [15] and independently by McCuaig [12] who stated polynomial time algorithms for one of the polynomially equivalent problems, and hence also for the even dicycle problem. One of these polynomially equivalent problems makes use of the following definition. Definition 1.1 ([18]). Let D be a digraph. We call D non-even, if there exists a set J of directed edges in D such that every directed cycle C in D intersects J in an odd number of edges.
    [Show full text]
  • Small Space Stream Summary for Matroid Center
    Small Space Stream Summary for Matroid Center Sagar Kale EPFL [email protected] Abstract In the matroid center problem, which generalizes the k-center problem, we need to pick a set of cen- ters that is an independent set of a matroid with rank r. We study this problem in streaming, where ele- ments of the ground set arrive in the stream. We first show that any randomized one-pass streaming al- gorithm that computes a better than ∆-approximation for partition-matroid center must use Ω(r2) bits of space, where ∆ is the aspect ratio of the metric and can be arbitrarily large. This shows a quadratic separation between matroid center and k-center, for which the Doubling algorithm [CCFM97] gives an 8-approximation using O(k)-space and one pass. To complement this, we give a one-pass algorithm for matroid center that stores at most O(r2 log(1/ε)/ε) points (viz., stream summary) among which a (7 + ε)-approximate solution exists, which can be found by brute force, or a (17 + ε)-approximation can be found with an efficient algorithm. If we are allowed a second pass, we can compute a (3 + ε)- approximation efficiently; this also achieves almost the known-best approximation ratio (of 3+ε) with total running time of O((nr + r3.5)log(1/ε)/ε + r2(log∆)/ε), where n is the number of input points. We also consider the problem of matroid center with z outliers and give a one-pass algorithm that outputs a set of O((r2 + rz)log(1/ε)/ε) points that contains a (15 + ε)-approximate solution.
    [Show full text]
  • Matroid Optimization and Algorithms Robert E. Bixby and William H
    Matroid Optimization and Algorithms Robert E. Bixby and William H. Cunningham June, 1990 TR90-15 MATROID OPTIMIZATION AND ALGORITHMS by Robert E. Bixby Rice University and William H. Cunningham Carleton University Contents 1. Introduction 2. Matroid Optimization 3. Applications of Matroid Intersection 4. Submodular Functions and Polymatroids 5. Submodular Flows and other General Models 6. Matroid Connectivity Algorithms 7. Recognition of Representability 8. Matroid Flows and Linear Programming 1 1. INTRODUCTION This chapter considers matroid theory from a constructive and algorithmic viewpoint. A substantial part of the developments in this direction have been motivated by opti­ mization. Matroid theory has led to a unification of fundamental ideas of combinatorial optimization as well as to the solution of significant open problems in the subject. In addition to its influence on this larger subject, matroid optimization is itself a beautiful part of matroid theory. The most basic optimizational property of matroids is that for any subset every max­ imal independent set contained in it is maximum. Alternatively, a trivial algorithm max­ imizes any {O, 1 }-valued weight function over the independent sets. Most of matroid op­ timization consists of attempts to solve successive generalizations of this problem. In one direction it is generalized to the problem of finding a largest common independent set of two matroids: the matroid intersection problem. This problem includes the matching problem for bipartite graphs, and several other combinatorial problems. In Edmonds' solu­ tion of it and the equivalent matroid partition problem, he introduced the notions of good characterization (intimately related to the NP class of problems) and matroid (oracle) algorithm.
    [Show full text]
  • A Gentle Introduction to Matroid Algorithmics
    A Gentle Introduction to Matroid Algorithmics Matthias F. Stallmann April 14, 2016 1 Matroid axioms The term matroid was first used in 1935 by Hassler Whitney [34]. Most of the material in this overview comes from the textbooks of Lawler [25] and Welsh [33]. A matroid is defined by a set of axioms with respect to a ground set S. There are several ways to do this. Among these are: Independent sets. Let I be a collection of subsets defined by • (I1) ; 2 I • (I2) if X 2 I and Y ⊆ X then Y 2 I • (I3) if X; Y 2 I with jXj = jY j + 1 then there exists x 2 X n Y such that Y [ x 2 I A classic example is a graphic matroid. Here S is the set of edges of a graph and X 2 I if X is a forest. (I1) and (I2) are clearly satisfied. (I3) is a simplified version of the key ingredient in the correctness proofs of most minimum spanning tree algorithms. If a set of edges X forms a forest then jV (X)j − jXj = jc(X)j, where V (x) is the set of endpoints of X and c(X) is the set of components induced by X. If jXj = jY j + 1, then jc(X)j = jc(Y )j − 1, meaning X must have at least one edge x that connects two of the components induced by Y . So Y [ x is also a forest. Another classic algorithmic example is the following uniprocessor scheduling problem, let's call it the scheduling matroid.
    [Show full text]
  • Survey of Matroid Secretary Problem Algorithms
    6.854 Fall 2012 Advanced Algorithms Matroid Secretary Problems Gautam Kamath Instructor: David Karger [email protected] 1 Introduction In 1963, Dynkin introduced the secretary problem [6]. In this problem, an algorithm is presented with n positive values, one by one. After each value, the algorithm must either accept or reject the value, where all decisions are final. The algorithm can only pick one value, and the goal is to pick the maximum value in the sequence. The name for this problem arises from a situation where n candidates are interviewed for a secretary position, and the interviewer wants to hire the best candidate. Other names for this problem include the best choice problem, the marriage problem, and the MIT dating problem. At first glance, it may seem impossible to hope to pick the greatest value - since the values are arbitrary, couldn't we always have a bigger element after we accept one? It turns out that one 1 can pick the best value with probability approaching e . The essential idea is to learn properties of the sequence based on the first several values, and then using what we learned in order to make a decision on the remainder of the sequence. This concept is very general, and can be applied on extensions of the original problem. Indeed, a number of extensions have studied - the most natural extension is the multiple-choice secretary problem [7], where the algorithm may choose k values, rather than just one. Other extensions include the submodular secretary problem [3] and the knapsack secretary problem [1].
    [Show full text]