The Unique Games Conjecture and Some of Its Implications on Inapproximability
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Zig-Zag Product
High Dimensional Expanders Lecture 12: Agreement Tests Lecture by: Irit Dinur Scribe: Irit Dinur In this lecture we describe agreement tests, which are a generalization of direct product tests and low degree tests, both of which play a role in PCP constructions. It turns out that the underlying structures that support agreement tests are invariably some kind of high dimensional expander, as we shall see. 1 General Setup It is a basic fact of computation that any global computation can be broken down into a sequence of local steps. The PCP theorem [AS98, ALM+98] says that moreover, this can be done in a robust fashion, so that as long as most steps are correct, the entire computation checks out. At the heart of this is a local-to-global argument that allows deducing a global property from local pieces that fit together only approximately. In an agreement test, a function is given by an ensemble of local restrictions. The agreement test checks that the restrictions agree when they overlap, and the main question is whether typical agreement of the local pieces implies that there exists a global function that agrees with most local restrictions. Let us describe the basic framework, consisting of a quadruple (V; X; fFSgS2X ; D). • Ground set: Let V be a set of n points (or vertices). • Collection of small subsets: Let X be a collection of subsets S ⊂ V , typically for each S 2 X we have jSj n. • Local functions: for each subset S 2 X, there is a space FS ⊂ ff : S ! Σg of functions on S. -
The Strongish Planted Clique Hypothesis and Its Consequences
The Strongish Planted Clique Hypothesis and Its Consequences Pasin Manurangsi Google Research, Mountain View, CA, USA [email protected] Aviad Rubinstein Stanford University, CA, USA [email protected] Tselil Schramm Stanford University, CA, USA [email protected] Abstract We formulate a new hardness assumption, the Strongish Planted Clique Hypothesis (SPCH), which postulates that any algorithm for planted clique must run in time nΩ(log n) (so that the state-of-the-art running time of nO(log n) is optimal up to a constant in the exponent). We provide two sets of applications of the new hypothesis. First, we show that SPCH implies (nearly) tight inapproximability results for the following well-studied problems in terms of the parameter k: Densest k-Subgraph, Smallest k-Edge Subgraph, Densest k-Subhypergraph, Steiner k-Forest, and Directed Steiner Network with k terminal pairs. For example, we show, under SPCH, that no polynomial time algorithm achieves o(k)-approximation for Densest k-Subgraph. This inapproximability ratio improves upon the previous best ko(1) factor from (Chalermsook et al., FOCS 2017). Furthermore, our lower bounds hold even against fixed-parameter tractable algorithms with parameter k. Our second application focuses on the complexity of graph pattern detection. For both induced and non-induced graph pattern detection, we prove hardness results under SPCH, improving the running time lower bounds obtained by (Dalirrooyfard et al., STOC 2019) under the Exponential Time Hypothesis. 2012 ACM Subject Classification Theory of computation → Problems, reductions and completeness; Theory of computation → Fixed parameter tractability Keywords and phrases Planted Clique, Densest k-Subgraph, Hardness of Approximation Digital Object Identifier 10.4230/LIPIcs.ITCS.2021.10 Related Version A full version of the paper is available at https://arxiv.org/abs/2011.05555. -
Lecture 20 — March 20, 2013 1 the Maximum Cut Problem 2 a Spectral
UBC CPSC 536N: Sparse Approximations Winter 2013 Lecture 20 | March 20, 2013 Prof. Nick Harvey Scribe: Alexandre Fr´echette We study the Maximum Cut Problem with approximation in mind, and naturally we provide a spectral graph theoretic approach. 1 The Maximum Cut Problem Definition 1.1. Given a (undirected) graph G = (V; E), the maximum cut δ(U) for U ⊆ V is the cut with maximal value jδ(U)j. The Maximum Cut Problem consists of finding a maximal cut. We let MaxCut(G) = maxfjδ(U)j : U ⊆ V g be the value of the maximum cut in G, and 0 MaxCut(G) MaxCut = jEj be the normalized version (note that both normalized and unnormalized maximum cut values are induced by the same set of nodes | so we can interchange them in our search for an actual maximum cut). The Maximum Cut Problem is NP-hard. For a long time, the randomized algorithm 1 1 consisting of uniformly choosing a cut was state-of-the-art with its 2 -approximation factor The simplicity of the algorithm and our inability to find a better solution were unsettling. Linear pro- gramming unfortunately didn't seem to help. However, a (novel) technique called semi-definite programming provided a 0:878-approximation algorithm [Goemans-Williamson '94], which is optimal assuming the Unique Games Conjecture. 2 A Spectral Approximation Algorithm Today, we study a result of [Trevison '09] giving a spectral approach to the Maximum Cut Problem with a 0:531-approximation factor. We will present a slightly simplified result with 0:5292-approximation factor. -
Tight Integrality Gaps for Lovasz-Schrijver LP Relaxations of Vertex Cover and Max Cut
Tight Integrality Gaps for Lovasz-Schrijver LP Relaxations of Vertex Cover and Max Cut Grant Schoenebeck∗ Luca Trevisany Madhur Tulsianiz Abstract We study linear programming relaxations of Vertex Cover and Max Cut arising from repeated applications of the \lift-and-project" method of Lovasz and Schrijver starting from the standard linear programming relaxation. For Vertex Cover, Arora, Bollobas, Lovasz and Tourlakis prove that the integrality gap remains at least 2 − " after Ω"(log n) rounds, where n is the number of vertices, and Tourlakis proves that integrality gap remains at least 1:5 − " after Ω((log n)2) rounds. Fernandez de la 1 Vega and Kenyon prove that the integrality gap of Max Cut is at most 2 + " after any constant number of rounds. (Their result also applies to the more powerful Sherali-Adams method.) We prove that the integrality gap of Vertex Cover remains at least 2 − " after Ω"(n) rounds, and that the integrality gap of Max Cut remains at most 1=2 + " after Ω"(n) rounds. 1 Introduction Lovasz and Schrijver [LS91] describe a method, referred to as LS, to tighten a linear programming relaxation of a 0/1 integer program. The method adds auxiliary variables and valid inequalities, and it can be applied several times sequentially, yielding a sequence (a \hierarchy") of tighter and tighter relaxations. The method is interesting because it produces relaxations that are both tightly constrained and efficiently solvable. For a linear programming relaxation K, denote by N(K) the relaxation obtained by the application of the LS method, and by N k(K) the relaxation obtained by applying the LS method k times. -
Optimal Inapproximability Results for MAX-CUT and Other 2-Variable Csps?
Electronic Colloquium on Computational Complexity, Report No. 101 (2005) Optimal Inapproximability Results for MAX-CUT and Other 2-Variable CSPs? Subhash Khot∗ Guy Kindlery Elchanan Mosselz College of Computing Theory Group Department of Statistics Georgia Tech Microsoft Research U.C. Berkeley [email protected] [email protected] [email protected] Ryan O'Donnell∗ Theory Group Microsoft Research [email protected] September 19, 2005 Abstract In this paper we show a reduction from the Unique Games problem to the problem of approximating MAX-CUT to within a factor of αGW + , for all > 0; here αGW :878567 denotes the approxima- tion ratio achieved by the Goemans-Williamson algorithm [25]. This≈implies that if the Unique Games Conjecture of Khot [36] holds then the Goemans-Williamson approximation algorithm is optimal. Our result indicates that the geometric nature of the Goemans-Williamson algorithm might be intrinsic to the MAX-CUT problem. Our reduction relies on a theorem we call Majority Is Stablest. This was introduced as a conjecture in the original version of this paper, and was subsequently confirmed in [42]. A stronger version of this conjecture called Plurality Is Stablest is still open, although [42] contains a proof of an asymptotic version of it. Our techniques extend to several other two-variable constraint satisfaction problems. In particular, subject to the Unique Games Conjecture, we show tight or nearly tight hardness results for MAX-2SAT, MAX-q-CUT, and MAX-2LIN(q). For MAX-2SAT we show approximation hardness up to a factor of roughly :943. This nearly matches the :940 approximation algorithm of Lewin, Livnat, and Zwick [40]. -
Calibrating Noise to Sensitivity in Private Data Analysis
Calibrating Noise to Sensitivity in Private Data Analysis Cynthia Dwork1, Frank McSherry1, Kobbi Nissim2, and Adam Smith3? 1 Microsoft Research, Silicon Valley. {dwork,mcsherry}@microsoft.com 2 Ben-Gurion University. [email protected] 3 Weizmann Institute of Science. [email protected] Abstract. We continue a line of research initiated in [10, 11] on privacy- preserving statistical databases. Consider a trusted server that holds a database of sensitive information. Given a query function f mapping databases to reals, the so-called true answer is the result of applying f to the database. To protect privacy, the true answer is perturbed by the addition of random noise generated according to a carefully chosen distribution, and this response, the true answer plus noise, is returned to the user. Previous work focused on the case of noisy sums, in which f = P i g(xi), where xi denotes the ith row of the database and g maps database rows to [0, 1]. We extend the study to general functions f, proving that privacy can be preserved by calibrating the standard devi- ation of the noise according to the sensitivity of the function f. Roughly speaking, this is the amount that any single argument to f can change its output. The new analysis shows that for several particular applications substantially less noise is needed than was previously understood to be the case. The first step is a very clean characterization of privacy in terms of indistinguishability of transcripts. Additionally, we obtain separation re- sults showing the increased value of interactive sanitization mechanisms over non-interactive. -
Approximating Cumulative Pebbling Cost Is Unique Games Hard
Approximating Cumulative Pebbling Cost Is Unique Games Hard Jeremiah Blocki Department of Computer Science, Purdue University, West Lafayette, IN, USA https://www.cs.purdue.edu/homes/jblocki [email protected] Seunghoon Lee Department of Computer Science, Purdue University, West Lafayette, IN, USA https://www.cs.purdue.edu/homes/lee2856 [email protected] Samson Zhou School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA https://samsonzhou.github.io/ [email protected] Abstract P The cumulative pebbling complexity of a directed acyclic graph G is defined as cc(G) = minP i |Pi|, where the minimum is taken over all legal (parallel) black pebblings of G and |Pi| denotes the number of pebbles on the graph during round i. Intuitively, cc(G) captures the amortized Space-Time complexity of pebbling m copies of G in parallel. The cumulative pebbling complexity of a graph G is of particular interest in the field of cryptography as cc(G) is tightly related to the amortized Area-Time complexity of the Data-Independent Memory-Hard Function (iMHF) fG,H [7] defined using a constant indegree directed acyclic graph (DAG) G and a random oracle H(·). A secure iMHF should have amortized Space-Time complexity as high as possible, e.g., to deter brute-force password attacker who wants to find x such that fG,H (x) = h. Thus, to analyze the (in)security of a candidate iMHF fG,H , it is crucial to estimate the value cc(G) but currently, upper and lower bounds for leading iMHF candidates differ by several orders of magnitude. Blocki and Zhou recently showed that it is NP-Hard to compute cc(G), but their techniques do not even rule out an efficient (1 + ε)-approximation algorithm for any constant ε > 0. -
A New Point of NP-Hardness for Unique Games
A new point of NP-hardness for Unique Games Ryan O'Donnell∗ John Wrighty September 30, 2012 Abstract 1 3 We show that distinguishing 2 -satisfiable Unique-Games instances from ( 8 + )-satisfiable instances is NP-hard (for all > 0). A consequence is that we match or improve the best known c vs. s NP-hardness result for Unique-Games for all values of c (except for c very close to 0). For these c, ours is the first hardness result showing that it helps to take the alphabet size larger than 2. Our NP-hardness reductions are quasilinear-size and thus show nearly full exponential time is required, assuming the ETH. ∗Department of Computer Science, Carnegie Mellon University. Supported by NSF grants CCF-0747250 and CCF-0915893, and by a Sloan fellowship. Some of this research performed while the author was a von Neumann Fellow at the Institute for Advanced Study, supported by NSF grants DMS-083537 and CCF-0832797. yDepartment of Computer Science, Carnegie Mellon University. 1 Introduction Thanks largely to the groundbreaking work of H˚astad[H˚as01], we have optimal NP-hardness of approximation results for several constraint satisfaction problems (CSPs), including 3Lin(Z2) and 3Sat. But for many others | including most interesting CSPs with 2-variable constraints | we lack matching algorithmic and NP-hardness results. Take the 2Lin(Z2) problem for example, in which there are Boolean variables with constraints of the form \xi = xj" and \xi 6= xj". The largest approximation ratio known to be achievable in polynomial time is roughly :878 [GW95], whereas it 11 is only known that achieving ratios above 12 ≈ :917 is NP-hard [H˚as01, TSSW00]. -
Lecture 16: Constraint Satisfaction Problems 1 Max-Cut SDP
A Theorist's Toolkit (CMU 18-859T, Fall 2013) Lecture 16: Constraint Satisfaction Problems 10/30/2013 Lecturer: Ryan O'Donnell Scribe: Neal Barcelo 1 Max-Cut SDP Approximation Recall the Max-Cut problem from last lecture. We are given a graph G = (V; E) and our goal is to find a function F : V ! {−1; 1g that maximizes Pr(i;j)∼E[F (vi) 6= F (vj)]. As we saw before, this is NP-hard to do optimally, but we had the following SDP relaxation. 1 1 max avg(i;j)2Ef 2 − 2 yijg s.t. yvv = 1 8v Y = (yij) is symmetric and PSD Note that the second condition is a relaxation of the constraint \9xi s.t. yij = xixj" which is the constraint needed to exactly model the max-cut problem. Further, recall that a matrix being Positive Semi Definitie (PSD) is equivalent to the following characterizations. 1. Y 2 Rn×m is PSD. 2. There exists a U 2 Rm×n such that Y = U T U. 3. There exists joint random variables X1;::: Xn such that yuv = E[XuXv]. ~ ~ ~ ~ 4. There exists vectors U1;:::; Un such that yvw = hUv; Uwi. Using the last characterization, we can rewrite our SDP relaxation as follows 1 1 max avg(i;j)2Ef 2 − 2 h~vi; ~vjig 2 s.t. k~vik2 = 1 Vector ~vi for each vi 2 V First let's make sure this is actually a relaxation of our original problem. To see this, ∗ ∗ consider an optimal cut F : V ! f1; −1g. Then, if we let ~vi = (F (vi); 0;:::; 0), all of the constraints are satisfied and the objective value remains the same. -
Call for Papers
Call For Papers The 6th Innovations in Theoretical Computer Science (ITCS) conference, sponsored by the ACM Special Interest Group on Algorithms and Computation Theory (SIGACT), will be held at the Weizmann Institute of Science, Israel, January 11-13, 2015. ITCS (previously known as ICS) seeks to promote research that carries a strong conceptual message (e.g., introducing a new concept or model, opening a new line of inquiry within traditional or cross-interdisciplinary areas, or introducing new techniques or new applications of known techniques). ITCS welcomes all submissions, whether aligned with current theory of computation research directions or deviating from them. Important Dates Paper Submission Deadline: Friday, August 8, 2014, 5PM PDT Notification to Authors: Monday, October 20, 2014 Camera ready papers due: Monday, November 24, 2014 Conference dates: Sunday-Tuesday, January 11-13, 2015 Program Committee: Benny Applebaum (Tel Aviv University) Avrim Blum (Carnegie Mellon University) Costis Daskalakis (MIT) Uriel Feige (Weizmann Institute) Vitaly Feldman (IBM Research - Almaden) Parikshit Gopalan (Microsoft Research) Bernhard Haeupler (Carnegie Mellon University and Microsoft Research) Stefano Leonardi (University of Rome La Sapienza) Tal Malkin (Columbia University) Nicole Megow (Technische Universitat Berlin) Michael Mitzenmacher (Harvard University) Noam Nisan (Hebrew University and Microsoft Research) Ryan O'Donnell (Carnegie Mellon University) Rafael Pass (Cornell and Cornell NYC Tech) Dana Ron (Tel Aviv University) Guy Rothblum -
Research Statement
RESEARCH STATEMENT SAM HOPKINS Introduction The last decade has seen enormous improvements in the practice, prevalence, and usefulness of machine learning. Deep nets give our phones ears; matrix completion gives our televisions good taste. Data-driven articial intelligence has become capable of the dicult—driving a car on a highway—and the nearly impossible—distinguishing cat pictures without human intervention [18]. This is the stu of science ction. But we are far behind in explaining why many algorithms— even now-commonplace ones for relatively elementary tasks—work so stunningly well as they do in the wild. In general, we can prove very little about the performance of these algorithms; in many cases provable performance guarantees for them appear to be beyond attack by our best proof techniques. This means that our current-best explanations for how such algorithms can be so successful resort eventually to (intelligent) guesswork and heuristic analysis. This is not only a problem for our intellectual honesty. As machine learning shows up in more and more mission- critical applications it is essential that we understand the guarantees and limits of our approaches with the certainty aorded only by rigorous mathematical analysis. The theory-of-computing community is only beginning to equip itself with the tools to make rigorous attacks on learning problems. The sophisticated learning pipelines used in practice are built layer-by-layer from simpler fundamental statistical inference algorithms. We can address these statistical inference problems rigorously, but only by rst landing on tractable problem de- nitions. In their traditional theory-of-computing formulations, many of these inference problems have intractable worst-case instances. -
Greedy Algorithms for the Maximum Satisfiability Problem: Simple Algorithms and Inapproximability Bounds∗
SIAM J. COMPUT. c 2017 Society for Industrial and Applied Mathematics Vol. 46, No. 3, pp. 1029–1061 GREEDY ALGORITHMS FOR THE MAXIMUM SATISFIABILITY PROBLEM: SIMPLE ALGORITHMS AND INAPPROXIMABILITY BOUNDS∗ MATTHIAS POLOCZEK†, GEORG SCHNITGER‡ , DAVID P. WILLIAMSON† , AND ANKE VAN ZUYLEN§ Abstract. We give a simple, randomized greedy algorithm for the maximum satisfiability 3 problem (MAX SAT) that obtains a 4 -approximation in expectation. In contrast to previously 3 known 4 -approximation algorithms, our algorithm does not use flows or linear programming. Hence we provide a positive answer to a question posed by Williamson in 1998 on whether such an algorithm 3 exists. Moreover, we show that Johnson’s greedy algorithm cannot guarantee a 4 -approximation, even if the variables are processed in a random order. Thereby we partially solve a problem posed by Chen, Friesen, and Zheng in 1999. In order to explore the limitations of the greedy paradigm, we use the model of priority algorithms of Borodin, Nielsen, and Rackoff. Since our greedy algorithm works in an online scenario where the variables arrive with their set of undecided clauses, we wonder if a better approximation ratio can be obtained by further fine-tuning its random decisions. For a particular information model we show that no priority algorithm can approximate Online MAX SAT 3 ε ε> within 4 + (for any 0). We further investigate the strength of deterministic greedy algorithms that may choose the variable ordering. Here we show that no adaptive priority algorithm can achieve 3 approximation ratio 4 . We propose two ways in which this inapproximability result can be bypassed.