The Parameterized Complexity of Counting Problems
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Computational Complexity Theory Introduction to Computational
31/03/2012 Computational complexity theory Introduction to computational complexity theory Complexity (computability) theory theory deals with two aspects: Algorithm’s complexity. Problem’s complexity. References S. Cook, « The complexity of Theorem Proving Procedures », 1971. Garey and Johnson, « Computers and Intractability, A guide to the theory of NP-completeness », 1979. J. Carlier et Ph. Chrétienne « Problèmes d’ordonnancements : algorithmes et complexité », 1988. 1 31/03/2012 Basic Notions • Some problem is a “question” characterized by parameters and needs an answer. – Parameters description; – Properties that a solutions must satisfy; – An instance is obtained when the parameters are fixed to some values. • An algorithm: a set of instructions describing how some task can be achieved or a problem can be solved. • A program : the computational implementation of an algorithm. Algorithm’s complexity (I) • There may exists several algorithms for the same problem • Raised questions: – Which one to choose ? – How they are compared ? – How measuring the efficiency ? – What are the most appropriate measures, running time, memory space ? 2 31/03/2012 Algorithm’s complexity (II) • Running time depends on: – The data of the problem, – Quality of program..., – Computer type, – Algorithm’s efficiency, – etc. • Proceed by analyzing the algorithm: – Search for some n characterizing the data. – Compute the running time in terms of n. – Evaluating the number of elementary operations, (elementary operation = simple instruction of a programming language). Algorithm’s evaluation (I) • Any algorithm is composed of two main stages: initialization and computing one • The complexity parameter is the size data n (binary coding). Definition: Let be n>0 andT(n) the running time of an algorithm expressed in terms of the size data n, T(n) is of O(f(n)) iff n0 and some constant c such that: n n0, we have T(n) c f(n). -
Submitted Thesis
Saarland University Faculty of Mathematics and Computer Science Bachelor’s Thesis Synthetic One-One, Many-One, and Truth-Table Reducibility in Coq Author Advisor Felix Jahn Yannick Forster Reviewers Prof. Dr. Gert Smolka Prof. Dr. Markus Bläser Submitted: September 15th, 2020 ii Eidesstattliche Erklärung Ich erkläre hiermit an Eides statt, dass ich die vorliegende Arbeit selbstständig ver- fasst und keine anderen als die angegebenen Quellen und Hilfsmittel verwendet habe. Statement in Lieu of an Oath I hereby confirm that I have written this thesis on my own and that I have not used any other media or materials than the ones referred to in this thesis. Einverständniserklärung Ich bin damit einverstanden, dass meine (bestandene) Arbeit in beiden Versionen in die Bibliothek der Informatik aufgenommen und damit veröffentlicht wird. Declaration of Consent I agree to make both versions of my thesis (with a passing grade) accessible to the public by having them added to the library of the Computer Science Department. Saarbrücken, September 15th, 2020 Abstract Reducibility is an essential concept for undecidability proofs in computability the- ory. The idea behind reductions was conceived by Turing, who introduced the later so-called Turing reduction based on oracle machines. In 1944, Post furthermore in- troduced with one-one, many-one, and truth-table reductions in comparison to Tur- ing reductions more specific reducibility notions. Post then also started to analyze the structure of the different reducibility notions and their computability degrees. Most undecidable problems were reducible from the halting problem, since this was exactly the method to show them undecidable. However, Post was able to con- struct also semidecidable but undecidable sets that do not one-one, many-one, or truth-table reduce from the halting problem. -
Module 34: Reductions and NP-Complete Proofs
Module 34: Reductions and NP-Complete Proofs This module 34 focuses on reductions and proof of NP-Complete problems. The module illustrates the ways of reducing one problem to another. Then, the module illustrates the outlines of proof of NP- Complete problems with some examples. The objectives of this module are To understand the concept of Reductions among problems To Understand proof of NP-Complete problems To overview proof of some Important NP-Complete problems. Computational Complexity There are two types of complexity theories. One is related to algorithms, known as algorithmic complexity theory and another related to complexity of problems called computational complexity theory [2,3]. Algorithmic complexity theory [3,4] aims to analyze the algorithms in terms of the size of the problem. In modules 3 and 4, we had discussed about these methods. The size is the length of the input. It can be recollected from module 3 that the size of a number n is defined to be the number of binary bits needed to write n. For example, Example: b(5) = b(1012) = 3. In other words, the complexity of the algorithm is stated in terms of the size of the algorithm. Asymptotic Analysis [1,2] refers to the study of an algorithm as the input size reaches a limit and analyzing the behaviour of the algorithms. The asymptotic analysis is also science of approximation where the behaviour of the algorithm is expressed in terms of notations such as big-oh, Big-Omega and Big- Theta. Computational complexity theory is different. Computational complexity aims to determine complexity of the problems itself. -
Fixed-Parameter Tractability Marko Samer and Stefan Szeider
Handbook of Satisfiability 363 Armin Biere, Marijn Heule, Hans van Maaren and Toby Walsch IOS Press, 2008 c 2008 Marko Samer and Stefan Szeider. All rights reserved. Chapter 13 Fixed-Parameter Tractability Marko Samer and Stefan Szeider 13.1. Introduction The propositional satisfiability problem (SAT) is famous for being the first prob- lem shown to be NP-complete—we cannot expect to find a polynomial-time al- gorithm for SAT. However, over the last decade, SAT-solvers have become amaz- ingly successful in solving formulas with thousands of variables that encode prob- lems arising from various application areas. Theoretical performance guarantees, however, are far from explaining this empirically observed efficiency. Actually, theorists believe that the trivial 2n time bound for solving SAT instances with n variables cannot be significantly improved, say to 2o(n) (see the end of Sec- tion 13.2). This enormous discrepancy between theoretical performance guaran- tees and the empirically observed performance of SAT solvers can be explained by the presence of a certain “hidden structure” in instances that come from applica- tions. This hidden structure greatly facilitates the propagation and simplification mechanisms of SAT solvers. Thus, for deriving theoretical performance guaran- tees that are closer to the actual performance of solvers one needs to take this hidden structure of instances into account. The literature contains several sug- gestions for making the vague term of a hidden structure explicit. For example, the hidden structure can be considered as the “tree-likeness” or “Horn-likeness” of the instance (below we will discuss how these notions can be made precise). -
Introduction to Parameterized Complexity
Introduction to Parameterized Complexity Ignasi Sau CNRS, LIRMM, Montpellier, France UFMG Belo Horizonte, February 2018 1/37 Outline of the talk 1 Why parameterized complexity? 2 Basic definitions 3 Kernelization 4 Some techniques 2/37 Next section is... 1 Why parameterized complexity? 2 Basic definitions 3 Kernelization 4 Some techniques 3/37 Karp (1972): list of 21 important NP-complete problems. Nowadays, literally thousands of problems are known to be NP-hard: unlessP = NP, they cannot be solved in polynomial time. But what does it mean for a problem to be NP-hard? No algorithm solves all instances optimally in polynomial time. Some history of complexity: NP-completeness Cook-Levin Theorem (1971): the SAT problem is NP-complete. 4/37 Nowadays, literally thousands of problems are known to be NP-hard: unlessP = NP, they cannot be solved in polynomial time. But what does it mean for a problem to be NP-hard? No algorithm solves all instances optimally in polynomial time. Some history of complexity: NP-completeness Cook-Levin Theorem (1971): the SAT problem is NP-complete. Karp (1972): list of 21 important NP-complete problems. 4/37 But what does it mean for a problem to be NP-hard? No algorithm solves all instances optimally in polynomial time. Some history of complexity: NP-completeness Cook-Levin Theorem (1971): the SAT problem is NP-complete. Karp (1972): list of 21 important NP-complete problems. Nowadays, literally thousands of problems are known to be NP-hard: unlessP = NP, they cannot be solved in polynomial time. 4/37 Some history of complexity: NP-completeness Cook-Levin Theorem (1971): the SAT problem is NP-complete. -
Exploiting C-Closure in Kernelization Algorithms for Graph Problems
Exploiting c-Closure in Kernelization Algorithms for Graph Problems Tomohiro Koana Technische Universität Berlin, Algorithmics and Computational Complexity, Germany [email protected] Christian Komusiewicz Philipps-Universität Marburg, Fachbereich Mathematik und Informatik, Marburg, Germany [email protected] Frank Sommer Philipps-Universität Marburg, Fachbereich Mathematik und Informatik, Marburg, Germany [email protected] Abstract A graph is c-closed if every pair of vertices with at least c common neighbors is adjacent. The c-closure of a graph G is the smallest number such that G is c-closed. Fox et al. [ICALP ’18] defined c-closure and investigated it in the context of clique enumeration. We show that c-closure can be applied in kernelization algorithms for several classic graph problems. We show that Dominating Set admits a kernel of size kO(c), that Induced Matching admits a kernel with O(c7k8) vertices, and that Irredundant Set admits a kernel with O(c5/2k3) vertices. Our kernelization exploits the fact that c-closed graphs have polynomially-bounded Ramsey numbers, as we show. 2012 ACM Subject Classification Theory of computation → Parameterized complexity and exact algorithms; Theory of computation → Graph algorithms analysis Keywords and phrases Fixed-parameter tractability, kernelization, c-closure, Dominating Set, In- duced Matching, Irredundant Set, Ramsey numbers Funding Frank Sommer: Supported by the Deutsche Forschungsgemeinschaft (DFG), project MAGZ, KO 3669/4-1. 1 Introduction Parameterized complexity [9, 14] aims at understanding which properties of input data can be used in the design of efficient algorithms for problems that are hard in general. The properties of input data are encapsulated in the notion of a parameter, a numerical value that can be attributed to each input instance I. -
Lower Bounds Based on the Exponential Time Hypothesis
Lower bounds based on the Exponential Time Hypothesis Daniel Lokshtanov∗ Dániel Marxy Saket Saurabhz Abstract In this article we survey algorithmic lower bound results that have been obtained in the field of exact exponential time algorithms and pa- rameterized complexity under certain assumptions on the running time of algorithms solving CNF-Sat, namely Exponential time hypothesis (ETH) and Strong Exponential time hypothesis (SETH). 1 Introduction The theory of NP-hardness gives us strong evidence that certain fundamen- tal combinatorial problems, such as 3SAT or 3-Coloring, are unlikely to be polynomial-time solvable. However, NP-hardness does not give us any information on what kind of super-polynomial running time is possible for NP-hard problems. For example, according to our current knowledge, the complexity assumption P 6= NP does not rule out the possibility of having an nO(log n) time algorithm for 3SAT or an nO(log log k) time algorithm for k- Clique, but such incredibly efficient super-polynomial algorithms would be still highly surprising. Therefore, in order to obtain qualitative lower bounds that rule out such algorithms, we need a complexity assumption stronger than P 6= NP. Impagliazzo, Paturi, and Zane [38, 37] introduced the Exponential Time Hypothesis (ETH) and the stronger variant, the Strong Exponential Time Hypothesis (SETH). These complexity assumptions state lower bounds on how fast satisfiability problems can be solved. These assumptions can be used as a basis for qualitative lower bounds for other concrete computational problems. ∗Dept. of Comp. Sc. and Engineering, University of California, USA. [email protected] yInstitut für Informatik, Humboldt-Universiät , Berlin, Germany. -
Decidable. (Recall It’S Not “Regular”.)
15-251: Great Theoretical Ideas in Computer Science Fall 2016 Lecture 6 September 15, 2016 Turing & the Uncomputable Comparing the cardinality of sets 퐴 ≤ 퐵 if there is an injection (one-to-one map) from 퐴 to 퐵 퐴 ≥ 퐵 if there is a surjection (onto map) from 퐴 to 퐵 퐴 = 퐵 if there is a bijection from 퐴 to 퐵 퐴 > |퐵| if there is no surjection from 퐵 to 퐴 (or equivalently, there is no injection from 퐴 to 퐵) Countable and uncountable sets countable countably infinite uncountable One slide guide to countability questions You are given a set 퐴 : is it countable or uncountable 퐴 ≤ |ℕ| or 퐴 > |ℕ| 퐴 ≤ |ℕ| : • Show directly surjection from ℕ to 퐴 • Show that 퐴 ≤ |퐵| where 퐵 ∈ {ℤ, ℤ x ℤ, ℚ, Σ∗, ℚ[x], …} 퐴 > |ℕ| : • Show directly using a diagonalization argument • Show that 퐴 ≥ | 0,1 ∞| Proving sets countable using computation For example, f(n) = ‘the nth prime’. You could write a program (Turing machine) to compute f. So this is a well-defined rule. Or: f(n) = the nth rational in our listing of ℚ. (List ℤ2 via the spiral, omit the terms p/0, omit rationals seen before…) You could write a program to compute this f. Poll Let 퐴 be the set of all languages over Σ = 1 ∗ Select the correct ones: - A is finite - A is infinite - A is countable - A is uncountable Another thing to remember from last week Encoding different objects with strings Fix some alphabet Σ . We use the ⋅ notation to denote the encoding of an object as a string in Σ∗ Examples: is the encoding a TM 푀 is the encoding a DFA 퐷 is the encoding of a pair of TMs 푀1, 푀2 is the encoding a pair 푀, 푥, where 푀 is a TM, and 푥 ∈ Σ∗ is an input to 푀 Uncountable to uncomputable The real number 1/7 is “computable”. -
12. Exponential Time Hypothesis COMP6741: Parameterized and Exact Computation
12. Exponential Time Hypothesis COMP6741: Parameterized and Exact Computation Serge Gaspers Semester 2, 2017 Contents 1 SAT and k-SAT 1 2 Subexponential time algorithms 2 3 ETH and SETH 2 4 Algorithmic lower bounds based on ETH 2 5 Algorithmic lower bounds based on SETH 3 6 Further Reading 4 1 SAT and k-SAT SAT SAT Input: A propositional formula F in conjunctive normal form (CNF) Parameter: n = jvar(F )j, the number of variables in F Question: Is there an assignment to var(F ) satisfying all clauses of F ? k-SAT Input: A CNF formula F where each clause has length at most k Parameter: n = jvar(F )j, the number of variables in F Question: Is there an assignment to var(F ) satisfying all clauses of F ? Example: (x1 _ x2) ^ (:x2 _ x3 _:x4) ^ (x1 _ x4) ^ (:x1 _:x3 _:x4) Algorithms for SAT • Brute-force: O∗(2n) • ... after > 50 years of SAT solving (SAT association, SAT conference, JSAT journal, annual SAT competitions, ...) • fastest known algorithm for SAT: O∗(2n·(1−1=O(log m=n))), where m is the number of clauses [Calabro, Impagli- azzo, Paturi, 2006] [Dantsin, Hirsch, 2009] • However: no O∗(1:9999n) time algorithm is known 1 • fastest known algorithms for 3-SAT: O∗(1:3303n) deterministic [Makino, Tamaki, Yamamoto, 2013] and O∗(1:3071n) randomized [Hertli, 2014] • Could it be that 3-SAT cannot be solved in 2o(n) time? • Could it be that SAT cannot be solved in O∗((2 − )n) time for any > 0? 2 Subexponential time algorithms NP-hard problems in subexponential time? • Are there any NP-hard problems that can be solved in 2o(n) time? • Yes. -
Computability on the Integers
Computability on the integers Laurent Bienvenu ( LIAFA, CNRS & Université de Paris 7 ) EJCIM 2011 Amiens, France 30 mars 2011 1. Basic objects of computability The formalization and study of the notion of computable function is what computability theory is about. As opposed to complexity theory, we do not care about efficiency, just about feasibility. Computable. functions What does it means for a function f : N ! N to be computable? 1. Basic objects of computability 3/79 As opposed to complexity theory, we do not care about efficiency, just about feasibility. Computable. functions What does it means for a function f : N ! N to be computable? The formalization and study of the notion of computable function is what computability theory is about. 1. Basic objects of computability 3/79 Computable. functions What does it means for a function f : N ! N to be computable? The formalization and study of the notion of computable function is what computability theory is about. As opposed to complexity theory, we do not care about efficiency, just about feasibility. 1. Basic objects of computability 3/79 But for us, it is now obvious that computable = realizable by a program / algorithm. Surprisingly, the first acceptable formalization (Turing machines) is still one of the best (if not the best) we know today. The. intuition At the time these questions were first considered (1930’s), computers did not exist (at least in the modern sense). 1. Basic objects of computability 4/79 Surprisingly, the first acceptable formalization (Turing machines) is still one of the best (if not the best) we know today. -
Computability Theory
CSC 438F/2404F Notes (S. Cook and T. Pitassi) Fall, 2019 Computability Theory This section is partly inspired by the material in \A Course in Mathematical Logic" by Bell and Machover, Chap 6, sections 1-10. Other references: \Introduction to the theory of computation" by Michael Sipser, and \Com- putability, Complexity, and Languages" by M. Davis and E. Weyuker. Our first goal is to give a formal definition for what it means for a function on N to be com- putable by an algorithm. Historically the first convincing such definition was given by Alan Turing in 1936, in his paper which introduced what we now call Turing machines. Slightly before Turing, Alonzo Church gave a definition based on his lambda calculus. About the same time G¨odel,Herbrand, and Kleene developed definitions based on recursion schemes. Fortunately all of these definitions are equivalent, and each of many other definitions pro- posed later are also equivalent to Turing's definition. This has lead to the general belief that these definitions have got it right, and this assertion is roughly what we now call \Church's Thesis". A natural definition of computable function f on N allows for the possibility that f(x) may not be defined for all x 2 N, because algorithms do not always halt. Thus we will use the symbol 1 to mean “undefined". Definition: A partial function is a function n f :(N [ f1g) ! N [ f1g; n ≥ 0 such that f(c1; :::; cn) = 1 if some ci = 1. In the context of computability theory, whenever we refer to a function on N, we mean a partial function in the above sense. -
On Miniaturized Problems in Parameterized Complexity Theory
On miniaturized problems in parameterized complexity theory Yijia Chen Institut fur¨ Informatik, Humboldt-Universitat¨ zu Berlin, Unter den Linden 6, 10099 Berlin, Germany. [email protected] Jorg¨ Flum Abteilung fur¨ Mathematische Logik, Universitat¨ Freiburg, Eckerstr. 1, 79104 Freiburg, Germany. [email protected] Abstract We introduce a general notion of miniaturization of a problem that comprises the different miniaturizations of concrete problems considered so far. We develop parts of the basic theory of miniaturizations. Using the appropriate logical formalism, we show that the miniaturization of a definable problem in W[t] lies in W[t], too. In particular, the miniaturization of the dominating set problem is in W[2]. Furthermore we investigate the relation between f(k) · no(k) time and subexponential time algorithms for the dominating set problem and for the clique problem. 1. Introduction Parameterized complexity theory provides a framework for a refined complexity analysis of algorithmic problems that are intractable in general. Central to the theory is the notion of fixed-parameter tractability, which relaxes the classical notion of tractability, polynomial time computability, by admitting algorithms whose runtime is exponential, but only in terms of some parameter that is usually expected to be small. Let FPT denote the class of all fixed-parameter tractable problems. A well-known example of a problem in FPT is the vertex cover problem, the parameter being the size of the vertex cover we ask for. As a complexity theoretic counterpart, a theory of parameterized intractability has been developed. In classical complex- ity, the notion of NP-completeness is central to a nice and simple theory for intractable problems.