On the Correspondence Between Arithmetic Theories and Propositional Proof Systems - a Survey
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The Top Eight Misconceptions About NP- Hardness
The top eight misconceptions about NP- hardness Zoltán Ádám Mann Budapest University of Technology and Economics Department of Computer Science and Information Theory Budapest, Hungary Abstract: The notions of NP-completeness and NP-hardness are used frequently in the computer science literature, but unfortunately, often in a wrong way. The aim of this paper is to show the most wide-spread misconceptions, why they are wrong, and why we should care. Keywords: F.1.3 Complexity Measures and Classes; F.2 Analysis of Algorithms and Problem Complexity; G.2.1.a Combinatorial algorithms; I.1.2.c Analysis of algorithms Published in IEEE Computer, volume 50, issue 5, pages 72-79, May 2017 1. Introduction Recently, I have been working on a survey about algorithmic approaches to virtual machine allocation in cloud data centers [11]. While doing so, I was shocked (again) to see how often the notions of NP-completeness or NP-hardness are used in a faulty, inadequate, or imprecise way in the literature. Here is an example from an – otherwise excellent – paper that appeared recently in IEEE Transactions on Computers, one of the most prestigious journals of the field: “Since the problem is NP-hard, no polynomial time optimal algorithm exists.” This claim, if it were shown to be true, would finally settle the famous P versus NP problem, which is one of the seven Millenium Prize Problems [10]; thus correctly solving it would earn the authors 1 million dollars. But of course, the authors do not attack this problem; the above claim is just an error in the paper. -
Circuit Complexity, Proof Complexity and Polynomial Identity Testing
Electronic Colloquium on Computational Complexity, Report No. 52 (2014) Circuit Complexity, Proof Complexity and Polynomial Identity Testing Joshua A. Grochow and Toniann Pitassi April 12, 2014 Abstract We introduce a new and very natural algebraic proof system, which has tight connections to (algebraic) circuit complexity. In particular, we show that any super-polynomial lower bound on any Boolean tautology in our proof system implies that the permanent does not have polynomial- size algebraic circuits (VNP 6= VP). As a corollary to the proof, we also show that super- polynomial lower bounds on the number of lines in Polynomial Calculus proofs (as opposed to the usual measure of number of monomials) imply the Permanent versus Determinant Conjecture. Note that, prior to our work, there was no proof system for which lower bounds on an arbitrary tautology implied any computational lower bound. Our proof system helps clarify the relationships between previous algebraic proof systems, and begins to shed light on why proof complexity lower bounds for various proof systems have been so much harder than lower bounds on the corresponding circuit classes. In doing so, we highlight the importance of polynomial identity testing (PIT) for understanding proof complex- ity. More specifically, we introduce certain propositional axioms satisfied by any Boolean circuit computing PIT. (The existence of efficient proofs for our PIT axioms appears to be somewhere in between the major conjecture that PIT2 P and the known result that PIT2 P=poly.) We use these PIT axioms to shed light on AC0[p]-Frege lower bounds, which have been open for nearly 30 years, with no satisfactory explanation as to their apparent difficulty. -
The Complexity Zoo
The Complexity Zoo Scott Aaronson www.ScottAaronson.com LATEX Translation by Chris Bourke [email protected] 417 classes and counting 1 Contents 1 About This Document 3 2 Introductory Essay 4 2.1 Recommended Further Reading ......................... 4 2.2 Other Theory Compendia ............................ 5 2.3 Errors? ....................................... 5 3 Pronunciation Guide 6 4 Complexity Classes 10 5 Special Zoo Exhibit: Classes of Quantum States and Probability Distribu- tions 110 6 Acknowledgements 116 7 Bibliography 117 2 1 About This Document What is this? Well its a PDF version of the website www.ComplexityZoo.com typeset in LATEX using the complexity package. Well, what’s that? The original Complexity Zoo is a website created by Scott Aaronson which contains a (more or less) comprehensive list of Complexity Classes studied in the area of theoretical computer science known as Computa- tional Complexity. I took on the (mostly painless, thank god for regular expressions) task of translating the Zoo’s HTML code to LATEX for two reasons. First, as a regular Zoo patron, I thought, “what better way to honor such an endeavor than to spruce up the cages a bit and typeset them all in beautiful LATEX.” Second, I thought it would be a perfect project to develop complexity, a LATEX pack- age I’ve created that defines commands to typeset (almost) all of the complexity classes you’ll find here (along with some handy options that allow you to conveniently change the fonts with a single option parameters). To get the package, visit my own home page at http://www.cse.unl.edu/~cbourke/. -
LSPACE VS NP IS AS HARD AS P VS NP 1. Introduction in Complexity
LSPACE VS NP IS AS HARD AS P VS NP FRANK VEGA Abstract. The P versus NP problem is a major unsolved problem in com- puter science. This consists in knowing the answer of the following question: Is P equal to NP? Another major complexity classes are LSPACE, PSPACE, ESPACE, E and EXP. Whether LSPACE = P is a fundamental question that it is as important as it is unresolved. We show if P = NP, then LSPACE = NP. Consequently, if LSPACE is not equal to NP, then P is not equal to NP. According to Lance Fortnow, it seems that LSPACE versus NP is easier to be proven. However, with this proof we show this problem is as hard as P versus NP. Moreover, we prove the complexity class P is not equal to PSPACE as a direct consequence of this result. Furthermore, we demonstrate if PSPACE is not equal to EXP, then P is not equal to NP. In addition, if E = ESPACE, then P is not equal to NP. 1. Introduction In complexity theory, a function problem is a computational problem where a single output is expected for every input, but the output is more complex than that of a decision problem [6]. A functional problem F is defined as a binary relation (x; y) 2 R over strings of an arbitrary alphabet Σ: R ⊂ Σ∗ × Σ∗: A Turing machine M solves F if for every input x such that there exists a y satisfying (x; y) 2 R, M produces one such y, that is M(x) = y [6]. -
Reflection Principles, Propositional Proof Systems, and Theories
Reflection principles, propositional proof systems, and theories In memory of Gaisi Takeuti Pavel Pudl´ak ∗ July 30, 2020 Abstract The reflection principle is the statement that if a sentence is provable then it is true. Reflection principles have been studied for first-order theories, but they also play an important role in propositional proof complexity. In this paper we will revisit some results about the reflection principles for propositional proofs systems using a finer scale of reflection principles. We will use the result that proving lower bounds on Resolution proofs is hard in Resolution. This appeared first in the recent article of Atserias and M¨uller [2] as a key lemma and was generalized and simplified in some spin- off papers [11, 13, 12]. We will also survey some results about arithmetical theories and proof systems associated with them. We will show a connection between a conjecture about proof complexity of finite consistency statements and a statement about proof systems associated with a theory. 1 Introduction arXiv:2007.14835v1 [math.LO] 29 Jul 2020 This paper is essentially a survey of some well-known results in proof complexity supple- mented with some observations. In most cases we will also sketch or give an idea of the proofs. Our aim is to focus on some interesting results rather than giving a complete ac- count of known results. We presuppose knowledge of basic concepts and theorems in proof complexity. An excellent source is Kraj´ıˇcek’s last book [19], where you can find the necessary definitions and read more about the results mentioned here. -
QP Versus NP
QP versus NP Frank Vega Joysonic, Belgrade, Serbia [email protected] Abstract. Given an instance of XOR 2SAT and a positive integer 2K , the problem exponential exclusive-or 2-satisfiability consists in deciding whether this Boolean formula has a truth assignment with at leat K satisfiable clauses. We prove exponential exclusive-or 2-satisfiability is in QP and NP-complete. In this way, we show QP ⊆ NP . Keywords: Complexity Classes · Completeness · Logarithmic Space · Nondeterministic · XOR 2SAT. 1 Introduction The P versus NP problem is a major unsolved problem in computer science [4]. This is considered by many to be the most important open problem in the field [4]. It is one of the seven Millennium Prize Problems selected by the Clay Mathematics Institute to carry a US$1,000,000 prize for the first correct solution [4]. It was essentially mentioned in 1955 from a letter written by John Nash to the United States National Security Agency [1]. However, the precise statement of the P = NP problem was introduced in 1971 by Stephen Cook in a seminal paper [4]. In 1936, Turing developed his theoretical computational model [14]. The deterministic and nondeterministic Turing machines have become in two of the most important definitions related to this theoretical model for computation [14]. A deterministic Turing machine has only one next action for each step defined in its program or transition function [14]. A nondeterministic Turing machine could contain more than one action defined for each step of its program, where this one is no longer a function, but a relation [14]. -
Contents Part 1. a Tour of Propositional Proof Complexity. 418 1. Is There A
The Bulletin of Symbolic Logic Volume 13, Number 4, Dec. 2007 THE COMPLEXITY OF PROPOSITIONAL PROOFS NATHAN SEGERLIND Abstract. Propositional proof complexity is the study of the sizes of propositional proofs, and more generally, the resources necessary to certify propositional tautologies. Questions about proof sizes have connections with computational complexity, theories of arithmetic, and satisfiability algorithms. This is article includes a broad survey of the field, and a technical exposition of some recently developed techniques for proving lower bounds on proof sizes. Contents Part 1. A tour of propositional proof complexity. 418 1. Is there a way to prove every tautology with a short proof? 418 2. Satisfiability algorithms and theories of arithmetic 420 3. A menagerie of Frege-like proof systems 427 4. Reverse mathematics of propositional principles 434 5. Feasible interpolation 437 6. Further connections with satisfiability algorithms 438 7. Beyond the Frege systems 441 Part 2. Some lower bounds on refutation sizes. 444 8. The size-width trade-off for resolution 445 9. The small restriction switching lemma 450 10. Expansion clean-up and random 3-CNFs 459 11. Resolution pseudowidth and very weak pigeonhole principles 467 Part 3. Open problems, further reading, acknowledgments. 471 Appendix. Notation. 472 References. 473 Received June 27, 2007. This survey article is based upon the author’s Sacks Prize winning PhD dissertation, however, the emphasis here is on context. The vast majority of results discussed are not the author’s, and several results from the author’s dissertation are omitted. c 2007, Association for Symbolic Logic 1079-8986/07/1304-0001/$7.50 417 418 NATHAN SEGERLIND Part 1. -
A Short History of Computational Complexity
The Computational Complexity Column by Lance FORTNOW NEC Laboratories America 4 Independence Way, Princeton, NJ 08540, USA [email protected] http://www.neci.nj.nec.com/homepages/fortnow/beatcs Every third year the Conference on Computational Complexity is held in Europe and this summer the University of Aarhus (Denmark) will host the meeting July 7-10. More details at the conference web page http://www.computationalcomplexity.org This month we present a historical view of computational complexity written by Steve Homer and myself. This is a preliminary version of a chapter to be included in an upcoming North-Holland Handbook of the History of Mathematical Logic edited by Dirk van Dalen, John Dawson and Aki Kanamori. A Short History of Computational Complexity Lance Fortnow1 Steve Homer2 NEC Research Institute Computer Science Department 4 Independence Way Boston University Princeton, NJ 08540 111 Cummington Street Boston, MA 02215 1 Introduction It all started with a machine. In 1936, Turing developed his theoretical com- putational model. He based his model on how he perceived mathematicians think. As digital computers were developed in the 40's and 50's, the Turing machine proved itself as the right theoretical model for computation. Quickly though we discovered that the basic Turing machine model fails to account for the amount of time or memory needed by a computer, a critical issue today but even more so in those early days of computing. The key idea to measure time and space as a function of the length of the input came in the early 1960's by Hartmanis and Stearns. -
Notes for Lecture 1
Stanford University | CS254: Computational Complexity Notes 1 Luca Trevisan January 7, 2014 Notes for Lecture 1 This course assumes CS154, or an equivalent course on automata, computability and computational complexity, as a prerequisite. We will assume that the reader is familiar with the notions of algorithm, running time, and of polynomial time reduction, as well as with basic notions of discrete math and probability. We will occasionally refer to Turing machines. In this lecture we give an (incomplete) overview of the field of Computational Complexity and of the topics covered by this course. Computational complexity is the part of theoretical computer science that studies 1. Impossibility results (lower bounds) for computations. Eventually, one would like the theory to prove its major conjectured lower bounds, and in particular, to prove Conjecture 1P 6= NP, implying that thousands of natural combinatorial problems admit no polynomial time algorithm 2. Relations between the power of different computational resources (time, memory ran- domness, communications) and the difficulties of different modes of computations (exact vs approximate, worst case versus average-case, etc.). A representative open problem of this type is Conjecture 2P = BPP, meaning that every polynomial time randomized algorithm admits a polynomial time deterministic simulation. Ultimately, one would like complexity theory to not only answer asymptotic worst-case complexity questions such as the P versus NP problem, but also address the complexity of specific finite-size instances, on average as well as in worst-case. Ideally, the theory would be eventually able to prove results such as • The smallest boolean circuit that solves 3SAT on formulas with 300 variables has size more than 270, or • The smallest boolean circuit that can factor more than half of the 2000-digit integers, has size more than 280. -
The P Versus Np Problem
THE P VERSUS NP PROBLEM STEPHEN COOK 1. Statement of the Problem The P versus NP problem is to determine whether every language accepted by some nondeterministic algorithm in polynomial time is also accepted by some (deterministic) algorithm in polynomial time. To define the problem precisely it is necessary to give a formal model of a computer. The standard computer model in computability theory is the Turing machine, introduced by Alan Turing in 1936 [37]. Although the model was introduced before physical computers were built, it nevertheless continues to be accepted as the proper computer model for the purpose of defining the notion of computable function. Informally the class P is the class of decision problems solvable by some algorithm within a number of steps bounded by some fixed polynomial in the length of the input. Turing was not concerned with the efficiency of his machines, rather his concern was whether they can simulate arbitrary algorithms given sufficient time. It turns out, however, Turing machines can generally simulate more efficient computer models (for example, machines equipped with many tapes or an unbounded random access memory) by at most squaring or cubing the computation time. Thus P is a robust class and has equivalent definitions over a large class of computer models. Here we follow standard practice and define the class P in terms of Turing machines. Formally the elements of the class P are languages. Let Σ be a finite alphabet (that is, a finite nonempty set) with at least two elements, and let Σ∗ be the set of finite strings over Σ. -
Proof Complexity in Algebraic Systems and Bounded Depth Frege Systems with Modular Counting
PROOF COMPLEXITY IN ALGEBRAIC SYSTEMS AND BOUNDED DEPTH FREGE SYSTEMS WITH MODULAR COUNTING S. Buss, R. Impagliazzo, J. Kraj¶³cek,· P. Pudlak,¶ A. A. Razborov and J. Sgall Abstract. We prove a lower bound of the form N (1) on the degree of polynomials in a Nullstellensatz refutation of the Countq polynomials over Zm, where q is a prime not dividing m. In addition, we give an explicit construction of a degree N (1) design for the Countq principle over Zm. As a corollary, using Beame et al. (1994) we obtain (1) a lower bound of the form 2N for the number of formulas in a constant-depth N Frege proof of the modular counting principle Countq from instances of the counting M principle Countm . We discuss the polynomial calculus proof system and give a method of converting tree-like polynomial calculus derivations into low degree Nullstellensatz derivations. Further we show that a lower bound for proofs in a bounded depth Frege system in the language with the modular counting connective MODp follows from a lower bound on the degree of Nullstellensatz proofs with a constant number of levels of extension axioms, where the extension axioms comprise a formalization of the approximation method of Razborov (1987), Smolensky (1987) (in fact, these two proof systems are basically equivalent). Introduction A propositional proof system is intuitively a system for establishing the validity of propo- sitional tautologies in some ¯xed complete language. The formal de¯nition of propositional proof system is that it is a polynomial time function f which maps strings over an alphabet § onto the set of propositional tautologies (Cook & Reckhow 1979). -
On Monotone Circuits with Local Oracles and Clique Lower Bounds
On monotone circuits with local oracles and clique lower bounds Jan Kraj´ıˇcek Igor C. Oliveira Faculty of Mathematics and Physics Charles University in Prague December 17, 2019 Abstract We investigate monotone circuits with local oracles [K., 2016], i.e., circuits containing additional inputs yi = yi(~x) that can perform unstructured computations on the input string ~x. Let µ [0, 1] be the locality of the circuit, a parameter that bounds the combined ∈ m strength of the oracle functions yi(~x), and Un,k, Vn,k 0, 1 be the set of k-cliques and the set of complete (k 1)-partite graphs, respectively⊆ (similarly { } to [Razborov, 1985]). Our results can be informally− stated as follows. (i) For an appropriate extension of depth-2 monotone circuits with local oracles, we show that the size of the smallest circuits separating Un,3 (triangles) and Vn,3 (complete bipartite graphs) undergoes two phase transitions according to µ. (ii) For 5 k(n) n1/4, arbitrary depth, and µ 1/50, we prove that the monotone ≤ ≤ ≤ Θ(√k) circuit size complexity of separating the sets Un,k and Vn,k is n , under a certain restrictive assumption on the local oracle gates. The second result, which concerns monotone circuits with restricted oracles, extends and provides a matching upper bound for the exponential lower bounds on the monotone circuit size complexity of k-clique obtained by Alon and Boppana (1987). 1 Introduction and motivation arXiv:1704.06241v3 [cs.CC] 16 Dec 2019 We establish initial lower bounds on the power of monotone circuits with local oracles (mono- tone CLOs), an extension of monotone circuits introduced in [Kra16] motivated by problems in proof complexity.