NTIME Versus NSPACE

Total Page:16

File Type:pdf, Size:1020Kb

NTIME Versus NSPACE NTIME versus NSPACE Dr. Satyadev Nandakumar Department of Computer Science and Engineering, IITK Aditya Huddedar Department of Computer Science and Engineering, IITK Abstract Analysis using pebbling games is a very old technique used for exploring the relations between different com- plexity classes. They are of special importance to us in calculating the space complexities in many cases. We study the feasibility of an approach based on pebbling techniques to the open question of the relation between NTIME and NSPACE. We discuss one of the basic pebbling techniques used by Paul, Hopcroft and Valiant to relate deterministic time and space classes. We explore an equivalence between the existence of a \good" pebbling strategy for nondeterministic computation and existence of sub-linear separators for the multi-pushdown graphs. We survey why the recent construction of Monotone expander graphs forbids such separators, showing that pebbling strategies are unlikely to settle the separation question between nondeterministic time and nondeterministic space. 1 Introduction A pebble game is a type of mathematical game played by moving \pebbles" on a directed acyclic graph. A variety of different pebble games exist. A game known as \Black Pebbling" was developed in 1970 by Paul, Hopcroft and Valiant. We discuss it in details in this report. There has been a lot of pondering over the strategy of using these games for finding a non-trivial relation between nondeterministic time and nondeterministic space. We review one of these, namely BWR game and discuss why it is unlikely to settle the separation question. 2 Block Respecting Turing Machine [2] Here we describe a computational model that will be used often in the later discussion. The model was introduced by Paul, Hopcroft and Valiant. They introduced a nondeterministic turing machine simulating a deterministic k−tape Turing machine of time complexity t. 2.1 Construction 2 1. Partition each tape into blocks of size t 3 . 2. Restrict the motion of the tape heads. Tape heads can cross boundaries between adjacent blocks only at times 2 which are integer multiples of t 3 . 3. Add another tape with one block marked on it. The head on this new tape can simply move back and forth over 2 the block. Thus, it serves as a clock to indicate multiples of t 3 units of time. 1 4. We want to construct the modified Turing Machie so that it runs slower by at most a constant factor c. 5. For this reason, we mark block boundaries on the tapes. If tape head on any of the original tapes attempts to 2 cross a block boundary at a time other than a multiple of t 3 , the simulation temporarily halts and waits until the clock tape indicates the next multiple. There arises a situation in which the simulation takes asymptotically greater time that t. We describe it in brief here. 2.2 Problem A tape head may only move back and forth between two adjacent tape cells which are in different blocks waiting for 2 the clock signals each time. In this case the simulation would be slower by a factor of t 3 in the worst case. The modified simulation removes this possibility as follows. 2.3 Solution Let w1 w2 w3 .... wr th 2 2 be an the tape-contents of the j tape of the original Turing machine after mt 3 steps where jwlj = t 3 - Then the 2 corresponding inscription for the block respecting Turing machine after cmt 3 steps is r r r X w1 w2 .... wr−1 w1 w2 w3 .... wr r r r w2 w3 w4 .... X r 2 where w stands for w reversed. The extra tracks are used to guarantee that t 3 moves can be simulated without crossing a block boundary. In this case, even if we have a situation like{ a tape head simply moving back and forth between two adjacent tape cells which are in different blocks, we can simulate such a cross-move directly without waiting for the clock, as now the adjacent tape cells can be traversed on different tracks of the tape in the appropriate direction. To illustrate this point further, lets take an example: Suppose tape head is currently at the end of block 2 i.e. the word w2. If we have a right move from here, we will 2 have to wait until the time is a multiple of t 3 . In the new constructed model, we can simply simulate this step by rd just moving left in the same block. So that the 3 element of the tuple (a1; a2; a3) i.e. the leftmost bit of w3 can be accessed without waiting for the clock pulse. Now the next problem of updating the changes in blocks everywhere comes into picture. But it can be done in the 2 next (c−1)t 3 moves. We update the tape contents only when we change the block in the modified multitrack machine. We need to treaverse the current block twice to copy the content of one track once. Also we need to store the whole 4 content somewhere because copying bit-by-bit would take O(t 3 ) time. We utilise the clock tape for this purpose as 2 the contents of that tape do not have any significance.Thus, we can update all the modified blocks in (c − 1)t 3 moves 2 2 for some constant c > 0. Thus we can process a block of size t 3 in ct 3 moves. Thus, we simutate the time bound turing machine in ct moves. 3 Computation Garph [2] Now, we define a computation graph which is used till date to simulate the computation of Block Respecting Turing machines. Let M be a block-respecting deterministic k-tape Turing machine of time complexity t and Let w be an input for M. Lets divide the computation of M on input w into time segments ∆ of size t2=3. 2 Since the original Turing machine makes at most t moves and the block size is t2=3, there are at most t1=3 time segments ∆. To define the computation graph, we need some more notations. 3.1 Some Notations • h(i; ∆) = the position of the tape head on the ith tape of M after time segment ∆ and h(∆) = [h(1; ∆):::::h(k; ∆)] • The sequence of head positions h = [h(1):::::h(t1=3)] • v(∆) = the vertex corresponding to time segment ∆. th • For each tape i, ∆i = the last time-segment prior to ∆ such that the i head was scanning the same block during the segment ∆i, as during the segment ∆. • The edges of G = fv(∆ − 1) ! v(∆)g [ fv(∆i) ! v(∆)j1 ≤ i ≤ kg • c(i; ∆) = the content after time segment ∆ of a block on the ith tape of M scanned by the tape head during the time segment ∆. c(∆) = [c(1; ∆); :::; c(k; ∆)]. • f(∆) = the initial contents of the blocks which are visited by M for the first time during the time segment ∆. • With each vertex v(∆) in G we can associate the information c(∆) and f(∆). • q(∆) = the state of M after time segment ∆ and q = [q(1); :::; q(t1=3)] 3.2 Observations: 1. Since there are only t1=3 time segments ∆, the above graph has at most t1=3 vertices. Hence, to write down a description of the graph requires space of the order of (k + 1)t1=3 log t since approximately log t bits are needed to specify each edge. 2. In order to determine the outcome of the original computation of M , we need to only determine the state q(∆) after the final time segment ∆. If it is an accepting state accept, otherwise reject. 3. Because M is deterministic and block respecting, the following holds for any ∆: (a) q(∆); h(∆); and c(∆)can be uniquely determined from q(∆ − 1); h(∆ − 1); c(∆1); ::::; c(∆k); and f(∆) by direct simulation of time segment ∆. The simulation requires O(kt2=3) space to store the contents of k blocks. (b) To store c(∆) requires space 0(kt2=3). (c) In order to carry out the whole simulation, we do not need to keep in memory the contents of the blocks corresponding to all vertices since we can go back and reconstruct certain blocks whenever we need them. Note that this is a crucial observation which is again highlighted in coming sections. (d) Thus, we will focus on finding a simulation which requires the minimum number of blocks to be stored at any one time. The answer to this question lies in a game on graphs. 3 4 Pebbling Game [2] Let Gk be the set of all finite directed acyclic graphs with indegree at most k. We call the vertices with indegree 0 as the input vertices. The game consists of placing pebbles on the vertices of such a graph G according to the following rules: 1. You can always place pebble can on an input vertex. 2. You can place a pebble on a vertex if all its parents have pebbles on them.. 3. A pebble can be removed at any time. Note that as mentioned in the introduction, this game allows \sliding" of pebbles. The goal of the game is to eventually place a pebble on a particular vertex v,designated in advance, by a scheme which minimizes the maximum number of pebbles simultaneously on the graph at any instance of time. Let Pk(n) be the maximum over all graphs in Gk with n vertices of the number of pebbles required to place a pebble on an arbitrary vertex of such a graph.
Recommended publications
  • CS601 DTIME and DSPACE Lecture 5 Time and Space Functions: T, S
    CS601 DTIME and DSPACE Lecture 5 Time and Space functions: t, s : N → N+ Definition 5.1 A set A ⊆ U is in DTIME[t(n)] iff there exists a deterministic, multi-tape TM, M, and a constant c, such that, 1. A = L(M) ≡ w ∈ U M(w)=1 , and 2. ∀w ∈ U, M(w) halts within c · t(|w|) steps. Definition 5.2 A set A ⊆ U is in DSPACE[s(n)] iff there exists a deterministic, multi-tape TM, M, and a constant c, such that, 1. A = L(M), and 2. ∀w ∈ U, M(w) uses at most c · s(|w|) work-tape cells. (Input tape is “read-only” and not counted as space used.) Example: PALINDROMES ∈ DTIME[n], DSPACE[n]. In fact, PALINDROMES ∈ DSPACE[log n]. [Exercise] 1 CS601 F(DTIME) and F(DSPACE) Lecture 5 Definition 5.3 f : U → U is in F (DTIME[t(n)]) iff there exists a deterministic, multi-tape TM, M, and a constant c, such that, 1. f = M(·); 2. ∀w ∈ U, M(w) halts within c · t(|w|) steps; 3. |f(w)|≤|w|O(1), i.e., f is polynomially bounded. Definition 5.4 f : U → U is in F (DSPACE[s(n)]) iff there exists a deterministic, multi-tape TM, M, and a constant c, such that, 1. f = M(·); 2. ∀w ∈ U, M(w) uses at most c · s(|w|) work-tape cells; 3. |f(w)|≤|w|O(1), i.e., f is polynomially bounded. (Input tape is “read-only”; Output tape is “write-only”.
    [Show full text]
  • Interactive Proof Systems and Alternating Time-Space Complexity
    Theoretical Computer Science 113 (1993) 55-73 55 Elsevier Interactive proof systems and alternating time-space complexity Lance Fortnow” and Carsten Lund** Department of Computer Science, Unicersity of Chicago. 1100 E. 58th Street, Chicago, IL 40637, USA Abstract Fortnow, L. and C. Lund, Interactive proof systems and alternating time-space complexity, Theoretical Computer Science 113 (1993) 55-73. We show a rough equivalence between alternating time-space complexity and a public-coin interactive proof system with the verifier having a polynomial-related time-space complexity. Special cases include the following: . All of NC has interactive proofs, with a log-space polynomial-time public-coin verifier vastly improving the best previous lower bound of LOGCFL for this model (Fortnow and Sipser, 1988). All languages in P have interactive proofs with a polynomial-time public-coin verifier using o(log’ n) space. l All exponential-time languages have interactive proof systems with public-coin polynomial-space exponential-time verifiers. To achieve better bounds, we show how to reduce a k-tape alternating Turing machine to a l-tape alternating Turing machine with only a constant factor increase in time and space. 1. Introduction In 1981, Chandra et al. [4] introduced alternating Turing machines, an extension of nondeterministic computation where the Turing machine can make both existential and universal moves. In 1985, Goldwasser et al. [lo] and Babai [l] introduced interactive proof systems, an extension of nondeterministic computation consisting of two players, an infinitely powerful prover and a probabilistic polynomial-time verifier. The prover will try to convince the verifier of the validity of some statement.
    [Show full text]
  • The Complexity of Space Bounded Interactive Proof Systems
    The Complexity of Space Bounded Interactive Proof Systems ANNE CONDON Computer Science Department, University of Wisconsin-Madison 1 INTRODUCTION Some of the most exciting developments in complexity theory in recent years concern the complexity of interactive proof systems, defined by Goldwasser, Micali and Rackoff (1985) and independently by Babai (1985). In this paper, we survey results on the complexity of space bounded interactive proof systems and their applications. An early motivation for the study of interactive proof systems was to extend the notion of NP as the class of problems with efficient \proofs of membership". Informally, a prover can convince a verifier in polynomial time that a string is in an NP language, by presenting a witness of that fact to the verifier. Suppose that the power of the verifier is extended so that it can flip coins and can interact with the prover during the course of a proof. In this way, a verifier can gather statistical evidence that an input is in a language. As we will see, the interactive proof system model precisely captures this in- teraction between a prover P and a verifier V . In the model, the computation of V is probabilistic, but is typically restricted in time or space. A language is accepted by the interactive proof system if, for all inputs in the language, V accepts with high probability, based on the communication with the \honest" prover P . However, on inputs not in the language, V rejects with high prob- ability, even when communicating with a \dishonest" prover. In the general model, V can keep its coin flips secret from the prover.
    [Show full text]
  • Lecture 14 (Feb 28): Probabilistically Checkable Proofs (PCP) 14.1 Motivation: Some Problems and Their Approximation Algorithms
    CMPUT 675: Computational Complexity Theory Winter 2019 Lecture 14 (Feb 28): Probabilistically Checkable Proofs (PCP) Lecturer: Zachary Friggstad Scribe: Haozhou Pang 14.1 Motivation: Some problems and their approximation algorithms Many optimization problems are NP-hard, and therefore it is unlikely to efficiently find optimal solutions. However, in some situations, finding a provably good-enough solution (approximation of the optimal) is also acceptable. We say an algorithm is an α-approximation algorithm for an optimization problem iff for every instance of the problem it can find a solution whose cost is a factor of α of the optimum solution cost. Here are some selected problems and their approximation algorithms: Example: Min Vertex Cover is the problem that given a graph G = (V; E), the goal is to find a vertex cover C ⊆ V of minimum size. The following algorithm gives a 2-approximation for Min Vertex Cover: Algorithm 1 Min Vertex Cover Algorithm Input: Graph G = (V; E) Output: a vertex cover C of G. C ; while some edge (u; v) has u; v2 = C do C C [ fu; vg end while return C Claim 1 Let C∗ be an optimal solution, the returned solution C satisfies jCj ≤ 2jC∗j. Proof. Let M be the edges considered by the algorithm in the loop, we have jCj = 2jMj. Also, jC∗j covers M and no two edges in M share an endpoint (M is a matching), so jC∗j ≥ jMj. Therefore, jCj = 2jMj ≤ 2jC∗j. Example: Max SAT is the problem that given a CNF formula, the goal is to find a assignment that satisfies as many clauses as possible.
    [Show full text]
  • Complexity Theory
    Complexity Theory Course Notes Sebastiaan A. Terwijn Radboud University Nijmegen Department of Mathematics P.O. Box 9010 6500 GL Nijmegen the Netherlands [email protected] Copyright c 2010 by Sebastiaan A. Terwijn Version: December 2017 ii Contents 1 Introduction 1 1.1 Complexity theory . .1 1.2 Preliminaries . .1 1.3 Turing machines . .2 1.4 Big O and small o .........................3 1.5 Logic . .3 1.6 Number theory . .4 1.7 Exercises . .5 2 Basics 6 2.1 Time and space bounds . .6 2.2 Inclusions between classes . .7 2.3 Hierarchy theorems . .8 2.4 Central complexity classes . 10 2.5 Problems from logic, algebra, and graph theory . 11 2.6 The Immerman-Szelepcs´enyi Theorem . 12 2.7 Exercises . 14 3 Reductions and completeness 16 3.1 Many-one reductions . 16 3.2 NP-complete problems . 18 3.3 More decision problems from logic . 19 3.4 Completeness of Hamilton path and TSP . 22 3.5 Exercises . 24 4 Relativized computation and the polynomial hierarchy 27 4.1 Relativized computation . 27 4.2 The Polynomial Hierarchy . 28 4.3 Relativization . 31 4.4 Exercises . 32 iii 5 Diagonalization 34 5.1 The Halting Problem . 34 5.2 Intermediate sets . 34 5.3 Oracle separations . 36 5.4 Many-one versus Turing reductions . 38 5.5 Sparse sets . 38 5.6 The Gap Theorem . 40 5.7 The Speed-Up Theorem . 41 5.8 Exercises . 43 6 Randomized computation 45 6.1 Probabilistic classes . 45 6.2 More about BPP . 48 6.3 The classes RP and ZPP .
    [Show full text]
  • Lecture 10: Space Complexity III
    Space Complexity Classes: NL and L Reductions NL-completeness The Relation between NL and coNL A Relation Among the Complexity Classes Lecture 10: Space Complexity III Arijit Bishnu 27.03.2010 Space Complexity Classes: NL and L Reductions NL-completeness The Relation between NL and coNL A Relation Among the Complexity Classes Outline 1 Space Complexity Classes: NL and L 2 Reductions 3 NL-completeness 4 The Relation between NL and coNL 5 A Relation Among the Complexity Classes Space Complexity Classes: NL and L Reductions NL-completeness The Relation between NL and coNL A Relation Among the Complexity Classes Outline 1 Space Complexity Classes: NL and L 2 Reductions 3 NL-completeness 4 The Relation between NL and coNL 5 A Relation Among the Complexity Classes Definition for Recapitulation S c NPSPACE = c>0 NSPACE(n ). The class NPSPACE is an analog of the class NP. Definition L = SPACE(log n). Definition NL = NSPACE(log n). Space Complexity Classes: NL and L Reductions NL-completeness The Relation between NL and coNL A Relation Among the Complexity Classes Space Complexity Classes Definition for Recapitulation S c PSPACE = c>0 SPACE(n ). The class PSPACE is an analog of the class P. Definition L = SPACE(log n). Definition NL = NSPACE(log n). Space Complexity Classes: NL and L Reductions NL-completeness The Relation between NL and coNL A Relation Among the Complexity Classes Space Complexity Classes Definition for Recapitulation S c PSPACE = c>0 SPACE(n ). The class PSPACE is an analog of the class P. Definition for Recapitulation S c NPSPACE = c>0 NSPACE(n ).
    [Show full text]
  • Probabilistic Proof Systems: a Primer
    Probabilistic Proof Systems: A Primer Oded Goldreich Department of Computer Science and Applied Mathematics Weizmann Institute of Science, Rehovot, Israel. June 30, 2008 Contents Preface 1 Conventions and Organization 3 1 Interactive Proof Systems 4 1.1 Motivation and Perspective ::::::::::::::::::::::: 4 1.1.1 A static object versus an interactive process :::::::::: 5 1.1.2 Prover and Veri¯er :::::::::::::::::::::::: 6 1.1.3 Completeness and Soundness :::::::::::::::::: 6 1.2 De¯nition ::::::::::::::::::::::::::::::::: 7 1.3 The Power of Interactive Proofs ::::::::::::::::::::: 9 1.3.1 A simple example :::::::::::::::::::::::: 9 1.3.2 The full power of interactive proofs ::::::::::::::: 11 1.4 Variants and ¯ner structure: an overview ::::::::::::::: 16 1.4.1 Arthur-Merlin games a.k.a public-coin proof systems ::::: 16 1.4.2 Interactive proof systems with two-sided error ::::::::: 16 1.4.3 A hierarchy of interactive proof systems :::::::::::: 17 1.4.4 Something completely di®erent ::::::::::::::::: 18 1.5 On computationally bounded provers: an overview :::::::::: 18 1.5.1 How powerful should the prover be? :::::::::::::: 19 1.5.2 Computational Soundness :::::::::::::::::::: 20 2 Zero-Knowledge Proof Systems 22 2.1 De¯nitional Issues :::::::::::::::::::::::::::: 23 2.1.1 A wider perspective: the simulation paradigm ::::::::: 23 2.1.2 The basic de¯nitions ::::::::::::::::::::::: 24 2.2 The Power of Zero-Knowledge :::::::::::::::::::::: 26 2.2.1 A simple example :::::::::::::::::::::::: 26 2.2.2 The full power of zero-knowledge proofs ::::::::::::
    [Show full text]
  • Computability and Complexity Time and Space Complexity Definition Of
    Time and Space Complexity For practical solutions to computational problems available time, and Computability and Complexity available memory are two main considerations. Lecture 14 We have already studied time complexity, now we will focus on Space Complexity space (memory) complexity. Savitch’s Theorem Space Complexity Class PSPACE Question: How do we measure space complexity of a Turing machine? given by Jiri Srba Answer: The largest number of tape cells a Turing machine visits on all inputs of a given length n. Lecture 14 Computability and Complexity 1/15 Lecture 14 Computability and Complexity 2/15 Definition of Space Complexity The Complexity Classes SPACE and NSPACE Definition (Space Complexity of a TM) Definition (Complexity Class SPACE(t(n))) Let t : N R>0 be a function. Let M be a deterministic decider. The space complexity of M is a → function def SPACE(t(n)) = L(M) M is a decider running in space O(t(n)) f : N N → { | } where f (n) is the maximum number of tape cells that M scans on Definition (Complexity Class NSPACE(t(n))) any input of length n. Then we say that M runs in space f (n). Let t : N R>0 be a function. → def Definition (Space Complexity of a Nondeterministic TM) NSPACE(t(n)) = L(M) M is a nondetermin. decider running in space O(t(n)) Let M be a nondeterministic decider. The space complexity of M { | } is a function In other words: f : N N → SPACE(t(n)) is the class (collection) of languages that are where f (n) is the maximum number of tape cells that M scans on decidable by TMs running in space O(t(n)), and any branch of its computation on any input of length n.
    [Show full text]
  • Lecture 4: Space Complexity II: NL=Conl, Savitch's Theorem
    COM S 6810 Theory of Computing January 29, 2009 Lecture 4: Space Complexity II Instructor: Rafael Pass Scribe: Shuang Zhao 1 Recall from last lecture Definition 1 (SPACE, NSPACE) SPACE(S(n)) := Languages decidable by some TM using S(n) space; NSPACE(S(n)) := Languages decidable by some NTM using S(n) space. Definition 2 (PSPACE, NPSPACE) [ [ PSPACE := SPACE(nc), NPSPACE := NSPACE(nc). c>1 c>1 Definition 3 (L, NL) L := SPACE(log n), NL := NSPACE(log n). Theorem 1 SPACE(S(n)) ⊆ NSPACE(S(n)) ⊆ DTIME 2 O(S(n)) . This immediately gives that N L ⊆ P. Theorem 2 STCONN is NL-complete. 2 Today’s lecture Theorem 3 N L ⊆ SPACE(log2 n), namely STCONN can be solved in O(log2 n) space. Proof. Recall the STCONN problem: given digraph (V, E) and s, t ∈ V , determine if there exists a path from s to t. 4-1 Define boolean function Reach(u, v, k) with u, v ∈ V and k ∈ Z+ as follows: if there exists a path from u to v with length smaller than or equal to k, then Reach(u, v, k) = 1; otherwise Reach(u, v, k) = 0. It is easy to verify that there exists a path from s to t iff Reach(s, t, |V |) = 1. Next we show that Reach(s, t, |V |) can be recursively computed in O(log2 n) space. For all u, v ∈ V and k ∈ Z+: • k = 1: Reach(u, v, k) = 1 iff (u, v) ∈ E; • k > 1: Reach(u, v, k) = 1 iff there exists w ∈ V such that Reach(u, w, dk/2e) = 1 and Reach(w, v, bk/2c) = 1.
    [Show full text]
  • 26 Space Complexity
    CS:4330 Theory of Computation Spring 2018 Computability Theory Space Complexity Haniel Barbosa Readings for this lecture Chapter 8 of [Sipser 1996], 3rd edition. Sections 8.1, 8.2, and 8.3. Space Complexity B We now consider the complexity of computational problems in terms of the amount of space, or memory, they require B Time and space are two of the most important considerations when we seek practical solutions to most problems B Space complexity shares many of the features of time complexity B It serves a further way of classifying problems according to their computational difficulty 1 / 22 Space Complexity Definition Let M be a deterministic Turing machine, DTM, that halts on all inputs. The space complexity of M is the function f : N ! N, where f (n) is the maximum number of tape cells that M scans on any input of length n. Definition If M is a nondeterministic Turing machine, NTM, wherein all branches of its computation halt on all inputs, we define the space complexity of M, f (n), to be the maximum number of tape cells that M scans on any branch of its computation for any input of length n. 2 / 22 Estimation of space complexity Let f : N ! N be a function. The space complexity classes, SPACE(f (n)) and NSPACE(f (n)), are defined by: B SPACE(f (n)) = fL j L is a language decided by an O(f (n)) space DTMg B NSPACE(f (n)) = fL j L is a language decided by an O(f (n)) space NTMg 3 / 22 Example SAT can be solved with the linear space algorithm M1: M1 =“On input h'i, where ' is a Boolean formula: 1.
    [Show full text]
  • Introduction to Complexity Theory Big O Notation Review Linear Function: R(N)=O(N)
    GS019 - Lecture 1 on Complexity Theory Jarod Alper (jalper) Introduction to Complexity Theory Big O Notation Review Linear function: r(n)=O(n). Polynomial function: r(n)=2O(1) Exponential function: r(n)=2nO(1) Logarithmic function: r(n)=O(log n) Poly-log function: r(n)=logO(1) n Definition 1 (TIME) Let t : . Define the time complexity class, TIME(t(n)) to be ℵ−→ℵ TIME(t(n)) = L DTM M which decides L in time O(t(n)) . { |∃ } Definition 2 (NTIME) Let t : . Define the time complexity class, NTIME(t(n)) to be ℵ−→ℵ NTIME(t(n)) = L NDTM M which decides L in time O(t(n)) . { |∃ } Example 1 Consider the language A = 0k1k k 0 . { | ≥ } Is A TIME(n2)? Is A ∈ TIME(n log n)? Is A ∈ TIME(n)? Could∈ we potentially place A in a smaller complexity class if we consider other computational models? Theorem 1 If t(n) n, then every t(n) time multitape Turing machine has an equivalent O(t2(n)) time single-tape turing≥ machine. Proof: see Theorem 7:8 in Sipser (pg. 232) Theorem 2 If t(n) n, then every t(n) time RAM machine has an equivalent O(t3(n)) time multi-tape turing machine.≥ Proof: optional exercise Conclusion: Linear time is model specific; polynomical time is model indepedent. Definition 3 (The Class P ) k P = [ TIME(n ) k Definition 4 (The Class NP) GS019 - Lecture 1 on Complexity Theory Jarod Alper (jalper) k NP = [ NTIME(n ) k Equivalent Definition of NP NP = L L has a polynomial time verifier .
    [Show full text]
  • Circuit Lower Bounds for Merlin-Arthur Classes
    Electronic Colloquium on Computational Complexity, Report No. 5 (2007) Circuit Lower Bounds for Merlin-Arthur Classes Rahul Santhanam Simon Fraser University [email protected] January 16, 2007 Abstract We show that for each k > 0, MA/1 (MA with 1 bit of advice) doesn’t have circuits of size nk. This implies the first superlinear circuit lower bounds for the promise versions of the classes MA AM ZPPNP , and k . We extend our main result in several ways. For each k, we give an explicit language in (MA ∩ coMA)/1 which doesn’t have circuits of size nk. We also adapt our lower bound to the average-case setting, i.e., we show that MA/1 cannot be solved on more than 1/2+1/nk fraction of inputs of length n by circuits of size nk. Furthermore, we prove that MA does not have arithmetic circuits of size nk for any k. As a corollary to our main result, we obtain that derandomization of MA with O(1) advice implies the existence of pseudo-random generators computable using O(1) bits of advice. 1 Introduction Proving circuit lower bounds within uniform complexity classes is one of the most fundamental and challenging tasks in complexity theory. Apart from clarifying our understanding of the power of non-uniformity, circuit lower bounds have direct relevance to some longstanding open questions. Proving super-polynomial circuit lower bounds for NP would separate P from NP. The weaker result that for each k there is a language in NP which doesn’t have circuits of size nk would separate BPP from NEXP, thus answering an important question in the theory of derandomization.
    [Show full text]