In Search of an Easy Witness: Exponential Time Vs. Probabilistic

Total Page:16

File Type:pdf, Size:1020Kb

In Search of an Easy Witness: Exponential Time Vs. Probabilistic In Search of an Easy Witness Exp onential Time vs Probabilistic Polynomial Time y Valentine Kabanets Russell Impagliazzo Department of Computer Science Department of Computer Science University of California San Diego University of California San Diego La Jolla CA La Jolla CA kabanetscsucsdedu russellcsucsdedu z Avi Wigderson Department of Computer Science Hebrew University Jerusalem Israel and Institute for Advanced Study Princeton NJ aviiasedu May Abstract n Restricting the search space f g to the set of truth tables of easy Bo olean functions e establish on log n variables as well as using some known hardnessrandomness tradeos w a numb er of results relating the complexityof exp onentialtime and probabilistic p olynomial time complexity classes In particular we show that NEXP Ppoly NEXP MA this can be interpreted as saying that no derandomization of MA and hence of promiseBPP is p ossible unless NEXP contains a hard Bo olean function Wealsoprove several downward closure results for ZPP RP BPP and MA eg weshow EXP BPP EE BPE where EE is the doubleexp onential time class and BPE is the exp onentialtime analogue of BPP Intro duction One of the most imp ortant question in complexity theory is whether probabilistic algorithms are more p owerful than their deterministic counterparts A concrete formulation is the op en question of whether BPP P Despite growing evidence that BPP can be derandomized ie simulated deterministically without a signicant increase in the running time so far it has not b een ruled out that NEXP BPP Research supp orted byNSFAward CCR and USAIsrael BSF Grant y This researchwas done while the author was at the Department of Computer Science UniversityofToronto and at the Scho ol of Mathematics Institute for Advanced Study Princeton At IAS the researchwas supp orted by NSF grant DMS z This research was supp orted by grant numb er of the Israel Science Foundation founded by the Israel Academy for Sciences and Humanities and partially supp orted by NSF grants CCR and CCR Anumb er of conditional derandomization results are known which are based on the assumption that EXP contains hard Bo olean functions ie those of high circuit complexity NW BFNW O n ACR IW STV For instance it is shown in IW that BPP P if DTIME contains n a language that requires Bo olean circuits of size Results of this form usually called hardness randomness tradeos are proved by showing that the truth table of a hard Bo olean function can b e used to construct a pseudorandom generator which is then used to derandomize BPP or some other probabilistic complexity class It is well known that such pseudorandom generators exist if and only if there exist hard Bo olean functions in EXP However it is not known whether the existence of hard Bo olean functions in EXP is actually necessary for derandomizing BPP That is it is not known if BPP SUBEXP EXP Ppoly Obtaining such an implication would yield a normal form for derandomization b ecause hardnessvsrandomness results actually conclude that BPP can b e derandomized in a very sp ecic way Think of a probabilistic algorithm after xing its input x as dening a Bo olean function f r on the random bits r Since the algorithm is fast we know f is easy ie has low x x circuit complexity For an algorithm accepting a language L in BPP f is either almost always x if x L or almost always otherwise To decide which it suces to approximate the fraction of r s with f r to within a constant additive error To do this the derandomization rst x computes all p ossible sequences that are outputs of a generator G say r r and tries f r t x i for each i If G is a pseudorandom generator the nal output is the ma jority of the bits f r x i Other constructions such as the hittingset derandomization from ACR are more complicated but have the same general form In particular We never use the acceptance probability guarantees for the algorithm on other inputs Thus w e can derandomize algorithms even when acceptance separations arent guaranteed for all inputs ie we can derandomize promise BPP For KRC Intuitively this means that randomized heuristics that only p erform well on some inputs can also be simulated by a deterministic algorithm that p erforms well on the same inputs as the randomized algorithm The derandomization pro cedure only uses f as an oracle Although its correctness relies on x the existence of a small circuit computing f the circuit itself is only used in a blackbox x fashion Derandomization along the lines ab ove is equivalent to proving circuit lower b ounds which seems dicult One might hop e to achieve derandomization unconditionally by relaxing the ab ove restrictions In particular one could hop e that it is easier to approximate the acceptance probability using the circuit itself than it is treating it as a black box In fact recent results of a circuit indicate that in general having access to the circuit computing a function is stronger than having the function as an oracle BGI Bar However we show that this hop e is illfounded for nondeterministic algorithms solving the approximation problem for circuit acceptance oracle access is just as powerful as access to the circuit In particular any even nondeterministic derandomization of promise BPP yields a circuit lower b ound for NEXP and hence a blackb ox circuit approximation algorithm running in non deterministic sub exp onential time Thus unconditional results in derandomization require either er b ound for NEXP making a distinction b etween BPP and promise BPP or proving a circuit low More precisely we show that NEXP Ppoly MA NEXP and hence no derandomization of MA is p ossible unless there are hard functions in NEXP Since derandomizing promise BPP also allows one to derandomize MA the conclusion is that no full derandomization result is p ossible without assuming or proving circuit lower b ounds for NEXP Another piece of evidence that it will be dicult to show EXP BPP or NEXP MA comes from the downward closure results for these classes It is a basic fact in computational complexity that the equalities of complexity classes translate upwards For example if NP P then NEXP EXP by a simple padding argument Thus a separation at a higher level implies a separation at a lower level which suggests that higherlevel separations are probably harder to prove We show that separating EXP from BPP is as hard as separating their higher time complexity analogues More precisely we show that EXP BPP i EE BPE where EE is the O n O n class of languages accepted in deterministic time and BPE is the time analogue of BPP Weprove similar downward closures for ZPP RPandMA Main Techniques One of the main ideas that we use to derive our results can be informally describ ed as the easy witness metho d invented by Kabanets Kab It consists in searching that have for a desired ob ject eg a witness in a NEXP search problem among those ob jects concise descriptions eg truth tables of Bo olean functions of low circuit complexity Since there are few binary strings with small descriptions such a search is more ecient than the exhaustive search On the other hand if our search fails then we obtain a certain hardness test an ecient algorithm that accepts only those binary strings which do not have small descriptions With such a hardness test we can guess a truth table of a hard Bo olean function and then use it as a source of pseudorandomness via known hardnessrandomness tradeos Recall that the problem SuccinctSAT is to decide whether a prop ositional formula is satisable when given a Bo olean circuit which enco des the formula eg the truth table of the Bo olean function computed by the circuit is an enco ding of the prop ositional formula it is easy to see that SuccinctSATisNEXP complete Thus the idea of reducing the search space for NEXP problems to easy witnesses is suggested by the following natural question Is it true that every satisable prop ositional formula that is describ ed by a small Bo olean circuit must have at least one satisfying assignment that can also b e describ ed by a small Bo olean circuit We will show that this is indeed the case if NEXP Pp oly This idea was applied in Kab to RP search problems in order to obtain certain uniform setting derandomization of RP In this pap er we consider NEXP search problems which allows us to prove our results in the standard setting Remainder of the pap er In Section we present the necessary background In Section we describ e our main technical to ols In particular as an application of the easy witness metho d we show that nontrivial derandomization of AM can be achieved under the uniform complexity assumption that NEXP EXP cf Theorem where the class AM is a probabilistic version of NP see the next section for the denitions In Section we proveseveral results on complexityof NEXP In particular Section contains the pro of of the equivalence NEXP Ppoly NEXP MA In Section we show that every polyn NEXP search problem can be solved in deterministic time if NEXP AM we also prove that if NEXP Ppoly then every language in NEXP has memb ership witnesses of p olynomial circuit complexity from Section for the Section contains several interesting implications of our main result circuit approximation problem and natural pro ofs In Section we establish our downward closure results for ZPP RP BPP and MA We also prove gap theorems for ZPE BPE and MA in particular our gap theorem for ZPE states that Such closure results were also obtained byFortnow and Miltersen
Recommended publications
  • COMPLEXITY THEORY Review Lecture 13: Space Hierarchy and Gaps
    COMPLEXITY THEORY Review Lecture 13: Space Hierarchy and Gaps Markus Krotzsch¨ Knowledge-Based Systems TU Dresden, 26th Dec 2018 Markus Krötzsch, 26th Dec 2018 Complexity Theory slide 2 of 19 Review: Time Hierarchy Theorems Time Hierarchy Theorem 12.12 If f , g : N N are such that f is time- → constructible, and g log g o(f ), then · ∈ DTime (g) ( DTime (f ) ∗ ∗ Nondeterministic Time Hierarchy Theorem 12.14 If f , g : N N are such that f → is time-constructible, and g(n + 1) o(f (n)), then A Hierarchy for Space ∈ NTime (g) ( NTime (f ) ∗ ∗ In particular, we find that P , ExpTime and NP , NExpTime: , L NL P NP PSpace ExpTime NExpTime ExpSpace ⊆ ⊆ ⊆ ⊆ ⊆ ⊆ ⊆ , Markus Krötzsch, 26th Dec 2018 Complexity Theory slide 3 of 19 Markus Krötzsch, 26th Dec 2018 Complexity Theory slide 4 of 19 Space Hierarchy Proving the Space Hierarchy Theorem (1) Space Hierarchy Theorem 13.1: If f , g : N N are such that f is space- → constructible, and g o(f ), then For space, we can always assume a single working tape: ∈ Tape reduction leads to a constant-factor increase in space • DSpace(g) ( DSpace(f ) Constant factors can be eliminated by space compression • Proof: Again, we construct a diagonalisation machine . We define a multi-tape TM Therefore, DSpacek(f ) = DSpace1(f ). D D for inputs of the form , w (other cases do not matter), assuming that , w = n hM i |hM i| Compute f (n) in unary to mark the available space on the working tape Space turns out to be easier to separate – we get: • Space Hierarchy Theorem 13.1: If f , g : N N are such that f is space- Initialise a separate countdown tape with the largest binary number that can be → • constructible, and g o(f ), then written in f (n) space ∈ Simulate on , w , making sure that only previously marked tape cells are • M hM i DSpace(g) ( DSpace(f ) used Time-bound the simulation using the content of the countdown tape by • Challenge: TMs can run forever even within bounded space.
    [Show full text]
  • The Complexity Zoo
    The Complexity Zoo Scott Aaronson www.ScottAaronson.com LATEX Translation by Chris Bourke [email protected] 417 classes and counting 1 Contents 1 About This Document 3 2 Introductory Essay 4 2.1 Recommended Further Reading ......................... 4 2.2 Other Theory Compendia ............................ 5 2.3 Errors? ....................................... 5 3 Pronunciation Guide 6 4 Complexity Classes 10 5 Special Zoo Exhibit: Classes of Quantum States and Probability Distribu- tions 110 6 Acknowledgements 116 7 Bibliography 117 2 1 About This Document What is this? Well its a PDF version of the website www.ComplexityZoo.com typeset in LATEX using the complexity package. Well, what’s that? The original Complexity Zoo is a website created by Scott Aaronson which contains a (more or less) comprehensive list of Complexity Classes studied in the area of theoretical computer science known as Computa- tional Complexity. I took on the (mostly painless, thank god for regular expressions) task of translating the Zoo’s HTML code to LATEX for two reasons. First, as a regular Zoo patron, I thought, “what better way to honor such an endeavor than to spruce up the cages a bit and typeset them all in beautiful LATEX.” Second, I thought it would be a perfect project to develop complexity, a LATEX pack- age I’ve created that defines commands to typeset (almost) all of the complexity classes you’ll find here (along with some handy options that allow you to conveniently change the fonts with a single option parameters). To get the package, visit my own home page at http://www.cse.unl.edu/~cbourke/.
    [Show full text]
  • The Polynomial Hierarchy
    ij 'I '""T', :J[_ ';(" THE POLYNOMIAL HIERARCHY Although the complexity classes we shall study now are in one sense byproducts of our definition of NP, they have a remarkable life of their own. 17.1 OPTIMIZATION PROBLEMS Optimization problems have not been classified in a satisfactory way within the theory of P and NP; it is these problems that motivate the immediate extensions of this theory beyond NP. Let us take the traveling salesman problem as our working example. In the problem TSP we are given the distance matrix of a set of cities; we want to find the shortest tour of the cities. We have studied the complexity of the TSP within the framework of P and NP only indirectly: We defined the decision version TSP (D), and proved it NP-complete (corollary to Theorem 9.7). For the purpose of understanding better the complexity of the traveling salesman problem, we now introduce two more variants. EXACT TSP: Given a distance matrix and an integer B, is the length of the shortest tour equal to B? Also, TSP COST: Given a distance matrix, compute the length of the shortest tour. The four variants can be ordered in "increasing complexity" as follows: TSP (D); EXACTTSP; TSP COST; TSP. Each problem in this progression can be reduced to the next. For the last three problems this is trivial; for the first two one has to notice that the reduction in 411 j ;1 17.1 Optimization Problems 413 I 412 Chapter 17: THE POLYNOMIALHIERARCHY the corollary to Theorem 9.7 proving that TSP (D) is NP-complete can be used with DP.
    [Show full text]
  • Relations and Equivalences Between Circuit Lower Bounds and Karp-Lipton Theorems*
    Electronic Colloquium on Computational Complexity, Report No. 75 (2019) Relations and Equivalences Between Circuit Lower Bounds and Karp-Lipton Theorems* Lijie Chen Dylan M. McKay Cody D. Murray† R. Ryan Williams MIT MIT MIT Abstract A frontier open problem in circuit complexity is to prove PNP 6⊂ SIZE[nk] for all k; this is a neces- NP sary intermediate step towards NP 6⊂ P=poly. Previously, for several classes containing P , including NP NP NP , ZPP , and S2P, such lower bounds have been proved via Karp-Lipton-style Theorems: to prove k C 6⊂ SIZE[n ] for all k, we show that C ⊂ P=poly implies a “collapse” D = C for some larger class D, where we already know D 6⊂ SIZE[nk] for all k. It seems obvious that one could take a different approach to prove circuit lower bounds for PNP that does not require proving any Karp-Lipton-style theorems along the way. We show this intuition is wrong: (weak) Karp-Lipton-style theorems for PNP are equivalent to fixed-polynomial size circuit lower NP NP k NP bounds for P . That is, P 6⊂ SIZE[n ] for all k if and only if (NP ⊂ P=poly implies PH ⊂ i.o.-P=n ). Next, we present new consequences of the assumption NP ⊂ P=poly, towards proving similar re- sults for NP circuit lower bounds. We show that under the assumption, fixed-polynomial circuit lower bounds for NP, nondeterministic polynomial-time derandomizations, and various fixed-polynomial time simulations of NP are all equivalent. Applying this equivalence, we show that circuit lower bounds for NP imply better Karp-Lipton collapses.
    [Show full text]
  • Ex. 8 Complexity 1. by Savich Theorem NL ⊆ DSPACE (Log2n)
    Ex. 8 Complexity 1. By Savich theorem NL ⊆ DSP ACE(log2n) ⊆ DSP ASE(n). we saw in one of the previous ex. that DSP ACE(n2f(n)) 6= DSP ACE(f(n)) we also saw NL ½ P SP ASE so we conclude NL ½ DSP ACE(n) ½ DSP ACE(n3) ½ P SP ACE but DSP ACE(n) 6= DSP ACE(n3) as needed. 2. (a) We emulate the running of the s ¡ t ¡ conn algorithm in parallel for two paths (that is, we maintain two pointers, both starts at s and guess separately the next step for each of them). We raise a flag as soon as the two pointers di®er. We stop updating each of the two pointers as soon as it gets to t. If both got to t and the flag is raised, we return true. (b) As A 2 NL by Immerman-Szelepsc¶enyi Theorem (NL = vo ¡ NL) we know that A 2 NL. As C = A \ s ¡ t ¡ conn and s ¡ t ¡ conn 2 NL we get that A 2 NL. 3. If L 2 NICE, then taking \quit" as reject we get an NP machine (when x2 = L we always reject) for L, and by taking \quit" as accept we get coNP machine for L (if x 2 L we always reject), and so NICE ⊆ NP \ coNP . For the other direction, if L is in NP \coNP , then there is one NP machine M1(x; y) solving L, and another NP machine M2(x; z) solving coL. We can therefore de¯ne a new machine that guesses the guess y for M1, and the guess z for M2.
    [Show full text]
  • ZPP (Zero Probabilistic Polynomial)
    ZPPa (Zero Probabilistic Polynomial) • The class ZPP is defined as RP ∩ coRP. • A language in ZPP has two Monte Carlo algorithms, one with no false positives (RP) and the other with no false negatives (coRP). • If we repeatedly run both Monte Carlo algorithms, eventually one definite answer will come (unlike RP). – A positive answer from the one without false positives. – A negative answer from the one without false negatives. aGill (1977). c 2016 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 579 The ZPP Algorithm (Las Vegas) 1: {Suppose L ∈ ZPP.} 2: {N1 has no false positives, and N2 has no false negatives.} 3: while true do 4: if N1(x)=“yes”then 5: return “yes”; 6: end if 7: if N2(x) = “no” then 8: return “no”; 9: end if 10: end while c 2016 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 580 ZPP (concluded) • The expected running time for the correct answer to emerge is polynomial. – The probability that a run of the 2 algorithms does not generate a definite answer is 0.5 (why?). – Let p(n) be the running time of each run of the while-loop. – The expected running time for a definite answer is ∞ 0.5iip(n)=2p(n). i=1 • Essentially, ZPP is the class of problems that can be solved, without errors, in expected polynomial time. c 2016 Prof. Yuh-Dauh Lyuu, National Taiwan University Page 581 Large Deviations • Suppose you have a biased coin. • One side has probability 0.5+ to appear and the other 0.5 − ,forsome0<<0.5.
    [Show full text]
  • The Weakness of CTC Qubits and the Power of Approximate Counting
    The weakness of CTC qubits and the power of approximate counting Ryan O'Donnell∗ A. C. Cem Sayy April 7, 2015 Abstract We present results in structural complexity theory concerned with the following interre- lated topics: computation with postselection/restarting, closed timelike curves (CTCs), and approximate counting. The first result is a new characterization of the lesser known complexity class BPPpath in terms of more familiar concepts. Precisely, BPPpath is the class of problems that can be efficiently solved with a nonadaptive oracle for the Approximate Counting problem. Similarly, PP equals the class of problems that can be solved efficiently with nonadaptive queries for the related Approximate Difference problem. Another result is concerned with the compu- tational power conferred by CTCs; or equivalently, the computational complexity of finding stationary distributions for quantum channels. Using the above-mentioned characterization of PP, we show that any poly(n)-time quantum computation using a CTC of O(log n) qubits may as well just use a CTC of 1 classical bit. This result essentially amounts to showing that one can find a stationary distribution for a poly(n)-dimensional quantum channel in PP. ∗Department of Computer Science, Carnegie Mellon University. Work performed while the author was at the Bo˘gazi¸ciUniversity Computer Engineering Department, supported by Marie Curie International Incoming Fellowship project number 626373. yBo˘gazi¸ciUniversity Computer Engineering Department. 1 Introduction It is well known that studying \non-realistic" augmentations of computational models can shed a great deal of light on the power of more standard models. The study of nondeterminism and the study of relativization (i.e., oracle computation) are famous examples of this phenomenon.
    [Show full text]
  • Introduction to Complexity Theory Big O Notation Review Linear Function: R(N)=O(N)
    GS019 - Lecture 1 on Complexity Theory Jarod Alper (jalper) Introduction to Complexity Theory Big O Notation Review Linear function: r(n)=O(n). Polynomial function: r(n)=2O(1) Exponential function: r(n)=2nO(1) Logarithmic function: r(n)=O(log n) Poly-log function: r(n)=logO(1) n Definition 1 (TIME) Let t : . Define the time complexity class, TIME(t(n)) to be ℵ−→ℵ TIME(t(n)) = L DTM M which decides L in time O(t(n)) . { |∃ } Definition 2 (NTIME) Let t : . Define the time complexity class, NTIME(t(n)) to be ℵ−→ℵ NTIME(t(n)) = L NDTM M which decides L in time O(t(n)) . { |∃ } Example 1 Consider the language A = 0k1k k 0 . { | ≥ } Is A TIME(n2)? Is A ∈ TIME(n log n)? Is A ∈ TIME(n)? Could∈ we potentially place A in a smaller complexity class if we consider other computational models? Theorem 1 If t(n) n, then every t(n) time multitape Turing machine has an equivalent O(t2(n)) time single-tape turing≥ machine. Proof: see Theorem 7:8 in Sipser (pg. 232) Theorem 2 If t(n) n, then every t(n) time RAM machine has an equivalent O(t3(n)) time multi-tape turing machine.≥ Proof: optional exercise Conclusion: Linear time is model specific; polynomical time is model indepedent. Definition 3 (The Class P ) k P = [ TIME(n ) k Definition 4 (The Class NP) GS019 - Lecture 1 on Complexity Theory Jarod Alper (jalper) k NP = [ NTIME(n ) k Equivalent Definition of NP NP = L L has a polynomial time verifier .
    [Show full text]
  • The Computational Complexity Column by V
    The Computational Complexity Column by V. Arvind Institute of Mathematical Sciences, CIT Campus, Taramani Chennai 600113, India [email protected] http://www.imsc.res.in/~arvind The mid-1960’s to the early 1980’s marked the early epoch in the field of com- putational complexity. Several ideas and research directions which shaped the field at that time originated in recursive function theory. Some of these made a lasting impact and still have interesting open questions. The notion of lowness for complexity classes is a prominent example. Uwe Schöning was the first to study lowness for complexity classes and proved some striking results. The present article by Johannes Köbler and Jacobo Torán, written on the occasion of Uwe Schöning’s soon-to-come 60th birthday, is a nice introductory survey on this intriguing topic. It explains how lowness gives a unifying perspective on many complexity results, especially about problems that are of intermediate complexity. The survey also points to the several questions that still remain open. Lowness results: the next generation Johannes Köbler Humboldt Universität zu Berlin [email protected] Jacobo Torán Universität Ulm [email protected] Abstract Our colleague and friend Uwe Schöning, who has helped to shape the area of Complexity Theory in many decisive ways is turning 60 this year. As a little birthday present we survey in this column some of the newer results related with the concept of lowness, an idea that Uwe translated from the area of Recursion Theory in the early eighties. Originally this concept was applied to the complexity classes in the polynomial time hierarchy.
    [Show full text]
  • Computational Complexity
    Computational complexity Plan Complexity of a computation. Complexity classes DTIME(T (n)). Relations between complexity classes. Complete problems. Domino problems. NP-completeness. Complete problems for other classes Alternating machines. – p.1/17 Complexity of a computation A machine M is T (n) time bounded if for every n and word w of size n, every computation of M has length at most T (n). A machine M is S(n) space bounded if for every n and word w of size n, every computation of M uses at most S(n) cells of the working tape. Fact: If M is time or space bounded then L(M) is recursive. If L is recursive then there is a time and space bounded machine recognizing L. DTIME(T (n)) = fL(M) : M is det. and T (n) time boundedg NTIME(T (n)) = fL(M) : M is T (n) time boundedg DSPACE(S(n)) = fL(M) : M is det. and S(n) space boundedg NSPACE(S(n)) = fL(M) : M is S(n) space boundedg . – p.2/17 Hierarchy theorems A function S(n) is space constructible iff there is S(n)-bounded machine that given w writes aS(jwj) on the tape and stops. A function T (n) is time constructible iff there is a machine that on a word of size n makes exactly T (n) steps. Thm: Let S2(n) be a space-constructible function and let S1(n) ≥ log(n). If S1(n) lim infn!1 = 0 S2(n) then DSPACE(S2(n)) − DSPACE(S1(n)) is nonempty. Thm: Let T2(n) be a time-constructible function and let T1(n) log(T1(n)) lim infn!1 = 0 T2(n) then DTIME(T2(n)) − DTIME(T1(n)) is nonempty.
    [Show full text]
  • Complexity Theory
    COMPLEXITY THEORY Lecture 13: Space Hierarchy and Gaps Markus Krotzsch¨ Knowledge-Based Systems TU Dresden, 5th Dec 2017 Review Markus Krötzsch, 5th Dec 2017 Complexity Theory slide 2 of 19 Review: Time Hierarchy Theorems Time Hierarchy Theorem 12.12 If f , g : N ! N are such that f is time- constructible, and g · log g 2 o(f ), then DTime∗(g) ( DTime∗(f ) Nondeterministic Time Hierarchy Theorem 12.14 If f , g : N ! N are such that f is time-constructible, and g(n + 1) 2 o(f (n)), then NTime∗(g) ( NTime∗(f ) In particular, we find that P , ExpTime and NP , NExpTime: , L ⊆ NL ⊆ P ⊆ NP ⊆ PSpace ⊆ ExpTime ⊆ NExpTime ⊆ ExpSpace , Markus Krötzsch, 5th Dec 2017 Complexity Theory slide 3 of 19 A Hierarchy for Space Markus Krötzsch, 5th Dec 2017 Complexity Theory slide 4 of 19 Space Hierarchy For space, we can always assume a single working tape: • Tape reduction leads to a constant-factor increase in space • Constant factors can be eliminated by space compression Therefore, DSpacek(f ) = DSpace1(f ). Space turns out to be easier to separate – we get: Space Hierarchy Theorem 13.1: If f , g : N ! N are such that f is space- constructible, and g 2 o(f ), then DSpace(g) ( DSpace(f ) Challenge: TMs can run forever even within bounded space. Markus Krötzsch, 5th Dec 2017 Complexity Theory slide 5 of 19 Proof: Again, we construct a diagonalisation machine D. We define a multi-tape TM D for inputs of the form hM, wi (other cases do not matter), assuming that jhM, wij = n • Compute f (n) in unary to mark the available space on the working tape • Initialise
    [Show full text]
  • Quantum Zero-Error Algorithms Cannot Be Composed∗
    CORE Metadata, citation and similar papers at core.ac.uk Provided by CERN Document Server Quantum Zero-Error Algorithms Cannot be Composed∗ Harry Buhrman Ronald de Wolf CWI and U. of Amsterdam CWI [email protected] [email protected] Abstract We exhibit two black-box problems, both of which have an efficient quantum algorithm with zero-error, yet whose composition does not have an efficient quantum algorithm with zero-error. This shows that quantum zero-error algorithms cannot be composed. In oracle terms, we give a relativized world where ZQPZQP = ZQP, while classically we always have ZPPZPP =ZPP. 6 Keywords: Analysis of algorithms. Quantum computing. Zero-error computation. 1 Introduction We can define a “zero-error” algorithm of complexity T in two different but essentially equivalent ways: either as an algorithm that always outputs the correct value with expected complexity T (expectation taken over the internal randomness of the algorithm), or as an algorithm that outputs the correct value with probability at least 1=2, never outputs an incorrect value, and runs in worst- case complexity T . Expectation is linear, so we can compose two classical algorithms that have an efficient expected complexity to get another algorithm with efficient expected complexity. If algorithm A uses an expected number of a applications of B and an expected number of a0 other operations, then using a subroutine for B that has an expected number of b operations gives A an expected number of a b + a operations. In terms of complexity classes, we have · 0 ZPPZPP =ZPP; where ZPP is the class of problems that can be solved by a polynomial-time classical zero-error algorithm.
    [Show full text]