Two-Way Quantum Finite Automata and the Word Problem Zachary Remscrim Department of Mathematics, MIT, Cambridge, MA, USA [email protected]
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Quantum Algorithms 1
Quantum Algorithms 1 Computer scientists usually find conventional expositions of quantum computation difficult, since a lot of new concepts have to be learned to- gether to make sense of it all. In this course, we take a different approach, starting from a point where every computer science student is comfortable, and introducing the new concepts one by one into the familiar framework. We hope it will be much easier this way. So let's start with deterministic finite automata! Chapter 1 Finite automata 1.1 Deterministic model You already know the stuff in Sipser's book. Note that the transitions of a DFA can be represented as 0-1 matrices, the states can be traced with vectors, and the execution on an input string boils down to multiplying this vector with the corresponding sequence of those matrices. 1.2 Probabilistic model We define real-time1 probabilistic finite automata (rtPFAs) by generalizing the matrix definition of rtDFAs to column stochastic transition matrices. Consider a transition graph representation for a moment. It is as if compu- tation flows through many arrows (associated with the same input symbol) parallelly in each step. Under this \zero-error" regime (exact computation) rtPFAs are identical to rtDFAs: Since none of these computational paths can arrive at an incorrect response at the end, tracing any single path is suffi- cient for seeing what the machine will do. So we convert each rtPFA matrix (which have to contain at least one nonzero entry in each column) to a rtDFA matrix with a 1 in just the position corresponding to the topmost nonzero entry of the rtPFA in that column. -
Relations and Equivalences Between Circuit Lower Bounds and Karp-Lipton Theorems*
Electronic Colloquium on Computational Complexity, Report No. 75 (2019) Relations and Equivalences Between Circuit Lower Bounds and Karp-Lipton Theorems* Lijie Chen Dylan M. McKay Cody D. Murray† R. Ryan Williams MIT MIT MIT Abstract A frontier open problem in circuit complexity is to prove PNP 6⊂ SIZE[nk] for all k; this is a neces- NP sary intermediate step towards NP 6⊂ P=poly. Previously, for several classes containing P , including NP NP NP , ZPP , and S2P, such lower bounds have been proved via Karp-Lipton-style Theorems: to prove k C 6⊂ SIZE[n ] for all k, we show that C ⊂ P=poly implies a “collapse” D = C for some larger class D, where we already know D 6⊂ SIZE[nk] for all k. It seems obvious that one could take a different approach to prove circuit lower bounds for PNP that does not require proving any Karp-Lipton-style theorems along the way. We show this intuition is wrong: (weak) Karp-Lipton-style theorems for PNP are equivalent to fixed-polynomial size circuit lower NP NP k NP bounds for P . That is, P 6⊂ SIZE[n ] for all k if and only if (NP ⊂ P=poly implies PH ⊂ i.o.-P=n ). Next, we present new consequences of the assumption NP ⊂ P=poly, towards proving similar re- sults for NP circuit lower bounds. We show that under the assumption, fixed-polynomial circuit lower bounds for NP, nondeterministic polynomial-time derandomizations, and various fixed-polynomial time simulations of NP are all equivalent. Applying this equivalence, we show that circuit lower bounds for NP imply better Karp-Lipton collapses. -
The Computational Complexity of Nash Equilibria in Concisely Represented Games∗
Electronic Colloquium on Computational Complexity, Report No. 52 (2005) The Computational Complexity of Nash Equilibria in Concisely Represented Games∗ Grant R. Schoenebeck y Salil P. Vadhanz May 4, 2005 Abstract Games may be represented in many different ways, and different representations of games affect the complexity of problems associated with games, such as finding a Nash equilibrium. The traditional method of representing a game is to explicitly list all the payoffs, but this incurs an exponential blowup as the number of agents grows. We study two models of concisely represented games: circuit games, where the payoffs are computed by a given boolean circuit, and graph games, where each agent's payoff is a function of only the strategies played by its neighbors in a given graph. For these two models, we study the complexity of four questions: determining if a given strategy is a Nash equilibrium, finding a Nash equilibrium, determining if there exists a pure Nash equilibrium, and determining if there exists a Nash equilibrium in which the payoffs to a player meet some given guarantees. In many cases, we obtain tight results, showing that the problems are complete for various complexity classes. 1 Introduction In recent years, there has been a surge of interest at the interface between computer science and game theory. On one hand, game theory and its notions of equilibria provide a rich framework for modelling the behavior of selfish agents in the kinds of distributed or networked environments that often arise in computer science, and offer mechanisms to achieve efficient and desirable global outcomes in spite of the selfish behavior. -
Towards Non-Black-Box Lower Bounds in Cryptography
Towards Non-Black-Box Lower Bounds in Cryptography Rafael Pass?, Wei-Lung Dustin Tseng, and Muthuramakrishnan Venkitasubramaniam Cornell University, {rafael,wdtseng,vmuthu}@cs.cornell.edu Abstract. We consider average-case strengthenings of the traditional assumption that coNP is not contained in AM. Under these assumptions, we rule out generic and potentially non-black-box constructions of various cryptographic primitives (e.g., one-way permutations, collision-resistant hash-functions, constant-round statistically hiding commitments, and constant-round black-box zero-knowledge proofs for NP) from one-way functions, assuming the security reductions are black-box. 1 Introduction In the past four decades, many cryptographic tasks have been put under rigorous treatment in an eort to realize these tasks under minimal assumptions. In par- ticular, one-way functions are widely regarded as the most basic cryptographic primitive; their existence is implied by most other cryptographic tasks. Presently, one-way functions are known to imply schemes such as private-key encryp- tion [GM84,GGM86,HILL99], pseudo-random generators [HILL99], statistically- binding commitments [Nao91], statistically-hiding commitments [NOVY98,HR07] and zero-knowledge proofs [GMW91]. At the same time, some other tasks still have no known constructions based on one-way functions (e.g., key agreement schemes or collision-resistant hash functions). Following the seminal paper by Impagliazzo and Rudich [IR88], many works have addressed this phenomenon by demonstrating black-box separations, which rules out constructions of a cryptographic task using the underlying primitive as a black-box. For instance, Impagliazzo and Rudich rule out black-box con- structions of key-agreement protocols (and thus also trapdoor predicates) from one-way functions; Simon [Sim98] rules out black-box constructions of collision- resistant hash functions from one-way functions. -
Implementation of Elliptic Curve Cryptography in DNA Computing
International Journal of Scientific & Engineering Research Volume 8, Issue 6, June-2017 49 ISSN 2229-5518 Implementation of Elliptic Curve Cryptography in DNA computing 1Sourav Sinha, 2Shubhi Gupta 1Student: Department of Computer Science, 2Assistant Professor Amity University (dit school of Engineering) Greater Noida, India Abstract— DNA computing is the recent and powerful aspect of computer science Iin near future DNA computing is going to replace today’s silicon-based computing. In this paper, we are going to propose a method to implement Elliptic Curve Cryptography in DNA computing. Keywords—DNA computing; Elliptic Curve Cryptography 1. INTRODUCTION (HEADING 1) 2.2 DNA computer The hardware limitation of today’s computer is a barrier The DNA computer is different from Modern day’s in a development in technology. DNA computers, also classic computers. The DNA computer is nothing just a test known as molecular computer, have proven beneficial in tube containing a DNA and solvents for better mobility. such cases. Recent developments have seen massive The operations are done by chemical process and protein progress in technologies that enables a DNA computer to synthesis. DNA does not have any operational capacities, solve Hamiltonian Path Problem [1]. Cryptography and but it can be used as a hard drive to store and transfer data. network security is the most important section in development. Data Encryption Standard(DES) can also be 3. DNA-BASED ELLIPTIC CURVE ALGORITHM broken in a DNA computer, due to its ability to process The Elliptic curve cryptography makes use of an Elliptic parallel[2]. Curve to get the value of Its variable coefficients. -
Automata Modeling for Cognitive Interference in Users' Relevance
Quantum Informatics for Cognitive, Social, and Semantic Processes: Papers from the AAAI Fall Symposium (FS-10-08) Automata Modeling for Cognitive Interference in Users’ Relevance Judgment Peng Zhang1, Dawei Song1, Yuexian Hou1,2, Jun Wang1, Peter Bruza3 1School of Computing, The Robert Gordon University, UK. 2School of Computer Sci. & Tec., Tianjin University, China. 3Queensland University of Technology, Australia Email: [email protected] Abstract This paper is a step further along the direction of inves- tigating the quantum-like interference in a central IR pro- Quantum theory has recently been employed to further ad- cess - user’s relevance judgment, where users are involved vance the theory of information retrieval (IR). A challenging to judge the relevance degree of documents with respect to research topic is to investigate the so called quantum-like in- a given query. In this process, users’ relevance judgment terference in users’ relevance judgment process, where users are involved to judge the relevance degree of each document for the current document is often interfered by the judgment with respect to a given query. In this process, users’ relevance for previously seen documents, as a result of cognitive in- judgment for the current document is often interfered by the terference. A typical scenario is that, after finding a more judgment for previous documents, due to the interference on relevant document, a user may lower the relevance degree users’ cognitive status. Research from cognitive science has of a previously judged document. For example, when a user demonstrated some initial evidence of quantum-like cogni- judges whether a document d0 about “the theory of relativ- tive interference in human decision making, which underpins ity” is relevant to the query “Albert Einstein”, the user might the user’s relevance judgment process. -
Lower Bounds on the Running Time of Two-Way Quantum Finite Automata and Sublogarithmic-Space Quantum Turing Machines
Lower Bounds on the Running Time of Two-Way Quantum Finite Automata and Sublogarithmic-Space Quantum Turing Machines Zachary Remscrim Department of Computer Science, The University of Chicago, IL, USA [email protected] Abstract The two-way finite automaton with quantum and classical states (2QCFA), defined by Ambainis and Watrous, is a model of quantum computation whose quantum part is extremely limited; however, as they showed, 2QCFA are surprisingly powerful: a 2QCFA with only a single-qubit can recognize the ∗ O(n) language Lpal = {w ∈ {a, b} : w is a palindrome} with bounded error in expected time 2 . We prove that their result cannot be improved upon: a 2QCFA (of any size) cannot recognize o(n) Lpal with bounded error in expected time 2 . This is the first example of a language that can be recognized with bounded error by a 2QCFA in exponential time but not in subexponential time. Moreover, we prove that a quantum Turing machine (QTM) running in space o(log n) and expected n1−Ω(1) time 2 cannot recognize Lpal with bounded error; again, this is the first lower bound of its kind. Far more generally, we establish a lower bound on the running time of any 2QCFA or o(log n)-space QTM that recognizes any language L in terms of a natural “hardness measure” of L. This allows us to exhibit a large family of languages for which we have asymptotically matching lower and upper bounds on the running time of any such 2QCFA or QTM recognizer. 2012 ACM Subject Classification Theory of computation → Formal languages and automata theory; Theory of computation → Quantum computation theory Keywords and phrases Quantum computation, Lower bounds, Finite automata Digital Object Identifier 10.4230/LIPIcs.ITCS.2021.39 Related Version Full version of the paper https://arxiv.org/abs/2003.09877. -
Probabilistic and Quantum Finite Automata with Postselection
Probabilistic and quantum finite automata with ⋆ ⋆⋆ postselection ’ Abuzer Yakaryılmaz and A. C. Cem Say Bo˘gazi¸ci University, Department of Computer Engineering, Bebek 34342 Istanbul,˙ Turkey abuzer,[email protected] June 26, 2018 Abstract. We prove that endowing a real-time probabilistic or quantum computer with the ability of postselection increases its computational power. For this purpose, we provide a new model of finite automata with postselection, and compare it with the model of L¯ace et al. We examine the related language classes, and also establish separations between the classical and quantum versions, and between the zero-error vs. bounded- error modes of recognition in this model. 1 Introduction The notion of postselection as a mode of computation was introduced by Aaronson [1]. Postselection is the (unrealistic) capability of discarding all branches of a computation in which a specific event does not occur, and focusing on the surviving branches for the final decision about the membership of the input string in the recognized language. Aaronson examined PostBQP, the class of languages recognized with bounded error by polynomial-time quantum computers with postselection, and showed it to be identical to the well-known classical complexity class PP. It is, arXiv:1102.0666v1 [cs.CC] 3 Feb 2011 however, still an open question whether postselection adds anything to the power of polynomial-time computation, since we do not even know whether P, the class of languages recognized by classical computers with zero error in polynomial time, equals PP or not. In this paper, we prove that postselection is useful for real-time computers with a constant space bound, that is, finite automata. -
Quantum Finite Automata and Their Simulations
Quantum Finite Automata and Their Simulations Gustaw Lippa Krzysztof Makie la Marcin Kuta [email protected] Department of Computer Science, AGH University of Science and Technology, Krak´ow, Poland Krak´owQuantum Informatics Seminar (KQIS) 13 October 2020 Outline Motivation Quantum finite automata Library for simulating quantum finite automata Existing libraries Java Formal Languages and Automata Package (JFLAP) Quirk Quantum++ Q# Qiskit ProjectQ Probabilistic automata Probabilistic Finite Automaton (PFA) Quantum automata Measure-Once Quantum Finite Automaton (MO-QFA) Measure-Many Quantum Finite Automaton (MM-QFA) General Quantum Finite Automaton (GQFA) Types of Finite Automata Classical automata Deterministic Finite Automaton (DFA) Nondeterministic Finite Automaton (NFA) Alternating Finite Automaton (AFA) Quantum automata Measure-Once Quantum Finite Automaton (MO-QFA) Measure-Many Quantum Finite Automaton (MM-QFA) General Quantum Finite Automaton (GQFA) Types of Finite Automata Classical automata Deterministic Finite Automaton (DFA) Nondeterministic Finite Automaton (NFA) Alternating Finite Automaton (AFA) Probabilistic automata Probabilistic Finite Automaton (PFA) Types of Finite Automata Classical automata Deterministic Finite Automaton (DFA) Nondeterministic Finite Automaton (NFA) Alternating Finite Automaton (AFA) Probabilistic automata Probabilistic Finite Automaton (PFA) Quantum automata Measure-Once Quantum Finite Automaton (MO-QFA) Measure-Many Quantum Finite Automaton (MM-QFA) General Quantum Finite Automaton (GQFA) Deterministic -
Quantum Reversibility and a New Model of Quantum Automaton
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by CERN Document Server Quantum reversibility and a new model of quantum automaton Massimo Pica Ciamarra Dipartimento di Scienze Fisiche, Universit`a di Napoli \Federico II", Napoli, Italy. E-mail: [email protected] Abstract This article is an attempt to generalize the classical theory of reversible com- puting, principally developed by Bennet [IBM J. Res. Develop., 17(1973)] and by Fredkin and Toffoli [Internat. J. Theoret. Phys., 21(1982)], to the quantum case. This is a fundamental step towards the construction of a quantum computer be- cause a time efficient quantum computation is a reversible physical process. The paper is organized as follows. The first section reviews the classical theory of re- versible computing. In the second section it is showed that the designs used in the classical framework to decrease the consumption of space cannot be generalized to the quantum case; it is also suggested that quantum computing is generally more demanding of space than classical computing. In the last section a new model of fully quantum and reversible automaton is proposed. The computational power of this automaton is at least equal to that of classical automata. Some conclusion are drawn in the last section. 1 Reversible computing: a glance The classical theory of reversible computing has been analyzed extensively ([?, ?]). The main results are the following: 1. Every irreversible computation f : x f(x)canbe effectively enclosed in a reversible computation F :(x; 0) (f(x);x). →F is reversible because its input (x; 0) is uniquely determined by its output→ (f(x);x). -
Circuit Lower Bounds for Merlin-Arthur Classes
Electronic Colloquium on Computational Complexity, Report No. 5 (2007) Circuit Lower Bounds for Merlin-Arthur Classes Rahul Santhanam Simon Fraser University [email protected] January 16, 2007 Abstract We show that for each k > 0, MA/1 (MA with 1 bit of advice) doesn’t have circuits of size nk. This implies the first superlinear circuit lower bounds for the promise versions of the classes MA AM ZPPNP , and k . We extend our main result in several ways. For each k, we give an explicit language in (MA ∩ coMA)/1 which doesn’t have circuits of size nk. We also adapt our lower bound to the average-case setting, i.e., we show that MA/1 cannot be solved on more than 1/2+1/nk fraction of inputs of length n by circuits of size nk. Furthermore, we prove that MA does not have arithmetic circuits of size nk for any k. As a corollary to our main result, we obtain that derandomization of MA with O(1) advice implies the existence of pseudo-random generators computable using O(1) bits of advice. 1 Introduction Proving circuit lower bounds within uniform complexity classes is one of the most fundamental and challenging tasks in complexity theory. Apart from clarifying our understanding of the power of non-uniformity, circuit lower bounds have direct relevance to some longstanding open questions. Proving super-polynomial circuit lower bounds for NP would separate P from NP. The weaker result that for each k there is a language in NP which doesn’t have circuits of size nk would separate BPP from NEXP, thus answering an important question in the theory of derandomization. -
Arxiv:1111.3965V3 [Quant-Ph] 25 Jun 2012 Setting
Quantum measurement occurrence is undecidable J. Eisert,1 M. P. Muller,¨ 2 and C. Gogolin1 1Qmio Group, Dahlem Center for Complex Quantum Systems, Freie Universitat¨ Berlin, 14195 Berlin, Germany 2Perimeter Institute for Theoretical Physics, 31 Caroline Street North, Waterloo, Ontario N2L 2Y5, Canada In this work, we show that very natural, apparently simple problems in quantum measurement theory can be undecidable even if their classical analogues are decidable. Undecidability hence appears as a genuine quantum property here. Formally, an undecidable problem is a decision problem for which one cannot construct a single algorithm that will always provide a correct answer in finite time. The problem we consider is to determine whether sequentially used identical Stern-Gerlach-type measurement devices, giving rise to a tree of possible outcomes, have outcomes that never occur. Finally, we point out implications for measurement-based quantum computing and studies of quantum many-body models and suggest that a plethora of problems may indeed be undecidable. At the heart of the field of quantum information theory is the insight that the computational complexity of similar tasks in quantum and classical settings may be crucially different. Here we present an extreme example of this phenomenon: an operationally defined problem that is undecidable in the quan- tum setting but decidable in an even slightly more general classical analog. While the early focus in the field was on the assessment of tasks of quantum information processing, it has become increasingly clear that studies in computational complexity are also very fruitful when approaching problems FIG. 1. (Color online) The setting of sequential application of Stern- Gerlach-type devices considered here, gives rise to a tree of possible outside the realm of actual information processing, for exam- outcomes.