Quantum Leap a New Proof Supports a 25-Year-Old Claim N of the Unique Power of Quantum Computing
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Simulating Quantum Field Theory with a Quantum Computer
Simulating quantum field theory with a quantum computer John Preskill Lattice 2018 28 July 2018 This talk has two parts (1) Near-term prospects for quantum computing. (2) Opportunities in quantum simulation of quantum field theory. Exascale digital computers will advance our knowledge of QCD, but some challenges will remain, especially concerning real-time evolution and properties of nuclear matter and quark-gluon plasma at nonzero temperature and chemical potential. Digital computers may never be able to address these (and other) problems; quantum computers will solve them eventually, though I’m not sure when. The physics payoff may still be far away, but today’s research can hasten the arrival of a new era in which quantum simulation fuels progress in fundamental physics. Frontiers of Physics short distance long distance complexity Higgs boson Large scale structure “More is different” Neutrino masses Cosmic microwave Many-body entanglement background Supersymmetry Phases of quantum Dark matter matter Quantum gravity Dark energy Quantum computing String theory Gravitational waves Quantum spacetime particle collision molecular chemistry entangled electrons A quantum computer can simulate efficiently any physical process that occurs in Nature. (Maybe. We don’t actually know for sure.) superconductor black hole early universe Two fundamental ideas (1) Quantum complexity Why we think quantum computing is powerful. (2) Quantum error correction Why we think quantum computing is scalable. A complete description of a typical quantum state of just 300 qubits requires more bits than the number of atoms in the visible universe. Why we think quantum computing is powerful We know examples of problems that can be solved efficiently by a quantum computer, where we believe the problems are hard for classical computers. -
The Multiplicative Weights Update Method: a Meta-Algorithm and Applications
THEORY OF COMPUTING, Volume 8 (2012), pp. 121–164 www.theoryofcomputing.org RESEARCH SURVEY The Multiplicative Weights Update Method: A Meta-Algorithm and Applications Sanjeev Arora∗ Elad Hazan Satyen Kale Received: July 22, 2008; revised: July 2, 2011; published: May 1, 2012. Abstract: Algorithms in varied fields use the idea of maintaining a distribution over a certain set and use the multiplicative update rule to iteratively change these weights. Their analyses are usually very similar and rely on an exponential potential function. In this survey we present a simple meta-algorithm that unifies many of these disparate algorithms and derives them as simple instantiations of the meta-algorithm. We feel that since this meta-algorithm and its analysis are so simple, and its applications so broad, it should be a standard part of algorithms courses, like “divide and conquer.” ACM Classification: G.1.6 AMS Classification: 68Q25 Key words and phrases: algorithms, game theory, machine learning 1 Introduction The Multiplicative Weights (MW) method is a simple idea which has been repeatedly discovered in fields as diverse as Machine Learning, Optimization, and Game Theory. The setting for this algorithm is the following. A decision maker has a choice of n decisions, and needs to repeatedly make a decision and obtain an associated payoff. The decision maker’s goal, in the long run, is to achieve a total payoff which is comparable to the payoff of that fixed decision that maximizes the total payoff with the benefit of ∗This project was supported by David and Lucile Packard Fellowship and NSF grants MSPA-MCS 0528414 and CCR- 0205594. -
Limits on Efficient Computation in the Physical World
Limits on Efficient Computation in the Physical World by Scott Joel Aaronson Bachelor of Science (Cornell University) 2000 A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the GRADUATE DIVISION of the UNIVERSITY of CALIFORNIA, BERKELEY Committee in charge: Professor Umesh Vazirani, Chair Professor Luca Trevisan Professor K. Birgitta Whaley Fall 2004 The dissertation of Scott Joel Aaronson is approved: Chair Date Date Date University of California, Berkeley Fall 2004 Limits on Efficient Computation in the Physical World Copyright 2004 by Scott Joel Aaronson 1 Abstract Limits on Efficient Computation in the Physical World by Scott Joel Aaronson Doctor of Philosophy in Computer Science University of California, Berkeley Professor Umesh Vazirani, Chair More than a speculative technology, quantum computing seems to challenge our most basic intuitions about how the physical world should behave. In this thesis I show that, while some intuitions from classical computer science must be jettisoned in the light of modern physics, many others emerge nearly unscathed; and I use powerful tools from computational complexity theory to help determine which are which. In the first part of the thesis, I attack the common belief that quantum computing resembles classical exponential parallelism, by showing that quantum computers would face serious limitations on a wider range of problems than was previously known. In partic- ular, any quantum algorithm that solves the collision problem—that of deciding whether a sequence of n integers is one-to-one or two-to-one—must query the sequence Ω n1/5 times. -
Richard C. Brower, Boston University with Many Slides from Kate Clark, NVIDIA
Past and Future of QCD on GPUs Richard C. Brower, Boston University with many slides from Kate Clark, NVIDIA High Performance Computing in High Energy Physics CCNU Wahan China, Sept 19, 2018 Optimize the Intersection Application: QCD Algorithms: Architecture: AMG GPU Question to address • How do we put Quantum Field on the computer? • How to Maximize Flops/$, bandwidth/$ at Min Energy? • How to implement fastest algorithms: Dirac Sovlers, Sympletic Integrators, etc ? Standard Lattice QCD Formulation i d3xdt[ F 2 + (@ iA + m) ] Path Integral = 2A(x) (x) e− g2 µ⌫ µ µ − µ D Z R 1.Complex time for probability it x ! 4 iaAµ 2. Lattice Finite Differences (@µ iAµ) (x) ( x+ˆµ e x)/a − ! − d d e xDxy(A) y Det[D] 3. Fermionic integral − ! Z Det[D] dφdφ e φx[1/D]xyφy 4.Bosonic (pseudo- Fermions) ! − Z ij 1 γµ Lattice Dirac (x) − U ab(x) (x +ˆµ) ia 2 µ jb Color Dimension: a = 1,2,3 μ=1,2,…,d Spin i = 1,2,3,4 x x+ ➔ µ axis 2 x SU(3) Gauge Links ab ab iAµ (x) Uµ (x)=e x1 axis ➔ Quarks Propagation on Hypercubic Lattice* • Dominate Linear Algebra : Matrix solver for Dirac operator. • Gauge Evolution: In the semi-implicit quark Hamiltonian evolution in Monte vacuum Carlo time. u,d,s, proto u, • Analysis: proto d * Others: SUSY(Catteral et al), Random Lattices(Christ et al), Smiplicial Sphere (Brower et al) Byte/flop in Dirac Solver is main bottleneck • Bandwidth to memory rather than raw floating point throughput. • Wilson Dirac/DW operator (single prec) : 1440 bytes per 1320 flops. -
CS286.2 Lectures 5-6: Introduction to Hamiltonian Complexity, QMA-Completeness of the Local Hamiltonian Problem
CS286.2 Lectures 5-6: Introduction to Hamiltonian Complexity, QMA-completeness of the Local Hamiltonian problem Scribe: Jenish C. Mehta The Complexity Class BQP The complexity class BQP is the quantum analog of the class BPP. It consists of all languages that can be decided in quantum polynomial time. More formally, Definition 1. A language L 2 BQP if there exists a classical polynomial time algorithm A that ∗ maps inputs x 2 f0, 1g to quantum circuits Cx on n = poly(jxj) qubits, where the circuit is considered a sequence of unitary operators each on 2 qubits, i.e Cx = UTUT−1...U1 where each 2 2 Ui 2 L C ⊗ C , such that: 2 i. Completeness: x 2 L ) Pr(Cx accepts j0ni) ≥ 3 1 ii. Soundness: x 62 L ) Pr(Cx accepts j0ni) ≤ 3 We say that the circuit “Cx accepts jyi” if the first output qubit measured in Cxjyi is 0. More j0i specifically, letting P1 = j0ih0j1 be the projection of the first qubit on state j0i, j0i 2 Pr(Cx accepts jyi) =k (P1 ⊗ In−1)Cxjyi k2 The Complexity Class QMA The complexity class QMA (or BQNP, as Kitaev originally named it) is the quantum analog of the class NP. More formally, Definition 2. A language L 2 QMA if there exists a classical polynomial time algorithm A that ∗ maps inputs x 2 f0, 1g to quantum circuits Cx on n + q = poly(jxj) qubits, such that: 2q i. Completeness: x 2 L ) 9jyi 2 C , kjyik2 = 1, such that Pr(Cx accepts j0ni ⊗ 2 jyi) ≥ 3 2q 1 ii. -
Taking the Quantum Leap with Machine Learning
Introduction Quantum Algorithms Current Research Conclusion Taking the Quantum Leap with Machine Learning Zack Barnes University of Washington Bellevue College Mathematics and Physics Colloquium Series [email protected] January 15, 2019 Zack Barnes University of Washington UW Quantum Machine Learning Introduction Quantum Algorithms Current Research Conclusion Overview 1 Introduction What is Quantum Computing? What is Machine Learning? Quantum Power in Theory 2 Quantum Algorithms HHL Quantum Recommendation 3 Current Research Quantum Supremacy(?) 4 Conclusion Zack Barnes University of Washington UW Quantum Machine Learning Introduction Quantum Algorithms Current Research Conclusion What is Quantum Computing? \Quantum computing focuses on studying the problem of storing, processing and transferring information encoded in quantum mechanical systems.\ [Ciliberto, Carlo et al., 2018] Unit of quantum information is the qubit, or quantum binary integer. Zack Barnes University of Washington UW Quantum Machine Learning Supervised Uses labeled examples to predict future events Unsupervised Not classified or labeled Introduction Quantum Algorithms Current Research Conclusion What is Machine Learning? \Machine learning is the scientific study of algorithms and statistical models that computer systems use to progressively improve their performance on a specific task.\ (Wikipedia) Zack Barnes University of Washington UW Quantum Machine Learning Uses labeled examples to predict future events Unsupervised Not classified or labeled Introduction Quantum -
Interactive Proofs for Quantum Computations
Innovations in Computer Science 2010 Interactive Proofs For Quantum Computations Dorit Aharonov Michael Ben-Or Elad Eban School of Computer Science, The Hebrew University of Jerusalem, Israel [email protected] [email protected] [email protected] Abstract: The widely held belief that BQP strictly contains BPP raises fundamental questions: Upcoming generations of quantum computers might already be too large to be simulated classically. Is it possible to experimentally test that these systems perform as they should, if we cannot efficiently compute predictions for their behavior? Vazirani has asked [21]: If computing predictions for Quantum Mechanics requires exponential resources, is Quantum Mechanics a falsifiable theory? In cryptographic settings, an untrusted future company wants to sell a quantum computer or perform a delegated quantum computation. Can the customer be convinced of correctness without the ability to compare results to predictions? To provide answers to these questions, we define Quantum Prover Interactive Proofs (QPIP). Whereas in standard Interactive Proofs [13] the prover is computationally unbounded, here our prover is in BQP, representing a quantum computer. The verifier models our current computational capabilities: it is a BPP machine, with access to few qubits. Our main theorem can be roughly stated as: ”Any language in BQP has a QPIP, and moreover, a fault tolerant one” (providing a partial answer to a challenge posted in [1]). We provide two proofs. The simpler one uses a new (possibly of independent interest) quantum authentication scheme (QAS) based on random Clifford elements. This QPIP however, is not fault tolerant. Our second protocol uses polynomial codes QAS due to Ben-Or, Cr´epeau, Gottesman, Hassidim, and Smith [8], combined with quantum fault tolerance and secure multiparty quantum computation techniques. -
Quantum Supremacy
Quantum Supremacy Practical QS: perform some computational task on a well-controlled quantum device, which cannot be simulated in a reasonable time by the best-known classical algorithms and hardware. Theoretical QS: perform a computational task efficiently on a quantum device, and prove that task cannot be efficiently classically simulated. Since proving seems to be beyond the capabilities of our current civilization, we lower the standards for theoretical QS. One seeks to provide formal evidence that classical simulation is unlikely. For example: 3-SAT is NP-complete, so it cannot be efficiently classical solved unless P = NP. Theoretical QS: perform a computational task efficiently on a quantum device, and prove that task cannot be efficiently classically simulated unless “the polynomial Heierarchy collapses to the 3nd level.” Quantum Supremacy A common feature of QS arguments is that they consider sampling problems, rather than decision problems. They allow us to characterize the complexity of sampling measurements of quantum states. Which is more difficult: Task A: deciding if a circuit outputs 1 with probability at least 2/3s, or at most 1/3s Task B: sampling from the output of an n-qubit circuit in the computational basis Sampling from distributions is generically more difficult than approximating observables, since we can use samples to estimate observables, but not the other way around. One can imagine quantum systems whose local observables are easy to classically compute, but for which sampling the full state is computationally complex. By moving from decision problems to sampling problems, we make the task of classical simulation much more difficult. -
Quantum Algorithms for Classical Lattice Models
Home Search Collections Journals About Contact us My IOPscience Quantum algorithms for classical lattice models This content has been downloaded from IOPscience. Please scroll down to see the full text. 2011 New J. Phys. 13 093021 (http://iopscience.iop.org/1367-2630/13/9/093021) View the table of contents for this issue, or go to the journal homepage for more Download details: IP Address: 147.96.14.15 This content was downloaded on 16/12/2014 at 15:54 Please note that terms and conditions apply. New Journal of Physics The open–access journal for physics Quantum algorithms for classical lattice models G De las Cuevas1,2,5, W Dür1, M Van den Nest3 and M A Martin-Delgado4 1 Institut für Theoretische Physik, Universität Innsbruck, Technikerstraße 25, A-6020 Innsbruck, Austria 2 Institut für Quantenoptik und Quanteninformation der Österreichischen Akademie der Wissenschaften, Innsbruck, Austria 3 Max-Planck-Institut für Quantenoptik, Hans-Kopfermann-Strasse 1, D-85748 Garching, Germany 4 Departamento de Física Teórica I, Universidad Complutense, 28040 Madrid, Spain E-mail: [email protected] New Journal of Physics 13 (2011) 093021 (35pp) Received 15 April 2011 Published 9 September 2011 Online at http://www.njp.org/ doi:10.1088/1367-2630/13/9/093021 Abstract. We give efficient quantum algorithms to estimate the partition function of (i) the six-vertex model on a two-dimensional (2D) square lattice, (ii) the Ising model with magnetic fields on a planar graph, (iii) the Potts model on a quasi-2D square lattice and (iv) the Z2 lattice gauge theory on a 3D square lattice. -
Readout Rebalancing for Near Term Quantum Computers
Readout Rebalancing for Near Term Quantum Computers Rebecca Hicks,1, ∗ Christian W. Bauer,2, † and Benjamin Nachman2, ‡ 1Physics Department, University of California, Berkeley, Berkeley, CA 94720, USA 2Physics Division, Lawrence Berkeley National Laboratory, Berkeley, CA 94720, USA (Dated: October 16, 2020) Readout errors are a significant source of noise for near term intermediate scale quantum computers. Mismeasuring a qubit as a |1i when it should be |0i occurs much less often than mismeasuring a qubit as a |0i when it should have been |1i. We make the simple observation that one can improve the readout fidelity of quantum computers by applying targeted X gates prior to performing a measurement. These X gates are placed so that the expected number of qubits in the |1i state is minimized. Classical post processing can undo the effect of the X gates so that the expectation value of any observable is unchanged. We show that the statistical uncertainty following readout error corrections is smaller when using readout rebalancing. The statistical advantage is circuit- and computer-dependent, and is demonstrated for the W state, a Grover search, and for a Gaussian state. The benefit in statistical precision is most pronounced (and nearly a factor of two in some cases) when states with many qubits in the excited state have high probability. I. INTRODUCTION the simple observation that one can improve the readout fidelity of quantum computers by applying targeted X Quantum computers hold great promise for a variety of gates prior to performing a measurement. These X gates scientific and industrial applications. However, existing are placed so that the expected number of qubits in the noisy intermediate-scale quantum (NISQ) computers [1] |1i state is minimized. -
Expander Flows, Geometric Embeddings and Graph Partitioning
Expander Flows, Geometric Embeddings and Graph Partitioning SANJEEV ARORA Princeton University SATISH RAO and UMESH VAZIRANI UC Berkeley We give a O(√log n)-approximation algorithm for the sparsest cut, edge expansion, balanced separator,andgraph conductance problems. This improves the O(log n)-approximation of Leighton and Rao (1988). We use a well-known semidefinite relaxation with triangle inequality constraints. Central to our analysis is a geometric theorem about projections of point sets in d, whose proof makes essential use of a phenomenon called measure concentration. We also describe an interesting and natural “approximate certificate” for a graph’s expansion, which involves embedding an n-node expander in it with appropriate dilation and congestion. We call this an expander flow. Categories and Subject Descriptors: F.2.2 [Theory of Computation]: Analysis of Algorithms and Problem Complexity; G.2.2 [Mathematics of Computing]: Discrete Mathematics and Graph Algorithms General Terms: Algorithms,Theory Additional Key Words and Phrases: Graph Partitioning,semidefinite programs,graph separa- tors,multicommodity flows,expansion,expanders 1. INTRODUCTION Partitioning a graph into two (or more) large pieces while minimizing the size of the “interface” between them is a fundamental combinatorial problem. Graph partitions or separators are central objects of study in the theory of Markov chains, geometric embeddings and are a natural algorithmic primitive in numerous settings, including clustering, divide and conquer approaches, PRAM emulation, VLSI layout, and packet routing in distributed networks. Since finding optimal separators is NP-hard, one is forced to settle for approximation algorithms (see Shmoys [1995]). Here we give new approximation algorithms for some of the important problems in this class. -
Computational Pseudorandomness, the Wormhole Growth Paradox, and Constraints on the Ads/CFT Duality
Computational Pseudorandomness, the Wormhole Growth Paradox, and Constraints on the AdS/CFT Duality Adam Bouland Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, 617 Soda Hall, Berkeley, CA 94720, U.S.A. [email protected] Bill Fefferman Department of Computer Science, University of Chicago, 5730 S Ellis Ave, Chicago, IL 60637, U.S.A. [email protected] Umesh Vazirani Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, 671 Soda Hall, Berkeley, CA 94720, U.S.A. [email protected] Abstract The AdS/CFT correspondence is central to efforts to reconcile gravity and quantum mechanics, a fundamental goal of physics. It posits a duality between a gravitational theory in Anti de Sitter (AdS) space and a quantum mechanical conformal field theory (CFT), embodied in a map known as the AdS/CFT dictionary mapping states to states and operators to operators. This dictionary map is not well understood and has only been computed on special, structured instances. In this work we introduce cryptographic ideas to the study of AdS/CFT, and provide evidence that either the dictionary must be exponentially hard to compute, or else the quantum Extended Church-Turing thesis must be false in quantum gravity. Our argument has its origins in a fundamental paradox in the AdS/CFT correspondence known as the wormhole growth paradox. The paradox is that the CFT is believed to be “scrambling” – i.e. the expectation value of local operators equilibrates in polynomial time – whereas the gravity theory is not, because the interiors of certain black holes known as “wormholes” do not equilibrate and instead their volume grows at a linear rate for at least an exponential amount of time.