Memory Effects in Quantum Dynamics Modelled by Quantum Renewal Processes
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Simulating Quantum Field Theory with a Quantum Computer
Simulating quantum field theory with a quantum computer John Preskill Lattice 2018 28 July 2018 This talk has two parts (1) Near-term prospects for quantum computing. (2) Opportunities in quantum simulation of quantum field theory. Exascale digital computers will advance our knowledge of QCD, but some challenges will remain, especially concerning real-time evolution and properties of nuclear matter and quark-gluon plasma at nonzero temperature and chemical potential. Digital computers may never be able to address these (and other) problems; quantum computers will solve them eventually, though I’m not sure when. The physics payoff may still be far away, but today’s research can hasten the arrival of a new era in which quantum simulation fuels progress in fundamental physics. Frontiers of Physics short distance long distance complexity Higgs boson Large scale structure “More is different” Neutrino masses Cosmic microwave Many-body entanglement background Supersymmetry Phases of quantum Dark matter matter Quantum gravity Dark energy Quantum computing String theory Gravitational waves Quantum spacetime particle collision molecular chemistry entangled electrons A quantum computer can simulate efficiently any physical process that occurs in Nature. (Maybe. We don’t actually know for sure.) superconductor black hole early universe Two fundamental ideas (1) Quantum complexity Why we think quantum computing is powerful. (2) Quantum error correction Why we think quantum computing is scalable. A complete description of a typical quantum state of just 300 qubits requires more bits than the number of atoms in the visible universe. Why we think quantum computing is powerful We know examples of problems that can be solved efficiently by a quantum computer, where we believe the problems are hard for classical computers. -
Front Matter of the Book Contains a Detailed Table of Contents, Which We Encourage You to Browse
Cambridge University Press 978-1-107-00217-3 - Quantum Computation and Quantum Information: 10th Anniversary Edition Michael A. Nielsen & Isaac L. Chuang Frontmatter More information Quantum Computation and Quantum Information 10th Anniversary Edition One of the most cited books in physics of all time, Quantum Computation and Quantum Information remains the best textbook in this exciting field of science. This 10th Anniversary Edition includes a new Introduction and Afterword from the authors setting the work in context. This comprehensive textbook describes such remarkable effects as fast quantum algorithms, quantum teleportation, quantum cryptography, and quantum error-correction. Quantum mechanics and computer science are introduced, before moving on to describe what a quantum computer is, how it can be used to solve problems faster than “classical” computers, and its real-world implementation. It concludes with an in-depth treatment of quantum information. Containing a wealth of figures and exercises, this well-known textbook is ideal for courses on the subject, and will interest beginning graduate students and researchers in physics, computer science, mathematics, and electrical engineering. MICHAEL NIELSEN was educated at the University of Queensland, and as a Fulbright Scholar at the University of New Mexico. He worked at Los Alamos National Laboratory, as the Richard Chace Tolman Fellow at Caltech, was Foundation Professor of Quantum Information Science and a Federation Fellow at the University of Queensland, and a Senior Faculty Member at the Perimeter Institute for Theoretical Physics. He left Perimeter Institute to write a book about open science and now lives in Toronto. ISAAC CHUANG is a Professor at the Massachusetts Institute of Technology, jointly appointed in Electrical Engineering & Computer Science, and in Physics. -
Quantum State Complexity in Computationally Tractable Quantum Circuits
PRX QUANTUM 2, 010329 (2021) Quantum State Complexity in Computationally Tractable Quantum Circuits Jason Iaconis * Department of Physics and Center for Theory of Quantum Matter, University of Colorado, Boulder, Colorado 80309, USA (Received 28 September 2020; revised 29 December 2020; accepted 26 January 2021; published 23 February 2021) Characterizing the quantum complexity of local random quantum circuits is a very deep problem with implications to the seemingly disparate fields of quantum information theory, quantum many-body physics, and high-energy physics. While our theoretical understanding of these systems has progressed in recent years, numerical approaches for studying these models remains severely limited. In this paper, we discuss a special class of numerically tractable quantum circuits, known as quantum automaton circuits, which may be particularly well suited for this task. These are circuits that preserve the computational basis, yet can produce highly entangled output wave functions. Using ideas from quantum complexity theory, especially those concerning unitary designs, we argue that automaton wave functions have high quantum state complexity. We look at a wide variety of metrics, including measurements of the output bit-string distribution and characterization of the generalized entanglement properties of the quantum state, and find that automaton wave functions closely approximate the behavior of fully Haar random states. In addition to this, we identify the generalized out-of-time ordered 2k-point correlation functions as a particularly use- ful probe of complexity in automaton circuits. Using these correlators, we are able to numerically study the growth of complexity well beyond the scrambling time for very large systems. As a result, we are able to present evidence of a linear growth of design complexity in local quantum circuits, consistent with conjectures from quantum information theory. -
Quantum Computation and Complexity Theory
Quantum computation and complexity theory Course given at the Institut fÈurInformationssysteme, Abteilung fÈurDatenbanken und Expertensysteme, University of Technology Vienna, Wintersemester 1994/95 K. Svozil Institut fÈur Theoretische Physik University of Technology Vienna Wiedner Hauptstraûe 8-10/136 A-1040 Vienna, Austria e-mail: [email protected] December 5, 1994 qct.tex Abstract The Hilbert space formalism of quantum mechanics is reviewed with emphasis on applicationsto quantum computing. Standardinterferomeric techniques are used to construct a physical device capable of universal quantum computation. Some consequences for recursion theory and complexity theory are discussed. hep-th/9412047 06 Dec 94 1 Contents 1 The Quantum of action 3 2 Quantum mechanics for the computer scientist 7 2.1 Hilbert space quantum mechanics ..................... 7 2.2 From single to multiple quanta Ð ªsecondº ®eld quantization ...... 15 2.3 Quantum interference ............................ 17 2.4 Hilbert lattices and quantum logic ..................... 22 2.5 Partial algebras ............................... 24 3 Quantum information theory 25 3.1 Information is physical ........................... 25 3.2 Copying and cloning of qbits ........................ 25 3.3 Context dependence of qbits ........................ 26 3.4 Classical versus quantum tautologies .................... 27 4 Elements of quantum computatability and complexity theory 28 4.1 Universal quantum computers ....................... 30 4.2 Universal quantum networks ........................ 31 4.3 Quantum recursion theory ......................... 35 4.4 Factoring .................................. 36 4.5 Travelling salesman ............................. 36 4.6 Will the strong Church-Turing thesis survive? ............... 37 Appendix 39 A Hilbert space 39 B Fundamental constants of physics and their relations 42 B.1 Fundamental constants of physics ..................... 42 B.2 Conversion tables .............................. 43 B.3 Electromagnetic radiation and other wave phenomena ......... -
Bohmian Mechanics Versus Madelung Quantum Hydrodynamics
Ann. Univ. Sofia, Fac. Phys. Special Edition (2012) 112-119 [arXiv 0904.0723] Bohmian mechanics versus Madelung quantum hydrodynamics Roumen Tsekov Department of Physical Chemistry, University of Sofia, 1164 Sofia, Bulgaria It is shown that the Bohmian mechanics and the Madelung quantum hy- drodynamics are different theories and the latter is a better ontological interpre- tation of quantum mechanics. A new stochastic interpretation of quantum me- chanics is proposed, which is the background of the Madelung quantum hydro- dynamics. Its relation to the complex mechanics is also explored. A new complex hydrodynamics is proposed, which eliminates completely the Bohm quantum po- tential. It describes the quantum evolution of the probability density by a con- vective diffusion with imaginary transport coefficients. The Copenhagen interpretation of quantum mechanics is guilty for the quantum mys- tery and many strange phenomena such as the Schrödinger cat, parallel quantum and classical worlds, wave-particle duality, decoherence, etc. Many scientists have tried, however, to put the quantum mechanics back on ontological foundations. For instance, Bohm [1] proposed an al- ternative interpretation of quantum mechanics, which is able to overcome some puzzles of the Copenhagen interpretation. He developed further the de Broglie pilot-wave theory and, for this reason, the Bohmian mechanics is also known as the de Broglie-Bohm theory. At the time of inception of quantum mechanics Madelung [2] has demonstrated that the Schrödinger equa- tion can be transformed in hydrodynamic form. This so-called Madelung quantum hydrodynam- ics is a less elaborated theory and usually considered as a precursor of the Bohmian mechanics. The scope of the present paper is to show that these two theories are different and the Made- lung hydrodynamics is a better interpretation of quantum mechanics than the Bohmian me- chanics. -
Bounds on Errors in Observables Computed from Molecular Dynamics Simulations
Bounds on Errors in Observables Computed from Molecular Dynamics Simulations Vikram Sundar [email protected] Advisor: Dr. David Gelbwaser and Prof. Alan´ Aspuru-Guzik (Chemistry) Shadow Advisor: Prof. Martin Nowak Submitted in partial fulfillment of the honors requirements for the degree of Bachelor of Arts in Mathematics March 19, 2018 Contents Abstract iii Acknowledgments iv 1 Introduction1 1.1 Molecular Dynamics................................1 1.1.1 Why Molecular Dynamics?........................1 1.1.2 What is Molecular Dynamics?......................2 1.1.2.1 Force Fields and Integrators..................2 1.1.2.2 Observables...........................3 1.2 Structure of this Thesis...............................4 1.2.1 Sources of Error...............................4 1.2.2 Key Results of this Thesis.........................5 2 The St¨ormer-Verlet Method and Energy Conservation7 2.1 Symplecticity....................................8 2.1.1 Hamiltonian and Lagrangian Mechanics................8 2.1.2 Symplectic Structures on Manifolds...................9 2.1.3 Hamiltonian Flows are Symplectic.................... 10 2.1.4 Symplecticity of the Stormer-Verlet¨ Method............... 12 2.2 Backward Error Analysis.............................. 12 2.2.1 Symplectic Methods and Backward Error Analysis.......... 14 2.2.2 Bounds on Energy Conservation..................... 14 3 Integrable Systems and Bounds on Sampling Error 16 3.1 Integrable Systems and the Arnold-Liouville Theorem............. 17 3.1.1 Poisson Brackets and First Integrals................... 17 3.1.2 Liouville’s Theorem............................ 18 3.1.3 Action-Angle Coordinates......................... 20 3.1.4 Integrability and MD Force Fields.................... 22 3.2 Sampling Error................................... 23 3.2.1 Equivalence of Spatial and Time Averages............... 23 3.2.2 Bounding Sampling Error......................... 24 3.2.3 Reducing Sampling Error with Filter Functions........... -
Chapter 4 Information Theory
Chapter Information Theory Intro duction This lecture covers entropy joint entropy mutual information and minimum descrip tion length See the texts by Cover and Mackay for a more comprehensive treatment Measures of Information Information on a computer is represented by binary bit strings Decimal numb ers can b e represented using the following enco ding The p osition of the binary digit 3 2 1 0 Bit Bit Bit Bit Decimal Table Binary encoding indicates its decimal equivalent such that if there are N bits the ith bit represents N i the decimal numb er Bit is referred to as the most signicant bit and bit N as the least signicant bit To enco de M dierent messages requires log M bits 2 Signal Pro cessing Course WD Penny April Entropy The table b elow shows the probability of o ccurrence px to two decimal places of i selected letters x in the English alphab et These statistics were taken from Mackays i b o ok on Information Theory The table also shows the information content of a x px hx i i i a e j q t z Table Probability and Information content of letters letter hx log i px i which is a measure of surprise if we had to guess what a randomly chosen letter of the English alphab et was going to b e wed say it was an A E T or other frequently o ccuring letter If it turned out to b e a Z wed b e surprised The letter E is so common that it is unusual to nd a sentence without one An exception is the page novel Gadsby by Ernest Vincent Wright in which -
Guide for the Use of the International System of Units (SI)
Guide for the Use of the International System of Units (SI) m kg s cd SI mol K A NIST Special Publication 811 2008 Edition Ambler Thompson and Barry N. Taylor NIST Special Publication 811 2008 Edition Guide for the Use of the International System of Units (SI) Ambler Thompson Technology Services and Barry N. Taylor Physics Laboratory National Institute of Standards and Technology Gaithersburg, MD 20899 (Supersedes NIST Special Publication 811, 1995 Edition, April 1995) March 2008 U.S. Department of Commerce Carlos M. Gutierrez, Secretary National Institute of Standards and Technology James M. Turner, Acting Director National Institute of Standards and Technology Special Publication 811, 2008 Edition (Supersedes NIST Special Publication 811, April 1995 Edition) Natl. Inst. Stand. Technol. Spec. Publ. 811, 2008 Ed., 85 pages (March 2008; 2nd printing November 2008) CODEN: NSPUE3 Note on 2nd printing: This 2nd printing dated November 2008 of NIST SP811 corrects a number of minor typographical errors present in the 1st printing dated March 2008. Guide for the Use of the International System of Units (SI) Preface The International System of Units, universally abbreviated SI (from the French Le Système International d’Unités), is the modern metric system of measurement. Long the dominant measurement system used in science, the SI is becoming the dominant measurement system used in international commerce. The Omnibus Trade and Competitiveness Act of August 1988 [Public Law (PL) 100-418] changed the name of the National Bureau of Standards (NBS) to the National Institute of Standards and Technology (NIST) and gave to NIST the added task of helping U.S. -
Quantum Speedup for Aeroscience and Engineering
NASA/TM-2020-220590 Quantum Speedup for Aeroscience and Engineering Peyman Givi University of Pittsburgh, Pittsburgh, Pennsylvania Andrew J. Daley University of Strathclyde, Glasgow, United Kingdom Dimitri Mavriplis University of Wyoming, Laramie, Wyoming Mujeeb Malik Langley Research Center, Hampton, Virginia May 2020 NASA STI Program . in Profile Since its founding, NASA has been dedicated to • CONFERENCE PUBLICATION. Collected the advancement of aeronautics and space science. papers from scientific and technical The NASA scientific and technical information conferences, symposia, seminars, or other (STI) program plays a key part in helping NASA meetings sponsored or co-sponsored by maintain this important role. NASA. The NASA STI program operates under the • SPECIAL PUBLICATION. Scientific, auspices of the Agency Chief Information Officer. It technical, or historical information from collects, organizes, provides for archiving, and NASA programs, projects, and missions, often disseminates NASA’s STI. The NASA STI program concerned with subjects having substantial provides access to the NASA Aeronautics and Space public interest. Database and its public interface, the NASA Technical Report Server, thus providing one of the • TECHNICAL TRANSLATION. English- largest collections of aeronautical and space science language translations of foreign scientific and STI in the world. Results are published in both non- technical material pertinent to NASA’s NASA channels and by NASA in the NASA STI mission. Report Series, which includes the following report types: Specialized services also include creating custom thesauri, building customized databases, • TECHNICAL PUBLICATION. Reports of and organizing and publishing research results. completed research or a major significant phase of research that present the results of For more information about the NASA STI NASA programs and include extensive data or program, see the following: theoretical analysis. -
Understanding Shannon's Entropy Metric for Information
Understanding Shannon's Entropy metric for Information Sriram Vajapeyam [email protected] 24 March 2014 1. Overview Shannon's metric of "Entropy" of information is a foundational concept of information theory [1, 2]. Here is an intuitive way of understanding, remembering, and/or reconstructing Shannon's Entropy metric for information. Conceptually, information can be thought of as being stored in or transmitted as variables that can take on different values. A variable can be thought of as a unit of storage that can take on, at different times, one of several different specified values, following some process for taking on those values. Informally, we get information from a variable by looking at its value, just as we get information from an email by reading its contents. In the case of the variable, the information is about the process behind the variable. The entropy of a variable is the "amount of information" contained in the variable. This amount is determined not just by the number of different values the variable can take on, just as the information in an email is quantified not just by the number of words in the email or the different possible words in the language of the email. Informally, the amount of information in an email is proportional to the amount of “surprise” its reading causes. For example, if an email is simply a repeat of an earlier email, then it is not informative at all. On the other hand, if say the email reveals the outcome of a cliff-hanger election, then it is highly informative. -
Stochastic Quantum Dynamics and Relativity
Stochastic quantum dynamics and relativity Autor(en): Gisin, N. Objekttyp: Article Zeitschrift: Helvetica Physica Acta Band (Jahr): 62 (1989) Heft 4 PDF erstellt am: 25.09.2021 Persistenter Link: http://doi.org/10.5169/seals-116034 Nutzungsbedingungen Die ETH-Bibliothek ist Anbieterin der digitalisierten Zeitschriften. Sie besitzt keine Urheberrechte an den Inhalten der Zeitschriften. Die Rechte liegen in der Regel bei den Herausgebern. Die auf der Plattform e-periodica veröffentlichten Dokumente stehen für nicht-kommerzielle Zwecke in Lehre und Forschung sowie für die private Nutzung frei zur Verfügung. Einzelne Dateien oder Ausdrucke aus diesem Angebot können zusammen mit diesen Nutzungsbedingungen und den korrekten Herkunftsbezeichnungen weitergegeben werden. Das Veröffentlichen von Bildern in Print- und Online-Publikationen ist nur mit vorheriger Genehmigung der Rechteinhaber erlaubt. Die systematische Speicherung von Teilen des elektronischen Angebots auf anderen Servern bedarf ebenfalls des schriftlichen Einverständnisses der Rechteinhaber. Haftungsausschluss Alle Angaben erfolgen ohne Gewähr für Vollständigkeit oder Richtigkeit. Es wird keine Haftung übernommen für Schäden durch die Verwendung von Informationen aus diesem Online-Angebot oder durch das Fehlen von Informationen. Dies gilt auch für Inhalte Dritter, die über dieses Angebot zugänglich sind. Ein Dienst der ETH-Bibliothek ETH Zürich, Rämistrasse 101, 8092 Zürich, Schweiz, www.library.ethz.ch http://www.e-periodica.ch Helvetica Physica Acta, Vol. 62 (1989) 363-371 0018-0238/89/040363-09$l .50 + 0.20/0 © 1989 Birkhäuser Verlag, Basel Stochastic quantum dynamics and relativity By N. Gisin Groupe de Physique Appliquée, Université de Genève, 20 rue de l'Ecole de Médecine, 1211 Genève 4, Switzerland (13. I. 1989) Abstract. Our aim is to unify the Schrödinger dynamics and the projection postulate. -
Information, Entropy, and the Motivation for Source Codes
MIT 6.02 DRAFT Lecture Notes Last update: September 13, 2012 CHAPTER 2 Information, Entropy, and the Motivation for Source Codes The theory of information developed by Claude Shannon (MIT SM ’37 & PhD ’40) in the late 1940s is one of the most impactful ideas of the past century, and has changed the theory and practice of many fields of technology. The development of communication systems and networks has benefited greatly from Shannon’s work. In this chapter, we will first develop the intution behind information and formally define it as a mathematical quantity, then connect it to a related property of data sources, entropy. These notions guide us to efficiently compress a data source before communicating (or storing) it, and recovering the original data without distortion at the receiver. A key under lying idea here is coding, or more precisely, source coding, which takes each “symbol” being produced by any source of data and associates it with a codeword, while achieving several desirable properties. (A message may be thought of as a sequence of symbols in some al phabet.) This mapping between input symbols and codewords is called a code. Our focus will be on lossless source coding techniques, where the recipient of any uncorrupted mes sage can recover the original message exactly (we deal with corrupted messages in later chapters). ⌅ 2.1 Information and Entropy One of Shannon’s brilliant insights, building on earlier work by Hartley, was to realize that regardless of the application and the semantics of the messages involved, a general definition of information is possible.