Newsletter of the Topical Group on Quantum Information American Physical Society
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Limits on Efficient Computation in the Physical World
Limits on Efficient Computation in the Physical World by Scott Joel Aaronson Bachelor of Science (Cornell University) 2000 A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in Computer Science in the GRADUATE DIVISION of the UNIVERSITY of CALIFORNIA, BERKELEY Committee in charge: Professor Umesh Vazirani, Chair Professor Luca Trevisan Professor K. Birgitta Whaley Fall 2004 The dissertation of Scott Joel Aaronson is approved: Chair Date Date Date University of California, Berkeley Fall 2004 Limits on Efficient Computation in the Physical World Copyright 2004 by Scott Joel Aaronson 1 Abstract Limits on Efficient Computation in the Physical World by Scott Joel Aaronson Doctor of Philosophy in Computer Science University of California, Berkeley Professor Umesh Vazirani, Chair More than a speculative technology, quantum computing seems to challenge our most basic intuitions about how the physical world should behave. In this thesis I show that, while some intuitions from classical computer science must be jettisoned in the light of modern physics, many others emerge nearly unscathed; and I use powerful tools from computational complexity theory to help determine which are which. In the first part of the thesis, I attack the common belief that quantum computing resembles classical exponential parallelism, by showing that quantum computers would face serious limitations on a wider range of problems than was previously known. In partic- ular, any quantum algorithm that solves the collision problem—that of deciding whether a sequence of n integers is one-to-one or two-to-one—must query the sequence Ω n1/5 times. -
Closed Timelike Curves Make Quantum and Classical Computing Equivalent
Closed Timelike Curves Make Quantum and Classical Computing Equivalent Scott Aaronson∗ John Watrous† MIT University of Waterloo Abstract While closed timelike curves (CTCs) are not known to exist, studying their consequences has led to nontrivial insights in general relativity, quantum information, and other areas. In this paper we show that if CTCs existed, then quantum computers would be no more powerful than classical computers: both would have the (extremely large) power of the complexity class PSPACE, consisting of all problems solvable by a conventional computer using a polynomial amount of memory. This solves an open problem proposed by one of us in 2005, and gives an essentially complete understanding of computational complexity in the presence of CTCs. Following the work of Deutsch, we treat a CTC as simply a region of spacetime where a “causal consistency” condition is imposed, meaning that Nature has to produce a (probabilistic or quantum) fixed-point of some evolution operator. Our conclusion is then a consequence of the following theorem: given any quantum circuit (not necessarily unitary), a fixed-point of the circuit can be (implicitly) computed in polynomial space. This theorem might have independent applications in quantum information. 1 Introduction The possibility of closed timelike curves (CTCs) within general relativity and quantum gravity theories has been studied for almost a century [11, 15, 13]. A different line of research has sought to understand the implications of CTCs, supposing they existed, for quantum mechanics, computation, and information [9, 8, 5]. In this paper we contribute to the latter topic, by giving the first complete characterization of the computational power of CTCs. -
Computational Pseudorandomness, the Wormhole Growth Paradox, and Constraints on the Ads/CFT Duality
Computational Pseudorandomness, the Wormhole Growth Paradox, and Constraints on the AdS/CFT Duality Adam Bouland Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, 617 Soda Hall, Berkeley, CA 94720, U.S.A. [email protected] Bill Fefferman Department of Computer Science, University of Chicago, 5730 S Ellis Ave, Chicago, IL 60637, U.S.A. [email protected] Umesh Vazirani Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, 671 Soda Hall, Berkeley, CA 94720, U.S.A. [email protected] Abstract The AdS/CFT correspondence is central to efforts to reconcile gravity and quantum mechanics, a fundamental goal of physics. It posits a duality between a gravitational theory in Anti de Sitter (AdS) space and a quantum mechanical conformal field theory (CFT), embodied in a map known as the AdS/CFT dictionary mapping states to states and operators to operators. This dictionary map is not well understood and has only been computed on special, structured instances. In this work we introduce cryptographic ideas to the study of AdS/CFT, and provide evidence that either the dictionary must be exponentially hard to compute, or else the quantum Extended Church-Turing thesis must be false in quantum gravity. Our argument has its origins in a fundamental paradox in the AdS/CFT correspondence known as the wormhole growth paradox. The paradox is that the CFT is believed to be “scrambling” – i.e. the expectation value of local operators equilibrates in polynomial time – whereas the gravity theory is not, because the interiors of certain black holes known as “wormholes” do not equilibrate and instead their volume grows at a linear rate for at least an exponential amount of time. -
NP-Complete Problems and Physical Reality
NP-complete Problems and Physical Reality Scott Aaronson∗ Abstract Can NP-complete problems be solved efficiently in the physical universe? I survey proposals including soap bubbles, protein folding, quantum computing, quantum advice, quantum adia- batic algorithms, quantum-mechanical nonlinearities, hidden variables, relativistic time dilation, analog computing, Malament-Hogarth spacetimes, quantum gravity, closed timelike curves, and “anthropic computing.” The section on soap bubbles even includes some “experimental” re- sults. While I do not believe that any of the proposals will let us solve NP-complete problems efficiently, I argue that by studying them, we can learn something not only about computation but also about physics. 1 Introduction “Let a computer smear—with the right kind of quantum randomness—and you create, in effect, a ‘parallel’ machine with an astronomical number of processors . All you have to do is be sure that when you collapse the system, you choose the version that happened to find the needle in the mathematical haystack.” —From Quarantine [31], a 1992 science-fiction novel by Greg Egan If I had to debate the science writer John Horgan’s claim that basic science is coming to an end [48], my argument would lean heavily on one fact: it has been only a decade since we learned that quantum computers could factor integers in polynomial time. In my (unbiased) opinion, the showdown that quantum computing has forced—between our deepest intuitions about computers on the one hand, and our best-confirmed theory of the physical world on the other—constitutes one of the most exciting scientific dramas of our time. -
Hybrid Qubits Solve Key Hurdle to Quantum Computing 28 December 2018
Hybrid qubits solve key hurdle to quantum computing 28 December 2018 In 1998, Daniel Loss, one of the authors of the current study, came up with a proposal, along with David DiVincenzo of IBM, to build a quantum computer by using the spins of electrons embedded in a quantum dot—a small particle that behaves like an atom, but that can be manipulated, so that they are sometimes called "artificial atoms." In the time since then, Loss and his team have endeavored to build practical devices. There are a number of barriers to developing practical devices in terms of speed. First, the device must be able to be initialized quickly. Initialization is the process of putting a qubit into a certain state, and if that cannot be done rapidly it slows down the device. Second, it must maintain coherence for a time long enough to make a measurement. Coherence refers to the Schematic of the device. Credit: RIKEN entanglement between two quantum states, and ultimately this is used to make the measurement, so if qubits become decoherent due to environmental noise, for example, the device Spin-based quantum computers have the potential becomes worthless. And finally, the ultimate state to tackle difficult mathematical problems that of the qubit must be able to be quickly read out. cannot be solved using ordinary computers, but many problems remain in making these machines While a number of methods have been proposed scalable. Now, an international group of for building a quantum computer, the one proposed researchers led by the RIKEN Center for Emergent by Loss and DiVincenzo remains one of the most Matter Science have crafted a new architecture for practically feasible, as it is based on quantum computing. -
The Quantum Times) in Producing the First in Their List of Guidelines Nor Was It Reported As Operating Laser, Are All Interesting and Enlightening
TThhee QQuuaannttuumm TTiimmeess APS Topical Group on Quantum Information, Concepts, and Computation Summer 2007 Volume 2, Number 2 Advances & Challenges in Quantum Key Distribution Quantum Key Distribution (QKD) is the flagship success of quantum communication. Since Bennett and Brassard’s 1984 publication [2] it has given a new direction to the field. (For a review see e.g. [3].) Up to then in quantum communication, the properties of quantum mechanics were used, for example, to improve on the through-put of optical channels. With the arrival of QKD an application has been created that achieves what Inside… cannot be achieved with classical communication alone: … you will find a little bit of provably secure protocols with no assumption about the everything. We begin with an computational power of an adversary. (This scenario is referred article from Norbert Lütkenhaus on to as unconditional security, a technical term in cryptography. quantum key distribution (QKD) Still, the devices of sender and receiver are assumed to be that grew out of a news article from perfect and out of bound of the adversary, an assumption without our last issue. Norbert speaks from which no cryptography is possible at all.) It is sufficient to the standpoint of someone at the prepare non-orthogonal states as signals and to measure them at forefront of quantum cryptography the receiver’s end. Then, using an authenticated public channel, and his article should be read by both parties can distill a secret key from the data. The procedure anyone interested in the current needs to be kick-started with some secret key in order to state and future direction of QKD. -
Issues in Physics & Astronomy
Issues in Physics & Astronomy Board on Physics and Astronomy • The National Academies • Washington, D.C. • 202-334-3520 • national-academies.org/bpa • Summer 2008 Understanding the Impact of Selling the Helium Reserve Michael H. Moloney, BPA Staff nder the sponsorship of the monatomic element, it passes easily containing 0.3 percent helium is consid- Bureau of Land Management at through tiny orifices and is therefore used ered economically viable. A few gas Uthe Department of the Interior, for leak detection in many scientific and deposits contain as much as 8 percent the BPA, in cooperation with the Na- technical applications. Its density is only helium. By comparison, the atmosphere tional Materials Advisory Board, has 15 percent that of air, making it useful as a contains only about 0.0005 percent. He- initiated a study to understand the lifting gas for aerostats and other devices. lium from wells that produce impact on the scientific community of Its high heat capacity, along with its uneconomically low concentrations, or the continuing sale of the U.S. helium inertness, makes it the preferred quench- from wells that produce higher concentra- reserve and recent developments in the ing medium for many applications in tions but do not flow through an extrac- helium market. materials processing, such as the produc- tion plant, is often vented to the atmo- The element helium has unique tion of high-quality superalloy powders. It sphere when the natural gas is burned. A properties. Liquefying near absolute is the preferred carrier gas for gas chroma- relatively minor amount of helium also is zero, it is the only option for many tography, a widely applied technique for vented at extraction plants that have no cryogenic applications, such as cooling chemical separations. -
Quantum Lower Bound for Recursive Fourier Sampling
Quantum Lower Bound for Recursive Fourier Sampling Scott Aaronson Institute for Advanced Study, Princeton [email protected] Abstract One of the earliest quantum algorithms was discovered by Bernstein and Vazirani, for a problem called Recursive Fourier Sampling. This paper shows that the Bernstein-Vazirani algorithm is not far from optimal. The moral is that the need to “uncompute” garbage can impose a fundamental limit on efficient quantum computation. The proof introduces a new parameter of Boolean functions called the “nonparity coefficient,” which might be of independent interest. Like a classical algorithm, a quantum algorithm can solve problems recursively by calling itself as a sub- routine. When this is done, though, the algorithm typically needs to call itself twice for each subproblem to be solved. The second call’s purpose is to uncompute ‘garbage’ left over by the first call, and thereby enable interference between different branches of the computation. Of course, a factor of 2 increase in running time hardly seems like a big deal, when set against the speedups promised by quantum computing. The problem is that these factors of 2 multiply, with each level of recursion producing an additional factor. Thus, one might wonder whether the uncomputing step is really necessary, or whether a cleverly designed algorithm might avoid it. This paper gives the first nontrivial example in which recursive uncomputation is provably necessary. The example concerns a long-neglected problem called Recursive Fourier Sampling (henceforth RFS), which was introduced by Bernstein and Vazirani [5] in 1993 to prove the first oracle separation between BPP and BQP. Many surveys on quantum computing pass directly from the Deutsch-Jozsa algorithm [8] to the dramatic results of Simon [14] and Shor [13], without even mentioning RFS. -
Topological Phases and Applications to Quantum Information Processing" ______List of Organizers
Proposed title: "Topological Phases and Applications to Quantum Information Processing" __________________________________________________________ List of organizers: Nicholas E. Bonesteel (Florida State University and NHMFL) Phone: (850) 644-7805, E-mail: [email protected] James P. Eisenstein (Caltech) Phone: (626) 395-4649, E-mail: [email protected] Michael H. Freedman (Microsoft Research) Phone: (805) 893-6313, E-mail: [email protected] Kirill Shtengel (UC Riverside) Phone: (951) 827-1058, E-mail: [email protected] Steven H. Simon (Lucent Technologies, Bell Labs) Phone: (908) 582-6006, E-mail: [email protected] __________________________________________________________ Proposed length: 4 weeks, preferably July 2 through July 29 or June 25 through July 22. However, anytime between June 18 and Sept 2, 2007 is acceptable. Pushing it toward earlier or later dates will result in conflicts with teaching for many of the potential attendees. __________________________________________________________ Abstract: Quantum computers, if realized in practice, promise exponential speed-up of some of the computational tasks that conventional classical algorithms are intrinsically incapable of handling efficiently. Perhaps even more intriguing possibility arises from being able to use quantum computers to simulate the behavior of other physical systems -- an exciting idea dating back to Feynman. Unfortunately, realizing a quantum computer in practice proves to be very difficult, chiefly due to the debilitating effects of decoherence plaguing all possible schemes which use microscopic degrees of freedom (such as nuclear or electronic spins) as basic building blocks. >From this perspective, topological quantum computing offers an attractive alternative by encoding quantum information in nonlocal topological degrees of freedom that are intrinsically protected from decoherence due to local noise. -
Quantum Computation with Spins in Quantum Dots
Quantum Computation With Spins In Quantum Dots Fikeraddis Damtie June 10, 2013 1 Introduction Physical implementation of quantum computating has been an active area of research since the last couple of decades. In classical computing, the transmission and manipulation of classical information is carried out by physical machines (computer hardwares, etc.). In theses machines the manipulation and transmission of information can be described using the laws of classical physics. Since Newtonian mechanics is a special limit of quantum mechanics, computers making use of the laws of quantum mechanics have greater computational power than classical computers. This need to create a powerful computing machine is the driving motor for research in the field of quantum computing. Until today, there are a few different schemes for implementing a quantum computer based on the David Divincenzo chriterias. Among these are: Spectral hole burning, Trapped ion, e-Helium, Gated qubits, Nuclear Magnetic Resonance, Optics, Quantum dots, Neutral atom, superconductors and doped silicon. In this project only the quantum dot scheme will be discussed. In the year 1997 Daniel Loss and David P.DiVincenzo proposed a spin-qubit quantum computer also called The Loss-DiVincenzo quantum computer. This proposal is now considered to be one of the most promising candidates for quantum computation in the solid state. The main idea of the proposal was to use the intrinsic spin-1=2 degrees of freedom of individual elecrons confined in semiconductor quantum dots. The proposal was made in a way to satisfy the five requirements for quantum computing by David diVincenzo which will be described in sec. -
Local and Distributed Quantum Computation
Local and Distributed Quantum Computation Rodney Van Meter and Simon J. Devitt May 24, 2016 Abstract Experimental groups are now fabricating quantum processors powerful enough to execute small instances of quantum algorithms and definitively demonstrate quantum error correction that extends the lifetime of quantum data, adding ur- gency to architectural investigations. Although other options continue to be ex- plored, effort is coalescing around topological coding models as the most prac- tical implementation option for error correction on realizable microarchitectures. Scalability concerns have also motivated architects to propose distributed memory multicomputer architectures, with experimental efforts demonstrating some of the basic building blocks to make such designs possible. We compile the latest re- sults from a variety of different systems aiming at the construction of a scalable quantum computer. arXiv:1605.06951v1 [quant-ph] 23 May 2016 1 1 Introduction Quantum computers and networks look like increasingly inevitable extensions to our already astonishing classical computing and communication capabilities [1, 2]. How do they work, and once built, what capabilities will they bring? Quantum computation and communication can be understood through seven key con- cepts (see sidebar). Each concept is simple, but collectively they imply that our clas- sical notion of computation is incomplete, and that quantum effects can be used to efficiently solve some previously intractable problems. In the 1980s and 90s, a handful of algorithms were developed and the foundations of quantum computational complexity were laid, but the full range of capabilities and the process of creating new algorithms were poorly understood [1, 3, 4, 5, 6]. Over the last fifteen years, a deeper understanding of this process has developed, and the num- ber of proposed algorithms has exploded 1. -
The Power of Unentanglement
The Power of Unentanglement Scott Aaronson∗ Salman Beigi Andrew Drucker Bill Fefferman MIT MIT MIT University of Chicago Peter Shor MIT Abstract The class QMA (k), introduced by Kobayashi et al., consists of all languages that can be verified using k unentangled quantum proofs. Many of the simplest questions about this class have remained embarrassingly open: for example, can we give any evidence that k quantum proofs are more powerful than one? Does QMA (k) = QMA (2) for k 2? Can QMA (k) protocols be amplified to exponentially small error? ≥ In this paper, we make progress on all of the above questions. We give a protocol by which a verifier can be convinced that a 3Sat formula of size m is • satisfiable, with constant soundness, given O (√m) unentangled quantum witnesses with O (log m) qubits each. Our protocol relies on the existence of very short PCPs. We show that assuming a weak version of thee Additivity Conjecture from quantum infor- • mation theory, any QMA (2) protocol can be amplified to exponentially small error, and QMA (k)= QMA (2) for all k 2. ≥ We prove the nonexistence of “perfect disentanglers” for simulating multiple Merlins with • one. 1 Introduction Quantum entanglement is often described as a complicated, hard-to-understand resource. But ironically, many questions in quantum computing are easiest to answer assuming unlimited entan- glement, and become much more difficult if entanglement is not allowed! One way to understand this is that Hilbert space—the space of all quantum states—has extremely useful linear-algebraic properties, and when we restrict to the set of separable states we lose many of those properties.