An Executive Primer on Artificial General Intelligence

Total Page:16

File Type:pdf, Size:1020Kb

An Executive Primer on Artificial General Intelligence Operations Practice An executive primer on artificial general intelligence While human-like artificial general intelligence may not be imminent, substantial advances may be possible in the coming years. Executives can prepare by recognizing the early signs of progress. by Federico Berruti, Pieter Nel, and Rob Whiteman © Getty Images April 2020 Headlines sounding the alarms that artificial University (NYU), is much more direct: “It’s hard intelligence (AI) will lead humanity to a dystopian to explain to non-specialists that AGI is not a future seem to be everywhere. Prominent thought ‘thing’, and that most venues that have AGI in their leaders, from Silicon Valley figures to legendary name deal in highly speculative and theoretical scientists, have warned that should AI evolve into issues...” artificial general intelligence (AGI)—AI that is as capable of learning intellectual tasks as humans Still, many academics and researchers maintain are—civilization will be under serious threat. that there is at least a chance that human-level artificial intelligence could be achieved in the Few seeing these warnings, stories, and images next decade. Richard Sutton, professor of could be blamed for believing that the arrival of AGI computer science of the University of Alberta, is imminent. Little surprise, then, that so many media stated in a 2017 talk: “Understanding human- stories and business presentations about machine level AI will be a profound scientific achievement learning are accompanied by unsettling illustrations (and economic boon) and may well happen by featuring humanoid robots. 2030 (25% chance), or by 2040 (50% chance)— or never (10% chance).” Many of the most respected researchers and academics see things differently, however. They What should executives take away from this argue that we are decades away from realizing AGI, debate? Even a small probability of achieving AGI and some even predict that we won’t see AGI in in the next decade justifies paying attention to this century. With so much uncertainty, why should developments in the field, given the potentially executives care about AGI today? The answer dramatic inflection point that AGI could bring is that, while the timing of AGI is uncertain, the about in society. As LeCun explains: “There is disruptive effects it could have on society cannot be a thin domain of research that, while having understated. ambitious goals of making progress towards human-level intelligence, is also sufficiently Much has already been written about the likely grounded in science and engineering impact of AI and the importance of carefully methodologies to bring real progress in managing the transition to a more automated world. technology. That’s the sweet spot.” The purpose of this article is to provide an AGI primer to help executives understand the path to machines For business leaders, it is critical to identify achieving human-level intelligence, indicators by those researchers who operate in this sweet which to measure progress, and actions the reader spot. In this executive’s guide to AGI, we aim to can take to begin preparations today. help readers make that assessment by reviewing the history of the field (see sidebar, “A brief history of AI”), the problems that must be solved How imminent is AGI? before researchers can claim they are close to In predicting that AGI won’t arrive until the year developing human-level artificial intelligence, 2300, Rodney Brooks, an MIT roboticist and and what executives should do given these co-founder of iRobot, doesn’t mince words: “It is insights. a fraught time understanding the true promise and dangers of AI. Most of what we read in the headlines… is, I believe, completely off the mark.” What capabilities would turn AI into AGI? Brooks is far from being a lone voice of dissent. To understand the complexity of achieving true Leading AI researchers such as Geoffrey Hinton and human-level intelligence, it is worthwhile to look Demis Hassabis have stated that AGI is nowhere at some the capabilities that true AGI will need to close to reality. In responding to one of Brooks’ master. posts, Yann LeCun, a professor at the Courant Institute of Mathematical Sciences at New York Sensory perception. Whereas deep learning has 2 An executive primer on artificial general intelligence Sidebar A brief history of AI The term “artificial intelligence” was around computer vision, natural lan- knowledge of the world through rela- coined by John McCarthy in the research guage understanding, and neural nets tionships between these symbols. One proposal for a 1956 workshop at Dart- are, in many cases, several decades old. familiar example is the knowledge that mouth that would kick off humanity’s a German shepherd is a dog, which is a efforts on this topic. The AI topics that Within these AI research communities, mammal; all mammals are warm-blood- McCarthy outlined in the introduction there has been substantial success ed; therefore, a German shepherd included how to get a computer to use and advancement in what the field now should be warm-blooded. human language; how to arrange “neuron terms “narrow AI” applications. Narrow nets” (which had been invented in 1943) AI is the application of AI techniques to a The limitation is that humans still provide so that they can form concepts; how a specific and well-defined problem, such the ground truth by encoding our knowl- machine can improve itself (that is, learn as chatbots that help customers resolve edge of the world, rather than allowing or evolve); how a machine could form ab- issues with their phone bills, algorithms an AI system to observe and encode stractions by using its sensors to observe that spot fraud in credit-card transac- these relationships itself. Symbolic AI the world; and how to make computers tions, and natural-language-processing was the dominant paradigm of AI re- think creatively. In essence, McCarthy engines that quickly process thousands search from the mid-1950s until the late was describing in 1956 what we now of legal documents. Applying narrow AI 1980s. One ongoing effort to provide a call AGI. solutions in use cases across industries solution to common-sense reasoning can generate tremendous economic through symbolic AI is the Cyc Project, McCarthy was certainly not the first to benefits. Our colleagues’ research launched in 1984 to collect knowledge talk about machines and “intelligence.” shows that the potential value of ap- represented declaratively in the form Alan Turing’s 1950 paper on “Comput- plying this sort of deep learning could of logical assertions, of which it had ing machinery and intelligence” intro- range from $3.5 trillion to $5.8 trillion collected 25 million by 2017. duced the “imitation game,” a test of a annually. machine’s ability to exhibit intelligent Neural networks (1954, 1969, 1986, behavior, and which became known as To differentiate themselves from re- 2012). Recent advances in speech the “Turing test.” Turing optimistically searchers solving narrow AI problems, recognition (now widely used in devices estimated that, in the year 2000, a com- a few research teams have claimed an such as smart speakers), and in com- puter with 128Mb of memory would have almost proprietary interest in producing puter vision (such as limited self-driving a 70 percent chance of fooling a person. human-level intelligence (or more) under abilities in cars) are all due to deep In his earlier 1948 paper on “Intelligent the name “artificial general intelligence.” neural networks. The artificial neuron as machinery,” he describes what we today Some have adopted the term “super- a computational model of the “nerve net” call computers, as well as a machine that intelligence” to describe AGI systems of the brain was proposed as early as fully imitates a person. He points out that by themselves could rapidly design 1943.¹ In the decades since then, it has that our ability to build adequate sensors even more capable systems, with those been through multiple highs and lows in and actuators might not be sufficient for systems further evolving to develop ca- its popularity as a tool for AI. some time and that our efforts are best pabilities that far exceed any possessed invested in the aspect of intelligence that by humans. MIT’s Marvin Minsky and Seymour relates to games and cryptography. The Papert put a damper on this research in clear aspiration, however, has always MIT roboticist Rodney Brooks describes their 1969 book “Perceptrons,” where been to achieve human-level intelligence. the four previous attempts at AGI, along they mathematically demonstrated that with their approximate start dates, in neural networks could only perform very The early work, like Turing suggested, discussions that provide important con- basic tasks. They also discussed the dif- revolved around subject areas that do text for understanding their progress. ficulty of training multi-layer networks. not require too much sensing and action, In 1986, however, Geoffrey Hinton and such as those of games and language Symbolic AI (1956). The key concept is some colleagues solved this problem translation. Research communities the use of symbols and the encoding of with the publication of the back-prop- 1 Warren McCulloch and Walter Pitts, “A logical calculus of the ideas immanent in nervous activity,” Bulletin of Mathematical Biophysics, volume 5, 1943. An executive primer on artificial general intelligence 3 Sidebar cont. agation algorithm. In the 1990’s, work vision to identify and navigate through Behavior-based robotics (1985). by Yann LeCun made major advances in their environments or to understand Insects navigate well in the real world neural networks’ use in computer vision, the geometry of objects and maneuver with very few neurons. This observation and Jürgen Schmidhuber similarly them, such as moving around blocks of inspired researchers to investigate how advanced the application of recurrent various shapes and colors.
Recommended publications
  • Lecture Note on Deep Learning and Quantum Many-Body Computation
    Lecture Note on Deep Learning and Quantum Many-Body Computation Jin-Guo Liu, Shuo-Hui Li, and Lei Wang∗ Institute of Physics, Chinese Academy of Sciences Beijing 100190, China November 23, 2018 Abstract This note introduces deep learning from a computa- tional quantum physicist’s perspective. The focus is on deep learning’s impacts to quantum many-body compu- tation, and vice versa. The latest version of the note is at http://wangleiphy.github.io/. Please send comments, suggestions and corrections to the email address in below. ∗ [email protected] CONTENTS 1 introduction2 2 discriminative learning4 2.1 Data Representation 4 2.2 Model: Artificial Neural Networks 6 2.3 Cost Function 9 2.4 Optimization 11 2.4.1 Back Propagation 11 2.4.2 Gradient Descend 13 2.5 Understanding, Visualization and Applications Beyond Classification 15 3 generative modeling 17 3.1 Unsupervised Probabilistic Modeling 17 3.2 Generative Model Zoo 18 3.2.1 Boltzmann Machines 19 3.2.2 Autoregressive Models 22 3.2.3 Normalizing Flow 23 3.2.4 Variational Autoencoders 25 3.2.5 Tensor Networks 28 3.2.6 Generative Adversarial Networks 29 3.3 Summary 32 4 applications to quantum many-body physics and more 33 4.1 Material and Chemistry Discoveries 33 4.2 Density Functional Theory 34 4.3 “Phase” Recognition 34 4.4 Variational Ansatz 34 4.5 Renormalization Group 35 4.6 Monte Carlo Update Proposals 36 4.7 Tensor Networks 37 4.8 Quantum Machine Leanring 38 4.9 Miscellaneous 38 5 hands on session 39 5.1 Computation Graph and Back Propagation 39 5.2 Deep Learning Libraries 41 5.3 Generative Modeling using Normalizing Flows 42 5.4 Restricted Boltzmann Machine for Image Restoration 43 5.5 Neural Network as a Quantum Wave Function Ansatz 43 6 challenges ahead 45 7 resources 46 BIBLIOGRAPHY 47 1 1 INTRODUCTION Deep Learning (DL) ⊂ Machine Learning (ML) ⊂ Artificial Intelli- gence (AI).
    [Show full text]
  • AI Computer Wraps up 4-1 Victory Against Human Champion Nature Reports from Alphago's Victory in Seoul
    The Go Files: AI computer wraps up 4-1 victory against human champion Nature reports from AlphaGo's victory in Seoul. Tanguy Chouard 15 March 2016 SEOUL, SOUTH KOREA Google DeepMind Lee Sedol, who has lost 4-1 to AlphaGo. Tanguy Chouard, an editor with Nature, saw Google-DeepMind’s AI system AlphaGo defeat a human professional for the first time last year at the ancient board game Go. This week, he is watching top professional Lee Sedol take on AlphaGo, in Seoul, for a $1 million prize. It’s all over at the Four Seasons Hotel in Seoul, where this morning AlphaGo wrapped up a 4-1 victory over Lee Sedol — incidentally, earning itself and its creators an honorary '9-dan professional' degree from the Korean Baduk Association. After winning the first three games, Google-DeepMind's computer looked impregnable. But the last two games may have revealed some weaknesses in its makeup. Game four totally changed the Go world’s view on AlphaGo’s dominance because it made it clear that the computer can 'bug' — or at least play very poor moves when on the losing side. It was obvious that Lee felt under much less pressure than in game three. And he adopted a different style, one based on taking large amounts of territory early on rather than immediately going for ‘street fighting’ such as making threats to capture stones. This style – called ‘amashi’ – seems to have paid off, because on move 78, Lee produced a play that somehow slipped under AlphaGo’s radar. David Silver, a scientist at DeepMind who's been leading the development of AlphaGo, said the program estimated its probability as 1 in 10,000.
    [Show full text]
  • Accelerators and Obstacles to Emerging ML-Enabled Research Fields
    Life at the edge: accelerators and obstacles to emerging ML-enabled research fields Soukayna Mouatadid Steve Easterbrook Department of Computer Science Department of Computer Science University of Toronto University of Toronto [email protected] [email protected] Abstract Machine learning is transforming several scientific disciplines, and has resulted in the emergence of new interdisciplinary data-driven research fields. Surveys of such emerging fields often highlight how machine learning can accelerate scientific discovery. However, a less discussed question is how connections between the parent fields develop and how the emerging interdisciplinary research evolves. This study attempts to answer this question by exploring the interactions between ma- chine learning research and traditional scientific disciplines. We examine different examples of such emerging fields, to identify the obstacles and accelerators that influence interdisciplinary collaborations between the parent fields. 1 Introduction In recent decades, several scientific research fields have experienced a deluge of data, which is impacting the way science is conducted. Fields such as neuroscience, psychology or social sciences are being reshaped by the advances in machine learning (ML) and the data processing technologies it provides. In some cases, new research fields have emerged at the intersection of machine learning and more traditional research disciplines. This is the case for example, for bioinformatics, born from collaborations between biologists and computer scientists in the mid-80s [1], which is now considered a well-established field with an active community, specialized conferences, university departments and research groups. How do these transformations take place? Do they follow similar paths? If yes, how can the advances in such interdisciplinary fields be accelerated? Researchers in the philosophy of science have long been interested in these questions.
    [Show full text]
  • The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design
    The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design Jeffrey Dean Google Research [email protected] Abstract The past decade has seen a remarkable series of advances in machine learning, and in particular deep learning approaches based on artificial neural networks, to improve our abilities to build more accurate systems across a broad range of areas, including computer vision, speech recognition, language translation, and natural language understanding tasks. This paper is a companion paper to a keynote talk at the 2020 International Solid-State Circuits Conference (ISSCC) discussing some of the advances in machine learning, and their implications on the kinds of computational devices we need to build, especially in the post-Moore’s Law-era. It also discusses some of the ways that machine learning may also be able to help with some aspects of the circuit design process. Finally, it provides a sketch of at least one interesting direction towards much larger-scale multi-task models that are sparsely activated and employ much more dynamic, example- and task-based routing than the machine learning models of today. Introduction The past decade has seen a remarkable series of advances in machine learning (ML), and in particular deep learning approaches based on artificial neural networks, to improve our abilities to build more accurate systems across a broad range of areas [LeCun et al. 2015]. Major areas of significant advances ​ ​ include computer vision [Krizhevsky et al. 2012, Szegedy et al. 2015, He et al. 2016, Real et al. 2017, Tan ​ ​ ​ ​ ​ ​ and Le 2019], speech recognition [Hinton et al.
    [Show full text]
  • The Power and Promise of Computers That Learn by Example
    Machine learning: the power and promise of computers that learn by example MACHINE LEARNING: THE POWER AND PROMISE OF COMPUTERS THAT LEARN BY EXAMPLE 1 Machine learning: the power and promise of computers that learn by example Issued: April 2017 DES4702 ISBN: 978-1-78252-259-1 The text of this work is licensed under the terms of the Creative Commons Attribution License which permits unrestricted use, provided the original author and source are credited. The license is available at: creativecommons.org/licenses/by/4.0 Images are not covered by this license. This report can be viewed online at royalsociety.org/machine-learning Cover image © shulz. 2 MACHINE LEARNING: THE POWER AND PROMISE OF COMPUTERS THAT LEARN BY EXAMPLE Contents Executive summary 5 Recommendations 8 Chapter one – Machine learning 15 1.1 Systems that learn from data 16 1.2 The Royal Society’s machine learning project 18 1.3 What is machine learning? 19 1.4 Machine learning in daily life 21 1.5 Machine learning, statistics, data science, robotics, and AI 24 1.6 Origins and evolution of machine learning 25 1.7 Canonical problems in machine learning 29 Chapter two – Emerging applications of machine learning 33 2.1 Potential near-term applications in the public and private sectors 34 2.2 Machine learning in research 41 2.3 Increasing the UK’s absorptive capacity for machine learning 45 Chapter three – Extracting value from data 47 3.1 Machine learning helps extract value from ‘big data’ 48 3.2 Creating a data environment to support machine learning 49 3.3 Extending the lifecycle
    [Show full text]
  • ARCHITECTS of INTELLIGENCE for Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS of INTELLIGENCE
    MARTIN FORD ARCHITECTS OF INTELLIGENCE For Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS OF INTELLIGENCE THE TRUTH ABOUT AI FROM THE PEOPLE BUILDING IT MARTIN FORD ARCHITECTS OF INTELLIGENCE Copyright © 2018 Packt Publishing All rights reserved. No part of this book may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, without the prior written permission of the publisher, except in the case of brief quotations embedded in critical articles or reviews. Every effort has been made in the preparation of this book to ensure the accuracy of the information presented. However, the information contained in this book is sold without warranty, either express or implied. Neither the author, nor Packt Publishing or its dealers and distributors, will be held liable for any damages caused or alleged to have been caused directly or indirectly by this book. Packt Publishing has endeavored to provide trademark information about all of the companies and products mentioned in this book by the appropriate use of capitals. However, Packt Publishing cannot guarantee the accuracy of this information. Acquisition Editors: Ben Renow-Clarke Project Editor: Radhika Atitkar Content Development Editor: Alex Sorrentino Proofreader: Safis Editing Presentation Designer: Sandip Tadge Cover Designer: Clare Bowyer Production Editor: Amit Ramadas Marketing Manager: Rajveer Samra Editorial Director: Dominic Shakeshaft First published: November 2018 Production reference: 2201118 Published by Packt Publishing Ltd. Livery Place 35 Livery Street Birmingham B3 2PB, UK ISBN 978-1-78913-151-2 www.packt.com Contents Introduction ........................................................................ 1 A Brief Introduction to the Vocabulary of Artificial Intelligence .......10 How AI Systems Learn ........................................................11 Yoshua Bengio .....................................................................17 Stuart J.
    [Show full text]
  • Technology Innovation and the Future of Air Force Intelligence Analysis
    C O R P O R A T I O N LANCE MENTHE, DAHLIA ANNE GOLDFELD, ABBIE TINGSTAD, SHERRILL LINGEL, EDWARD GEIST, DONALD BRUNK, AMANDA WICKER, SARAH SOLIMAN, BALYS GINTAUTAS, ANNE STICKELLS, AMADO CORDOVA Technology Innovation and the Future of Air Force Intelligence Analysis Volume 2, Technical Analysis and Supporting Material RR-A341-2_cover.indd All Pages 2/8/21 12:20 PM For more information on this publication, visit www.rand.org/t/RRA341-2 Library of Congress Cataloging-in-Publication Data is available for this publication. ISBN: 978-1-9774-0633-0 Published by the RAND Corporation, Santa Monica, Calif. © Copyright 2021 RAND Corporation R® is a registered trademark. Cover: U.S. Marine Corps photo by Cpl. William Chockey; faraktinov, Adobe Stock. Limited Print and Electronic Distribution Rights This document and trademark(s) contained herein are protected by law. This representation of RAND intellectual property is provided for noncommercial use only. Unauthorized posting of this publication online is prohibited. Permission is given to duplicate this document for personal use only, as long as it is unaltered and complete. Permission is required from RAND to reproduce, or reuse in another form, any of its research documents for commercial use. For information on reprint and linking permissions, please visit www.rand.org/pubs/permissions. The RAND Corporation is a research organization that develops solutions to public policy challenges to help make communities throughout the world safer and more secure, healthier and more prosperous. RAND is nonprofit, nonpartisan, and committed to the public interest. RAND’s publications do not necessarily reflect the opinions of its research clients and sponsors.
    [Show full text]
  • Much Has Been Written About the Turing Test in the Last Few Years, Some of It
    1 Much has been written about the Turing Test in the last few years, some of it preposterously off the mark. People typically mis-imagine the test by orders of magnitude. This essay is an antidote, a prosthesis for the imagination, showing how huge the task posed by the Turing Test is, and hence how unlikely it is that any computer will ever pass it. It does not go far enough in the imagination-enhancement department, however, and I have updated the essay with a new postscript. Can Machines Think?1 Can machines think? This has been a conundrum for philosophers for years, but in their fascination with the pure conceptual issues they have for the most part overlooked the real social importance of the answer. It is of more than academic importance that we learn to think clearly about the actual cognitive powers of computers, for they are now being introduced into a variety of sensitive social roles, where their powers will be put to the ultimate test: In a wide variety of areas, we are on the verge of making ourselves dependent upon their cognitive powers. The cost of overestimating them could be enormous. One of the principal inventors of the computer was the great 1 Originally appeared in Shafto, M., ed., How We Know (San Francisco: Harper & Row, 1985). 2 British mathematician Alan Turing. It was he who first figured out, in highly abstract terms, how to design a programmable computing device--what we not call a universal Turing machine. All programmable computers in use today are in essence Turing machines.
    [Show full text]
  • Turing Test Does Not Work in Theory but in Practice
    Int'l Conf. Artificial Intelligence | ICAI'15 | 433 Turing test does not work in theory but in practice Pertti Saariluoma1 and Matthias Rauterberg2 1Information Technology, University of Jyväskylä, Jyväskylä, Finland 2Industrial Design, Eindhoven University of Technology, Eindhoven, The Netherlands Abstract - The Turing test is considered one of the most im- Essentially, the Turing test is an imitation game. Like any portant thought experiments in the history of AI. It is argued good experiment, it has two conditions. In the case of control that the test shows how people think like computers, but is this conditions, it is assumed that there is an opaque screen. On actually true? In this paper, we discuss an entirely new per- one side of the screen is an interrogator whose task is to ask spective. Scientific languages have their foundational limita- questions and assess the nature of the answers. On the other tions, for example, in their power of expression. It is thus pos- side, there are two people, A and B. The task of A and B is to sible to discuss the limitations of formal concepts and theory answer the questions, and the task of the interrogator is to languages. In order to represent real world phenomena in guess who has given the answer. In the case of experimental formal concepts, formal symbols must be given semantics and conditions, B is replaced by a machine (computer) and again, information contents; that is, they must be given an interpreta- the interrogator must decide whether it was the human or the tion. They key argument is that it is not possible to express machine who answered the questions.
    [Show full text]
  • THE TURING TEST RELIES on a MISTAKE ABOUT the BRAIN For
    THE TURING TEST RELIES ON A MISTAKE ABOUT THE BRAIN K. L. KIRKPATRICK Abstract. There has been a long controversy about how to define and study intel- ligence in machines and whether machine intelligence is possible. In fact, both the Turing Test and the most important objection to it (called variously the Shannon- McCarthy, Blockhead, and Chinese Room arguments) are based on a mistake that Turing made about the brain in his 1948 paper \Intelligent Machinery," a paper that he never published but whose main assumption got embedded in his famous 1950 paper and the celebrated Imitation Game. In this paper I will show how the mistake is a false dichotomy and how it should be fixed, to provide a solid foundation for a new understanding of the brain and a new approach to artificial intelligence. In the process, I make an analogy between the brain and the ribosome, machines that translate information into action, and through this analogy I demonstrate how it is possible to go beyond the purely information-processing paradigm of computing and to derive meaning from syntax. For decades, the metaphor of the brain as an information processor has dominated both neuroscience and artificial intelligence research and allowed the fruitful appli- cations of Turing's computation theory and Shannon's information theory in both fields. But this metaphor may be leading us astray, because of its limitations and the deficiencies in our understanding of the brain and AI. In this paper I will present a new metaphor to consider, of the brain as both an information processor and a producer of physical effects.
    [Show full text]
  • Machine Learning Conference Report
    Part of the conference series Breakthrough science and technologies Transforming our future Machine learning Conference report Machine learning report – Conference report 1 Introduction On 22 May 2015, the Royal Society hosted a unique, high level conference on the subject of machine learning. The conference brought together scientists, technologists and experts from across academia, industry and government to learn about the state-of-the-art in machine learning from world and UK experts. The presentations covered four main areas of machine learning: cutting-edge developments in software, the underpinning hardware, current applications and future social and economic impacts. This conference is the first in a new series organised This report is not a verbatim record, but summarises by the Royal Society, entitled Breakthrough Science the discussions that took place during the day and the and Technologies: Transforming our Future, which will key points raised. Comments and recommendations address the major scientific and technical challenges reflect the views and opinions of the speakers and of the next decade. Each conference will focus on one not necessarily that of the Royal Society. Full versions technology and cover key issues including the current of the presentations can be found on our website at: state of the UK industry sector, future direction of royalsociety.org/events/2015/05/breakthrough-science- research and the wider social and economic implications. technologies-machine-learning The conference series is being organised through the Royal Society’s Science and Industry programme, which demonstrates our commitment to reintegrate science and industry at the Society and to promote science and its value, build relationships and foster translation.
    [Show full text]
  • Arxiv:1709.02779V1 [Quant-Ph] 8 Sep 2017 I
    Machine learning & artificial intelligence in the quantum domain Vedran Dunjko Institute for Theoretical Physics, University of Innsbruck, Innsbruck 6020, Austria Max Planck Institute of Quantum Optics, Garching 85748, Germany Email: [email protected] Hans J. Briegel Institute for Theoretical Physics, University of Innsbruck Innsbruck 6020, Austria Department of Philosophy, University of Konstanz, Konstanz 78457, Germany Email: [email protected] Abstract. Quantum information technologies, on the one side, and intelligent learning systems, on the other, are both emergent technologies that will likely have a transforming impact on our society in the future. The respective underlying fields of basic research { quantum information (QI) versus machine learning and artificial intelligence (AI) { have their own specific questions and challenges, which have hitherto been investigated largely independently. However, in a growing body of recent work, researchers have been prob- ing the question to what extent these fields can indeed learn and benefit from each other. QML explores the interaction between quantum computing and machine learning, inves- tigating how results and techniques from one field can be used to solve the problems of the other. In recent time, we have witnessed significant breakthroughs in both directions of influence. For instance, quantum computing is finding a vital application in providing speed-ups for machine learning problems, critical in our \big data" world. Conversely, machine learning already permeates many cutting-edge technologies, and may become instrumental in advanced quantum technologies. Aside from quantum speed-up in data analysis, or classical machine learning optimization used in quantum experiments, quan- tum enhancements have also been (theoretically) demonstrated for interactive learning tasks, highlighting the potential of quantum-enhanced learning agents.
    [Show full text]