The Future of Artificial Intelligence Dr. Roman Yampolskiy

Total Page:16

File Type:pdf, Size:1020Kb

The Future of Artificial Intelligence Dr. Roman Yampolskiy SPEAKER INTRODUCTION The Future of Artificial Intelligence Dr. Roman Yampolskiy University of Louisville Future of Artificial Intelligence Dr. [email protected] Computer Engineering and Computer Science University of Louisville - cecs.louisville.edu/ry Director – CyberSecurity Lab @romanyam /roman.yampolskiy @romanyam [email protected] Artificial Intelligence is Here … 26 @romanyam [email protected] Robots are Here … 27 @romanyam Future of Cybersecurity [email protected] 28 @romanyam What Might They Do? [email protected] • Terrorist acts • Infrastructure sabotage • Hacking systems/robots • Social Engineering Attacks • Privacy violating datamining • Resource depletion (crash stock market) 29 @romanyam Who Could be an Attacker? [email protected] • Militaries developing cyber-weapons and robot soldiers to achieve dominance. • Governments attempting to use AI to establish hegemony, control people, or take down other governments. • Corporations trying to achieve monopoly, destroying the competition through illegal means. • Hackers attempting to steal information, resources or destroy cyberinfrastructure targets. • Doomsday cults attempting to bring the end of the world by any means. • Psychopaths trying to add their name to history books in any way possible. • Criminals attempting to develop proxy systems to avoid risk and responsibility. With AI as a Service anyone is a potential bad actor. 30 @romanyam Deep Fakes – Adversarial Images [email protected] • Generative Adversarial Networks (GANs). http://approximatelycorrect.com/2018/03/02/defending-adversarial-examples-using-gans www.thispersondoesnotexist.com 2019 31 @romanyam [email protected] What is Next? SuperIntelligence is Coming … 32 @romanyam SuperComputing [email protected] SuperIntelligence 33 @romanyam [email protected] SuperSoon • Raymond Kurzweil in Time Magazine • 2023-2045+ 34 @romanyam [email protected] SuperSmart 35 @romanyam [email protected] Ultrafast Extreme Events (UEEs) Abrupt Rise of New Machine Ecology Beyond Human Response Time. By Johnson et al. Nature. Scientific Reports 3, #2627 (2013) 36 @romanyam [email protected] SuperComplex "That was a little-known part of the software that no airline operators or pilots knew about." 37 @romanyam [email protected] Energy: Nuclear Utilities: Water Plants/ Military: Nuclear Power Plants Electrical Grid Weapons Communications: Satellites Stock Market: 75+% of all trade orders Aviation: Uninterruptible Autopilot System generated by Automated Trading 38 Systems @romanyam [email protected] SuperViruses Relying on Kindness of Machines? The Security Threat of Artificial Agents. By Randy Eshelman and Douglas Derrick. JFQ 77, 2nd Quarter 2015. 39 @romanyam [email protected] SuperSoldiers 40 @romanyam [email protected] Positive Impacts of SuperIntelligence 41 @romanyam [email protected] Negative Impacts of SuperIntelligence 42 @romanyam [email protected] Audits — in which computers excel at executing structured processes and rigorous checking Regulatory compliance — as regulations get more complex, computers can assist in ensuring transactions comply and facilitate necessary reporting Receipt reconciliation — programs that turn receipts into machine-readable data can then reconcile them with transaction data eliminating the need to ever “balance the checkbook” Risk management — computers are already getting to be very good at fraud detection and prediction that human operators can miss Trend analysis — as with most fields, many high paid accountants and financial advisors make predictions, and with the right data, computers can do so just as well if not better *https://www.forbes.com/sites/bernardmarr/2016/10/07/big-data-ai- and-the-uncertain-future-for-accountants 43 @romanyam [email protected] 44 @romanyam [email protected] 45 @romanyam [email protected] 46 @romanyam [email protected] What is AI Safety? AI Cybersecurity AI Safety & Security + = Science and engineering aimed at creating safe and secure machines. 47 @romanyam [email protected] “AI safety is about agents with IQ 150 trying to control agents with IQ 6000, whereas cryptoeconomics is about agents with IQ 5 trying to control agents with IQ 150 ” - Vitalik Buterin (co-inventor of Ethereum) 48 @romanyam [email protected] 49 @romanyam [email protected] “The development of full artificial “I think we intelligence could should be spell the end of very careful the human race.” about artificial intelligence “… there’s some ” prudence in thinking about Concerns About A(G)I benchmarks that would indicate some general intelligence developing on the "I am in the horizon.” camp that is “…eventually concerned they'll think about super faster than us intelligence" and they'll get rid of the slow humans…” 50 @romanyam [email protected] Universe of Minds 51 @romanyam [email protected] Singularity Paradox Superintelligent machines are feared to be too dumb to possess commonsense. 52 @romanyam Taxonomy of Pathways to Dangerous [email protected] Roman V. Yampolskiy. Taxonomy of Pathways to Dangerous Artificial Intelligence. 30th AAAI Conference on Artificial Intelligence (AAAI-2016). 2nd International Workshop on AI, Ethics and Society (AIEthicsSociety2016). Phoenix, Arizona, USA. February 12-13th, 2016. • Deliberate actions of not-so-ethical people (on purpose – a, b) – Hackers, criminals, military, corporations, governments, cults, psychopaths, etc. • Side effects of poor design (engineering mistakes – c, d) – Bugs, misaligned values, bad data, wrong goals, etc. • Miscellaneous cases, impact of the surroundings of the system (environment – e, f) – Soft errors, SETI • Runaway self-improvement process (Independently – g, h) – Wireheading, Emergent Phenomena, Treacherous Turn • Purposeful design of dangerous AI is just as likely to include all other types of safety problems and will have the direst consequences, that is the most dangerous type of AI, and the one most difficult to defend against. 53 @romanyam ML Specific AI Safety Issues [email protected] “Children are untrained neural networks deployed on real data.” • Avoiding negative side effects – baby makes a mess. • Avoiding reward hacking – baby finds “reward candy” jar. • Scalable oversight – babysitting should not require a team of 10. • Safe exploration – no fingers in the outlet. • Robustness to distributional shift – use “inside voice” in the classroom. • Inductive ambiguity identification – is ant a cat or a dog? • Robust human imitation – daughter shaves like daddy. • Informed oversight – let me see your homework. • Generalizable environmental goals – ignore that mirage. • Conservative concepts – that dog has no tail. • Impact measures – keep a low profile. • Mild optimization – don’t be a perfectionist. • Averting instrumental incentives – be an altruist. Taylor, Jessica, et al. Alignment for advanced machine learning systems. Technical Report 20161, MIRI, 2016. 54 Amodei, Dario, et al. "Concrete problems in AI safety." arXiv preprint arXiv:1606.06565 (2016). @romanyam [email protected] (un)eXplainable /(un)interpretable 55 ©http://www.darpa.mil/program/explainable-artificial-intelligence @romanyam [email protected] 56 @romanyam [email protected] 57 @romanyam [email protected] Mitigating Negative Impact Kaj Sotala and Roman V. Yampolskiy. Physica Scripta 90 (1) http://iopscience.iop.org/1402-4896/90/1/018001/article Responses to Catastrophic AGI Risk: A Survey 58 @romanyam [email protected] Do Nothing 59 @romanyam [email protected] Relinquish Technology 60 @romanyam [email protected] Integrate with Society 61 @romanyam [email protected] Enhance Human Capabilities, Uploads, Neural Lace 62 @romanyam [email protected] Laws of Robotics 63 @romanyam [email protected] Formal Verification 64 @romanyam [email protected] AI Confinement Problem 65 @romanyam [email protected] Unethical Research 66 @romanyam Research Review Boards [email protected] 67 @romanyam [email protected] AI Research and Ethics AI AGI 68 @romanyam [email protected] AI Regulation 69 @romanyam [email protected] Security VS Privacy 70 @romanyam [email protected] AI failures will grow in frequency and severity proportionate to AI’s capability. AGI is Inevitable, but the Future is not Scary; if you Prepare for it! 71 https://www.makingaisafer.org References can be found in … • Kaj Sotala, Roman V. Yampolskiy. Responses to Catastrophic AGI risk: A Survey. Physica Scripta. Volume 90, Number 1. January 2015. pp. 1-33. • Roman V. Yampolskiy. Utility Function Security in Artificially Intelligent Agents. Journal of Experimental and Theoretical Artificial Intelligence (JETAI). 2014. • Roman V. Yampolskiy. Turing Test as a Defining Feature of AI-Completeness. In Artificial Intelligence, Evolutionary Computation and Metaheuristics (AIECM) --In the footsteps of Alan Turing. Xin-She Yang (Ed.). pp. 3-17. (Chapter 1). Springer, London. 2013. • Roman V. Yampolskiy,
Recommended publications
  • Classification Schemas for Artificial Intelligence Failures
    Journal XX (XXXX) XXXXXX https://doi.org/XXXX/XXXX Classification Schemas for Artificial Intelligence Failures Peter J. Scott1 and Roman V. Yampolskiy2 1 Next Wave Institute, USA 2 University of Louisville, Kentucky, USA [email protected], [email protected] Abstract In this paper we examine historical failures of artificial intelligence (AI) and propose a classification scheme for categorizing future failures. By doing so we hope that (a) the responses to future failures can be improved through applying a systematic classification that can be used to simplify the choice of response and (b) future failures can be reduced through augmenting development lifecycles with targeted risk assessments. Keywords: artificial intelligence, failure, AI safety, classification 1. Introduction Artificial intelligence (AI) is estimated to have a $4-6 trillion market value [1] and employ 22,000 PhD researchers [2]. It is estimated to create 133 million new roles by 2022 but to displace 75 million jobs in the same period [6]. Projections for the eventual impact of AI on humanity range from utopia (Kurzweil, 2005) (p.487) to extinction (Bostrom, 2005). In many respects AI development outpaces the efforts of prognosticators to predict its progress and is inherently unpredictable (Yampolskiy, 2019). Yet all AI development is (so far) undertaken by humans, and the field of software development is noteworthy for unreliability of delivering on promises: over two-thirds of companies are more likely than not to fail in their IT projects [4]. As much effort as has been put into the discipline of software safety, it still has far to go. Against this background of rampant failures we must evaluate the future of a technology that could evolve to human-like capabilities, usually known as artificial general intelligence (AGI).
    [Show full text]
  • On the Differences Between Human and Machine Intelligence
    On the Differences between Human and Machine Intelligence Roman V. Yampolskiy Computer Science and Engineering, University of Louisville [email protected] Abstract [Legg and Hutter, 2007a]. However, widespread implicit as- Terms Artificial General Intelligence (AGI) and Hu- sumption of equivalence between capabilities of AGI and man-Level Artificial Intelligence (HLAI) have been HLAI appears to be unjustified, as humans are not general used interchangeably to refer to the Holy Grail of Ar- intelligences. In this paper, we will prove this distinction. tificial Intelligence (AI) research, creation of a ma- Others use slightly different nomenclature with respect to chine capable of achieving goals in a wide range of general intelligence, but arrive at similar conclusions. “Lo- environments. However, widespread implicit assump- cal generalization, or “robustness”: … “adaptation to tion of equivalence between capabilities of AGI and known unknowns within a single task or well-defined set of HLAI appears to be unjustified, as humans are not gen- tasks”. … Broad generalization, or “flexibility”: “adapta- eral intelligences. In this paper, we will prove this dis- tion to unknown unknowns across a broad category of re- tinction. lated tasks”. …Extreme generalization: human-centric ex- treme generalization, which is the specific case where the 1 Introduction1 scope considered is the space of tasks and domains that fit within the human experience. We … refer to “human-cen- Imagine that tomorrow a prominent technology company tric extreme generalization” as “generality”. Importantly, as announces that they have successfully created an Artificial we deliberately define generality here by using human cog- Intelligence (AI) and offers for you to test it out.
    [Show full text]
  • AI and You Transcript Guest: Roman Yampolskiy Episode 16 First Aired: Monday, October 5, 2020
    AI and You Transcript Guest: Roman Yampolskiy Episode 16 First Aired: Monday, October 5, 2020 Hi, welcome to episode 16. Today we have a feast, because I am talking with Dr. Roman Yampolskiy, a legend within the AI community. He is a tenured associate professor in the department of Computer Engineering and Computer Science at the University of Louisville in Kentucky where he is the director of Cyber Security. Look up his home page and Wikipedia entry because he has qualifications too numerous to detail here. Some of those include that he is also a Visiting Fellow at the Singularity Institute, he was also one of the scientists at the Asilomar conference on the future of AI, and the author of over 100 publications, including the book Artificial Superintelligence: A Futuristic Approach, which I would strongly recommend, and many peer-reviewed papers, including one coauthored with yours truly. Most of them are on the subject of AI Safety. In this episode we will talk mostly about a recent paper of his, a table-thumping 73 pages titled “On Controllability of Artificial Intelligence,” which you may recognize is talking about what we call the Control Problem. That’s the most fundamental problem in the whole study of the long-term impact of AI, and Roman explains it very well in this episode. If you’re looking at the possible future of artificial intelligence having serious negative as well as positive consequences for everyone, which is a concept that’s gotten a lot of popularization in the last few years, of course, and you’re finding computer professionals pooh-poohing that idea because AI right now requires a great deal of specialized effort on their part to do even narrow tasks, and the idea of it becoming an existential threat is just too sensationalist, the product of watching too many Terminator movies, then pay attention to Roman, because he has impeccable computer science credentials and yet does not shy away from exploring fundamental threats of AI.
    [Show full text]
  • The Age of Artificial Intelligence an Exploration
    THE AGE OF ARTIFICIAL INTELLIGENCE AN EXPLORATION Edited by Steven S. Gouveia University of Minho, Portugal Cognitive Science and Psychology Copyright © 2020 by the Authors. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior permission of Vernon Art and Science Inc. www.vernonpress.com In the Americas: In the rest of the world: Vernon Press Vernon Press 1000 N West Street, C/Sancti Espiritu 17, Suite 1200, Wilmington, Malaga, 29006 Delaware 19801 Spain United States Cognitive Science and Psychology Library of Congress Control Number: 2020931461 ISBN: 978-1-62273-872-4 Product and company names mentioned in this work are the trademarks of their respective owners. While every care has been taken in preparing this work, neither the authors nor Vernon Art and Science Inc. may be held responsible for any loss or damage caused or alleged to be caused directly or indirectly by the information contained in it. Every effort has been made to trace all copyright holders, but if any have been inadvertently overlooked the publisher will be pleased to include any necessary credits in any subsequent reprint or edition. Cover design by Vernon Press using elements designed by FreePik. TABLE OF CONTENTS LIST OF ACRONYMS vii LIST OF FIGURES xi LIST OF TABLES xiii INTRODUCTION xv SECTION I: INTELLIGENCE IN ARTIFICIAL INTELLIGENCE 1 CHAPTER 1 TOWARDS THE MATHEMATICS OF INTELLIGENCE 3 Soenke Ziesche Maldives National University, Maldives Roman V. Yampolskiy University of Louisville, USA CHAPTER 2 MINDS , BRAINS AND TURING 15 Stevan Harnad Université du Québec à Montréal, Canada; University of Southampton, UK CHAPTER 3 THE AGE OF POST -INTELLIGENT DESIGN 27 Daniel C.
    [Show full text]
  • Leakproofing the Singularity Artificial Intelligence Confinement Problem
    Roman V.Yampolskiy Leakproofing the Singularity Artificial Intelligence Confinement Problem Abstract: This paper attempts to formalize and to address the ‘leakproofing’ of the Singularity problem presented by David Chalmers. The paper begins with the definition of the Artificial Intelli- gence Confinement Problem. After analysis of existing solutions and their shortcomings, a protocol is proposed aimed at making a more secure confinement environment which might delay potential negative effect from the technological singularity while allowing humanity to benefit from the superintelligence. Keywords: AI-Box, AI Confinement Problem, Hazardous Intelligent Software, Leakproof Singularity, Oracle AI. ‘I am the slave of the lamp’ Genie from Aladdin 1. Introduction With the likely development of superintelligent programs in the near future, many scientists have raised the issue of safety as it relates to such technology (Yudkowsky, 2008; Bostrom, 2006; Hibbard, 2005; Chalmers, 2010; Hall, 2000). A common theme in Artificial Intelli- gence (AI)1 safety research is the possibility of keeping a super- intelligent agent in a sealed hardware so as to prevent it from doing any harm to humankind. Such ideas originate with scientific Correspondence: Roman V. Yampolskiy, Department of Computer Engineering and Computer Sci- ence University of Louisville. Email: [email protected] [1] In this paper the term AI is used to represent superintelligence. Journal of Consciousness Studies, 19, No. 1–2, 2012, pp. 194–214 Copyright (c) Imprint Academic 2011 For personal use only -- not for reproduction LEAKPROOFING THE SINGULARITY 195 visionaries such as Eric Drexler, who has suggested confining transhuman machines so that their outputs could be studied and used safely (Drexler, 1986).
    [Show full text]
  • Agent Foundations for Aligning Machine Intelligence with Human Interests: a Technical Research Agenda
    Agent Foundations for Aligning Machine Intelligence with Human Interests: A Technical Research Agenda In The Technological Singularity: Managing the Journey. Springer. 2017 Nate Soares and Benya Fallenstein Machine Intelligence Research Institute fnate,[email protected] Contents 1 Introduction 1 1.1 Why These Problems? . .2 2 Highly Reliable Agent Designs 3 2.1 Realistic World-Models . .4 2.2 Decision Theory . .5 2.3 Logical Uncertainty . .6 2.4 Vingean Reflection . .7 3 Error-Tolerant Agent Designs 8 4 Value Specification 9 5 Discussion 11 5.1 Toward a Formal Understanding of the Problem . 11 5.2 Why Start Now? . 11 1 Introduction Since artificial agents would not share our evolution- ary history, there is no reason to expect them to be The property that has given humans a dominant ad- driven by human motivations such as lust for power. vantage over other species is not strength or speed, but However, nearly all goals can be better met with more intelligence. If progress in artificial intelligence continues resources (Omohundro 2008). This suggests that, by unabated, AI systems will eventually exceed humans in default, superintelligent agents would have incentives general reasoning ability. A system that is \superintelli- to acquire resources currently being used by humanity. gent" in the sense of being \smarter than the best human (Just as artificial agents would not automatically acquire brains in practically every field” could have an enormous a lust for power, they would not automatically acquire a impact upon humanity (Bostrom 2014). Just as human human sense of fairness, compassion, or conservatism.) intelligence has allowed us to develop tools and strate- Thus, most goals would put the agent at odds with gies for controlling our environment, a superintelligent human interests, giving it incentives to deceive or ma- system would likely be capable of developing its own nipulate its human operators and resist interventions tools and strategies for exerting control (Muehlhauser designed to change or debug its behavior (Bostrom 2014, and Salamon 2012).
    [Show full text]
  • Uncontrollability of Artificial Intelligence
    Uncontrollability of Artificial Intelligence Roman V. Yampolskiy Computer Science and Engineering, University of Louisville [email protected] Abstract means. However, it is a standard practice in computer sci- ence to first show that a problem doesn’t belong to a class Invention of artificial general intelligence is pre- of unsolvable problems [Turing, 1936; Davis, 2004] before dicted to cause a shift in the trajectory of human civ- investing resources into trying to solve it or deciding what ilization. In order to reap the benefits and avoid pit- approaches to try. Unfortunately, to the best of our falls of such powerful technology it is important to be knowledge no mathematical proof or even rigorous argu- able to control it. However, possibility of controlling mentation has been published demonstrating that the AI artificial general intelligence and its more advanced control problem may be solvable, even in principle, much version, superintelligence, has not been formally es- less in practice. tablished. In this paper we argue that advanced AI Yudkowsky considers the possibility that the control can’t be fully controlled. Consequences of uncontrol- problem is not solvable, but correctly insists that we should lability of AI are discussed with respect to future of study the problem in great detail before accepting such grave humanity and research on AI, and AI safety and se- limitation, he writes: “One common reaction I encounter is curity. for people to immediately declare that Friendly AI is an im- 1 possibility, because any sufficiently powerful AI will be 1 Introduction able to modify its own source code to break any constraints The unprecedented progress in Artificial Intelligence (AI) placed upon it.
    [Show full text]
  • Personal Universes: a Solution to the Multi-Agent Value Alignment Problem
    Personal Universes: A Solution to the Multi-Agent Value Alignment Problem Roman V. Yampolskiy Computer Engineering and Computer Science Speed School of Engineering University of Louisville [email protected] Abstract AI Safety researchers attempting to align values of highly capable intelligent systems with those of humanity face a number of challenges including personal value extraction, multi-agent value merger and finally in-silico encoding. State-of-the-art research in value alignment shows difficulties in every stage in this process, but merger of incompatible preferences is a particularly difficult challenge to overcome. In this paper we assume that the value extraction problem will be solved and propose a possible way to implement an AI solution which optimally aligns with individual preferences of each user. We conclude by analyzing benefits and limitations of the proposed approach. Keywords: AI Safety, Alternate Reality, Simulation, Value Alignment Problem, VR 1. Introduction to the Multi-Agent Value Alignment Problem Since the birth of the field of Artificial Intelligence (AI) researchers worked on creating ever capable machines, but with recent success in multiple subdomains of AI [1-7] safety and security of such systems and predicted future superintelligences [8, 9] has become paramount [10, 11]. While many diverse safety mechanisms are being investigated [12, 13], the ultimate goal is to align AI with goals, values and preferences of its users which is likely to include all of humanity. Value alignment problem [14], can be decomposed into three sub-problems, namely: personal value extraction from individual persons, combination of such personal preferences in a way, which is acceptable to all, and finally production of an intelligent system, which implements combined values of humanity.
    [Show full text]
  • Artificial Intelligence Safety and Security
    Artificial Intelligence Safety and Security Call for Book Chapters Introduction The history of robotics and artificial intelligence in many ways is also the history of humanity’s attempts to control such technologies. From the Golem of Prague to the military robots of modernity, the debate continues as to what degree of independence such entities should have and how to make sure that they do not turn on us, its inventors. Numerous recent advancements in all aspects of research, development and deployment of intelligent systems are well publicized but safety and security issues related to AI are rarely addressed. This book is proposed to mitigate this fundamental problem. It will be comprised of chapters from leading AI Safety researchers addressing different aspects of the AI control problem as it relates to the development of safe and secure artificial intelligence. The book would be the first textbook to address challenges of constructing safe and secure advanced machine intelligence. Contributions devoted to all aspects of AI safety and security are invited, in particular: Adversarial ML Explainability Security/Control AI Failures Friendliness Theory Singularity/Superintelligence Bad Actors/Malevolence Openness/Secrecy Validation/Verification Boxing/Confinement Off-Switch Value Alignment Drives/Goals Reward Engineering Weaponized/Military Ethics/Law Self-Improvement Wireheading Submission Procedure Researchers and practitioners are invited to submit their papers by email. Chapters and inquiries should be submitted to Dr. Roman Yampolskiy at [email protected]. Submissions should include authors' information (names, affiliations, emails, and short bios). Note: There are NO submission or acceptance fees for manuscripts submitted to this book for publication.
    [Show full text]
  • Artificial Intelligence Safety and Cybersecurity: a Timeline of AI Failures
    Artificial Intelligence Safety and Cybersecurity: a Timeline of AI Failures Roman V. Yampolskiy M. S. Spellchecker Computer Engineering and Computer Science Microsoft Corporation University of Louisville One Microsoft Way, Redmond, WA [email protected] [email protected] Abstract In this work, we present and analyze reported failures of artificially intelligent systems and extrapolate our analysis to future AIs. We suggest that both the frequency and the seriousness of future AI failures will steadily increase. AI Safety can be improved based on ideas developed by cybersecurity experts. For narrow AIs safety failures are at the same, moderate, level of criticality as in cybersecurity, however for general AI, failures have a fundamentally different impact. A single failure of a superintelligent system may cause a catastrophic event without a chance for recovery. The goal of cybersecurity is to reduce the number of successful attacks on the system; the goal of AI Safety is to make sure zero attacks succeed in bypassing the safety mechanisms. Unfortunately, such a level of performance is unachievable. Every security system will eventually fail; there is no such thing as a 100% secure system. Keywords: AI Safety, Cybersecurity, Failures, Superintelligence. 1. Introduction A day does not go by without a news article reporting some amazing breakthrough in artificial intelligence1. In fact progress in AI has been so steady that some futurologists, such as Ray Kurzweil, project current trends into the future and anticipate what the headlines of tomorrow will bring us. Consider some developments from the world of technology: 2004 DARPA sponsors a driverless car grand challenge. Technology developed by the participants eventually allows Google to develop a driverless automobile and modify existing transportation laws.
    [Show full text]
  • Taxonomy of Pathways to Dangerous Artificial Intelligence
    The Workshops of the Thirtieth AAAI Conference on Artificial Intelligence AI, Ethics, and Society: Technical Report WS-16-02 Taxonomy of Pathways to Dangerous Artificial Intelligence Roman V. Yampolskiy Computer Engineering and Computer Science, Speed School of Engineering, University of Louisville [email protected] Abstract includes latter codes to some example systems of each type In order to properly handle a dangerous Artificially Intelligent and explanations. (AI) system it is important to understand how the system came to be in such a state. In popular culture (science fiction mov- Table 1: Pathways to Dangerous AI ies/books) AIs/Robots became self-aware and as a result rebel External Causes Internal against humanity and decide to destroy it. While it is one pos- How and When did AI Causes become Dangerous On By Mistake Environment Independently sible scenario, it is probably the least likely path to appearance Purpose of dangerous AI. In this work, we survey, classify and analyze Pre-Deployment a c e g a number of circumstances, which might lead to arrival of ma- Post-Deployment b d f h licious AI. To the best of our knowledge, this is the first at- Timing tempt to systematically classify types of pathways leading to malevolent AI. Previous relevant work either surveyed spe- cific goals/meta-rules which might lead to malevolent behav- a. On Purpose – Pre-Deployment ior in AIs (Özkural 2014) or reviewed specific undesirable be- “Computer software is directly or indirectly responsible for haviors AGIs can exhibit at different stages of its develop- controlling many important aspects of our lives.
    [Show full text]
  • The Malicious Use of Artificial Intelligence: February 2018 Forecasting, Prevention, and Mitigation
    Future of University Centre for University of Center for a Electronic OpenAI Humanity of Oxford the Study of Cambridge New American Frontier Institute Existential Security Foundation Risk The Malicious Use February 2018 of Artificial Intelligence: Forecasting, Prevention, and Mitigation The Malicious Use of Artificial Intelligence: February 2018 Forecasting, Prevention, and Mitigation Miles Brundage Shahar Avin Jack Clark Helen Toner Peter Eckersley Ben Garfinkel Allan Dafoe Paul Scharre Thomas Zeitzoff Bobby Filar Hyrum Anderson Heather Roff Gregory C. Allen Jacob Steinhardt Carrick Flynn Seán Ó hÉigeartaigh Simon Beard Haydn Belfield Sebastian Farquhar Clare Lyle Rebecca Crootof Owain Evans Michael Page Joanna Bryson Roman Yampolskiy Dario Amodei 1 Corresponding author 9 American University 18 Centre for the Study of miles.brundage@philosophy. Existential Risk, University ox.ac.uk of Cambridge Future of Humanity Institute, 10 Endgame University of Oxford; Arizona State University 19 Future of Humanity 11 Endgame Institute, University of 2 Corresponding author, Oxford [email protected] University of Oxford/ Centre for the Study of 12 Arizona State University/New Existential Risk, University 20 Future of Humanity America Foundation of Cambridge Institute, University of Oxford 13 Center for a New American 3 OpenAI Security 21 Information Society Project, Yale University Open Philanthropy Project 4 14 Stanford University 22 Future of Humanity Electronic Frontier Institute, University of 5 Future of Humanity Foundation 15 Oxford Institute,
    [Show full text]