Artificial Intelligence Safety and Cybersecurity: a Timeline of AI Failures
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Safety of Artificial Superintelligence
Environment. Technology. Resources. Rezekne, Latvia Proceedings of the 12th International Scientific and Practical Conference. Volume II, 180-183 Safety of Artificial Superintelligence Aleksejs Zorins Peter Grabusts Faculty of Engineering Faculty of Engineering Rezekne Academy of Technologies Rezekne Academy of Technologies Rezekne, Latvia Rezekne, Latvia [email protected] [email protected] Abstract—The paper analyses an important problem of to make a man happy it will do it with the fastest and cyber security from human safety perspective which is usu- cheapest (in terms of computational resources) without ally described as data and/or computer safety itself with- using a common sense (for example killing of all people out mentioning the human. There are numerous scientific will lead to the situation that no one is unhappy or the predictions of creation of artificial superintelligence, which decision to treat human with drugs will also make him could arise in the near future. That is why the strong neces- sity for protection of such a system from causing any farm happy etc.). Another issue is that we want computer to arises. This paper reviews approaches and methods already that we want, but due to bugs in the code computer will presented for solving this problem in a single article, analy- do what the code says, and in the case of superintelligent ses its results and provides future research directions. system, this may be a disaster. Keywords—Artificial Superintelligence, Cyber Security, The next sections of the paper will show the possible Singularity, Safety of Artificial Intelligence. solutions of making superintelligence safe to humanity. -
Liability Law for Present and Future Robotics Technology Trevor N
Liability Law for Present and Future Robotics Technology Trevor N. White and Seth D. Baum Global Catastrophic Risk Institute http://gcrinstitute.org * http://sethbaum.com Published in Robot Ethics 2.0, edited by Patrick Lin, George Bekey, Keith Abney, and Ryan Jenkins, Oxford University Press, 2017, pages 66-79. This version 10 October 2017 Abstract Advances in robotics technology are causing major changes in manufacturing, transportation, medicine, and a number of other sectors. While many of these changes are beneficial, there will inevitably be some harms. Who or what is liable when a robot causes harm? This paper addresses how liability law can and should account for robots, including robots that exist today and robots that potentially could be built at some point in the near or distant future. Already, robots have been implicated in a variety of harms. However, current and near-future robots pose no significant challenge for liability law: they can be readily handled with existing liability law or minor variations thereof. We show this through examples from medical technology, drones, and consumer robotics. A greater challenge will arise if it becomes possible to build robots that merit legal personhood and thus can be held liable. Liability law for robot persons could draw on certain precedents, such as animal liability. However, legal innovations will be needed, in particular for determining which robots merit legal personhood. Finally, a major challenge comes from the possibility of future robots that could cause major global catastrophe. As with other global catastrophic risks, liability law could not apply, because there would be no post- catastrophe legal system to impose liability. -
Classification Schemas for Artificial Intelligence Failures
Journal XX (XXXX) XXXXXX https://doi.org/XXXX/XXXX Classification Schemas for Artificial Intelligence Failures Peter J. Scott1 and Roman V. Yampolskiy2 1 Next Wave Institute, USA 2 University of Louisville, Kentucky, USA [email protected], [email protected] Abstract In this paper we examine historical failures of artificial intelligence (AI) and propose a classification scheme for categorizing future failures. By doing so we hope that (a) the responses to future failures can be improved through applying a systematic classification that can be used to simplify the choice of response and (b) future failures can be reduced through augmenting development lifecycles with targeted risk assessments. Keywords: artificial intelligence, failure, AI safety, classification 1. Introduction Artificial intelligence (AI) is estimated to have a $4-6 trillion market value [1] and employ 22,000 PhD researchers [2]. It is estimated to create 133 million new roles by 2022 but to displace 75 million jobs in the same period [6]. Projections for the eventual impact of AI on humanity range from utopia (Kurzweil, 2005) (p.487) to extinction (Bostrom, 2005). In many respects AI development outpaces the efforts of prognosticators to predict its progress and is inherently unpredictable (Yampolskiy, 2019). Yet all AI development is (so far) undertaken by humans, and the field of software development is noteworthy for unreliability of delivering on promises: over two-thirds of companies are more likely than not to fail in their IT projects [4]. As much effort as has been put into the discipline of software safety, it still has far to go. Against this background of rampant failures we must evaluate the future of a technology that could evolve to human-like capabilities, usually known as artificial general intelligence (AGI). -
On the Differences Between Human and Machine Intelligence
On the Differences between Human and Machine Intelligence Roman V. Yampolskiy Computer Science and Engineering, University of Louisville [email protected] Abstract [Legg and Hutter, 2007a]. However, widespread implicit as- Terms Artificial General Intelligence (AGI) and Hu- sumption of equivalence between capabilities of AGI and man-Level Artificial Intelligence (HLAI) have been HLAI appears to be unjustified, as humans are not general used interchangeably to refer to the Holy Grail of Ar- intelligences. In this paper, we will prove this distinction. tificial Intelligence (AI) research, creation of a ma- Others use slightly different nomenclature with respect to chine capable of achieving goals in a wide range of general intelligence, but arrive at similar conclusions. “Lo- environments. However, widespread implicit assump- cal generalization, or “robustness”: … “adaptation to tion of equivalence between capabilities of AGI and known unknowns within a single task or well-defined set of HLAI appears to be unjustified, as humans are not gen- tasks”. … Broad generalization, or “flexibility”: “adapta- eral intelligences. In this paper, we will prove this dis- tion to unknown unknowns across a broad category of re- tinction. lated tasks”. …Extreme generalization: human-centric ex- treme generalization, which is the specific case where the 1 Introduction1 scope considered is the space of tasks and domains that fit within the human experience. We … refer to “human-cen- Imagine that tomorrow a prominent technology company tric extreme generalization” as “generality”. Importantly, as announces that they have successfully created an Artificial we deliberately define generality here by using human cog- Intelligence (AI) and offers for you to test it out. -
The Dangers of Human-Like Bias in Machine-Learning Algorithms
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Missouri University of Science and Technology (Missouri S&T): Scholars' Mine Missouri S&T’s Peer to Peer Volume 2 Issue 1 Article 1 May 2018 The Dangers of Human-Like Bias in Machine-Learning Algorithms Daniel James Fuchs Missouri University of Science and Technology Follow this and additional works at: https://scholarsmine.mst.edu/peer2peer Part of the Computer Sciences Commons Recommended Citation Fuchs, Daniel J.. 2018. "The Dangers of Human-Like Bias in Machine-Learning Algorithms." Missouri S&T’s Peer to Peer 2, (1). https://scholarsmine.mst.edu/peer2peer/vol2/iss1/1 This Article - Journal is brought to you for free and open access by Scholars' Mine. It has been accepted for inclusion in Missouri S&T’s Peer to Peer by an authorized administrator of Scholars' Mine. This work is protected by U. S. Copyright Law. Unauthorized use including reproduction for redistribution requires the permission of the copyright holder. For more information, please contact [email protected]. Fuchs: Dangers of Human-Like Bias in MLAGs Machine learning (ML), frequently used in constructing artificial intelligence, relies on observing trends in data and forming relationships through pattern recognition. Machine learning algorithms, or MLAGs, use these relationships to solve various complex problems. Applications can range from Google's "Cleverbot" to résumé evaluation, to predicting the risk of a convicted criminal reoffending (Temming 2017). Naturally, by learning through data observation rather than being explicitly programmed to perform a certain way, MLAGs will develop biases towards certain types of input. -
Beyond Dyadic Interactions: Considering Chatbots As Community Members
CHI 2019 Paper CHI 2019, May 4–9, 2019, Glasgow, Scotland, UK Beyond Dyadic Interactions: Considering Chatbots as Community Members Joseph Seering Michal Luria Carnegie Mellon University Carnegie Mellon University Pittsburgh, Pennsylvania, USA Pittsburgh, Pennsylvania, USA [email protected] [email protected] Geoff Kaufman Jessica Hammer Carnegie Mellon University Carnegie Mellon University Pittsburgh, Pennsylvania, USA Pittsburgh, Pennsylvania, USA [email protected] [email protected] ABSTRACT KEYWORDS Chatbots have grown as a space for research and develop- Chatbots; social identity; online communities; dyadic com- ment in recent years due both to the realization of their munication commercial potential and to advancements in language pro- ACM Reference Format: cessing that have facilitated more natural conversations. Joseph Seering, Michal Luria, Geoff Kaufman, and Jessica Hammer. However, nearly all chatbots to date have been designed 2019. Beyond Dyadic Interactions: Considering Chatbots as Com- for dyadic, one-on-one communication with users. In this munity Members. In CHI Conference on Human Factors in Computing paper we present a comprehensive review of research on Systems Proceedings (CHI 2019), May 4–9, 2019, Glasgow, Scotland chatbots supplemented by a review of commercial and in- UK. ACM, New York, NY, USA, 13 pages. https://doi.org/10.1145/ dependent chatbots. We argue that chatbots’ social roles 3290605.3300680 and conversational capabilities beyond dyadic interactions have been underexplored, and that expansion into this de- 1 INTRODUCTION sign space could support richer social interactions in online From ELIZA to Tay, for half a century chatbots have mixed communities and help address the longstanding challenges provision of services with an attempt to emulate human con- of maintaining, moderating, and growing these communities. -
A Feminist Critique of Algorithmic Fairness
Redistribution and Rekognition: A Feminist Critique of Algorithmic Fairness Sarah Myers West AI Now Institute at New York University [email protected] Abstract Computer scientists, and artificial intelligence researchers in particular, have a predisposition for adopting precise, fixed definitions to serve as classifiers (Agre, 1997; Broussard, 2018). But classification is an enactment of power; it orders human interaction in ways that produce advantage and suffering (Bowker & Star, 1999). In so doing, it attempts to create order out of the messiness of human life, masking the work of the people involved in training machine learning systems, and hiding the uneven distribution of its impacts on communities (A. Taylor, 2018; Gray, 2019; Roberts, 2019). Feminist scholars, and particularly feminist scholars of color, have made powerful critiques of the ways in which artificial intelligence systems formalize, classify, and amplify historical forms of discrimination and act to reify and amplify existing forms of social inequality (Eubanks, 2017; Benjamin, 2019; Noble, 2018). In response, the machine learning community has begun to address claims of algorithmic bias under the rubric of fairness, accountability, and transparency. But it has dealt with these claims largely using computational approaches that obscure difference. Inequality is reflected and amplified in algorithmic systems in ways that exceed the capacity of statistical methods alone. This article examines how patterns of exclusion and erasure in algorithmic systems recapitulate and magnify a history of discrimination and erasure in the field of artificial intelligence, and in society more broadly. Shifting from individualized notions of fairness to more situated modeling of algorithmic remediation might create spaces of possibility for new forms of solidarity and West, Sarah Myers (2020). -
Essentializing Femininity in AI Linguistics
Student Publications Student Scholarship Fall 2019 Designing Women: Essentializing Femininity in AI Linguistics Ellianie S. Vega Gettysburg College Follow this and additional works at: https://cupola.gettysburg.edu/student_scholarship Part of the Artificial Intelligence and Robotics Commons, Feminist, Gender, and Sexuality Studies Commons, and the Linguistics Commons Share feedback about the accessibility of this item. Recommended Citation Vega, Ellianie S., "Designing Women: Essentializing Femininity in AI Linguistics" (2019). Student Publications. 797. https://cupola.gettysburg.edu/student_scholarship/797 This open access student research paper is brought to you by The Cupola: Scholarship at Gettysburg College. It has been accepted for inclusion by an authorized administrator of The Cupola. For more information, please contact [email protected]. Designing Women: Essentializing Femininity in AI Linguistics Abstract Since the eighties, feminists have considered technology a force capable of subverting sexism because of technology’s ability to produce unbiased logic. Most famously, Donna Haraway’s “A Cyborg Manifesto” posits that the cyborg has the inherent capability to transcend gender because of its removal from social construct and lack of loyalty to the natural world. But while humanoids and artificial intelligence have been imagined as inherently subversive to gender, current artificial intelligence perpetuates gender divides in labor and language as their programmers imbue them with traits considered “feminine.” A majority of 21st century AI and humanoids are programmed to fit emalef stereotypes as they fulfill emotional labor and perform pink-collar tasks, whether through roles as therapists, query-fillers, or companions. This paper examines four specific chat-based AI --ELIZA, XiaoIce, Sophia, and Erica-- and examines how their feminine linguistic patterns are used to maintain the illusion of emotional understanding in regards to the tasks that they perform. -
AI and You Transcript Guest: Roman Yampolskiy Episode 16 First Aired: Monday, October 5, 2020
AI and You Transcript Guest: Roman Yampolskiy Episode 16 First Aired: Monday, October 5, 2020 Hi, welcome to episode 16. Today we have a feast, because I am talking with Dr. Roman Yampolskiy, a legend within the AI community. He is a tenured associate professor in the department of Computer Engineering and Computer Science at the University of Louisville in Kentucky where he is the director of Cyber Security. Look up his home page and Wikipedia entry because he has qualifications too numerous to detail here. Some of those include that he is also a Visiting Fellow at the Singularity Institute, he was also one of the scientists at the Asilomar conference on the future of AI, and the author of over 100 publications, including the book Artificial Superintelligence: A Futuristic Approach, which I would strongly recommend, and many peer-reviewed papers, including one coauthored with yours truly. Most of them are on the subject of AI Safety. In this episode we will talk mostly about a recent paper of his, a table-thumping 73 pages titled “On Controllability of Artificial Intelligence,” which you may recognize is talking about what we call the Control Problem. That’s the most fundamental problem in the whole study of the long-term impact of AI, and Roman explains it very well in this episode. If you’re looking at the possible future of artificial intelligence having serious negative as well as positive consequences for everyone, which is a concept that’s gotten a lot of popularization in the last few years, of course, and you’re finding computer professionals pooh-poohing that idea because AI right now requires a great deal of specialized effort on their part to do even narrow tasks, and the idea of it becoming an existential threat is just too sensationalist, the product of watching too many Terminator movies, then pay attention to Roman, because he has impeccable computer science credentials and yet does not shy away from exploring fundamental threats of AI. -
The Age of Artificial Intelligence an Exploration
THE AGE OF ARTIFICIAL INTELLIGENCE AN EXPLORATION Edited by Steven S. Gouveia University of Minho, Portugal Cognitive Science and Psychology Copyright © 2020 by the Authors. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior permission of Vernon Art and Science Inc. www.vernonpress.com In the Americas: In the rest of the world: Vernon Press Vernon Press 1000 N West Street, C/Sancti Espiritu 17, Suite 1200, Wilmington, Malaga, 29006 Delaware 19801 Spain United States Cognitive Science and Psychology Library of Congress Control Number: 2020931461 ISBN: 978-1-62273-872-4 Product and company names mentioned in this work are the trademarks of their respective owners. While every care has been taken in preparing this work, neither the authors nor Vernon Art and Science Inc. may be held responsible for any loss or damage caused or alleged to be caused directly or indirectly by the information contained in it. Every effort has been made to trace all copyright holders, but if any have been inadvertently overlooked the publisher will be pleased to include any necessary credits in any subsequent reprint or edition. Cover design by Vernon Press using elements designed by FreePik. TABLE OF CONTENTS LIST OF ACRONYMS vii LIST OF FIGURES xi LIST OF TABLES xiii INTRODUCTION xv SECTION I: INTELLIGENCE IN ARTIFICIAL INTELLIGENCE 1 CHAPTER 1 TOWARDS THE MATHEMATICS OF INTELLIGENCE 3 Soenke Ziesche Maldives National University, Maldives Roman V. Yampolskiy University of Louisville, USA CHAPTER 2 MINDS , BRAINS AND TURING 15 Stevan Harnad Université du Québec à Montréal, Canada; University of Southampton, UK CHAPTER 3 THE AGE OF POST -INTELLIGENT DESIGN 27 Daniel C. -
Leakproofing the Singularity Artificial Intelligence Confinement Problem
Roman V.Yampolskiy Leakproofing the Singularity Artificial Intelligence Confinement Problem Abstract: This paper attempts to formalize and to address the ‘leakproofing’ of the Singularity problem presented by David Chalmers. The paper begins with the definition of the Artificial Intelli- gence Confinement Problem. After analysis of existing solutions and their shortcomings, a protocol is proposed aimed at making a more secure confinement environment which might delay potential negative effect from the technological singularity while allowing humanity to benefit from the superintelligence. Keywords: AI-Box, AI Confinement Problem, Hazardous Intelligent Software, Leakproof Singularity, Oracle AI. ‘I am the slave of the lamp’ Genie from Aladdin 1. Introduction With the likely development of superintelligent programs in the near future, many scientists have raised the issue of safety as it relates to such technology (Yudkowsky, 2008; Bostrom, 2006; Hibbard, 2005; Chalmers, 2010; Hall, 2000). A common theme in Artificial Intelli- gence (AI)1 safety research is the possibility of keeping a super- intelligent agent in a sealed hardware so as to prevent it from doing any harm to humankind. Such ideas originate with scientific Correspondence: Roman V. Yampolskiy, Department of Computer Engineering and Computer Sci- ence University of Louisville. Email: [email protected] [1] In this paper the term AI is used to represent superintelligence. Journal of Consciousness Studies, 19, No. 1–2, 2012, pp. 194–214 Copyright (c) Imprint Academic 2011 For personal use only -- not for reproduction LEAKPROOFING THE SINGULARITY 195 visionaries such as Eric Drexler, who has suggested confining transhuman machines so that their outputs could be studied and used safely (Drexler, 1986). -
Agent Foundations for Aligning Machine Intelligence with Human Interests: a Technical Research Agenda
Agent Foundations for Aligning Machine Intelligence with Human Interests: A Technical Research Agenda In The Technological Singularity: Managing the Journey. Springer. 2017 Nate Soares and Benya Fallenstein Machine Intelligence Research Institute fnate,[email protected] Contents 1 Introduction 1 1.1 Why These Problems? . .2 2 Highly Reliable Agent Designs 3 2.1 Realistic World-Models . .4 2.2 Decision Theory . .5 2.3 Logical Uncertainty . .6 2.4 Vingean Reflection . .7 3 Error-Tolerant Agent Designs 8 4 Value Specification 9 5 Discussion 11 5.1 Toward a Formal Understanding of the Problem . 11 5.2 Why Start Now? . 11 1 Introduction Since artificial agents would not share our evolution- ary history, there is no reason to expect them to be The property that has given humans a dominant ad- driven by human motivations such as lust for power. vantage over other species is not strength or speed, but However, nearly all goals can be better met with more intelligence. If progress in artificial intelligence continues resources (Omohundro 2008). This suggests that, by unabated, AI systems will eventually exceed humans in default, superintelligent agents would have incentives general reasoning ability. A system that is \superintelli- to acquire resources currently being used by humanity. gent" in the sense of being \smarter than the best human (Just as artificial agents would not automatically acquire brains in practically every field” could have an enormous a lust for power, they would not automatically acquire a impact upon humanity (Bostrom 2014). Just as human human sense of fairness, compassion, or conservatism.) intelligence has allowed us to develop tools and strate- Thus, most goals would put the agent at odds with gies for controlling our environment, a superintelligent human interests, giving it incentives to deceive or ma- system would likely be capable of developing its own nipulate its human operators and resist interventions tools and strategies for exerting control (Muehlhauser designed to change or debug its behavior (Bostrom 2014, and Salamon 2012).