Good and Safe Uses of AI Oracles
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Evidence-Based Policy Cause Area Report
Evidence-Based Policy Cause Area Report AUTHOR: 11/18 MARINELLA CAPRIATI, DPHIL 1 — Founders Pledge Animal Welfare Executive Summary By supporting increased use of evidence in the governments of low- and middle-income countries, donors can dramatically increase their impact on the lives of people living in poverty. This report explores how focusing on evidence-based policy provides an opportunity for leverage, and presents the most promising organisation we identified in this area. A high-risk/high-return opportunity for leverage In the 2013 report ‘The State of the Poor’, the World Bank reported that, as of 2010, roughly 83% of people in extreme poverty lived in countries classified as ‘lower-middle income’ or below. By far the most resources spent on tackling poverty come from local governments. American think tank the Brookings Institution found that, in 2011, $2.3 trillion of the $2.8 trillion spent on financing development came from domestic government revenues in the countries affected. There are often large differences in the effectiveness and cost-effectiveness of social programs— the amount of good done per dollar spent can vary significantly across programs. Employing evidence allows us to identify the most cost-effective social programs. This is useful information for donors choosing which charity to support, but also for governments choosing which programs to implement, and how. This suggests that employing philanthropic funding to improve the effectiveness of policymaking in low- and middle-income countries is likely to constitute an exceptional opportunity for leverage: by supporting the production and use of evidence in low- and middle-income countries, donors can potentially enable policy makers to implement more effective policies, thereby reaching many more people than direct interventions. -
An Evolutionary Heuristic for Human Enhancement
18 TheWisdomofNature: An Evolutionary Heuristic for Human Enhancement Nick Bostrom and Anders Sandberg∗ Abstract Human beings are a marvel of evolved complexity. Such systems can be difficult to enhance. When we manipulate complex evolved systems, which are poorly understood, our interventions often fail or backfire. It can appear as if there is a ‘‘wisdom of nature’’ which we ignore at our peril. Sometimes the belief in nature’s wisdom—and corresponding doubts about the prudence of tampering with nature, especially human nature—manifest as diffusely moral objections against enhancement. Such objections may be expressed as intuitions about the superiority of the natural or the troublesomeness of hubris, or as an evaluative bias in favor of the status quo. This chapter explores the extent to which such prudence-derived anti-enhancement sentiments are justified. We develop a heuristic, inspired by the field of evolutionary medicine, for identifying promising human enhancement interventions. The heuristic incorporates the grains of truth contained in ‘‘nature knows best’’ attitudes while providing criteria for the special cases where we have reason to believe that it is feasible for us to improve on nature. 1.Introduction 1.1. The wisdom of nature, and the special problem of enhancement We marvel at the complexity of the human organism, how its various parts have evolved to solve intricate problems: the eye to collect and pre-process ∗ Oxford Future of Humanity Institute, Faculty of Philosophy and James Martin 21st Century School, Oxford University. Forthcoming in Enhancing Humans, ed. Julian Savulescu and Nick Bostrom (Oxford: Oxford University Press) 376 visual information, the immune system to fight infection and cancer, the lungs to oxygenate the blood. -
ARCHITECTS of INTELLIGENCE for Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS of INTELLIGENCE
MARTIN FORD ARCHITECTS OF INTELLIGENCE For Xiaoxiao, Elaine, Colin, and Tristan ARCHITECTS OF INTELLIGENCE THE TRUTH ABOUT AI FROM THE PEOPLE BUILDING IT MARTIN FORD ARCHITECTS OF INTELLIGENCE Copyright © 2018 Packt Publishing All rights reserved. No part of this book may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, without the prior written permission of the publisher, except in the case of brief quotations embedded in critical articles or reviews. Every effort has been made in the preparation of this book to ensure the accuracy of the information presented. However, the information contained in this book is sold without warranty, either express or implied. Neither the author, nor Packt Publishing or its dealers and distributors, will be held liable for any damages caused or alleged to have been caused directly or indirectly by this book. Packt Publishing has endeavored to provide trademark information about all of the companies and products mentioned in this book by the appropriate use of capitals. However, Packt Publishing cannot guarantee the accuracy of this information. Acquisition Editors: Ben Renow-Clarke Project Editor: Radhika Atitkar Content Development Editor: Alex Sorrentino Proofreader: Safis Editing Presentation Designer: Sandip Tadge Cover Designer: Clare Bowyer Production Editor: Amit Ramadas Marketing Manager: Rajveer Samra Editorial Director: Dominic Shakeshaft First published: November 2018 Production reference: 2201118 Published by Packt Publishing Ltd. Livery Place 35 Livery Street Birmingham B3 2PB, UK ISBN 978-1-78913-151-2 www.packt.com Contents Introduction ........................................................................ 1 A Brief Introduction to the Vocabulary of Artificial Intelligence .......10 How AI Systems Learn ........................................................11 Yoshua Bengio .....................................................................17 Stuart J. -
Heidegger, Personhood and Technology
Comparative Philosophy Volume 2, No. 2 (2011): 23-49 Open Access / ISSN 2151-6014 www.comparativephilosophy.org THE FUTURE OF HUMANITY: HEIDEGGER, PERSONHOOD AND TECHNOLOGY MAHON O‟BRIEN ABSTRACT: This paper argues that a number of entrenched posthumanist positions are seriously flawed as a result of their dependence on a technical interpretive approach that creates more problems than it solves. During the course of our discussion we consider in particular the question of personhood. After all, until we can determine what it means to be a person we cannot really discuss what it means to improve a person. What kinds of enhancements would even constitute improvements? This in turn leads to an examination of the technical model of analysis and the recurring tendency to approach notions like personhood using this technical model. In looking to sketch a Heideggerian account of personhood, we are reaffirming what we take to be a Platonic skepticism concerning technical models of inquiry when it comes to certain subjects. Finally we examine the question as to whether the posthumanist looks to apply technology‟s benefits in ways that we have reflectively determined to be useful or desirable or whether it is technology itself (or to speak as Heidegger would – the “essence” of technology) which prompts many posthumanists to rely on an excessively reductionist view of the human being. Keywords: Heidegger, posthumanism, technology, personhood, temporality A significant number of Posthumanists1 advocate the techno-scientific enhancement of various human cognitive and physical capacities. Recent trends in posthumanist theory have witnessed the collective emergence, in particular, of a series of analytically oriented philosophers as part of the Future of Humanity Institute at O‟BRIEN, MAHON: IRCHSS Postdoctoral Research Fellow, School of Philosophy, University College Dublin, Ireland. -
Cognitive Biases Nn Discussing Cognitive Biases with Colleagues and Pupils
The RSA: an enlightenment organisation committed to finding innovative practical solutions to today’s social challenges. Through its ideas, research and 27,000-strong Fellowship it seeks to understand and enhance human capability so we can close the gap between today’s reality and people’s hopes for a better world. EVERYONE STARTS WITH AN ‘A’ Applying behavioural insight to narrow the socioeconomic attainment gap in education NATHALIE SPENCER, JONATHAN ROWSON, LOUISE BAMFIELD MARCH 2014 8 John Adam Street London WC2N 6EZ +44 (0) 20 7930 5115 Registered as a charity in England and Wales IN COOPERATION WITH THE no. 212424 VODAFONE FOUNDATION GERMANY Copyright © RSA 2014 www.thersa.org www.thersa.org For use in conjunction with Everyone Starts with an “A” 3 ways to use behavioural insight in the classroom By Spencer, Rowson, www.theRSA.org Bamfield (2014), available www.vodafone-stiftung.de A guide for teachers and school leaders at: www.thersa.org/startswitha www.lehrerdialog.net n Praising pupils for effort instead of intelligence to help instil the Whether you and your pupils believe that academic idea that effort is key and intelligence is not a fixed trait. For ability is an innate trait (a ‘fixed mindset’) or can be example, try “great, you kept practicing” instead of “great, you’re strengthened through effort and practice like a muscle really clever”. (a ‘growth mindset’) affects learning, resilience to n Becoming the lead learner. Educators can shape mindset setbacks, and performance. Mindset through modelling it for the pupils. The way that you (and parents) give feedback to Think about ability like a muscle Try: n Giving a “not yet” grade instead of a “fail” to set the expectation pupils can reinforce or attenuate a given mindset. -
The Transhumanist Reader Is an Important, Provocative Compendium Critically Exploring the History, Philosophy, and Ethics of Transhumanism
TH “We are in the process of upgrading the human species, so we might as well do it E Classical and Contemporary with deliberation and foresight. A good first step is this book, which collects the smartest thinking available concerning the inevitable conflicts, challenges and opportunities arising as we re-invent ourselves. It’s a core text for anyone making TRA Essays on the Science, the future.” —Kevin Kelly, Senior Maverick for Wired Technology, and Philosophy “Transhumanism has moved from a fringe concern to a mainstream academic movement with real intellectual credibility. This is a great taster of some of the best N of the Human Future emerging work. In the last 10 years, transhumanism has spread not as a religion but as a creative rational endeavor.” SHU —Julian Savulescu, Uehiro Chair in Practical Ethics, University of Oxford “The Transhumanist Reader is an important, provocative compendium critically exploring the history, philosophy, and ethics of transhumanism. The contributors anticipate crucial biopolitical, ecological and planetary implications of a radically technologically enhanced population.” M —Edward Keller, Director, Center for Transformative Media, Parsons The New School for Design A “This important book contains essays by many of the top thinkers in the field of transhumanism. It’s a must-read for anyone interested in the future of humankind.” N —Sonia Arrison, Best-selling author of 100 Plus: How The Coming Age of Longevity Will Change Everything IS The rapid pace of emerging technologies is playing an increasingly important role in T overcoming fundamental human limitations. The Transhumanist Reader presents the first authoritative and comprehensive survey of the origins and current state of transhumanist Re thinking regarding technology’s impact on the future of humanity. -
Classification Schemas for Artificial Intelligence Failures
Journal XX (XXXX) XXXXXX https://doi.org/XXXX/XXXX Classification Schemas for Artificial Intelligence Failures Peter J. Scott1 and Roman V. Yampolskiy2 1 Next Wave Institute, USA 2 University of Louisville, Kentucky, USA [email protected], [email protected] Abstract In this paper we examine historical failures of artificial intelligence (AI) and propose a classification scheme for categorizing future failures. By doing so we hope that (a) the responses to future failures can be improved through applying a systematic classification that can be used to simplify the choice of response and (b) future failures can be reduced through augmenting development lifecycles with targeted risk assessments. Keywords: artificial intelligence, failure, AI safety, classification 1. Introduction Artificial intelligence (AI) is estimated to have a $4-6 trillion market value [1] and employ 22,000 PhD researchers [2]. It is estimated to create 133 million new roles by 2022 but to displace 75 million jobs in the same period [6]. Projections for the eventual impact of AI on humanity range from utopia (Kurzweil, 2005) (p.487) to extinction (Bostrom, 2005). In many respects AI development outpaces the efforts of prognosticators to predict its progress and is inherently unpredictable (Yampolskiy, 2019). Yet all AI development is (so far) undertaken by humans, and the field of software development is noteworthy for unreliability of delivering on promises: over two-thirds of companies are more likely than not to fail in their IT projects [4]. As much effort as has been put into the discipline of software safety, it still has far to go. Against this background of rampant failures we must evaluate the future of a technology that could evolve to human-like capabilities, usually known as artificial general intelligence (AGI). -
Race in Academia
I, SOCIETY AI AND INEQUALITY: NEW DECADE, MORE CHALLENGES Mona Sloane, Ph.D. New York University [email protected] AI TECHNOLOGIES are becoming INTEGRAL to SOCIETY © DotEveryone SURVEILLANCE CAPITALISM TRADING in BEHAVIOURAL FUTURES © The Verge © The Innovation Scout TOTALIZING NARRATIVE of AI AI as a master technology = often either utopia or dystopia AI = human hubris? ”In 2014, the eminent scientists Stephen Hawking, Stuart Russell, Max Tegmark and Frank Wilczek wrote: ‘The potential benefits are huge; everything that civilisation has to offer is a product of human intelligence; we cannot predict what we might achieve when this intelligence is magnified by the tools that AI may provide, but the eradication of war, disease, and poverty would be high on anyone's list. Success in creating AI would be the biggest event in human history’.” Artificial General Intelligence? CORRELATION CAUSATION Race Race Gender Class INTERSECTIONAL INEQUALITY ETHICS EVERYWHERE EVERYWHERE EVERYWHERE EVERYWHERE EVERYWHERE EVERYWHERE EVERYWHERE ETHICS as POLICY APPROACH or as TECHNOLOGY APPROACH The ‘SOCIAL’ Problem ETHICS IS A SMOKESCREEN WHOSE HEAD – OR “INTELLIGENCE” – ARE TRYING TO MODEL, ANYWAY? THE FOUNDERS OF THE ”BIG 5 TECH COMPANIES”: AMAZON, FACEBOOK, ALPHABET, APPLE, MICROSOFT BIG TECH, BIG BUSINESS https://www.visualcapitalist.com/how-big-tech-makes-their-billions-2020/ 1956 DARTMOUTH CONFRENCE THE DIVERSITY ARGUMENT IS IMPORTANT. BUT IT IS INSUFFICIENT. EXTRACTION as a social super structure Prediction Dependency Extraction EXTRACTION Capitalism Racism White Supremacy Settler Colonialism Carcerality Heteronormativity … SCIENCE = ONTOLOGICAL AND EPISTEMOLOGICAL ENGINE Real world consequences of classification systems: racial classification in apartheid South Africa Ruha Benjamin, Alondra Nelson, Charlton McIlwain, Meredith Broussard, Andre Brock, Safiya Noble, Mutale Nkonde, Rediet Abebe, Timnit Gebru, Joy Buolamwini, Jessie Daniels, Sasha Costanza- Chock, Marie Hicks, Anna Lauren Hoffmann, Laura Forlano, Danya Glabau, Kadija Ferryman, Vincent Southerland – and more. -
Global Catastrophic Risks Survey
GLOBAL CATASTROPHIC RISKS SURVEY (2008) Technical Report 2008/1 Published by Future of Humanity Institute, Oxford University Anders Sandberg and Nick Bostrom At the Global Catastrophic Risk Conference in Oxford (17‐20 July, 2008) an informal survey was circulated among participants, asking them to make their best guess at the chance that there will be disasters of different types before 2100. This report summarizes the main results. The median extinction risk estimates were: Risk At least 1 million At least 1 billion Human extinction dead dead Number killed by 25% 10% 5% molecular nanotech weapons. Total killed by 10% 5% 5% superintelligent AI. Total killed in all 98% 30% 4% wars (including civil wars). Number killed in 30% 10% 2% the single biggest engineered pandemic. Total killed in all 30% 10% 1% nuclear wars. Number killed in 5% 1% 0.5% the single biggest nanotech accident. Number killed in 60% 5% 0.05% the single biggest natural pandemic. Total killed in all 15% 1% 0.03% acts of nuclear terrorism. Overall risk of n/a n/a 19% extinction prior to 2100 These results should be taken with a grain of salt. Non‐responses have been omitted, although some might represent a statement of zero probability rather than no opinion. 1 There are likely to be many cognitive biases that affect the result, such as unpacking bias and the availability heuristic‒‐well as old‐fashioned optimism and pessimism. In appendix A the results are plotted with individual response distributions visible. Other Risks The list of risks was not intended to be inclusive of all the biggest risks. -
The Time for a National Study on AI Is NOW
The Time For a National Study On AI is NOW Subscribe Past Issues Translate RSS View this email in your browser "I think we should be very careful about artificial intelligence. If I were to guess at what our biggest existential threat is, it’s probably that. With artificial intelligence, we are summoning the demon." Elon Musk, CEO, Tesla Motors & SpaceX ARTIFICIAL INTELLIGENCE STUDY: AI TODAY, AI TOMORROW We are following up on our last message about a proposed National Study on Artificial Intelligence. We very much appreciate the responses and encouragement we have received, and welcome input and questions from any League and/or its members on this subject. http://mailchi.mp/35a449e1979f/the-time-for-a-national-study-on-ai-is-now?e=[UNIQID][12/20/2017 6:29:08 PM] The Time For a National Study On AI is NOW We believe that the pace of development of AI and the implications of its advancement make this an extremely timely subject. AI is becoming an important part of our society - from medical uses to transportation to banking to national defense. Putting off this study will put us behind in the progress being made in AI. We again request that your League consider supporting this study at your program planning meeting. The reason to propose this study now is because there are currently no positions that address many of the issues that fall under the umbrella of this relatively new and powerful technology. For several years now private groups and even countries (like Russia, China and Korea) have made it a priority to develop mechanical brains/computers that can function intelligently http://mailchi.mp/35a449e1979f/the-time-for-a-national-study-on-ai-is-now?e=[UNIQID][12/20/2017 6:29:08 PM] The Time For a National Study On AI is NOW and command robots or other machines to perform some task. -
“Is Cryonics an Ethical Means of Life Extension?” Rebekah Cron University of Exeter 2014
1 “Is Cryonics an Ethical Means of Life Extension?” Rebekah Cron University of Exeter 2014 2 “We all know we must die. But that, say the immortalists, is no longer true… Science has progressed so far that we are morally bound to seek solutions, just as we would be morally bound to prevent a real tsunami if we knew how” - Bryan Appleyard 1 “The moral argument for cryonics is that it's wrong to discontinue care of an unconscious person when they can still be rescued. This is why people who fall unconscious are taken to hospital by ambulance, why they will be maintained for weeks in intensive care if necessary, and why they will still be cared for even if they don't fully awaken after that. It is a moral imperative to care for unconscious people as long as there remains reasonable hope for recovery.” - ALCOR 2 “How many cryonicists does it take to screw in a light bulb? …None – they just sit in the dark and wait for the technology to improve” 3 - Sterling Blake 1 Appleyard 2008. Page 22-23 2 Alcor.org: ‘Frequently Asked Questions’ 2014 3 Blake 1996. Page 72 3 Introduction Biologists have known for some time that certain organisms can survive for sustained time periods in what is essentially a death"like state. The North American Wood Frog, for example, shuts down its entire body system in winter; its heart stops beating and its whole body is frozen, until summer returns; at which point it thaws and ‘comes back to life’ 4. -
The Definition of Effective Altruism
OUP CORRECTED PROOF – FINAL, 19/08/19, SPi 1 The Definition of Effective Altruism William MacAskill There are many problems in the world today. Over 750 million people live on less than $1.90 per day (at purchasing power parity).1 Around 6 million children die each year of easily preventable causes such as malaria, diarrhea, or pneumonia.2 Climate change is set to wreak environmental havoc and cost the economy tril- lions of dollars.3 A third of women worldwide have suffered from sexual or other physical violence in their lives.4 More than 3,000 nuclear warheads are in high-alert ready-to-launch status around the globe.5 Bacteria are becoming antibiotic- resistant.6 Partisanship is increasing, and democracy may be in decline.7 Given that the world has so many problems, and that these problems are so severe, surely we have a responsibility to do something about them. But what? There are countless problems that we could be addressing, and many different ways of addressing each of those problems. Moreover, our resources are scarce, so as individuals and even as a globe we can’t solve all these problems at once. So we must make decisions about how to allocate the resources we have. But on what basis should we make such decisions? The effective altruism movement has pioneered one approach. Those in this movement try to figure out, of all the different uses of our resources, which uses will do the most good, impartially considered. This movement is gathering con- siderable steam. There are now thousands of people around the world who have chosen