Euron Roboethics Roadmap

Total Page:16

File Type:pdf, Size:1020Kb

Euron Roboethics Roadmap EURON ROBOETHICS ROADMAP EURON Roboethics Roadmap Gianmarco Veruggio (*) Release 1.2 (January 2007) 1 of 42 EURON ROBOETHICS ROADMAP SUMMARY 1. INTRODUCTION....................................................................................................................5 1.1. Ethics & Robotics.............................................................................................................5 1.2. EURON (European Robotics Research Network) ...........................................................5 1.3. EURON Robotics Research Roadmap.............................................................................5 1.4. The Roboethics Atelier.....................................................................................................6 1.5. The Roboethics Roadmap ................................................................................................6 1.5.1. Scope: Near Future Urgency....................................................................................7 1.5.2. Target: Human Centred Ethics.................................................................................7 1.5.3. Methodology: Open Work .......................................................................................7 1.6. Pre-history ........................................................................................................................8 1.7. Documentary and background material ...........................................................................8 2. ETHIC AND ETHICS .............................................................................................................9 2.1. Ethical Issues..................................................................................................................10 2.2. Principles to be followed in Roboethics:........................................................................10 2.3. Ethical Issues in an ICT society .....................................................................................11 2.4. The precautionary principle ...........................................................................................12 3. ETHICS IN SCIENCE AND TECHNOLOGY.....................................................................13 3.1. Oath & pledge ................................................................................................................13 3.2. Code & guideline ...........................................................................................................13 3.3. Appeal ............................................................................................................................14 3.4. Recommendation............................................................................................................14 3.5. Manifesto........................................................................................................................15 3.6. Statement & declaration.................................................................................................15 3.7. Resolution.......................................................................................................................15 3.8. Convention .....................................................................................................................15 3.9. Charter............................................................................................................................15 3.10. Law.............................................................................................................................15 3.11. Conditions for Implementation ..................................................................................16 3.12. Operativeness of the Principles..................................................................................16 4. UNIVERSALLY ADOPTED ETHICAL PRINCIPLES.......................................................16 4.1. United Nations................................................................................................................16 4.2. Unesco............................................................................................................................17 4.3. Ilo - International Labor Organization ...........................................................................17 4.4. The Nuremberg Code.....................................................................................................17 4.5. World Medical Association............................................................................................18 4.6. World Summit on the Information Society....................................................................18 4.7. European Union..............................................................................................................18 4.8. Charter Of Fundamental Rights of the European Union (2000) ....................................19 4.9. General rules regarding ethics in EU research activities................................................19 Release 1.2 (January 2007) 2 of 42 EURON ROBOETHICS ROADMAP 5. Intercultural understanding and dialogue...............................................................................20 5.1. International Center for Information Ethics...................................................................20 5.2. International Society for Ethics and Information Technology.......................................21 5.3. Cultural Attitude Towards Technology and Communication Conference (CATaC) ....21 5.4. The Pugwash Conference on Science and World Affairs ..............................................21 5.5. Computer Professional for Social Responsibility (CPSR).............................................21 5.6. Makkula Center..............................................................................................................21 5.7. The Centre for Responsible Nanotechnology ................................................................21 5.8. Union of Concerned Scientists.......................................................................................22 5.9. The International Institute of Humanitarian Law...........................................................22 5.10. The World Transhumanist Association......................................................................22 6. ROBOTICS AND ETHICS ...................................................................................................22 6.1. Specificity of Robotics...................................................................................................22 6.2. About Roboethics...........................................................................................................23 6.3. What is a Robot? ............................................................................................................24 6.3.1. Robots are nothing but machines ...........................................................................24 6.3.2. Robots have ethical dimensions .............................................................................24 6.3.3. Robots as moral agents...........................................................................................24 6.3.4. Robots, evolution of a new specie.........................................................................24 6.4. Main positions on Roboethics........................................................................................24 6.4.1. Not interested in ethics...........................................................................................25 6.4.2. Interested in short-term ethical questions...............................................................25 6.4.3. Interested in long-term ethical concerns. ...............................................................25 6.5. About the Name .............................................................................................................25 6.6. Disciplines involved in Roboethics................................................................................25 7. ROBOETHICS TAXONOMY ..............................................................................................26 7.1. Humanoids .....................................................................................................................26 7.1.1. Artificial Mind........................................................................................................27 7.1.2. Artificial Body........................................................................................................28 7.1.3. Benefits...................................................................................................................28 7.1.4. Problems.................................................................................................................28 7.1.5. Recommendations ..................................................................................................29 7.2. Advanced production systems........................................................................................29 7.2.1. Industrial robotics...................................................................................................29 7.2.2. Benefits...................................................................................................................30 7.2.3. Problems.................................................................................................................30 7.2.4. Recommendations ..................................................................................................30
Recommended publications
  • Introduction to Robotics. Sensors and Actuators
    Introduction to Computer Vision and Robotics Florian Teich and Tomas Kulvicius* Introduction to Robotics. Sensors and Actuators Large portion of slides are adopted from Florentin Wörgötter, John Hallam and Simon Reich *[email protected] Perception-Action loop Environment Object Eyes Action Perception Arm Brain Perception-Action loop (cont.) Environment Sense Object Cameras Action Perception Robot-arm Computer Act Plan Outline of the course • L1.CV1: Introduction to Computer Vision. Thresholding, Filtering & Connected Coomponents • L2.CV2: Bilateral Filtering, Morphological Operators & Edge Detection • L3.CV3: Corner Detection & Non-Local Filtering • L4.R1: Introduction to Robotics. Sensors and actuators • L5.R2: Movement generation methods • L6.R3: Path planning algorithms • L7.CV4: Line/Circle Detection, Template Matching & Feature Detection • L8.CV5: Segmentation • L9.CV6: Fate Detection, Pedestrian Tracking & Computer Vision in 3D • L10.R4: Robot kinematics and control • L11.R5: Learning algorithms in robotics I: Supervised and unsupervised learning • L12.R6: Learning algorithms in robotics II: Reinforcement learning and genetic algorithms Introduction to Robotics History of robotics • Karel Čapek was one of the most influential Czech writers of the 20th century and a Nobel Prize nominee (1936). • He introduced and made popular the frequently used international word robot, which first appeared in his play R.U.R. (Rossum's Universal Robots) in 1921. 1890-1938 • “Robot” comes from the Czech word “robota”, meaning “forced labor” • Karel named his brother Josef Čapek as the true inventor of the word robot. History of robotics (cont.) • The word "robotics" also comes from science fiction - it first appeared in the short story "Runaround" (1942) by American writer Isaac Asimov.
    [Show full text]
  • AI, Robots, and Swarms: Issues, Questions, and Recommended Studies
    AI, Robots, and Swarms Issues, Questions, and Recommended Studies Andrew Ilachinski January 2017 Approved for Public Release; Distribution Unlimited. This document contains the best opinion of CNA at the time of issue. It does not necessarily represent the opinion of the sponsor. Distribution Approved for Public Release; Distribution Unlimited. Specific authority: N00014-11-D-0323. Copies of this document can be obtained through the Defense Technical Information Center at www.dtic.mil or contact CNA Document Control and Distribution Section at 703-824-2123. Photography Credits: http://www.darpa.mil/DDM_Gallery/Small_Gremlins_Web.jpg; http://4810-presscdn-0-38.pagely.netdna-cdn.com/wp-content/uploads/2015/01/ Robotics.jpg; http://i.kinja-img.com/gawker-edia/image/upload/18kxb5jw3e01ujpg.jpg Approved by: January 2017 Dr. David A. Broyles Special Activities and Innovation Operations Evaluation Group Copyright © 2017 CNA Abstract The military is on the cusp of a major technological revolution, in which warfare is conducted by unmanned and increasingly autonomous weapon systems. However, unlike the last “sea change,” during the Cold War, when advanced technologies were developed primarily by the Department of Defense (DoD), the key technology enablers today are being developed mostly in the commercial world. This study looks at the state-of-the-art of AI, machine-learning, and robot technologies, and their potential future military implications for autonomous (and semi-autonomous) weapon systems. While no one can predict how AI will evolve or predict its impact on the development of military autonomous systems, it is possible to anticipate many of the conceptual, technical, and operational challenges that DoD will face as it increasingly turns to AI-based technologies.
    [Show full text]
  • History of Robotics: Timeline
    History of Robotics: Timeline This history of robotics is intertwined with the histories of technology, science and the basic principle of progress. Technology used in computing, electricity, even pneumatics and hydraulics can all be considered a part of the history of robotics. The timeline presented is therefore far from complete. Robotics currently represents one of mankind’s greatest accomplishments and is the single greatest attempt of mankind to produce an artificial, sentient being. It is only in recent years that manufacturers are making robotics increasingly available and attainable to the general public. The focus of this timeline is to provide the reader with a general overview of robotics (with a focus more on mobile robots) and to give an appreciation for the inventors and innovators in this field who have helped robotics to become what it is today. RobotShop Distribution Inc., 2008 www.robotshop.ca www.robotshop.us Greek Times Some historians affirm that Talos, a giant creature written about in ancient greek literature, was a creature (either a man or a bull) made of bronze, given by Zeus to Europa. [6] According to one version of the myths he was created in Sardinia by Hephaestus on Zeus' command, who gave him to the Cretan king Minos. In another version Talos came to Crete with Zeus to watch over his love Europa, and Minos received him as a gift from her. There are suppositions that his name Talos in the old Cretan language meant the "Sun" and that Zeus was known in Crete by the similar name of Zeus Tallaios.
    [Show full text]
  • Report of Comest on Robotics Ethics
    SHS/YES/COMEST-10/17/2 REV. Paris, 14 September 2017 Original: English REPORT OF COMEST ON ROBOTICS ETHICS Within the framework of its work programme for 2016-2017, COMEST decided to address the topic of robotics ethics building on its previous reflection on ethical issues related to modern robotics, as well as the ethics of nanotechnologies and converging technologies. At the 9th (Ordinary) Session of COMEST in September 2015, the Commission established a Working Group to develop an initial reflection on this topic. The COMEST Working Group met in Paris in May 2016 to define the structure and content of a preliminary draft report, which was discussed during the 9th Extraordinary Session of COMEST in September 2016. At that session, the content of the preliminary draft report was further refined and expanded, and the Working Group continued its work through email exchanges. The COMEST Working Group then met in Quebec in March 2017 to further develop its text. A revised text in the form of a draft report was submitted to COMEST and the IBC in June 2017 for comments. The draft report was then revised based on the comments received. The final draft of the report was further discussed and revised during the 10th (Ordinary) Session of COMEST, and was adopted by the Commission on 14 September 2017. This document does not pretend to be exhaustive and does not necessarily represent the views of the Member States of UNESCO. – 2 – REPORT OF COMEST ON ROBOTICS ETHICS EXECUTIVE SUMMARY I. INTRODUCTION II. WHAT IS A ROBOT? II.1. The complexity of defining a robot II.2.
    [Show full text]
  • Robots and Healthcare Past, Present, and Future
    ROBOTS AND HEALTHCARE PAST, PRESENT, AND FUTURE COMPILED BY HOWIE BAUM What do you think of when you hear the word “robot”? 2 Why Robotics? Areas that robots are used: Industrial robots Military, government and space robots Service robots for home, healthcare, laboratory Why are robots used? Dangerous tasks or in hazardous environments Repetitive tasks High precision tasks or those requiring high quality Labor savings Control technologies: Autonomous (self-controlled), tele-operated (remote control) 3 The term “robot" was first used in 1920 in a play called "R.U.R." Or "Rossum's universal robots" by the Czech writer Karel Capek. The acclaimed Czech playwright (1890-1938) made the first use of the word from the Czech word “Robota” for forced labor or serf. Capek was reportedly several times a candidate for the Nobel prize for his works and very influential and prolific as a writer and playwright. ROBOTIC APPLICATIONS EXPLORATION- – Space Missions – Robots in the Antarctic – Exploring Volcanoes – Underwater Exploration MEDICAL SCIENCE – Surgical assistant Health Care ASSEMBLY- factories Parts Handling - Assembly - Painting - Surveillance - Security (bomb disposal, etc) - Home help (home sweeping (Roomba), grass cutting, or nursing) 7 Isaac Asimov, famous write of Science Fiction books, proposed his three "Laws of Robotics", and he later added a 'zeroth law'. Law Zero: A robot may not injure humanity, or, through inaction, allow humanity to come to harm. Law One: A robot may not injure a human being, or, through inaction, allow a human being to come to harm, unless this would violate a higher order law. Law Two: A robot must obey orders given it by human beings, except where such orders would conflict with a higher order law.
    [Show full text]
  • Dr. Asimov's Automatons
    Dr. Asimov’s Automatons Take on a Life of their Own Twenty years after his death, author Isaac Asimov’s robot fiction offers a blueprint to our robotic future...and the problems we could face by Alan S. Brown HIS PAST April, the University of Miami School These became the Three Laws. Today, they are the starting of Law held We Robot, the first-ever legal and point for any serious conversation about how humans and policy issues conference about robots. The name robots will behave around one another. of the conference, which brought together lawyers, As the mere fact of lawyers discussing robot law shows, engineers, and technologists, played on the title the issue is no longer theoretical. If robots are not yet of the most famous book intelligent, they are increasingly t ever written about robots, autonomous in how they carry out I, Robot, by Isaac Asimov. tasks. At the most basic level, every The point was underscored by day millions of robotic Roombas de- Laurie Silvers, president of Holly- cide how to navigate tables, chairs, wood Media Corp., which sponsored sofas, toys, and even pets as they the event. In 1991, Silvers founded vacuum homes. SyFy, a cable channel that specializ- At a more sophisticated level, es in science fiction. Within moments, autonomous robots help select in- she too had dropped Asimov’s name. ventory in warehouses, move and Silvers turned to Asimov for ad- position products in factories, and vice before launching SyFy. It was a care for patients in hospitals. South natural choice. Asimov was one of the Korea is testing robotic jailers.
    [Show full text]
  • The Three Laws of Robotics Machine Ethics (Or Machine Morality) Is A
    The Three Laws of Robotics Machine ethics (or machine morality) is a part of the ethics of AI concerned with the moral behavior of artificially intelligent beings. Machine ethics contrasts with roboethics, which is concerned with the moral behavior of humans as they design, construct, use and treat such beings. Machine ethics should not be confused with computer ethics, which focuses on professional behavior towards computers and information. Isaac Asimov considered the issue in the 1950s in his novel: I, Robot. He proposed the Three Laws of Robotics to govern artificially intelligent systems. Much of his work was then spent testing the boundaries of his three laws to see where they would break down, or where they would create paradoxical or unanticipated behavior. His work suggests that no set of fixed laws can sufficiently anticipate all possible circumstances. Asimov’s laws are still mentioned as a template for guiding our development of robots ; but given how much robotics has changed, since they appeared, and will continue to grow in the future, we need to ask how these rules could be updated for a 21st century version of artificial intelligence. Asimov’s suggested laws were devised to protect humans from interactions with robots. They are: A robot may not injure a human being or, through inaction, allow a human being to come to harm A robot must obey the orders given it by human beings except where such orders would conflict with the First Law A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws As mentioned, one of the obvious issues is that robots today appear to be far more varied than those in Asimov’s stories, including some that are far more simple.
    [Show full text]
  • Three Laws of Robotics
    Three Laws of Robotics From Wikipedia, the free encyclopedia The Three Laws of Robotics (often shortened to The Three Laws or Three Laws) are a set of rules devised by the science fiction author Isaac Asimov. The rules were introduced in his 1942 short story "Runaround", although they had been foreshadowed in a few earlier stories. The Three Laws are: 1. A robot may not injure a human being or, through inaction, allow a human being to come to harm. 2. A robot must obey the orders given to it by human beings, except where such orders would conflict with the First Law. 3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws. These form an organizing principle and unifying theme for Asimov's robotic-based fiction, appearing in his Robot series, the stories linked to it, and his Lucky Starr series of young-adult fiction. The Laws are incorporated into almost all of the positronic robots appearing in his fiction, and cannot be bypassed, being intended as a safety feature. Many of Asimov's robot-focused stories involve robots behaving in unusual and counter-intuitive ways as an unintended consequence of how the robot applies the Three Laws to the situation in which it finds itself. Other authors working in Asimov's fictional universe have adopted them and references, often parodic, appear throughout science fiction as well as in other genres. This cover of I, Robot illustrates the story "Runaround", the first to list all Three Laws of Robotics.
    [Show full text]
  • The First Law of Robotics ( a Call to Arms )
    :-0 appear, AAAI-94 The First Law of Robotics ( a call to arms ) Daniel Weld Oren Etzioni* Department of Computer Science and Engineering University of Washington Seattle, WA 98195 {weld, etzioni}~cs.washington.edu Abstract should not slavishly obey human commands-its fore- Even before the advent of Artificial Intelligence, sci- most goal should be to avoid harming humans. Con- ence fiction writer Isaac Asimov recognized that an sider the following scenarios: agent must place the protection of humans from harm .A construction robot is instructed to fill a pothole at a higher priority than obeying human orders. In- in the road. Although the robot repairs the cavity, spired by Asimov, we pose the following fundamental it leavesthe steam roller, chunks of tar, and an oil questions: (I) How should one formalize the rich, but slick in the middle of a busy highway. informal, notion of "harm"? (2) How can an agent avoid performing harmful actions, and do so in a com- .A softbot (software robot) is instructed to reduce putationally tractable manner? (3) How should an disk utilization below 90%. It succeeds, but in- agent resolve conflict between its goals and the need spection revealsthat the agent deleted irreplaceable to avoid harm? (4) When should an agent prevent a IjcTEXfiles without backing them up to tape. human from harming herself? While we address some of these questions in technical detail, the primary goal While less dramatic than Asimov's stories, the sce- of this paper is to focus attention on Asimov's concern: narios illustrate his point: not all ways of satisfying a society will reject autonomous agents unless we have human order are equally good; in fact, sometimesit is some credible means of making them safe! better not to satisfy the order at all.
    [Show full text]
  • Kush Agrawal the Ethics of Robotics Kush Agrawal, Delhi Public School
    Kush Agrawal The Ethics of Robotics Kush Agrawal, Delhi Public School, R.K Puram 2010 Kush Agrawal Abstract “1. A robot may not injure a human being or, through inaction, allow a human being to come to harm. 2. A robot must obey any orders given to it by human beings, except where such orders would conflict with the First Law. 3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.” [16] - Isaac Asimov These three laws first appeared together in Isaac Asimov’s story ‘Runaround’[15] after being mentioned in some form or the other in previous works by Asimov. These three laws commonly known as the three laws of robotics are the earliest forms of depiction for the needs of ethics in Robotics. In simplistic language Isaac Asimov is able to explain what rules a robot must confine itself to in order to maintain societal sanctity. However, even though they are outdated they still represent some of our innate fears which are beginning to resurface in present day 21st Century. Our society is on the advent of a new revolution; a revolution led by advances in Computer Science, Artificial Intelligence & Nanotechnology. Some of our advances have been so phenomenal that we surpassed what was predicted by the Moore’s law. With these advancements comes the fear that our future may be at the mercy of these androids. Humans today are scared that we, ourselves, might create something which we cannot control. We may end up creating something which can not only learn much faster than anyone of us can, but also evolve faster than what the theory of evolution has allowed us to.
    [Show full text]
  • Why the Three Laws of Robotics Do Not Work
    International Journal of Research in Engineering and Innovation Vol-2, Issue-2 (2018), 121-126 __________________________________________________________________________________________________________________________________ International Journal of Research in Engineering and Innovation (IJREI) journal home page: http://www.ijrei.com ISSN (Online): 2456-6934 _______________________________________________________________________________________ Why the three laws of robotics do not work Chris Stokes School of Philosophy, Wuhan University, Wuhan, China _______________________________________________________________________________________ Abstract This paper will be exploring the issue of safeguards around artificial intelligence. AI is a technological innovation that could potentially be created in the next few decades. There must be have controls in place before the creation of 'strong', sentient AI to avoid potentially catastrophic risks. Many AI researchers and computer engineers believe that the 'Three Laws of Robotics', written by Isaac Asimov, are sufficient controls. This paper aims to show that the Three Laws are actually inadequate to the task. This paper will look at the Three Laws of Robotics and explain why they are insufficient in terms of the safeguards that are required to protect humanity from rogue or badly programmed AI. It looks at each law individually and explain why it fails. The First Law fails because of ambiguity in language, and because of complicated ethical problems that are too complex to have a simple yes or no answer. The Second Law fails because of the unethical nature of having a law that requires sentient beings to remain as slaves. The Third Law fails because it results in a permanent social stratification, with the vast amount of potential exploitation built into this system of laws. The ‘Zeroth’ Law, like the first, fails because of ambiguous ideology.
    [Show full text]
  • Self-Reconfigurable Robots
    Self-Reconfigurable Robots An Introduction Kasper Stoy David Brandt David J. Christensen The MIT Press Cambridge, Massachusetts London, England 6 2010 Massachusetts Institute of Technology All rights reserved. No part of this book may be reproduced in any form by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher. For information about special quantity discounts, please email [email protected] This book was set in Times New Roman on 3B2 by Asco Typesetters, Hong Kong. Printed and bound in the United States of America. Library of Congress Cataloging-in-Publication Data Stoy, Kasper, 1973– Self-reconfigurable robots : an introduction / Kasper Stoy, David Brandt, and David J. Christensen. p. cm. — (Intelligent robotics and autonomous agents series) Includes bibliographical references and index. ISBN 978-0-262-01371-0 (hardcover : alk. paper) 1. Robotics. 2. Robots. I. Brandt, David, 1979– II. Christensen, David J. III. Title. TJ211.S794 2010 629.8092—dc22 2009021060 10987654321 1Self-Reconfigurable Robots Imagine that an unexpected event happens on a Mars planetary mission: a Mars rover has located a cave from which a strange light is coming. The entrance to the cave is too narrow for the rover to enter, but scientists are eager to examine the cave and find the light source. Luckily, they have brought a self-reconfigurable robot along. In the safe environment of their Martian habitat they assemble a snakelike robot and send it out on the surface of the planet through an airlock. The robot bites its own tail and forms a rolling track and rolls toward the cave at high speed.
    [Show full text]