Dissertation Ist an Der TU Wien Bibliothek Verfügbar
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Artificial Intelligence and the Ethics of Self-Learning Robots Shannon Vallor Santa Clara University, [email protected]
Santa Clara University Scholar Commons Philosophy College of Arts & Sciences 10-3-2017 Artificial Intelligence and the Ethics of Self-learning Robots Shannon Vallor Santa Clara University, [email protected] George A. Bekey Follow this and additional works at: http://scholarcommons.scu.edu/phi Part of the Philosophy Commons Recommended Citation Vallor, S., & Bekey, G. A. (2017). Artificial Intelligence and the Ethics of Self-learning Robots. In P. Lin, K. Abney, & R. Jenkins (Eds.), Robot Ethics 2.0 (pp. 338–353). Oxford University Press. This material was originally published in Robot Ethics 2.0 edited by Patrick Lin, Keith Abney, and Ryan Jenkins, and has been reproduced by permission of Oxford University Press. For permission to reuse this material, please visit http://www.oup.co.uk/academic/rights/permissions. This Book Chapter is brought to you for free and open access by the College of Arts & Sciences at Scholar Commons. It has been accepted for inclusion in Philosophy by an authorized administrator of Scholar Commons. For more information, please contact [email protected]. ARTIFICIAL INTELLIGENCE AND 22 THE ETHICS OF SELF - LEARNING ROBOTS Shannon Va ll or and George A. Bekey The convergence of robotics technology with the science of artificial intelligence (or AI) is rapidly enabling the development of robots that emulate a wide range of intelligent human behaviors.1 Recent advances in machine learning techniques have produced significant gains in the ability of artificial agents to perform or even excel in activities for merly thought to be the exclusive province of human intelligence, including abstract problem-solving, perceptual recognition, social interaction, and natural language use. -
Introduction to Robotics. Sensors and Actuators
Introduction to Computer Vision and Robotics Florian Teich and Tomas Kulvicius* Introduction to Robotics. Sensors and Actuators Large portion of slides are adopted from Florentin Wörgötter, John Hallam and Simon Reich *[email protected] Perception-Action loop Environment Object Eyes Action Perception Arm Brain Perception-Action loop (cont.) Environment Sense Object Cameras Action Perception Robot-arm Computer Act Plan Outline of the course • L1.CV1: Introduction to Computer Vision. Thresholding, Filtering & Connected Coomponents • L2.CV2: Bilateral Filtering, Morphological Operators & Edge Detection • L3.CV3: Corner Detection & Non-Local Filtering • L4.R1: Introduction to Robotics. Sensors and actuators • L5.R2: Movement generation methods • L6.R3: Path planning algorithms • L7.CV4: Line/Circle Detection, Template Matching & Feature Detection • L8.CV5: Segmentation • L9.CV6: Fate Detection, Pedestrian Tracking & Computer Vision in 3D • L10.R4: Robot kinematics and control • L11.R5: Learning algorithms in robotics I: Supervised and unsupervised learning • L12.R6: Learning algorithms in robotics II: Reinforcement learning and genetic algorithms Introduction to Robotics History of robotics • Karel Čapek was one of the most influential Czech writers of the 20th century and a Nobel Prize nominee (1936). • He introduced and made popular the frequently used international word robot, which first appeared in his play R.U.R. (Rossum's Universal Robots) in 1921. 1890-1938 • “Robot” comes from the Czech word “robota”, meaning “forced labor” • Karel named his brother Josef Čapek as the true inventor of the word robot. History of robotics (cont.) • The word "robotics" also comes from science fiction - it first appeared in the short story "Runaround" (1942) by American writer Isaac Asimov. -
Evidence Francis Quinn Was a Politician of the New School. That, Of
Evidence Francis Quinn was a politician of the new school. That, of course, is a meaningless expression, as are all expressions of the sort. Most of the “new schools” we have were duplicated in the social life of ancient Greece, and perhaps, if we knew more about it, in the social life of ancient Sumeria and in the lake dwellings of prehistoric Switzerland as well. But, to get out from under what promises to be a dull and complicated beginning, it might be best to state hastily that Quinn neither ran for office nor canvassed for votes, made no speeches and stuffed no ballot boxes. Any more than Napoleon pulled a trigger at Austerlitz. And since politics makes strange bedfellows, Alfred Lanning sat at the other side of the desk with his ferocious white eyebrows bent far forward over eyes in which chronic impatience had sharpened to acuity. He was not pleased. The fact, if known to Quinn, would have annoyed him not the least. His voice was friendly, perhaps professionally so. “I assume you know Stephen Byerley, Dr. Lanning.” “I have heard of him. So have many people.” “Yes, so have I. Perhaps you intend voting for him at the next election.” “I couldn’t say.” There was an unmistakable trace of acidity here. “I have not followed the political currents, so I’m not aware that he is running for office.” “He may be our next mayor. Of course, he is only a lawyer now, but great oaks--” “Yes,” interrupted Lanning, “I have heard the phrase before. But I wonder if we can get to the business at hand.” “We are at the business at hand, Dr. -
AI, Robots, and Swarms: Issues, Questions, and Recommended Studies
AI, Robots, and Swarms Issues, Questions, and Recommended Studies Andrew Ilachinski January 2017 Approved for Public Release; Distribution Unlimited. This document contains the best opinion of CNA at the time of issue. It does not necessarily represent the opinion of the sponsor. Distribution Approved for Public Release; Distribution Unlimited. Specific authority: N00014-11-D-0323. Copies of this document can be obtained through the Defense Technical Information Center at www.dtic.mil or contact CNA Document Control and Distribution Section at 703-824-2123. Photography Credits: http://www.darpa.mil/DDM_Gallery/Small_Gremlins_Web.jpg; http://4810-presscdn-0-38.pagely.netdna-cdn.com/wp-content/uploads/2015/01/ Robotics.jpg; http://i.kinja-img.com/gawker-edia/image/upload/18kxb5jw3e01ujpg.jpg Approved by: January 2017 Dr. David A. Broyles Special Activities and Innovation Operations Evaluation Group Copyright © 2017 CNA Abstract The military is on the cusp of a major technological revolution, in which warfare is conducted by unmanned and increasingly autonomous weapon systems. However, unlike the last “sea change,” during the Cold War, when advanced technologies were developed primarily by the Department of Defense (DoD), the key technology enablers today are being developed mostly in the commercial world. This study looks at the state-of-the-art of AI, machine-learning, and robot technologies, and their potential future military implications for autonomous (and semi-autonomous) weapon systems. While no one can predict how AI will evolve or predict its impact on the development of military autonomous systems, it is possible to anticipate many of the conceptual, technical, and operational challenges that DoD will face as it increasingly turns to AI-based technologies. -
An Ethical Framework for Smart Robots Mika Westerlund
An Ethical Framework for Smart Robots Mika Westerlund Never underestimate a droid. Leia Organa Star Wars: The Rise of Skywalker This article focuses on “roboethics” in the age of growing adoption of smart robots, which can now be seen as a new robotic “species”. As autonomous AI systems, they can collaborate with humans and are capable of learning from their operating environment, experiences, and human behaviour feedback in human-machine interaction. This enables smart robots to improve their performance and capabilities. This conceptual article reviews key perspectives to roboethics, as well as establishes a framework to illustrate its main ideas and features. Building on previous literature, roboethics has four major types of implications for smart robots: 1) smart robots as amoral and passive tools, 2) smart robots as recipients of ethical behaviour in society, 3) smart robots as moral and active agents, and 4) smart robots as ethical impact-makers in society. The study contributes to current literature by suggesting that there are two underlying ethical and moral dimensions behind these perspectives, namely the “ethical agency of smart robots” and “object of moral judgment”, as well as what this could look like as smart robots become more widespread in society. The article concludes by suggesting how scientists and smart robot designers can benefit from a framework, discussing the limitations of the present study, and proposing avenues for future research. Introduction capabilities (Lichocki et al., 2011; Petersen, 2007). Hence, Lin et al. (2011) define a “robot” as an Robots are becoming increasingly prevalent in our engineered machine that senses, thinks, and acts, thus daily, social, and professional lives, performing various being able to process information from sensors and work and household tasks, as well as operating other sources, such as an internal set of rules, either driverless vehicles and public transportation systems programmed or learned, that enables the machine to (Leenes et al., 2017). -
Arxiv:2009.09068V1 [Cs.CY] 16 Sep 2020
Hacking with God: a Common Programming Language of Robopsychology and Robophilosophy Norbert Bátfai∗ Department of Information Technology University of Debrecen, Hungary September 22, 2020 Abstract This note is a sketch of how the concept of robopsychology and robophi- losophy could be reinterpreted and repositioned in the spirit of the original vocation of psychology and philosophy. The notion of the robopsychology as a fictional science and a fictional occupation was introduced by Asimov in the middle of the last century. The robophilosophy, on the other hand, is only a few years old today. But at this moment, none of these new emerging disciplines focus on the fundamental and overall issues of the development of artificial general intelligence. Instead, they focus only on issues that, although are extremely important, play a complementary role, such as moral or ethical ones, rather than the big questions of life. We try to outline a conception in which the robophilosophy and robopsychology will be able to play a similar leading rule in the progress of artificial intel- ligence than the philosophy and psychology have done in the progress of human intelligence. To facilitate this, we outline the idea of a visual artifi- cial language and interactive theorem prover-based computer application called Prime Convo Assistant. The question to be decided in the future is whether we can develop such an application. And if so, can we build a computer game on it, or even an esport game? It may be an interesting question in order for this game will be able to transform human thinking on the widest possible social scale and will be able to develop a stan- dard mathematical logic-based communication channel between human and machine intelligence. -
Nudging for Good: Robots and the Ethical Appropriateness of Nurturing Empathy and Charitable Behavior
Nudging for Good: Robots and the Ethical Appropriateness of Nurturing Empathy and Charitable Behavior Jason Borenstein* and Ron Arkin** Predictions are being commonly voiced about how robots are going to become an increasingly prominent feature of our day-to-day lives. Beyond the military and industrial sectors, they are in the process of being created to function as butlers, nannies, housekeepers, and even as companions (Wallach and Allen 2009). The design of these robotic technologies and their use in these roles raises numerous ethical issues. Indeed, entire conferences and books are now devoted to the subject (Lin et al. 2014).1 One particular under-examined aspect of human-robot interaction that requires systematic analysis is whether to allow robots to influence a user’s behavior for that person’s own good. However, an even more controversial practice is on the horizon and warrants attention, which is the ethical acceptability of allowing a robot to “nudge” a user’s behavior for the good of society. For the purposes of this paper, we will examine the feasibility of creating robots that would seek to nurture a user’s empathy towards other human beings. We specifically draw attention to whether it would be ethically appropriate for roboticists to pursue this type of design pathway. In our prior work, we examined the ethical aspects of encoding Rawls’ Theory of Social Justice into robots in order to encourage users to act more socially just towards other humans (Borenstein and Arkin 2016). Here, we primarily limit the focus to the performance of charitable acts, which could shed light on a range of socially just actions that a robot could potentially elicit from a user and what the associated ethical concerns may be. -
History of Robotics: Timeline
History of Robotics: Timeline This history of robotics is intertwined with the histories of technology, science and the basic principle of progress. Technology used in computing, electricity, even pneumatics and hydraulics can all be considered a part of the history of robotics. The timeline presented is therefore far from complete. Robotics currently represents one of mankind’s greatest accomplishments and is the single greatest attempt of mankind to produce an artificial, sentient being. It is only in recent years that manufacturers are making robotics increasingly available and attainable to the general public. The focus of this timeline is to provide the reader with a general overview of robotics (with a focus more on mobile robots) and to give an appreciation for the inventors and innovators in this field who have helped robotics to become what it is today. RobotShop Distribution Inc., 2008 www.robotshop.ca www.robotshop.us Greek Times Some historians affirm that Talos, a giant creature written about in ancient greek literature, was a creature (either a man or a bull) made of bronze, given by Zeus to Europa. [6] According to one version of the myths he was created in Sardinia by Hephaestus on Zeus' command, who gave him to the Cretan king Minos. In another version Talos came to Crete with Zeus to watch over his love Europa, and Minos received him as a gift from her. There are suppositions that his name Talos in the old Cretan language meant the "Sun" and that Zeus was known in Crete by the similar name of Zeus Tallaios. -
Lio - a Personal Robot Assistant for Human-Robot Interaction and Care Applications
Lio - A Personal Robot Assistant for Human-Robot Interaction and Care Applications Justinas Miseikis,ˇ Pietro Caroni, Patricia Duchamp, Alina Gasser, Rastislav Marko, Nelija Miseikienˇ e,˙ Frederik Zwilling, Charles de Castelbajac, Lucas Eicher, Michael Fruh,¨ Hansruedi Fruh¨ Abstract— Lio is a mobile robot platform with a multi- careers [4]. A possible staff shortage of 500’000 healthcare functional arm explicitly designed for human-robot interaction employees is estimated in Europe by the year of 2030 [5]. and personal care assistant tasks. The robot has already Care robotics is not an entirely new field. There has been deployed in several health care facilities, where it is functioning autonomously, assisting staff and patients on an been significant development in this direction. One of the everyday basis. Lio is intrinsically safe by having full coverage most known robots is Pepper by SoftBank Robotics, which in soft artificial-leather material as well as having collision was created for interaction and entertainment tasks. It is detection, limited speed and forces. Furthermore, the robot has capable of voice interactions with humans, face and mood a compliant motion controller. A combination of visual, audio, recognition. In the healthcare sector Pepper is used for laser, ultrasound and mechanical sensors are used for safe navigation and environment understanding. The ROS-enabled interaction with dementia patients [6]. setup allows researchers to access raw sensor data as well as Another example is the robot RIBA by RIKEN. It is have direct control of the robot. The friendly appearance of designed to carry around patients. The robot is capable of Lio has resulted in the robot being well accepted by health localising a voice source and lifting patients weighing up to care staff and patients. -
An Architecture for Ethical Robots
1 An architecture for ethical robots Dieter Vanderelst & Alan Winfield Bristol Robotics Laboratory, University of the West of England T Block, Frenchay Campus, Coldharbour Lane, Bristol, BS16 1QY, United Kingdom Abstract—Robots are becoming ever more autonomous. This Anderson [2] and our previous work [27] are the only instances expanding ability to take unsupervised decisions renders it im- of robots having been equipped with a set of moral principles. perative that mechanisms are in place to guarantee the safety of So far, most work has been either theoretical [e.g., 25] or behaviours executed by the robot. Moreover, smart autonomous robots should be more than safe; they should also be explicitly simulation based [e.g., 3]. ethical – able to both choose and justify actions that prevent The approach taken by Anderson and Anderson [1, 2] and harm. Indeed, as the cognitive, perceptual and motor capabilities others [25, 3] is complementary with our research goals. These of robots expand, they will be expected to have an improved authors focus on developing methods to extract ethical rules capacity for making moral judgements. We present a control for robots. Conversely, our work concerns the development architecture that supplements existing robot controllers. This so-called Ethical Layer ensures robots behave according to a of a control architecture that supplements the existing robot predetermined set of ethical rules by predicting the outcomes of controller, ensuring robots behave according to a predeter- possible actions and evaluating the predicted outcomes against mined set of ethical rules [27, 26]. In other words, we are those rules. To validate the proposed architecture, we implement concerned with methods to enforce the rules once these have it on a humanoid robot so that it behaves according to Asimov’s been established [10]. -
Acknowledgements Acknowl
2161 Acknowledgements Acknowl. B.21 Actuators for Soft Robotics F.58 Robotics in Hazardous Applications by Alin Albu-Schäffer, Antonio Bicchi by James Trevelyan, William Hamel, The authors of this chapter have used liberally of Sung-Chul Kang work done by a group of collaborators involved James Trevelyan acknowledges Surya Singh for de- in the EU projects PHRIENDS, VIACTORS, and tailed suggestions on the original draft, and would also SAPHARI. We want to particularly thank Etienne Bur- like to thank the many unnamed mine clearance experts det, Federico Carpi, Manuel Catalano, Manolo Gara- who have provided guidance and comments over many bini, Giorgio Grioli, Sami Haddadin, Dominic Lacatos, years, as well as Prof. S. Hirose, Scanjack, Way In- Can zparpucu, Florian Petit, Joshua Schultz, Nikos dustry, Japan Atomic Energy Agency, and Total Marine Tsagarakis, Bram Vanderborght, and Sebastian Wolf for Systems for providing photographs. their substantial contributions to this chapter and the William R. Hamel would like to acknowledge work behind it. the US Department of Energy’s Robotics Crosscut- ting Program and all of his colleagues at the na- C.29 Inertial Sensing, GPS and Odometry tional laboratories and universities for many years by Gregory Dudek, Michael Jenkin of dealing with remote hazardous operations, and all We would like to thank Sarah Jenkin for her help with of his collaborators at the Field Robotics Center at the figures. Carnegie Mellon University, particularly James Os- born, who were pivotal in developing ideas for future D.36 Motion for Manipulation Tasks telerobots. by James Kuffner, Jing Xiao Sungchul Kang acknowledges Changhyun Cho, We acknowledge the contribution that the authors of the Woosub Lee, Dongsuk Ryu at KIST (Korean Institute first edition made to this chapter revision, particularly for Science and Technology), Korea for their provid- Sect. -
Report of Comest on Robotics Ethics
SHS/YES/COMEST-10/17/2 REV. Paris, 14 September 2017 Original: English REPORT OF COMEST ON ROBOTICS ETHICS Within the framework of its work programme for 2016-2017, COMEST decided to address the topic of robotics ethics building on its previous reflection on ethical issues related to modern robotics, as well as the ethics of nanotechnologies and converging technologies. At the 9th (Ordinary) Session of COMEST in September 2015, the Commission established a Working Group to develop an initial reflection on this topic. The COMEST Working Group met in Paris in May 2016 to define the structure and content of a preliminary draft report, which was discussed during the 9th Extraordinary Session of COMEST in September 2016. At that session, the content of the preliminary draft report was further refined and expanded, and the Working Group continued its work through email exchanges. The COMEST Working Group then met in Quebec in March 2017 to further develop its text. A revised text in the form of a draft report was submitted to COMEST and the IBC in June 2017 for comments. The draft report was then revised based on the comments received. The final draft of the report was further discussed and revised during the 10th (Ordinary) Session of COMEST, and was adopted by the Commission on 14 September 2017. This document does not pretend to be exhaustive and does not necessarily represent the views of the Member States of UNESCO. – 2 – REPORT OF COMEST ON ROBOTICS ETHICS EXECUTIVE SUMMARY I. INTRODUCTION II. WHAT IS A ROBOT? II.1. The complexity of defining a robot II.2.