Humans, Animals, and Robots: a Phenomenological Approach to Human-Robot Relations

Total Page:16

File Type:pdf, Size:1020Kb

Humans, Animals, and Robots: a Phenomenological Approach to Human-Robot Relations Int J Soc Robot (2011) 3: 197–204 DOI 10.1007/s12369-010-0075-6 Humans, Animals, and Robots: A Phenomenological Approach to Human-Robot Relations Mark Coeckelbergh Accepted: 22 August 2010 / Published online: 4 September 2010 © The Author(s) 2010. This article is published with open access at Springerlink.com Abstract This paper argues that our understanding of many bots may be presented as partners, as in the film The Step- human-robot relations can be enhanced by comparisons with ford Wives (2004), or in other substitute roles, as in Bicen- human-animal relations and by a phenomenological ap- tennial Man (1999) and I, Robot (2004). These images do proach which highlights the significance of how robots ap- not entirely belong to the domain of fiction: some humans pear to humans. Some potential gains of this approach are might engage in relationships with human-like robots [1] explored by discussing the concept of alterity, diversity and and as humanoid robotics develops further other practices change in human-robot relations, Heidegger’s claim that an- will emerge. However, it is more likely that in the near fu- imals are ‘poor in world’, and the issue of robot-animal rela- ture relations between humans and robots will mainly take tions. These philosophical reflections result in a perspective the form of strong attachments to robots that do not appear on human-robot relations that may guide robot design and human (although they might have some humanoid aspects). inspire more empirical human-robot relations research that They may be given an animal-like appearance (zoomor- is sensitive to how robots appear to humans in different con- phic) or a different appearance. Many current robots are texts at different times. zoomorphic: consider for example the iCat (Philips), Hug- gable (MIT Media Lab), and the ‘huggable animal-like ro- Keywords Human-robot relations · Human-animal relations · Phenomenology · Appearance · Robot-animal bot’ Probo (Free University Brussels) [2]. relations A turn away from anthropomorphism is more likely for several reasons. First, because human appearance is not re- quired it is easier to design the robots and make them avail- 1 Introduction able. Second, they avoid the problem of the so-called ‘un- canny valley’. The Uncanny Valley hypothesis suggests that The introduction of robots in the personal sphere—now and if a robot looks too much like a human (but when it is still in the near-future—gives rise to new personal relations be- clear that it is not a human), it appears ‘uncanny’ [3]. Feel- tween humans and robots. How can we better understand ings of uneasy do not occur when the robot looks very dif- these relations from a philosophical perspective? ferent or when we encounter a human; only the appearance The idea of personal relations between humans and ro- of the nearly-human has this effect. Robots that look differ- bots often invokes science-fiction images of robots that ap- ently do not incur this problem. Third, since humans already pear human and that are used as substitutes for humans. Ro- have relations with non-robotic non-humans such as animals and – to some extent – computers and other electronic equip- ment such as mobile phones, we can expect that the intro- A first version of this paper has been presented at the 2nd International Conference on Human-Robot Personal Relations, duction of non-humanoid robots will be less controversial Tilburg, June 11–12, 2009. than, for instance, humanoid sex robots. While the account developed here is relevant to relations with humanoid ro- M. Coeckelbergh () University of Twente, Enschede, The Netherlands bots, I am also concerned with relations with other robots, e-mail: [email protected] in particular intense, long-term relations with non-humanoid 198 Int J Soc Robot (2011) 3: 197–204 robots such as those that may develop between children and the technology is between us and the world. Consider the their toy robot or between adults and pet robots. role played by scientific instruments such as telescopes and In this paper, I propose to understand such relations microscopes: our perception is always mediated by these by taking a phenomenological approach and by comparing technologies. For example, research at the nano level is im- human-robot and human-animal relations. I support my ar- possible without such mediations. If we had only our human gument by drawing on insights from philosophy of technol- eyes, we could neither see any nano particles nor manipulate ogy and by briefly exploring the phenomenology of animals. them. The technology that shows things at that level is indis- I show how this approach can enhance our understanding of pensable, not as a mere ‘instrument’ but as a way of seeing human-robot relations. (and manipulating) the world. There are at least two senses When I use the term ‘phenomenology’ in this paper I re- in which robots could fulfil such a role. First, they can me- fer to (1) a general orientation that gives attention to how ro- diate between us and the world in a rather literal sense, for bots appear to human consciousness in order to better under- instance we can imagine a disabled or even locked-in per- stand human-robot relations (hermeneutics rather than sci- son using a robotic device to give her information and to do ence), and (2) two specific currents within the philosophical things in the world. Or we can think of remote controlled tradition of phenomenology: the work of Martin Heidegger robots on other planets that enable us to see and manipulate and the ‘postphenomenology’ of the American philosopher the world through the eyes of the robot. of technology Don Ihde. However, the third concept, alterity relations, is the most First, I use Ihde’s work to characterize personal human- interesting for the purpose of understanding relations to per- robot relations as alterity relations and show how this con- sonal robots: sometimes we relate to technology as an other. cept could contribute to existing studies of human-robot in- What does this mean? Using Ihde’s framework, we can un- teraction. Then I discuss the importance of appearance for derstand personal human-robot relations as alterity relations understanding that interaction and call attention to the va- if and to the extent that in these relations appear to humans riety of human-robot relations in different contexts. I make as an other or ‘quasi-other’ [6]. The robot is neither part of a comparison between human-robot and human-animal re- me (embodiment relation) nor something that mainly medi- lations, considering in particular robots that represent ani- ates my relation to the world. Instead, in our interaction with mals. I respond to Heidegger’s work to discuss the relevance the robot ‘it’ appears to us as more than a thing: an other to of phenomenological similarities and differences between which I relate [6, p. 107]. humans, animals, and robots for human-robot relations. Fi- This approach to the status of the robot stands in contrast nally, I raise the issue of robot-animal relations. to most traditional philosophical accounts (including tradi- tional phenomenology), which would require that the robot has intentionality, consciousness, and so on. Instead I argue 2 Human-Robot Relations as Alterity Relations here and elsewhere that what matters for understanding and evaluating human-robot relations is how the robot appears to In philosophy of technology, Ihde developed a (post)phe- us. Therefore, even if—based on traditional accounts – we nomenological framework to discuss human-technology re- are not prepared to ascribe to robots the character of alterity, lations [4]. Applying the concept of intentionality to human- we can speak of quasi-alterity in those cases where the robot technology relations, he distinguishes three structural fea- appears to us as an other. Whatever the ‘real’ status of the ro- appearance tures of technological intentionality: embodiment relations, bot may be, it is its that is relevant to how the human-robot relation is experienced and constructed. (I will hermeneutic relations, and alterity relations. Let me explain continue this argument in the next section.) these relations and apply the concepts to robots. Moreover, to view relations with robots in this way does Embodiment relations refer to the amplification of bod- not imply that there is a reciprocal relationship of ‘selves’ or ily perception: technology comes to be experienced as being that we need to seek recognition from the ‘android partner’, part of us; we do no longer notice it. For instance, usually I as Ramey suggests in his arguments for a phenomenologi- do not notice that I wear glasses. Most robots will not in- cal approach to the ethics of human-android interaction [7]. vite an embodiment relation. Exceptions are perhaps partic- Moreover, while I sympathize with Ramey’s rejection of sci- ular non-autonomous robots such as robot suits (for instance entism and with his recommendation that we attend to possi- HAL), robot arms, or robot ‘exo-skeletons’ (see for example ble changes in the (human) self as a result of interaction with the performances of the Australian performance artist Stel- android robots, I propose that we understand the robot other larc): such robots might come to be experienced as part of as an other that is experienced as external. However much the body. truth there may be in Ramey’s Heideggerian claim1 that we The hermeneutic relation refers to the role technology plays as an in-between that allows us to interpret the world [4, 5]. We do not have direct access to the world; instead, 1Ramey is influenced by Olafson’s interpretation of Heidegger [8].
Recommended publications
  • I Panel –The Relationship Between Law and Ethics
    I Panel –The relationship between law and ethics Tatjana Evas is a Legal and Policy Officer at the European Commission and Associate Professor of EU and Comparative law at Tallinn University of Technology. She has extensive professional experience including at the Columbia Law School, New York; Riga Graduate School of Law; Center for Policy Studies, Budapest; Jean Monnet Centre for European Studies, Bremen and European University Institute, Florence. She is author of several publications on prestigious national and international journals. Her current research work focuses on the regulation of new technologies, use of AI technologies in courts, and methodology for impact assessment. Before that, she was Policy Analyst in the European Parliamentary Research Service. In 2011, she received Bremen Studienpreis Award for best PhD Thesis is social science and humanities. Before working at the Commission, Tatjana was a Policy Analyst in the European Parliamentary Research Service, where – among other things – she scientifically coordinated European Parliament’s public consultation on Robotics and Artificial Intelligence (2017) and published European Added Value Assessment on Liability of Autonomous Vehicles (2018). European framework on ethical aspects of artificial intelligence, robotics and related technologies – The EU can become a global standard-setter in the area of artificial intelligence (AI) ethics. Common EU legislative action on ethical aspects of AI could boost the internal market and establish an important strategic advantage. While numerous public and private actors around the globe have produced ethical guidelines in this field, there is currently no comprehensive legal framework. The EU can profit from the absence of a competing global governance model and gain full 'first mover' advantages.
    [Show full text]
  • Information Technology and Law Series
    Information Technology and Law Series Volume 27 Editor-in-chief Simone van der Hof, eLaw (Center for Law and Digital Technologies), Leiden University, Leiden, The Netherlands Series editor Bibi van den Berg, eLaw (Center for Law and Digital Technologies), Leiden University, Leiden, The Netherlands Eleni Kosta, ICRI, KU Leuven, Leuven, Belgium Ben Van Rompuy, T.M.C. Asser Institute, The Hague, The Netherlands Ulrich Sieber, for Foreign and International Crimi, Max Planck Institute, Freiburg, Germany More information about this series at http://www.springer.com/series/8857 Bart Custers Editor The Future of Drone Use Opportunities and Threats from Ethical and Legal Perspectives 1 3 Editor Bart Custers Faculty of Law, eLaw Leiden University Leiden The Netherlands ISSN 1570-2782 ISSN 2215-1966 (electronic) Information Technology and Law Series ISBN 978-94-6265-131-9 ISBN 978-94-6265-132-6 (eBook) DOI 10.1007/978-94-6265-132-6 Library of Congress Control Number: 2016945865 Published by T.M.C. ASSER PRESS, The Hague, The Netherlands www.asserpress.nl Produced and distributed for T.M.C. ASSER PRESS by Springer-Verlag Berlin Heidelberg © T.M.C. ASSER PRESS and the authors 2016 No part of this work may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, microfilming, recording or otherwise, without written permission from the Publisher, with the exception of any material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. The use of general descriptive names, registered names, trademarks, service marks, etc.
    [Show full text]
  • Ethically Aligned Design, First Edition
    ETHICALLY ALIGNED DESIGN First Edition A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems Table of Contents The views and opinions expressed in this collaborative work are those of the authors and do not necessarily reflect the official policy or position of their respective institutions or of the Institute of Electrical and Electronics Engineers (IEEE). This work is published under the auspices of the IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems for the purposes of furthering public understanding of the importance of addressing ethical considerations in the design of autonomous and intelligent systems. Please see page 290, How the Document Was Prepared, for more details regarding the preparation of this document. This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 United States License. The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems Table of Contents Introduction 2 Executive Summary 3-6 Acknowledgements 7-8 Ethically Aligned Design From Principles to Practice 9-16 General Principles 17-35 Classical Ethics in A/IS 36-67 Well-being 68-89 Affective Computing 90-109 Personal Data and Individual Agency 110-123 Methods to Guide Ethical Research and Design 124-139 A/IS for Sustainable Development 140-168 Embedding Values into Autonomous and Intelligent Systems 169-197 Polic y 198-210 L a w 211-281 About Ethically Aligned Design The Mission and Results of The IEEE Global Initiative 282 From Principles to Practice—Results of Our Work to Date 283-284 IEEE P7000™ Approved Standardization Projects 285-286 Who We Are 287 Our Process 288-289 How the Document was Prepared 290 How to Cite Ethically Aligned Design 290 Key References 291 This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 United States License.
    [Show full text]
  • Mark Coeckelbergh
    The Language of Technology Invitation to the Inaugural Lecture Faculty of Philosophy and Education Wednesday, 10 May 2017 Mark Coeckelbergh Professor of Philosophy of Media and Technology Mark Coeckelbergh is Professor of Philosophy of Media and Technology at the Faculty of Philosophy and Education at the University of Vienna (since December 2015). Born in 1975 in Leuven, Belgium, Mark Coeckelbergh studied Social Sciences and Political Sciences at the University of Leuven, Belgium and Social Philosophy at the University of East Anglia, UK. 2003 Ph.D. at the University of Birmingham, UK. 2003 – 2006 Lecturer in Philosophy at Maastricht University, Netherlands. 2007 – 2014 Assistant Professor at the University of Twente, Netherlands. 2012 – 2014 Managing Director at the 3TU. (now 4TU.) Centre for Ethics and Technology, Netherlands. Since 2014 Full Professor of Technology and Social Responsibility at the Centre for Computing and Social Responsibility at De Montfort University, UK (part-time since December 2015). 2014 – 2016 Leader of work packages in European projects (FP7 projects DREAM and SATORI). Since 2016 Vice-President/President-Elect of the Society for Philosophy and Technology. Current research areas: Philosophy of technology, including philosophy of robotics, automation and artificial intelligence; Ethics, such as virtue ethics, ethics of technology, ethics of robotics, ethics of information technology, computer ethics, health care ethics, environmental ethics, animal ethics; Moral philosophy; Philosophy of language; Epistemology;
    [Show full text]
  • Enactivism As a Postphenomenological Metaphysics
    1 Human-Technology Relations: Postphenomenology and Philosophy of Technology 11-13 July 2018 DesignLab University of Twente, The Netherlands 2 DEAR PARTICIPANTS, Welcome to the Philosophy of Human-Technology Relations conference 2018! We are very happy to have such a great group of people presenting papers, showing work in art, design and engineering, and discussing each other’s work. The number of people sending in abstracts and registering to participate is much larger than we had dared to expect, which made it a true pleasure to organize the conference. While focusing on the philosophy of human- technology relations, the conference reaches out to art, design, engineering, and Science and Technology Studies. We have paper presentations, demonstrations, hands-on workshops, book panels, and a book exhibit. Participants come from all over the world, so we hope the conference will bring about many new connections. Our home base will be the DesignLab of the University of Twente, which brings technology together with the social sciences and humanities, focusing on responsible design. For the conference dinner, on Thursday evening, we will move to the city of Enschede, where we will have dinner in The Museum Factory: and old textile factory (Twente used to be the main Dutch textile industry region) which was turned into a museum after the Enschede Fireworks disaster in 2000, and which currently has an exposition on Frankenstein and Human-Technology Relations. If there are any questions, please don’t hesitate to contact the organization: there will always be people in the PhilosophyLab, students in the organization can be recognized by their t-shirt, and the members of the organizing committee will be around during the entire conference.
    [Show full text]
  • The Tragedy of the Master: Automation, Vulnerability, and Distance
    Author's personal copy Ethics Inf Technol (2015) 17:219–229 DOI 10.1007/s10676-015-9377-6 ORIGINAL PAPER The tragedy of the master: automation, vulnerability, and distance Mark Coeckelbergh1 Published online: 3 September 2015 Ó Springer Science+Business Media Dordrecht 2015 Abstract Responding to long-standing warnings that Introduction robots and AI will enslave humans, I argue that the main problem we face is not that automation might turn us into Echoing a long tradition going back at least to Aristotle’s slaves but, rather, that we remain masters. First I construct comments in the Nicomachean Ethics about the slave as a an argument concerning what I call ‘the tragedy of the ‘living tool’ and the tool as a ‘lifeless slave’ (Aristotle, master’: using the master–slave dialectic, I argue that 1161b2-7), thinking about technology in terms of masters automation technologies threaten to make us vulnerable, and slaves is more popular than ever. Current discussions alienated, and automated masters. I elaborate the implica- about automation, robotics, and AI in the robotics/AI tions for power, knowledge, and experience. Then I criti- community and beyond seem to be am especially fertile cally discuss and question this argument but also the very ground for such a thinking. For instance, there have been thinking in terms of masters and slaves, which fuels both many recent warnings about AI or robots taking over, arguments. I question the discourse about slavery and indeed becoming our masters rather than the other way object to the assumptions made about human–technology around. Consider for instance claims about AI in the media relations.
    [Show full text]
  • Robots: Ethical Or Unethical?
    Dutch side-event: 10 Nov 2017 – UNESCO Paris – Room X – 1-2.30 pm Robots: ethical or unethical? Intro Ambassador Lionel Veer & ADG SHS Nada Al-Nashif What change caused by robots can we expect? Prof. Vanessa Evers How do robots concretely change our lives? With her robots Prof. Evers will show some examples. Which are the ethical dilemmas? Prof. Mark Coeckelbergh The ‘Frankenstein risk’: who will be in charge, man or robot? Furthermore: is robotic behavior ethical and how does it affect humanity, for example our social life. How can we ensure that innovation is ethical? Prof. Peter-Paul Verbeek How do humans and robots interact, and how can development and innovation in robotics be shaped in an ethical way? Prof. Verbeek will present UNESCO’s COMEST report which advises Governments on these challenges. Ask a question to the speakers by using the hashtag #UnescoRobots! Nada Al-Nashif is UNESCO’s Assistant Director-General for Social and Human Sciences since 16 February 2015. Prior to joining UNESCO, she served as Assistant Director-General/Regional Director of the International Labour Organization’s Regional Office for Arab States, based in Beirut, Lebanon. Previously she worked at UNDP, where she started her UN career in 1991, serving in Libya, Lebanon, Iraq and at Headquarters in New York. As a development economist and practitioner, she serves in an advisory capacity on several boards, notably the Boards of Trustees of Birzeit University and the Human Development NGO, Taawon. Vanessa Evers is a full Professor of Human Media Interaction at the University of Twente. Her research focuses on Human Interaction with Autonomous Agents such as robots or machine learning systems and cultural aspects of Human Computer Interaction.
    [Show full text]
  • Can We Trust Robots?
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Springer - Publisher Connector Ethics Inf Technol (2012) 14:53–60 DOI 10.1007/s10676-011-9279-1 ORIGINAL PAPER Can we trust robots? Mark Coeckelbergh Published online: 3 September 2011 Ó The Author(s) 2011. This article is published with open access at Springerlink.com Abstract Can we trust robots? Responding to the litera- they rise up against us? But there is a broader and certainly ture on trust and e-trust, this paper asks if the question of more urgent issue about trust in intelligent autonomous trust is applicable to robots, discusses different approaches technologies that are already available today or will be to trust, and analyses some preconditions for trust. In the available soon. Robotics for entertainment, sex, health course of the paper a phenomenological-social approach to care, and military applications are fast developing fields of trust is articulated, which provides a way of thinking about research, autonomous cars are being developed, remote trust that puts less emphasis on individual choice and controlled robots are used in medicine and in the military, control than the contractarian-individualist approach. In and we already heavily rely on semi-robots such as auto- addition, the argument is made that while robots are neither pilot-airplanes. And of course we (often heavily) rely on human nor mere tools, we have sufficient functional, computers and other information technology in our daily agency-based, appearance-based, social-relational, and lives. The more we rely on these technologies, the more existential criteria left to evaluate trust in robots.
    [Show full text]
  • New Romantic Cyborgs
    New Romantic Cyborgs New Romantic Cyborgs Romanticism, Information Technology, and the End of the Machine Mark Coeckelbergh The MIT Press Cambridge, Massachusetts London, England © 2017 Massachusetts Institute of Technology All rights reserved. No part of this book may be reproduced in any form by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher. This book was set in Stone Sans and Stone Serif by Toppan Best-set Premedia Limited. Printed and bound in the United States of America. Library of Congress Cataloging-in-Publication Data Names: Coeckelbergh, Mark, author. Title: New romantic cyborgs : romanticism, information technology, and the end of the machine / Mark Coeckelbergh. Description: Cambridge, MA : MIT Press, 2017. | Includes bibliographical references and index. Identifiers: LCCN 2016021510 | ISBN 9780262035460 (hardcover : alk. paper) Subjects: LCSH: Technology--Philosophy. | Human-machine systems--Philosophy. | Information technology--Philosophy. | Cyborgs--Philosophy. | Romanticism. Classification: LCC T14 .C5726 2017 | DDC 601--dc23 LC record available at https://lccn.loc.gov/2016021510 10 9 8 7 6 5 4 3 2 1 To Tim O’Hagan and Nicholas Dent, from whom I learned a lot about Rousseau Contents Acknowledgments ix 1 Introduction: The Question Concerning Technology and Romanticism 1 I Romanticism against the Machine 19 2 Romanticism 21 3 Romanticism against the Machine? 71 II Romanticism with the Machine 95 4 Romanticism with
    [Show full text]
  • The Moral Consideration of Artificial Entities: a Literature Review
    The Moral Consideration of Artificial Entities: A Literature Review Jamie Harris: Sentience Institute, [email protected] Jacy Reese Anthis: Sentience Institute & Department of Sociology, University of Chicago Abstract Ethicists, policy-makers, and the general public have questioned whether artificial entities such as robots warrant rights or other forms of moral consideration. There is little synthesis of the research on this topic so far. We identify 294 relevant research or discussion items in our literature review of this topic. There is widespread agreement among scholars that some artificial entities could warrant moral consideration in the future, if not also the present. The reasoning varies, such as concern for the effects on artificial entities and concern for the effects on human society. Beyond the conventional consequentialist, deontological, and virtue ethicist ethical frameworks, some scholars encourage “information ethics” and “social-relational” approaches, though there are opportunities for more in-depth ethical research on the nuances of moral consideration of artificial entities. There is limited relevant empirical data collection, primarily in a few psychological studies on current moral and social attitudes of humans towards robots and other artificial entities. This suggests an important gap for social science research on how artificial entities will be integrated into society and the factors that will determine how the interests of sentient artificial entities are considered. Keywords: Artificial Intelligence, Robots, Rights, Ethics, Philosophy of Technology, Suffering Risk Introduction Humans could continue to exist for many centuries on earth or expand to far greater interstellar populations in the coming millennia (e.g. very approximately 1038 human lives per century if we colonize the Virgo Supercluster (Bostrom 2003)).
    [Show full text]
  • You, Robot: on the Linguistic Construction of Artificial Others
    AI & Soc (2011) 26:61–69 DOI 10.1007/s00146-010-0289-z ORIGINAL ARTICLE You, robot: on the linguistic construction of artificial others Mark Coeckelbergh Received: 20 March 2010 / Accepted: 26 July 2010 / Published online: 10 August 2010 Ó The Author(s) 2010. This article is published with open access at Springerlink.com Abstract How can we make sense of the idea of ‘per- 1 Introduction sonal’ or ‘social’ relations with robots? Starting from a social and phenomenological approach to human–robot I love you. Do you love me? relations, this paper explores how we can better under- (sentence addressed to a robotic doll, reported in Turkle et al. stand and evaluate these relations by attending to the ways 2006, p. 357). our conscious experience of the robot and the human– robot relation is mediated by language. It is argued that The robots are coming. But if they enter your home, our talk about and to robots is not a mere representation they may not kill you; there is a good chance that they of an objective robotic or social-interactive reality, but simply want a hug. Robots are no longer confined to fac- rather interprets and co-shapes our relation to these arti- tories, laboratories, and—increasingly—battlefields. They ficial quasi-others. Our use of language also changes as a gradually enter people’s daily lives, offering companion- result of our experiences and practices. This happens ship, entertainment, sex, or health care. Some people prefer when people start talking to robots. In addition, this paper artificial friends or even artificial partners.
    [Show full text]
  • Why Care About Robots? Empathy, Moral Standing, and the Language of Suffering
    Kairos. Journal of Philosophy & Science 20, 2018 Center for the Philosophy of Sciences of Lisbon University Why Care About Robots? Empathy, Moral Standing, and the Language of Suffering Mark Coeckelbergh University of Vienna and De Montfort University (UK) [email protected] Abstract. This paper tries to understand the phenomenon that humans are able to empathize with robots and the intuition that there might be something wrong with “abusing” robots by discussing the question regarding the moral standing of robots. After a review of some relevant work in empirical psychology and a discussion of the ethics of empathizing with robots, a philosophical argument concerning the moral standing of robots is made that questions distant and uncritical moral reasoning about entities’ properties and that recommends first trying to understand the issue by means of philosophical and artistic work that shows how ethics is always relatio- nal and historical, and that highlights the importance of language and appearance in moral reasoning and moral psychology. It is concluded that attention to relationality and to verbal and non-verbal languages of suffering is key to understand the pheno- menon under investigation, and that in robot ethics we need less certainty and more caution and patience when it comes to thinking about moral standing. Keywords: moral standing, robots; empathy; relations, language, art, phenomenolo- gy, hermeneutics, Wittgenstein. DOI 10.2478/kjps-2018-0007 Open Access. © 2018 M. Coeckelbergh, published by Sciendo. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. Why Care About Robots?Empathy, Moral Standing, and the Language of Suffering 1.
    [Show full text]