Can We Trust Robots?

Total Page:16

File Type:pdf, Size:1020Kb

Can We Trust Robots? View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Springer - Publisher Connector Ethics Inf Technol (2012) 14:53–60 DOI 10.1007/s10676-011-9279-1 ORIGINAL PAPER Can we trust robots? Mark Coeckelbergh Published online: 3 September 2011 Ó The Author(s) 2011. This article is published with open access at Springerlink.com Abstract Can we trust robots? Responding to the litera- they rise up against us? But there is a broader and certainly ture on trust and e-trust, this paper asks if the question of more urgent issue about trust in intelligent autonomous trust is applicable to robots, discusses different approaches technologies that are already available today or will be to trust, and analyses some preconditions for trust. In the available soon. Robotics for entertainment, sex, health course of the paper a phenomenological-social approach to care, and military applications are fast developing fields of trust is articulated, which provides a way of thinking about research, autonomous cars are being developed, remote trust that puts less emphasis on individual choice and controlled robots are used in medicine and in the military, control than the contractarian-individualist approach. In and we already heavily rely on semi-robots such as auto- addition, the argument is made that while robots are neither pilot-airplanes. And of course we (often heavily) rely on human nor mere tools, we have sufficient functional, computers and other information technology in our daily agency-based, appearance-based, social-relational, and lives. The more we rely on these technologies, the more existential criteria left to evaluate trust in robots. It is also urgent becomes the issue of trust. As Taddeo writes in her argued that such evaluations must be sensitive to cultural introduction to a special issue devoted to trust in differences, which impact on how we interpret the criteria technology: and how we think of trust in robots. Finally, it is suggested As the outsourcing to (informational) artefacts that when it comes to shaping conditions under which becomes more pervasive, the trust and the depen- humans can trust robots, fine-tuning human expectations dence of the users on such artefacts also grow, and robotic appearances is advisable. bringing to the fore [issues] like the nature of trust, the necessary requirements for its occurrence, whe- Keywords Trust Á Ethics Á Robots Á Social relations Á ther trust can be developed toward an artefact or can Phenomenology Á Culture only concern human beings (…). (Taddeo 2010b, p. 283) Introduction In other words, we already delegate tasks to machines and apparently we already trust them (I will return to this point To frame the ethical question concerning robotics in terms below). This seems also the case with robots. But do we of trust may easily suggest science-fiction scenarios like have good reasons to trust robots?1 Do we need to develop the story in the film ‘I, robot’ in which robots become what Wallach and Allen call ‘moral machines’ (Wallach artificially intelligent to such an extent that humans wonder and Allen 2008) in order to make robots trustworthy and if they can be trusted—which is usually interpreted as: Will avoid misuse of our trust? Does that mean we have to develop rational artificial agents? Is it appropriate at all to talk about trust here? M. Coeckelbergh (&) Department of Philosophy, University of Twente, Postbus 217, 7500 AE Enschede, Netherlands 1 We should also ask: Do they have reasons to trust us? I will discuss e-mail: [email protected] this question elsewhere. 123 54 M. Coeckelbergh In this paper, I postpone a direct analysis of the conditions But does this mean that trusting artefacts is only about under which we can trust robots. Rather, I discuss what reliance? Or does it mean that, as Pitt argues (Pitt 2010), trust means in relation to robots. First I provide a more the question about trust in technology comes down to general, brief analysis of trust: I discuss different approa- trusting people (to do this or that)?2 ches to analysing the conditions of trust (worthiness) in relation to artefacts and to people in general. This gives me some preconditions for trust. In the process I also comment Trusting people on (other) discussions of trust in the literature on trust and e-trust. Then I explore if we can apply these preconditions In relation to people, the question of trust is different and to robots and how different approaches shed light on this more complicated than reliance. Trust is generally regarded question. I argue that in so far as robots appear human, as something that develops within (or establishes) a rela- trusting them requires that they fulfil demanding criteria tion between humans (usually called a trustor and a trus- concerning the appearance of language use, freedom, and tee) that has ethical aspects (or creates an ethical social relations. This kind of virtual trust develops in sci- dimension). ence-fiction scenarios and can even become a question of ‘mutual’ trust. However, I also show that even if robots do Trust not appear human and/or do not fulfil these criteria, as is the case today, we have sufficient functional, agency- Before I discuss the relation between trust and ethics, let based, social-relational, and existential criteria left to talk me first say something about trust and how it is usually about, and evaluate, trust in robots. Finally, I argue that all approached, which influences the interpretation of the criteria depend to some degree on the culture in which one definition just given. lives and that therefore to evaluate trust in robots we have I propose that we distinguish between a contractarian- to attend to, and understand, differences in cultural attitude individualist and a phenomenological-social approach to towards technology (and robots in particular) and, more trust and its relation to ethics. In the former approach, there generally, cultural differences in ways of seeing and ways are ‘first’ individuals who ‘then’ create a social relation (in of doing. particular social expectations) and hence trust between By proceeding in this way, I hope not only to contribute them. In the latter approach, the social or the community is to the normative-ethical discussion about trust in robots, prior to the individual, which means that when we talk but also to discussions about how to approach information about trust in the context of a given relation between ethics and ethics of robotics. humans, it is presupposed rather than created. Here trust cannot be captured in a formula and is something given, not entirely within anyone’s control. Trusting artefacts I take most standard accounts of trust to belong to the contractarian-individualist category. For example, in phi- Although the notion of trust is usually applied to human– losophy of information technology Taddeo’s overview of human relations, we sometimes say of an artefact that we the debate on trust and e-trust, and indeed her own work on (do not) trust it. What we mean then is that we expect the e-trust, is formulated within a contractarian-individualist artefact to function, that is, to do what it is meant to do as framework (Taddeo 2009, 2010a, b). an instrument to attain goals set by humans. Although we In her overview (Taddeo 2009) she discusses Luhmann do not have full epistemic certainty that the instrument will (1979) and Gambetta (1988). Luhmann’s analysis starts actually function, we expect it do so. For example, one may from the problem of co-ordination. Trust is seen as a trust a cleaning robot to do what it’s supposed to do— response to that problem: it is needed to reduce complexity cleaning. This kind of trust is what sometimes is called and uncertainty. Thus, the analysis starts from individuals ‘trust as reliance’. and then it is questioned how the social can come into Related to this type of direct trust in artefacts is indirect existence (if at all). Similarly, when Gambetta defines trust trust in the humans related to the technology: we trust that the as ‘a particular level of the subjective probability with designer has done a good job to avoid bad outcomes and—if which an agent assesses that another agent or group of we are not ourselves the users—we trust that the users will agents will perform a particular action’ (Gambetta 1988, make good use of it, that is, that they will use the artefact for p. 216), we find ourselves in a kind of game-theoretical morally justifiable purposes. For example, military use of robots and artificially intelligent systems may be controversial 2 In a later section of this paper, I will question these instrumentalist because the (human) aims may be controversial (military and one-sided views of trust in technology (and trust in robots in action in general or particular uses, actions and targets). particular). 123 Can we trust robots? 55 setting typical of contractarian views of the social. Whether social human condition. But is this still about ethics? Is this or not to trust someone else is a matter of decision and still about trust? choice. The agent calculates (at least in Gambetta’s view) This leads me to the question: How do the different and cooperative relations are the result of these choices and approaches conceptualize the relation between trust and calculations. ethics? Taddeo’s account of e-trust differs from these views but does not question the contractarian-individualist paradigm Trust and ethics itself. E-trust may arise between agents under certain conditions (Taddeo 2009; see also Turili et al.
Recommended publications
  • I Panel –The Relationship Between Law and Ethics
    I Panel –The relationship between law and ethics Tatjana Evas is a Legal and Policy Officer at the European Commission and Associate Professor of EU and Comparative law at Tallinn University of Technology. She has extensive professional experience including at the Columbia Law School, New York; Riga Graduate School of Law; Center for Policy Studies, Budapest; Jean Monnet Centre for European Studies, Bremen and European University Institute, Florence. She is author of several publications on prestigious national and international journals. Her current research work focuses on the regulation of new technologies, use of AI technologies in courts, and methodology for impact assessment. Before that, she was Policy Analyst in the European Parliamentary Research Service. In 2011, she received Bremen Studienpreis Award for best PhD Thesis is social science and humanities. Before working at the Commission, Tatjana was a Policy Analyst in the European Parliamentary Research Service, where – among other things – she scientifically coordinated European Parliament’s public consultation on Robotics and Artificial Intelligence (2017) and published European Added Value Assessment on Liability of Autonomous Vehicles (2018). European framework on ethical aspects of artificial intelligence, robotics and related technologies – The EU can become a global standard-setter in the area of artificial intelligence (AI) ethics. Common EU legislative action on ethical aspects of AI could boost the internal market and establish an important strategic advantage. While numerous public and private actors around the globe have produced ethical guidelines in this field, there is currently no comprehensive legal framework. The EU can profit from the absence of a competing global governance model and gain full 'first mover' advantages.
    [Show full text]
  • Information Technology and Law Series
    Information Technology and Law Series Volume 27 Editor-in-chief Simone van der Hof, eLaw (Center for Law and Digital Technologies), Leiden University, Leiden, The Netherlands Series editor Bibi van den Berg, eLaw (Center for Law and Digital Technologies), Leiden University, Leiden, The Netherlands Eleni Kosta, ICRI, KU Leuven, Leuven, Belgium Ben Van Rompuy, T.M.C. Asser Institute, The Hague, The Netherlands Ulrich Sieber, for Foreign and International Crimi, Max Planck Institute, Freiburg, Germany More information about this series at http://www.springer.com/series/8857 Bart Custers Editor The Future of Drone Use Opportunities and Threats from Ethical and Legal Perspectives 1 3 Editor Bart Custers Faculty of Law, eLaw Leiden University Leiden The Netherlands ISSN 1570-2782 ISSN 2215-1966 (electronic) Information Technology and Law Series ISBN 978-94-6265-131-9 ISBN 978-94-6265-132-6 (eBook) DOI 10.1007/978-94-6265-132-6 Library of Congress Control Number: 2016945865 Published by T.M.C. ASSER PRESS, The Hague, The Netherlands www.asserpress.nl Produced and distributed for T.M.C. ASSER PRESS by Springer-Verlag Berlin Heidelberg © T.M.C. ASSER PRESS and the authors 2016 No part of this work may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, microfilming, recording or otherwise, without written permission from the Publisher, with the exception of any material supplied specifically for the purpose of being entered and executed on a computer system, for exclusive use by the purchaser of the work. The use of general descriptive names, registered names, trademarks, service marks, etc.
    [Show full text]
  • Ethically Aligned Design, First Edition
    ETHICALLY ALIGNED DESIGN First Edition A Vision for Prioritizing Human Well-being with Autonomous and Intelligent Systems The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems Table of Contents The views and opinions expressed in this collaborative work are those of the authors and do not necessarily reflect the official policy or position of their respective institutions or of the Institute of Electrical and Electronics Engineers (IEEE). This work is published under the auspices of the IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems for the purposes of furthering public understanding of the importance of addressing ethical considerations in the design of autonomous and intelligent systems. Please see page 290, How the Document Was Prepared, for more details regarding the preparation of this document. This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 United States License. The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems Table of Contents Introduction 2 Executive Summary 3-6 Acknowledgements 7-8 Ethically Aligned Design From Principles to Practice 9-16 General Principles 17-35 Classical Ethics in A/IS 36-67 Well-being 68-89 Affective Computing 90-109 Personal Data and Individual Agency 110-123 Methods to Guide Ethical Research and Design 124-139 A/IS for Sustainable Development 140-168 Embedding Values into Autonomous and Intelligent Systems 169-197 Polic y 198-210 L a w 211-281 About Ethically Aligned Design The Mission and Results of The IEEE Global Initiative 282 From Principles to Practice—Results of Our Work to Date 283-284 IEEE P7000™ Approved Standardization Projects 285-286 Who We Are 287 Our Process 288-289 How the Document was Prepared 290 How to Cite Ethically Aligned Design 290 Key References 291 This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 United States License.
    [Show full text]
  • Mark Coeckelbergh
    The Language of Technology Invitation to the Inaugural Lecture Faculty of Philosophy and Education Wednesday, 10 May 2017 Mark Coeckelbergh Professor of Philosophy of Media and Technology Mark Coeckelbergh is Professor of Philosophy of Media and Technology at the Faculty of Philosophy and Education at the University of Vienna (since December 2015). Born in 1975 in Leuven, Belgium, Mark Coeckelbergh studied Social Sciences and Political Sciences at the University of Leuven, Belgium and Social Philosophy at the University of East Anglia, UK. 2003 Ph.D. at the University of Birmingham, UK. 2003 – 2006 Lecturer in Philosophy at Maastricht University, Netherlands. 2007 – 2014 Assistant Professor at the University of Twente, Netherlands. 2012 – 2014 Managing Director at the 3TU. (now 4TU.) Centre for Ethics and Technology, Netherlands. Since 2014 Full Professor of Technology and Social Responsibility at the Centre for Computing and Social Responsibility at De Montfort University, UK (part-time since December 2015). 2014 – 2016 Leader of work packages in European projects (FP7 projects DREAM and SATORI). Since 2016 Vice-President/President-Elect of the Society for Philosophy and Technology. Current research areas: Philosophy of technology, including philosophy of robotics, automation and artificial intelligence; Ethics, such as virtue ethics, ethics of technology, ethics of robotics, ethics of information technology, computer ethics, health care ethics, environmental ethics, animal ethics; Moral philosophy; Philosophy of language; Epistemology;
    [Show full text]
  • Enactivism As a Postphenomenological Metaphysics
    1 Human-Technology Relations: Postphenomenology and Philosophy of Technology 11-13 July 2018 DesignLab University of Twente, The Netherlands 2 DEAR PARTICIPANTS, Welcome to the Philosophy of Human-Technology Relations conference 2018! We are very happy to have such a great group of people presenting papers, showing work in art, design and engineering, and discussing each other’s work. The number of people sending in abstracts and registering to participate is much larger than we had dared to expect, which made it a true pleasure to organize the conference. While focusing on the philosophy of human- technology relations, the conference reaches out to art, design, engineering, and Science and Technology Studies. We have paper presentations, demonstrations, hands-on workshops, book panels, and a book exhibit. Participants come from all over the world, so we hope the conference will bring about many new connections. Our home base will be the DesignLab of the University of Twente, which brings technology together with the social sciences and humanities, focusing on responsible design. For the conference dinner, on Thursday evening, we will move to the city of Enschede, where we will have dinner in The Museum Factory: and old textile factory (Twente used to be the main Dutch textile industry region) which was turned into a museum after the Enschede Fireworks disaster in 2000, and which currently has an exposition on Frankenstein and Human-Technology Relations. If there are any questions, please don’t hesitate to contact the organization: there will always be people in the PhilosophyLab, students in the organization can be recognized by their t-shirt, and the members of the organizing committee will be around during the entire conference.
    [Show full text]
  • The Tragedy of the Master: Automation, Vulnerability, and Distance
    Author's personal copy Ethics Inf Technol (2015) 17:219–229 DOI 10.1007/s10676-015-9377-6 ORIGINAL PAPER The tragedy of the master: automation, vulnerability, and distance Mark Coeckelbergh1 Published online: 3 September 2015 Ó Springer Science+Business Media Dordrecht 2015 Abstract Responding to long-standing warnings that Introduction robots and AI will enslave humans, I argue that the main problem we face is not that automation might turn us into Echoing a long tradition going back at least to Aristotle’s slaves but, rather, that we remain masters. First I construct comments in the Nicomachean Ethics about the slave as a an argument concerning what I call ‘the tragedy of the ‘living tool’ and the tool as a ‘lifeless slave’ (Aristotle, master’: using the master–slave dialectic, I argue that 1161b2-7), thinking about technology in terms of masters automation technologies threaten to make us vulnerable, and slaves is more popular than ever. Current discussions alienated, and automated masters. I elaborate the implica- about automation, robotics, and AI in the robotics/AI tions for power, knowledge, and experience. Then I criti- community and beyond seem to be am especially fertile cally discuss and question this argument but also the very ground for such a thinking. For instance, there have been thinking in terms of masters and slaves, which fuels both many recent warnings about AI or robots taking over, arguments. I question the discourse about slavery and indeed becoming our masters rather than the other way object to the assumptions made about human–technology around. Consider for instance claims about AI in the media relations.
    [Show full text]
  • Robots: Ethical Or Unethical?
    Dutch side-event: 10 Nov 2017 – UNESCO Paris – Room X – 1-2.30 pm Robots: ethical or unethical? Intro Ambassador Lionel Veer & ADG SHS Nada Al-Nashif What change caused by robots can we expect? Prof. Vanessa Evers How do robots concretely change our lives? With her robots Prof. Evers will show some examples. Which are the ethical dilemmas? Prof. Mark Coeckelbergh The ‘Frankenstein risk’: who will be in charge, man or robot? Furthermore: is robotic behavior ethical and how does it affect humanity, for example our social life. How can we ensure that innovation is ethical? Prof. Peter-Paul Verbeek How do humans and robots interact, and how can development and innovation in robotics be shaped in an ethical way? Prof. Verbeek will present UNESCO’s COMEST report which advises Governments on these challenges. Ask a question to the speakers by using the hashtag #UnescoRobots! Nada Al-Nashif is UNESCO’s Assistant Director-General for Social and Human Sciences since 16 February 2015. Prior to joining UNESCO, she served as Assistant Director-General/Regional Director of the International Labour Organization’s Regional Office for Arab States, based in Beirut, Lebanon. Previously she worked at UNDP, where she started her UN career in 1991, serving in Libya, Lebanon, Iraq and at Headquarters in New York. As a development economist and practitioner, she serves in an advisory capacity on several boards, notably the Boards of Trustees of Birzeit University and the Human Development NGO, Taawon. Vanessa Evers is a full Professor of Human Media Interaction at the University of Twente. Her research focuses on Human Interaction with Autonomous Agents such as robots or machine learning systems and cultural aspects of Human Computer Interaction.
    [Show full text]
  • Humans, Animals, and Robots: a Phenomenological Approach to Human-Robot Relations
    Int J Soc Robot (2011) 3: 197–204 DOI 10.1007/s12369-010-0075-6 Humans, Animals, and Robots: A Phenomenological Approach to Human-Robot Relations Mark Coeckelbergh Accepted: 22 August 2010 / Published online: 4 September 2010 © The Author(s) 2010. This article is published with open access at Springerlink.com Abstract This paper argues that our understanding of many bots may be presented as partners, as in the film The Step- human-robot relations can be enhanced by comparisons with ford Wives (2004), or in other substitute roles, as in Bicen- human-animal relations and by a phenomenological ap- tennial Man (1999) and I, Robot (2004). These images do proach which highlights the significance of how robots ap- not entirely belong to the domain of fiction: some humans pear to humans. Some potential gains of this approach are might engage in relationships with human-like robots [1] explored by discussing the concept of alterity, diversity and and as humanoid robotics develops further other practices change in human-robot relations, Heidegger’s claim that an- will emerge. However, it is more likely that in the near fu- imals are ‘poor in world’, and the issue of robot-animal rela- ture relations between humans and robots will mainly take tions. These philosophical reflections result in a perspective the form of strong attachments to robots that do not appear on human-robot relations that may guide robot design and human (although they might have some humanoid aspects). inspire more empirical human-robot relations research that They may be given an animal-like appearance (zoomor- is sensitive to how robots appear to humans in different con- phic) or a different appearance.
    [Show full text]
  • New Romantic Cyborgs
    New Romantic Cyborgs New Romantic Cyborgs Romanticism, Information Technology, and the End of the Machine Mark Coeckelbergh The MIT Press Cambridge, Massachusetts London, England © 2017 Massachusetts Institute of Technology All rights reserved. No part of this book may be reproduced in any form by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher. This book was set in Stone Sans and Stone Serif by Toppan Best-set Premedia Limited. Printed and bound in the United States of America. Library of Congress Cataloging-in-Publication Data Names: Coeckelbergh, Mark, author. Title: New romantic cyborgs : romanticism, information technology, and the end of the machine / Mark Coeckelbergh. Description: Cambridge, MA : MIT Press, 2017. | Includes bibliographical references and index. Identifiers: LCCN 2016021510 | ISBN 9780262035460 (hardcover : alk. paper) Subjects: LCSH: Technology--Philosophy. | Human-machine systems--Philosophy. | Information technology--Philosophy. | Cyborgs--Philosophy. | Romanticism. Classification: LCC T14 .C5726 2017 | DDC 601--dc23 LC record available at https://lccn.loc.gov/2016021510 10 9 8 7 6 5 4 3 2 1 To Tim O’Hagan and Nicholas Dent, from whom I learned a lot about Rousseau Contents Acknowledgments ix 1 Introduction: The Question Concerning Technology and Romanticism 1 I Romanticism against the Machine 19 2 Romanticism 21 3 Romanticism against the Machine? 71 II Romanticism with the Machine 95 4 Romanticism with
    [Show full text]
  • The Moral Consideration of Artificial Entities: a Literature Review
    The Moral Consideration of Artificial Entities: A Literature Review Jamie Harris: Sentience Institute, [email protected] Jacy Reese Anthis: Sentience Institute & Department of Sociology, University of Chicago Abstract Ethicists, policy-makers, and the general public have questioned whether artificial entities such as robots warrant rights or other forms of moral consideration. There is little synthesis of the research on this topic so far. We identify 294 relevant research or discussion items in our literature review of this topic. There is widespread agreement among scholars that some artificial entities could warrant moral consideration in the future, if not also the present. The reasoning varies, such as concern for the effects on artificial entities and concern for the effects on human society. Beyond the conventional consequentialist, deontological, and virtue ethicist ethical frameworks, some scholars encourage “information ethics” and “social-relational” approaches, though there are opportunities for more in-depth ethical research on the nuances of moral consideration of artificial entities. There is limited relevant empirical data collection, primarily in a few psychological studies on current moral and social attitudes of humans towards robots and other artificial entities. This suggests an important gap for social science research on how artificial entities will be integrated into society and the factors that will determine how the interests of sentient artificial entities are considered. Keywords: Artificial Intelligence, Robots, Rights, Ethics, Philosophy of Technology, Suffering Risk Introduction Humans could continue to exist for many centuries on earth or expand to far greater interstellar populations in the coming millennia (e.g. very approximately 1038 human lives per century if we colonize the Virgo Supercluster (Bostrom 2003)).
    [Show full text]
  • You, Robot: on the Linguistic Construction of Artificial Others
    AI & Soc (2011) 26:61–69 DOI 10.1007/s00146-010-0289-z ORIGINAL ARTICLE You, robot: on the linguistic construction of artificial others Mark Coeckelbergh Received: 20 March 2010 / Accepted: 26 July 2010 / Published online: 10 August 2010 Ó The Author(s) 2010. This article is published with open access at Springerlink.com Abstract How can we make sense of the idea of ‘per- 1 Introduction sonal’ or ‘social’ relations with robots? Starting from a social and phenomenological approach to human–robot I love you. Do you love me? relations, this paper explores how we can better under- (sentence addressed to a robotic doll, reported in Turkle et al. stand and evaluate these relations by attending to the ways 2006, p. 357). our conscious experience of the robot and the human– robot relation is mediated by language. It is argued that The robots are coming. But if they enter your home, our talk about and to robots is not a mere representation they may not kill you; there is a good chance that they of an objective robotic or social-interactive reality, but simply want a hug. Robots are no longer confined to fac- rather interprets and co-shapes our relation to these arti- tories, laboratories, and—increasingly—battlefields. They ficial quasi-others. Our use of language also changes as a gradually enter people’s daily lives, offering companion- result of our experiences and practices. This happens ship, entertainment, sex, or health care. Some people prefer when people start talking to robots. In addition, this paper artificial friends or even artificial partners.
    [Show full text]
  • Why Care About Robots? Empathy, Moral Standing, and the Language of Suffering
    Kairos. Journal of Philosophy & Science 20, 2018 Center for the Philosophy of Sciences of Lisbon University Why Care About Robots? Empathy, Moral Standing, and the Language of Suffering Mark Coeckelbergh University of Vienna and De Montfort University (UK) [email protected] Abstract. This paper tries to understand the phenomenon that humans are able to empathize with robots and the intuition that there might be something wrong with “abusing” robots by discussing the question regarding the moral standing of robots. After a review of some relevant work in empirical psychology and a discussion of the ethics of empathizing with robots, a philosophical argument concerning the moral standing of robots is made that questions distant and uncritical moral reasoning about entities’ properties and that recommends first trying to understand the issue by means of philosophical and artistic work that shows how ethics is always relatio- nal and historical, and that highlights the importance of language and appearance in moral reasoning and moral psychology. It is concluded that attention to relationality and to verbal and non-verbal languages of suffering is key to understand the pheno- menon under investigation, and that in robot ethics we need less certainty and more caution and patience when it comes to thinking about moral standing. Keywords: moral standing, robots; empathy; relations, language, art, phenomenolo- gy, hermeneutics, Wittgenstein. DOI 10.2478/kjps-2018-0007 Open Access. © 2018 M. Coeckelbergh, published by Sciendo. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. Why Care About Robots?Empathy, Moral Standing, and the Language of Suffering 1.
    [Show full text]