The Perceived Ethics of Artificial Intelligence
Total Page:16
File Type:pdf, Size:1020Kb
Markets, Globalization & Development Review Volume 5 Number 2 Article 3 2020 The Perceived Ethics of Artificial Intelligence Ross Murray University of Texas Rio Grande Valley Follow this and additional works at: https://digitalcommons.uri.edu/mgdr Part of the Business Law, Public Responsibility, and Ethics Commons, Economics Commons, Marketing Commons, Other Business Commons, Science and Technology Studies Commons, and the Sociology Commons Recommended Citation Murray, Ross (2020) "The Perceived Ethics of Artificial Intelligence," Markets, Globalization & Development Review: Vol. 5: No. 2, Article 3. DOI: 10.23860/MGDR-2020-05-02-03 Available at: https://digitalcommons.uri.edu/mgdr/vol5/iss2/3https://digitalcommons.uri.edu/mgdr/vol5/ iss2/3 This Article is brought to you for free and open access by DigitalCommons@URI. It has been accepted for inclusion in Markets, Globalization & Development Review by an authorized editor of DigitalCommons@URI. For more information, please contact [email protected]. The Perceived Ethics of Artificial Intelligence This article is available in Markets, Globalization & Development Review: https://digitalcommons.uri.edu/mgdr/vol5/ iss2/3 Murray: Ethics of Artificial Intelligence The Perceived Ethics of Artificial Intelligence Introduction Artificial Intelligence (AI) has been embedded in consumer products and services in thousands of products such as Apple’s iPhone, Amazon’s Alexa, Tesla’s autonomous vehicles, Facebook’s algorithms that attempt to increase click-through optimization, and smart vacuum cleaners. As products and services attempt to imitate the intelligence of humans – although the products and services are not making decisions based upon their own moral values – the moral values of the employees and business ethics of corporations that create the products and services are being coded into the technology that is evidenced in these products. Social researchers are becoming aware of and commenting on the biases and moral positions that are embedded – often unconsciously – in new technologies (see, e.g., Benjamin 2019; Dholakia et al. 2020; Noble 2018). The use of such products by consumers, although seemingly harmless today, may pose concern as consumers come to the realization that – by using and interacting with AI – they may be revealing the embedded ethical and moral values that were placed there by the creators of the AI. As AI products and services come to be used increasingly, there is a need for a perceived morality and ethical framework. AI has had the moral values and ethical positions designed (embedded) either on purpose or solely because – in order to imitate human intelligence – the AI has to demonstrate an ethical position. Some of the AI in the market today has advanced toward human-likeness such as semi-autonomous vehicles and in-home chatbots like Siri, Alexa and Google Assistant, which are approaching sounding like humans. Samsung’s Neon, a fully simulated digital avatar, unveiled at the 2020 CES (Consumer Electronics Show), appears very close to a human. CES is a conference and trade show that is held annually by the Consumer Technology Association and is for “technology leaders to connect, collaborate, and propel consumer technology forward” (CES 2020). Some of the recent AI related products presented at CES have been prosthetics, Selfie type, Lululab’s Lumine, Whisk, Wiser, Vital Signs monitoring (Taulli 2020). All these innovations use some type of AI, such as deep learning. Extrapolation of beliefs about AI show support for the idea that AI will outperform humans in many activities in the future. Researchers predict AI will outperform humans in many activities in the next ten years, such as translating languages (by 2024), writing high-school essays (by 2026), driving a truck (by 2027), working in retail (by 2031), writing a bestselling book (by 2049), and working as a surgeon (by 2053) (Grace et Published by DigitalCommons@URI, 2020 1 Markets, Globalization & Development Review, Vol. 5 [2020], No. 2, Art. 3 al. 2018). Relating specifically to the voice of AI, Grace et al. (2018) found that respondents of their survey believe AI will be able to “take a written passage and output a recording that can’t be distinguished from a voice actor, by an expert listener” in 9 years (median). As this study's findings were two years prior to this writing, the date is quickly approaching where we may see AI voice that is indistinguishable from a human voice. As the technology of AI progresses, it is the goal of some for AI to progress to artificial general intelligence (AGI). Examples of companies interested in AGI are OpenAI, Deepmind, Google Brain, and Facebook AI Research (Neo 2020). These companies are investing millions of dollars in developing AGI. Simply put, accomplishing AGI would mean that the technology can do anything a human can do or better, possibly surpassing capabilities of a human. With millions of dollars being spent on both AI and AGI development, the purpose of this paper is to explore and discuss the potential problems of creating a technology that will make decisions in its interactions with humans based on business ethics that are coded into the technology by a programmer. These decisions will stem from coding possibly written by an employee of a corporation or a soldier (or agent) of a government. The coder would have incentives to write the code for the technology in ways that will allow the AI to make decisions quickly, often without reference to a moral framework and/or business ethics. Immanuel Kant advanced the concept of ‘deontology’, which focuses on underlying ethical code to make decisions, not the consequences of decisions. Teleology takes into consideration the consequences of actions to guide decisions, whereas deontology does not concern itself with consequences (for a review, in the business ethics context, see Whetstone 2001). The literature has also found that time, motivations, and resources affect the ethical decision-making process (Conway and Gawronski 2013). As the extant literature continues to explore the influences of morality on ethical judgements, we have entered a new age where we have begun to create machines (AI) embedded with human morality and ethics that are performing tasks and making decisions that are instilled with the moral values we give them. Two real-life scenarios are discussed later – in the sections “The Specific Case of Alexa” and “One Human Versus One Human” – that demonstrate that AI is currently performing tasks and making decisions based upon programming by employees. Previous studies have brought to light multiple issues with AI. These include the directive of not harming humans (Bostrom and Yudkowsky 2014), emergency stop capabilities (Arnold and Scheutz 2018), creating a code of ethics (Boddington 2017), racial profiling https://digitalcommons.uri.edu/mgdr/vol5/iss2/3 DOI: 10.23860/MGDR-2020-05-02-03 2 Murray: Ethics of Artificial Intelligence concerns (Bostrom and Yudkowsky 2014; see also Benjamin 2019), imbuing ethics into an autonomous machine (Bonnemains, Saurel and Tessier 2018), security (Belk 2020; Lin et al. 2011) and moral status (Bostrom and Yudkowsky 2014). Surprisingly, however, we have limited research that crystallizes an executable action for the forward progress in terms of the perceived morality of AI. In addition, previous work on digital personal assistants (DPAs) has focused on gender issues (Søndergaard and Hansen 2018), politics motivated by economic reasons (Ojeda 2019), regulating bots on speech rights (Lamo and Calo 2019), and chatbot dialogue systems (Bendel 2019). These studies have also not explored the morality and ethics underlying the information provided to consumers by these chatbots. Those coding the AI may use a deontological moral framework, a teleological one, or a mixture, depending upon the situation; or may not even be aware that how they code has any framework at all (Noble 2018) and may only be guided by the business ethics of the corporation’s ethics handbook. Mixing of moral frameworks and business ethics for AI adds complexity to the issue at hand. Corporations and governments making such decisions will have to make judgments prior to launching the new technology. Many AI projects are unregulated and propelled by the best and fastest innovators. In the extant literature, concerns about the advancement of AI have mainly been about military applications (Fast and Horvitz 2017; Lin et al. 2009; Petit 2017; Roff and Moyes 2016; Russell et al. 2015; Sparrow 2009; Wasilow and Thorpe 2019), loss of control (Fast and Horvitz 2017; Horvitz 2014), and impact on work (Acemoglu and Restrepo 2018; Fast and Horvitz 2017; Morikawa 2017; Scherer 2015). Limited research has been done (of course, with the exceptions cited) as to which moral framework or ethical position was used to create a technology with AI and how the consumer perceives the AI’s morality or ethical position. Although currently unbeknownst to the AI (as it is not currently conscious of its decisions), there is an implicit moral framework being used when evaluated by a human looking in on the technology. One researcher even argues that it is too early for regulation (Reed 2018). Technology is progressing incrementally; it goes unnoticed at times and may seduce consumers into its use without questioning the moral framework of the technology. This paper contributes to the nascent literature by unveiling and discussing the issue of corporate employees designing products that are artificially intelligent, that make decisions based upon the business ethics of the corporation and provide information to consumers without disclosing ethical positions to the end user. Published by DigitalCommons@URI, 2020 3 Markets, Globalization & Development Review, Vol. 5 [2020], No. 2, Art. 3 A focus of this paper is to further reveal and discuss the problems of the entities – particularly corporations and governments – overseeing the production of technological products and services that influence human society, and how these entities may or may not be transparent with the moral framework or ethics used with the global use of their technologies.