<<

Machine : Bottom-up and Top-down Approaches for Modeling Human Moral Faculties Wendell Wallach ([email protected]) ISPS, Interdisciplinary Center; 87 Trumbull Street New Haven, CT 06520-8209 USA

Colin Allen ([email protected]) Indiana University, Department of History and 1011 E. Third Street, Bloomington, IN 47405 USA

Iva Smit([email protected]) E&E Consultants Slanijkseweg 11, 7077 AM Netterden, The Netherlands

Abstract system, robot, or capable of making moral judg- ments would be an artificial moral agent (AMA) and the The implementation of moral decision-making abilities proper of such systems is perhaps the most im- in AI is a natural and necessary extension to the social portant and challenging task facing developers of fully mechanisms of autonomous software agents and robots. Engineers exploring design strategies for systems sen- autonomous systems (Allen et al., 2000). Rosalind Pi- sitive to moral considerations in their choices and ac- card, director of the Affective Computing Group at MIT, tions will need to determine what role ethical theory put it well when she wrote, “The greater the freedom should play in defining control architectures for such of a machine, the more it will need moral standards.” systems. The architectures for morally intelligent agents (Picard, 1997) The capacity to make moral judgments fall within two broad approaches: the top-down imposi- tion of ethical theories, and the bottom-up building of is far from simple, and computer scientists are a long systems that aim at specified goals or standards which way from substantiating the collection of affective and may or may not be specified in explicitly theoretical cognitive skills necessary for moral reasoning in artificial terms. In this paper we wish to provide some direc- intelligence (AI). While there are aspects of moral judg- tion for continued research by outlining the and limitations inherent in each of these approaches. ment that can be isolated and codified for tightly defined contexts, where all available options and variables are known, moral intelligence for autonomous entities is a Introduction complex activity dependent on the integration of a broad array of discrete faculties. Over the past few decades it Moral judgment in humans is a complex activity, and has become apparent that moral judgment in humans is it is a skill that many either fail to learn adequately much more than a capacity for abstract moral reasoning. or perform with limited mastery. Although there are Emotional intelligence (Damasio, 1995; Goleman, 1996), shared values that transcend cultural differences, cul- sociability, the ability to learn from experience and social tures and individuals differ in the details of their ethi- interactions, , the capacity to understand cal systems and mores. Hence it can be extremely dif- the semantic content of symbols, a theory of mind, and ficult to agree upon criteria for judging the adequacy the ability to “read the minds” of others all contribute of moral decisions. Difficult value questions also often to moral intelligence. Social mechanisms contribute to arise where information is inadequate, and when dealing various refinements of behavior that people expect from with situations where the results of actions can not be each other, and will also be necessary for robots that fully known in advance. Thus can seem to be a function to a high degree of competence in social con- fuzzy discipline, whose methods and application apply texts. The importance of these “supra-rational” facul- to the most confusing challenges we encounter. In some ties raises the question of the extent to which an ar- respects ethics is as far away from science as one can tificial agent must emulate human faculties to function get. While any claim that ethics can be reduced to a sci- as an adequate moral agent. Computer scientists and ence would at best be naive, we believe that determining philosophers have begun to consider the challenge of how to enhance the moral acumen of autonomous soft- developing computer systems capable of acting within ware agents is a challenge that will significantly advance moral guidelines, under various rubrics such as “compu- the understanding of human decision making and ethics. tational ethics,” “machine ethics,” and “artificial moral- The task of building autonomous systems ity.” One obvious question is whose morality or what that safeguard basic human values will force scientists morality should computer scientists try to implement in to break down moral decision making into its component AI? Should the systems’ decisions and actions conform parts, recognize what kinds of decisions can and cannot to religiously or philosophically inspired value systems, be codified and managed by essentially mechanical sys- such as Christian, Buddhist, utilitarian, or other sources tems, and learn how to design cognitive and affective of social norms? In AMAs could be Islamic systems capable of managing ambiguity and conflicting androids, Kantian software agents, or robots designed perspectives. This project will demand that human de- to exemplify Classical , but the kind of morality cision making is analyzed to a degree of specificity as yet people wish to implement will suggest radical differences unknown and, we believe, it has the potential to revo- in the underlying structure of the systems in which com- lutionize the philosophical study of ethics. A computer puter scientists implement that morality. The more cen- gence, and in order to give engineers tral question from the perspective of a computer scien- a chart outlining the kinds of emotions and their roles tist is how to implement moral decision-making capabil- in moral decision making. Whether an AMA is a com- ities in AI. What kinds of decisions can be substantiated puter system or robot, implementing moral faculties in within computational systems given the present state of AI will require that scientists and philosophers study and computer and the progress we anticipate over break down moral decision making in humans into com- the next five to ten years? Attention to what can be im- putationally manageable modules or components. An plemented may also help us discern what kind of moral interplay exists between, on the one hand, analyzing system is feasible for artificial systems. For example, and theorizing about moral behavior in humans and, some ethical theories seem to provide a kind of moral cal- on the other hand, the testing of these theories in com- culus which allows, in principle, the course of action that puter models. For example, in artificial life (ALife) ex- will maximize welfare to be quantitatively determined, periments that attempt to simulate evolution, there is while other ethical theories appear to require higher or- evidence that certain moral , such as recipro- der mental faculties that we are a long way from knowing cal and fairness, may have emerged naturally how to reproduce in AI. Reproducing human faculties is through the essentially mechanical unfolding of evolu- not the only route for implementing moral decision mak- tion. If specific values are indeed naturally selected dur- ing in computational systems. In fact, computers may be ing evolution and our genes development towards better than humans in making moral decisions in so far these values, this will have profound ramifications for as they may not be as limited by the bounded rational- the manner in which we build AMAs. In any case, it is ity that characterizes human decisions, and they need essential that these systems function so as to be fully sen- not be vulnerable to emotional hijacking (Allen, 2002; sitive to the range of ethical concerns for the health and Wallach, 2004). It is worth expanding on the role of well-being of humans and other entities (animals, cor- emotions in moral decision making because it illustrates porations, etc.) worthy of moral consideration. Given many of the complexities of designing AMAs and may the need for this range of sensitivities, AMAs have the help us understand why AMAs or robots could function potential to cause considerable discomfort to human be- differently from humans. Despite appearances, the ob- ings who are not, for example, used to having machines servation that AMAs need not be subject to emotional detect their emotional states. The ability to detect emo- hijacking does not introduce any contradiction with our tions has broad ethical ramifications that pose a partic- previous claim that emotional intelligence may be neces- ularly complex challenge to sociologists, computer scien- sary for moral decision making. First, a large component tists, roboticists, and engineers concerned with human of affective intelligence concerns the ability to recognize reactions to machines. Successful human-machine inter- and respond appropriately to emotional states in others actions may well require that we incorporate the entire (Picard, 1997) and it is an open question whether this value systems underlying such interactions. This is es- kind of intelligence itself requires a computer system to sential if humans are going to trust sophisticated ma- have the capacity for emotions of their own. Second, in chines, for trust depends on the felt belief that those humans, while emotions are beneficial in many circum- you are interacting with share your essential values and stances, this is compatible with certain emotions being concerns, or at least will function within the constraints disadvantageous or even dysfunctional in other circum- suggested by those values. Having set the stage, our stances. Recognizing this fact presents an opportunity task now is to explore some bottom-up and top-down for engineers to design AMAs whose moral faculties op- approaches to the development of artificial systems capa- erate in a way that makes them less susceptible to emo- ble of making moral decisions. By “top-down” we mean tional interference or dysfunctionality. The literature on to combine the two slightly different senses of this term, affective computing is burgeoning and provides a as it occurs in engineering and as it occurs in ethics. starting point for thinking about these issues, but it is In the engineering sense, a top-down approach analyzes also noteworthy that this literature has thus far failed to or decomposes a task into simpler subtasks that can be deal with the role of emotions in moral decisions. This directly implemented and hierarchically arranged to ob- may partially reflect the lasting influence of Stoicism on tain a desired outcome. In the ethical sense, a top-down scientific thinking about morality. The Stoics believed approach to ethics is one which takes an antecedently that moral reasoning should be dispassionate and free of specified general ethical theory (whether philosophically emotional prejudice, which has been presumed to mean derived, such as , or religiously motivated, that emotions should be banned entirely from moral re- such as the “Golden Rule”) and derives its consequences flection. (The Stoic is exemplified in such science for particular cases. In our merged sense, a top-down fiction icons as Mr. Spock in Star Trek.) The emotional approach to the design of AMAs is any approach that intelligence literature demonstrates that the capacity for takes the antecedently specified ethical theory and ana- moral reasoning is complex and that emotions must be lyzes its computational requirements to guide the design considered an integral aspect of that capacity. Some of of and subsystems capable of implementing this complexity has been explored for rather different that theory. By contrast, bottom-up approaches, if they ends in the philosophical moral psychology literature. use a prior theory at all, do so only as a way of speci- The development of AMAs will require us to combine fying the task for the system, but not as a way of spec- the insights from affective computing, emotional intelli- ifying an implementation method or control structure. In bottom-up engineering, tasks can also be specified crete subsystems is not itself explicitly guided by any atheoretically using some sort of performance measure ethical theory. Rather, it is hoped that by experiment- (such as winning chess games, passing the Turing Test, ing with the way in which these subsystems interact, walking across a room without stumbling, etc.). Various something that has suitable moral capacities can be cre- trial and error techniques are available to engineers for ated. Computer scientists and roboticists are already progressively tuning the performance of systems so that working on a variety of discrete AI-related skills that are they approach or surpass the performance criteria. High presumably relevant to moral capacities. The techniques levels of performance on many tasks can be achieved, provided by artificial life, genetic algorithms, connection- even though the engineer lacks a theory of the best way ism, learning algorithms, embodied or subsumptive ar- to decompose the task into subtasks. Post-hoc analysis chitecture, evolutionary and epigenetic robotics, associa- of the system can sometimes yield a theory or specifi- tive learning platforms, and even good old-fashioned AI, cation of the relevant subtasks, but the results of such all have strengths in modeling specific cognitive skills or analyses can also be quite surprising and typically do capacities. Discrete subsystems might be built around not correspond to the kind of decomposition suggested the most effective techniques for implementing discrete by a priori theorizing. In its ethical sense, a bottom-up cognitive capacities and social mechanisms. But com- approach to ethics is one that treats normative values puter scientists following such an approach are then con- as being implicit in the activity of agents rather than fronted with the difficult (and perhaps insurmountable) explicitly articulated (or even articulatable) in terms of challenge of assembling these discrete systems into a a general theory. In our use of the term, we wish to functional whole. This discrete-systems approach can recognize that this may provide an accurate account of be contrasted with more holistic attempts to reproduce the agents’ understanding of their own morality and the in artificial systems, such as the attempt to morality of others, while we remain neutral on the onto- evolve AMAs in an ALife environment, or training a sin- logical question of whether morality is the kind of con- gle large connectionist network to reproduce the input- cept for which an adequate general theory can be pro- output functions of real moral agents. While the former duced. In practice, engineers and roboticists typically approach seems promising, it is limited by our present build their most complex systems using both top-down inability to simulate environments of sufficient social and and bottom-up approaches, assembling components that physical complexity to favor the complex capacities re- fulfill specific functions guided by a theoretical top-down quired for real-world performance. The latter approach analysis that is typically incomplete. Problem solving has no real practitioners (to our knowledge) for reasons for engineers often involves breaking down a complex that illustrate our point that practical AMA develop- project into discrete tasks and then assembling compo- ment is going to require a more structured approach; for nents that perform those discrete functions in a manner it is surely naive to think any single current technique that adequately fulfills the goals specified by the original for training homogeneous artificial neural networks could project. Commonly there is more than one route to meet satisfactorily replicate the entire suite of moral behavior the project goals, and there is a dynamic interplay be- by actual human beings. The development of artificial tween analysis of the project’s structure and the testing systems that act with sensitivity to moral considerations of the system designed to meet the goals. Failures of the is presently still largely confined to designing systems system may, for example, reveal that secondary consid- with “operational morality,” that is, “ensuring that the erations have been overlooked in one’s original analysis AI system functions as designed” (Wallach, 2003). This of the challenge, requiring additional control systems or is primarily the extension of the traditional engineering software that accommodates complex relationships. Be- concern with safety into the design of smart machines cause the top-down/bottom-up dichotomy is too simplis- that can reliably perform a specified task, whether that tic for many complex engineering tasks, we should not entails navigating a hallway without damaging itself or expect the design of AMAs to be any different. Never- running into people, visually distinguishing the presence theless, it serves a useful purpose for highlighting two of a human from an inanimate object, or deciphering the potential roles of ethical theory for the design of AMAs, emotional state implicit in a facial expression. Thus en- in suggesting actual algorithms and control structures gineers and computer scientists are still in the process of and in providing goals and standards of evaluation. designing methods for adequately fulfilling the discrete tasks that might lead cumulatively to more complex ac- Engineering Challenges: Bottom-Up tivities and greater . The word “discrete” here Discrete Systems should not be taken to mean “isolated.” In fact, among the most promising approaches to robotics are those There is a range of approaches for building any artifi- which exploit dynamic interaction between the various cially intelligent system. In this paper we focus our at- subtasks of visual perception, moving, manipulating, tention on the subset of approaches which corresponds and understanding speech. For example, Roy (submit- best to current research in AI and robotics: the assem- ted) exploits such interactions in the development of the bly of subsystems implementing relatively discrete hu- seeing, hearing, and talking robotic arm he calls “Rip- man capacities with the goal of creating a system that ley.” Ripley’s speech understanding and comprehension is complex enough to provide a substrate for artificial systems develop in the context of carrying out human morality. Such approaches are “bottom-up” in our sense requests for actions related to identifying and manipu- because the development and deployment of these dis- lating objects within its fields of vision and reach, while new relationships and feel their way over time to deep- balancing internal requirements such as not allowing its ening levels of trust is helpful in understanding what we servos to overheat – a very real mechanical concern, as it mean by dynamic morality. We humans invest each re- turns out. Especially interesting for our purposes is that lationship with varying degrees of trust, but there is no Roy (2005) explicitly places his project in the context of simple formula for trust or no one method for establish- Asimov’s Three (Asimov, 1950): do- ing the degree to which a given person will trust a new ing no harm to humans, obeying humans, and preserv- acquaintance. A variety of social mechanisms including ing self. Roy depicts the various subsystems involved low risk experiments with cooperation, the reading of in speech processing, object recognition, movement, and another’s emotions in specific situations, estimations of recuperation, as nodes in a graph whose edges indicate the other’s character, and calculations regarding what interactions between the subsystems and their relation- we are willing to risk in a given relationship all feed into ships to the laws. It is obvious, for example, how speech the dynamic determination of trust. Each new social comprehension is essential to the second law, and motor interaction holds the prospect of altering the degree of cooling is essential to the third law. But the most telling trust invested in a relationship. The lesson for AI re- feature of Roy’s graph for our purposes is that the line search and robotics is that while AMAs should not enter representing an edge leading from Asimov’s first law to the world suspicious of all relationships, they will need a presumed implementation of moral capacities trails off the capacity to dynamically negotiate or feel their way into dots. Without necessarily endorsing Asimov’s laws, through to elevated levels of trust with the other humans one may understand the challenge to computer scien- or computer systems with which they interact. A poten- tists as how to connect the dots to a substantial account tial strength of complex bottom-up systems lies in the of ethical behavior. How do we get from discrete skills manner they dynamically integrate input from differing to systems that are capable of autonomously displaying social mechanisms. Their weakness traditionally lies in complex behavior, including moral behavior, and that not knowing which goals to use for evaluating choices and are capable of meeting challenges in new environmental actions as contexts and circumstances change. Bottom- contexts and in interaction with many agents? Some sci- up systems work best when they are directed at achiev- entists hope or presume that the aggregation of discrete ing one clear goal. When the goals are several or the skill sets will lead to the emergence of higher order cog- available information is confusing or incomplete, it is a nitive faculties including emotional intelligence, moral much more difficult task for bottom-up engineering to judgment, and consciousness. While “emergence” is a provide a clear course of action. Nevertheless, progress popular word among some scientists and philosophers of in this area is being made, allowing adaptive systems to science, it is still a rather vague concept, which implies deal more effectively with transitions between different that more complex activities will somehow arise syner- phases of behavior within a given task (e.g., Smith et gistically from the integration of discrete skills. Per- al., 2002) and with transitions between different tasks haps the self-organizing capacity of evolutionary algo- (Roy, 2005). A strength of bottom-up engineering lies rithms, learning systems or evolutionary robots provides in the assembling of components to achieve a goal. Pre- a technique that will facilitate emergence. Systems that suming, however, that a sophisticated capacity for moral learn and evolve might discover ways to integrate dis- judgment will just emerge from bottom-up engineering crete skills in pursuit of specific goals. It is important is unlikely to get us there, and this suggests that the to note that the goals here are not predefined as they analysis provided by top-down approaches will be neces- would be by a top-down analysis. For example, in evolu- sary. tionary systems, a major problem is how to design a fit- ness function that will lead to the emergence of AMAs, Top-Down Theories of Ethics without explicitly applying moral criteria. The slogan “survival of the most moral” highlights the problem of The philosophical study of ethics has focused on the saying what “most moral” amounts to in a non-circular question of whether there are top-down criteria that il- (and computationally tractable) fashion, such that the luminate moral decision making. Are there duties, prin- competition for survival in a complex physical and social ciples, rules, or goals to which the countless moral judg- environment will lead to the emergence of moral agents. ments we make daily are all subservient? In a changing The individual components from which bottom-up sys- world it is impossible to have rules, laws, or goals that tems are constructed tend to be mechanical and limited adequately cover every possible situation. One of the in the flexibility of the responses they can make. But functions of legislatures, courts, and even differing cul- when integrated successfully these components can give tures is to adjudicate and prioritize values as new chal- rise to complex dynamic systems with a range of choices lenges or unanticipated events arise. Top-down ethical or optional responses to external conditions and pres- systems come from a variety of sources including reli- sures. Bottom-up engineering thus offers a kind of dy- gion, philosophy, and literature. Examples include the namic morality, where the ongoing feedback from differ- Golden Rule, the Ten Commandments, consequentialist ent social mechanisms facilitates varied responses as con- or utilitarian ethics, Kant’s moral imperative and other ditions change. For example, humans may be born trust- duty based theories, legal codes, ’s virtues, and ing their parents and other immediate caregivers, but the Asimov’s three laws for robots. To date very little re- manner in which humans, both children and adults, test search has been done on the computerization of top- down ethical theories. Those few systems designed to analyze moral challenges are largely relegated to medical duties or might suffer the same problem as a list advisors that help doctors and other health practition- of commandments (in fact, some of the traditional com- ers weigh alternative courses of treatments or to eval- mandments are duties to others). For example, a duty uate whether to withhold treatment for the terminally to tell the truth might, in some circumstances, come ill. Computer-aided decision support is also popular for into conflict with a duty to respect another person’s pri- many business applications, but generally the actual de- vacy. One way to resolve these problems is to submit cision making is left to human managers. There are, all prima facie duties to a higher principle. Thus, for nevertheless, many questions as to whether such systems instance, it was Kant’s belief that all legitimate moral might undermine the of the human duties could be grounded in a single principle, the cate- decision makers (Friedman & Kahn, 1992). Will, for ex- gorical imperative, which could be stated in such a way ample, doctors and hospitals, feel free to override the as to guarantee logical consistency. It turns out that a advice of an expert system with a good track record, computational Kantian would also have to do much com- when there will always be a computerized audit trail puting to achieve a full moral evaluation of any action. available to enterprising lawyers, if the human’s decision This is because Kant’s approach to the moral evaluation goes awry? Throughout industry and in financial ap- of actions requires a full understanding of the motives plications values are built into systems that must make behind the action, and an assessment of whether there choices among available courses of actions. Generally would be any inconsistency if everyone acted on the same these are the explicit values of the institutions who use motive. This requires much understanding of psychol- the systems and the engineers who design the systems, ogy and of the effects of actions in the world. Other although many systems also substantiate often unrecog- general deontological principles that seek to resolve con- nized implicit values (Friedman & Nissenbaum, 1996). flicts among prima facie duties would face similar issues. One would be hard pressed to say that these systems Both consequentialist (e.g., utilitarian) and deontologi- are engaged in any form of explicit moral reasoning, al- cal (e.g., Kantian) approaches provide general principles, though they are capable of making choices that their whose relevance to the design of ethical robots was first designers might not have been able to anticipate. The noted by Gips (1995). These different approaches raise usefulness of such systems lies largely in their ability to their own specific computational problems, but they also manage large quantities of information and make deci- raise a common problem of whether any computer (or sions based on their institution’s goals and values, at a human, for that matter) could ever gather and compare speed that humans can not approach. Top-down eth- all the information that would be necessary for the the- ical theories are each meant to capture the essence of ories to be applied in real time. Of course humans apply moral judgment. Nevertheless, the history of moral phi- consequentialist and deontological reasoning to practi- losophy can be viewed as a long inquiry into the limi- cal problems without calculating endlessly the utility or tations inherent in each new top-down theory proposed. moral ramifications of an act in all possible situations. These known limitations also suggest specific challenges Our morality, just as our reasoning, is bounded by time, in implementing a top-down theory in AI. For exam- capacity, and inclination. Parameters might also be set ple, if you have many rules or laws, what should the to limit the extent to which a computational system an- system do when it encounters a challenge where two or alyzes the beneficial or imperatival consequences of a more rules conflict? Can laws or values be prioritized specific action. How might we set those limits on the or does their relative importance change with a change options considered by a computer system, and will the in context? To handle such problems philosophers have course of action taken by such a system in addressing a turned toward more general or abstract principles from specific challenge be satisfactory? In humans the limits which all the more specific or particular principles might on reflection are set by heuristics and affective controls. be derived. Among ethical theories, there are two “big Both heuristics and affect can at times be irrational but picture” rivals for what the general principle should be. also tend to embody the wisdom gained through expe- Utilitarians claim that ultimately morality is about max- rience. We may well be able to implement heuristics in imizing the total amount of “utility” (a measure of hap- computational systems, but affective controls on moral piness or well being) in the world. The best actions (or judgments represent a much more difficult challenge. the best specific rules to follow) are those that maxi- mize aggregate utility. Because utilitarians care about Our brief survey by no means exhausts the variety the consequences of actions for utility, their views are of top-down ethical theories that might be implemented called “consequentialist.” Consequentialists have much in AI. Rather, it has been our intention to use these computing to do, because they need to work out many, two representative theories to illustrate the kinds of is- if not all, of the consequences of the available alterna- sues that arise when one considers the challenges entailed tives in order to evaluate them morally. The competing in designing an AMA. comes in many “big picture” view of moral principles is that ethics is variations and each religious tradition or culture has its about duties to others, and, on the flip side of duties, own set of rules. Consequences and rules are combined the rights of individuals. Being concerned with duties in some theories. Arguably one of these variations might and rights, such theories fall under the heading of “deon- have built in heuristics that cut through some of the com- tology,” a 19th century word which is pseudo-Greek for plexity we have discussed. Rule consequentialists, for ex- the study of obligations. In general, any list of specific ample, determine whether an act is morally wrong based on a code of rules that have been selected solely in terms of their average or expected consequences. The conse- cerned with evaluating the morality of actions on the quences of adopting the rule must be more favourable basis solely of outcomes (consequentialism), or in terms than unfavourable to all parties involved. Presumably of rights and duties (deontology). theorists main- these rules, once selected, will cut down on much of the tain that morally good actions flow from the cultivation calculation that we described earlier for an act conse- of good character, which consists in the realization of quentialist who will have to evaluate all the consequences specific virtues. Virtues are more complex than mere of his action. But there is a trade-off. The rule conse- skills, as they involve characteristic patterns of motiva- quentialist must deal with conflicting rules, the prioriti- tion and desire. claimed that virtues couldn’t zation of rules, and any exceptions to the rules. When be misused, because if people have a certain virtue, it is thinking of rules for robots, Asimov’s laws come imme- impossible for them to act as if they did not have it. This diately to mind. On the surface these three laws, plus a led him to the conclusion that there is only one virtue, “zeroth” law that he added in 1985 to place humanity’s the power of right judgment, while Aristotle favored a interest above that of any individual, appear to be in- longer list. Just as utilitarians do not agree on how to tuitive, straightforward, and general enough in scope to measure utility, and deontologists do not agree on which capture a broad array of ethical concerns. But in story list of duties apply, virtue do not agree on a after story Asimov demonstrates problems of prioriti- standard list of virtues that any moral agent should ex- zation and potential deadlock inherent in implementing emplify. Rather than focus on these difference in this even this small set of rules (Clark, 1994). Apparently paper, our attention will be directed at the computa- Asimov concluded that his laws would not work, and tional tractability of : could one make use other theorists have extended this conclusion to encom- of virtues as a programming tool? Virtues affect how pass any rule based ethical system implemented in AI we deliberate and how we motivate our actions, but an (Lang, 2002). In addition, we will require some crite- explicit description of the relevant virtue rarely occurs ria for evaluating whether a moral judgment made by in the content of the deliberation. For instance, a kind a computational system is satisfactory. When dealing person does kind things, but typically will not explain with differences in values and ethical theories this is by this behavior in terms of her own kindness. Rather, a no means an easy matter, and even criteria such as a kind person will state motives focused on the beneficiary Moral Turing Test, designed to manage differences in of the kindness, such as “she needs it,” “it will cheer opinion as to what is ‘right’ and ‘good,’ have inherent him up,” or “it will stop the pain” (Williams, 1985). weaknesses (Allen et al., 2000). The strength of top- Besides revealing some of the complexities of virtue the- down theories lies in their defining ethical goals with a ory, this example also demonstrates that the boundaries breadth that subsumes countless specific challenges. But between the various ethical theories, in this case utili- this strength can come at a price: either the goals are tarianism and virtue based ethics, can be quite fuzzy. defined so vaguely or abstractly that their meaning and Indeed, the very process of developing one’s virtues is application is subject for debate, or they get defined in a hard to imagine independently of training oneself to act manner that is static and fails to accommodate or may for the right motives so as to produce good outcomes. even be hostile to new conditions. Aristotle contended that the moral virtues are distinct from practical wisdom and intellectual virtues, which can Merging Top-Down and Bottom-Up be taught. He believed that the moral virtues must be learned through habit and through practice. This em- As we indicated at the end of the introduction, the phasis on habit, learning, and character places virtue top-down/bottom-up dichotomy is somewhat simplistic. theory in between the top-down explicit values advo- Top-down analysis and bottom-up techniques for devel- cated by a culture, and the bottom-up traits discovered oping or evolving skills and mental faculties will un- or learned by an individual through practice. Building doubtedly both be required to engineer AMAs. To il- computers with character can be approached as either a lustrate the way in which top-down and bottom-up as- top-down implementation of virtues or the development pects interact, we consider two cases. First, we’ll look of character by a learning computer. The former ap- at the possibility of utilizing a connectionist network to proach views virtues as characteristics that can be pro- develop a computer system with good character traits or grammed into the system. The latter approach stems virtues and then we’ll look at attempts to develop com- from the recognition of a convergence between modern plex mental faculties and social mechanism, such as a connectionist approaches to AI and virtue-based ethical theory of mind for an embodied robotic system. We’ll systems, particularly that of Aristotle. Top-down im- also address the challenge of bringing it all together in an plementations of the virtues are especially challenged by agent capable of interacting morally in a dynamic social the fact that virtues comprise complex patterns of mo- context. tivation and desire, and manifest themselves indirectly. Virtue Ethics For example, the virtue of being kind can be projected to hundreds of different activities. If applying virtue- Public discussions of morality are not just about rights theory in a top-down fashion, an artificial agent would (deontology) and welfare (utility); they are often also have to have considerable knowledge of psychology to about issues of character. This third element of moral figure out which virtue, or which action representing the theory can be traced back to Aristotle, and what is now virtue, to call upon in a given situation. A virtue-based known as “virtue ethics.” Virtue ethicists are not con- AMA, like its deontological cohorts, could get stuck in interesting and suggestive to note the similarity between endless looping when checking if its actions are congruent Aristotelian ethics and connectionism, but the challenge with the prescribed virtues, and then reflecting upon the of implementing virtues within a neural network remains checking, and so on. The problems here bear some rela- a formidable one. Existing connectionist systems are tionship to the well-known Frame Problem, but may also a long way from tackling the kind of complex learning be related to other issues that frequently arise in AI and tasks we associate with moral development. Neverthe- database contexts concerning predictions of indirect ef- less, the prospect that artificial neural networks might fects over extended time periods. By linking the virtues be employed for at least some dimensions of moral rea- to function (as in the Greek tradition), and tailoring soning is an intriguing possibility that deserves extensive them sharply to the specific tasks of an AMA, perhaps consideration. some of these computational problems can be mitigated. In principle, the stability of virtues – that is, if one has Consciousness, Theory of Mind, and Other a virtue, one cannot behave as if one does not have it – Supra-Rational Faculties and Social is a very attractive feature considering the need for an Mechanisms AMA to maintain “loyalty” under pressure while dealing with various, not always legitimate, information sources. A socially viable robot will require a rich set of skills to In humans, the stability of virtue largely stems from the function properly as a moral agent in multi-agent con- emotional grounding of virtues, which motivates us to texts that are in a dynamic state of flux. The relation- uphold our image as honorable beings. The challenge ships between these agents will be constantly evolving as for a designer of AMAs is to find a way to implement will the social context within which they operate. Cus- the same stability in a ‘cold’ unemotional machine. A toms change. Particularly in regards to the AMAs them- virtuous robot may require emotions of its own as well selves, one might imagine increasing acceptance and lat- as emotionally rooted goals such as . Perhaps itude for their actions as humans come to feel that their the artificial simulation of an admirable goal or desire to behavior is trustworthy. Conversely, if AMAs fail to act meet the criterion of being virtuous will suffice, but in appropriately, the public will demand laws and prac- all likelihood we will only find out by going through the tices that add new restrictions upon the AMAs behavior. actual exercise of building a virtue-based computational Morality evolves and the AMAs will be active partici- system. Several writers have mentioned that connec- pants in working through new challenges in many realms tionism or parallel distributed processing has similari- of activity. Computers actually function quite well in ties to Aristotle’s discussion of virtue ethics (DeMoss, multi-agent environments where auctions, bargaining, or 1998; Howell, 1999). Connectionism provides a bottom- other forms of negotiation are the primary mode of in- up strategy for building complex capacities, recogniz- teractions. But each transaction can also change the ing patterns or building categories naturally by map- relationships between agents. Within specific contexts, ping statistical regularities in complex inputs. Through roles, status, wealth, and other forms of social privilege the gradual accumulation of data the network develops give form to relationships and the perception of what generalized responses that go beyond the particulars on kinds of behavior are acceptable. The AMA will need which it is trained. Rather than relying on abstract, to understand customs and when to honor and when linguistically-represented theoretical knowledge, connec- to ignore these prerogatives, which can easily change – tionist systems “seem to emphasize the immediate, the for example, when an individual person moves into a perceptual, and the non-symbolic” (Gips, 1991; see also new role over time or even many times during a sin- Churchland, 1995; and DeMoss, 1998). After stating his gle day. In recent years the focus on high-level decision virtue-based theory, Aristotle spends much of the Nico- making has been directed less at theories of ethics and machean Ethics discussing the problem of how one is to more toward particular faculties and social mechanisms know which habits will lead to the “good,” or happiness. that enhance the ability to reason in concrete situations, He is clear at the outset that there is no explicit rule especially emotions (Damasio 1995; Clark 1998). Navi- for pursuing this generalized end, which is only grasped gating a constantly changing environment with both hu- intuitively. The end is deduced from the particulars, man and non-human agents will require AMAs to have from making connections between means and ends, be- both sophisticated social mechanisms as well as rich in- tween those specific things we need to do and the goals formational input about the changes in the context, and we wish to pursue. Through intuition, induction and among other agents with which it interacts. Whether experience – asking good people about the good – our we think of these faculties – which include things such as generalized sense of the goal comes into focus, and we consciousness, emotions, embodied intelligence, and the- acquire practical wisdom and moral excellence. Howell ory of mind – as supra-rational, or as illuminating once (1999) argues that connectionism is capable of explain- hidden dimensions of reason, is not of critical concern. ing how human minds develop intuitions through the What is important is that agents without these tools unconscious assimilation of large amounts of experience; will either fail in their moral decision-making abilities or he writes: “Random experiences, connectionist learning be constrained to act in limited domains. Good habits, through exposure to instances, reinforcement learning, good character, supra-rational faculties and social mech- and advice-seeking and taking, all play a part in child- anisms, are all relevant to our current understanding of hood development, and make us what we are.” It is moral acumen. If each of these faculties and mechanisms is, in turn, a composite of lower level skills, designers of AMAs can expect to borrow ideas from other attempts crude, are more desirable than ethically “blind” systems to model other high-level mental capacities. For exam- that select actions without any sensitivity to moral con- ple, in Igor Aleksander et al.’s (2005) top-down analysis, siderations. Ultimately the degree of comfort the gen- consciousness can be broken down into five broad skills eral public will feel with autonomous systems depends on including imagining, attending to input, thinking ahead, the belief that these systems will not harm humans and emotions, and being in an out-there world. Each of these other entities worthy of moral consideration, and will skills is in turn a composite or set of more limited tools. honor basic human values or norms. Short of this com- Each is likely to be independently relevant to the design fort, political pressures to slow or stop the development of AMAs, but there is also the possibility that AMAs of autonomous systems will mount. Deep philosophical will be limited unless they are fully conscious, and it objections to both the possibility and desirability of cre- also seems likely that consciousness cannot be reduced ating artificial systems that make complex decisions are to just five broad skills. Scassellati’s (2001) work on a unlikely to slow the inexorable progression toward au- theory of mind, the ability of an entity to appreciate tonomous software agents and robots. Whether strong the existence of other minds or other complex entities artificial intelligence is possible remains an open ques- affected by its actions, illustrates the challenge of imple- tion, yet regardless of how intelligent AI systems may menting a complex social mechanism within the design of become, they will require some degree of moral sensitiv- a robot. Utilizing the theories of cognitive scientists who ity in the choices and actions they take. If there are lim- have broken down a theory of mind into discrete skills, itations in the extent to which scientists can implement Scassellati and other computer scientists have tried to moral decision making capabilities in AI, it is incumbent substantiate each of these skills in hardware. While col- to recognize those limitations, so that we humans do not lectively these skills might lead to a system that actu- rely inappropriately on artificial decision makers. Moral ally acts as if it had a theory of mind, to date they have judgment for a fully functioning AMA would require the had only limited success in substantiating a few of these integration of many discrete skills. In the meantime, sci- attributes in AI. The hard work of coordinating or inte- entists will build systems that test more limited goals grating these skill sets lies ahead. Perhaps advances in for specific applications. These more specialized appli- evolutionary robotics will facilitate this integration. To cations will not require the full range of social mecha- date we not only lack systems with a theory of mind, nisms and reasoning powers necessary for complicated but we do not yet know whether we have an adequate ethical judgments in social contexts. Emotional intel- hypothesis about the attributes that are necessary for a ligence would not be required for a computer system system to have a theory of mind. The development of managing the shutdown procedures of an electrical grid other faculties and social mechanisms for computer sys- during a power surge. While such systems are not usu- tems and robots, such as being embodied in the world, ally thought to be engaged in making moral decisions, emotional intelligence, the capacity to learn from expe- presumably they will be designed to calculate minimal rience, and the ability to ‘understand’ the semantic con- harm and maximize utility. On the other hand, a ser- tent of symbols are also each in similar primitive states vice robot tending the elderly would need to be sensitive of development. If the components of a system are well to the possibility of upsetting or frightening its clients, designed and integrated properly the breadth of choices and should take appropriate action if it senses that its open to an AMA in responding to challenges arising from behavior caused fear or any other form of emotional dis- its environment and social context will expand. Presum- turbance. We have suggested that thinking in terms of ably the top-down capacity to evaluate those options will top-down and bottom-up approaches to substantiating lead to selecting those actions which both meet its goals moral decision-making faculties in AI provides a useful and fall within acceptable social norms. But ultimately framework for grasping the multi-faceted dimensions of the test of a successful AMA will not rely on whether its this challenge. There are limitations inherent in viewing bottom-up components or top-down evaluative modules moral judgments as subsumed exclusively under either are individually satisfactory. The system must function bottom-up or top-down approaches. The capacity for as a moral agent, responding to both internal and ex- moral judgment in humans is a hybrid of both bottom- ternally arising challenges. The immediate and ongoing up mechanisms shaped by evolution and learning, and challenge for computer science and robotics is the ability top-down mechanisms capable of theory-driven reason- of individual components and modules to work together ing. Morally intelligent robots require a similar fusion harmoniously. Given that the complexity of this inte- of bottom-up propensities and discrete skills and the gration will grow exponentially as we bring in more and top-down evaluation of possible courses of action. Even- more systems, scientists and engineers should focus on tually, perhaps sooner rather than later, we will need evolutionary and other self-organizing techniques. AMAs which maintain the dynamic and flexible morality of bottom-up systems that accommodate diverse inputs, Conclusion while subjecting the evaluation of choices and actions to top-down principles that represent ideals we strive Autonomous systems must make choices in the course to meet. A full appreciation of the way in which these of flexibly fulfilling their missions, and some of those elements might be integrated leads into important meta- choices will have potentially harmful consequences for ethical issues concerning the nature of moral agency it- humans and other subjects of moral concern. Systems self that are beyond the scope of the present paper. Nev- that approximate moral acumen to some degree, even if ertheless, we hope to have indicated the rich and varied Goleman, D. (1995). Emotional intelligence. New York: sources of insights that can immediately be deployed by Bantam Books. scientists who are ready to face the challenge of creating Howell, S.R. (1999). Neural networks and phi- computer systems that function as AMAs. losophy: Why Aristotle was a connectionist. http://www.psychology.mcmaster.ca/beckerlab/ Acknowledgments showell/aristotle.pdf An earlier version of the paper was prepared for and Lang, C. (2002). Ethics for artificial intelligence. Avail- presented at the Android Science Cog Sci 2005 Workshop able at http://philosophy.wisc.edu/lang/ in Stresa, Italy. We wish to thank Karl MacDorman, co- AIEthics/index.htm organizer of workshop, for his encouragement, detailed comments, and editing suggestions. We are also grateful Picard, R. (1997). Affective computing. Cambridge, for the helpful comments provided by three anonymous MA: MIT Press. reviewers arranged by the Workshop organizers. Roy, D. (2005). Meaning machines. Talk given at Indi- ana University, March 7, 2005. References Roy, D. (submitted). Grounding language in the word: Aleksander, I., Lahstein, M., & Lee, R. (2005). Will Schema theory meets semiotics. http://web.media. and emotions: A machine model that shuns illusion. mit.edu/˜dkroy/papers/pdf/aij v5.pdf. In AISB ’05: Proceedings of the symposium on next Scassellati, B. (2001). Foundations for a theory of mind generation approaches to machine consciousness. for a humanoid robot. http://www.ai.mit.edu/ Allen, C., Varner, G., & Zinser, J. (2000). Prolegomena projects/lbr/hrg/2001/scassellati-phd.pdf. to any future artificial moral agent. Journal of Ex- Smith, T., Husbands, P., & Philippides, A. (2002). perimental and Theoretical Artificial Intelligence 12, Neuronal plasticity and temporal adaptivity: GasNet 251–261. robot control networks. Adaptive Behavior 10, 161- Allen, C. (2002). Calculated morality: Ethical comput- 183. ing in the limit. In Smit, I. & Lasker, G.E. (Eds.), Cog- Skyrms, B. (2000). Game theory, rationality and evo- nitive, emotive and ethical aspects of decision making lution of the social contract in evolutionary origins and human action Vol. I. Windsor, Canada: IIAS. of morality. In Katz, L. (Ed.), Evolutionary origins Asimov, I. (1950). I, Robot. Gnome Press. of morality (pp. 269–285). Exeter, UK: Imprint Aca- demic. Churchland, P.M. (1995). The engine of reason, the seat of the soul: A philosophical journey into the brain. Wallach, W. (2003). Robot morals and human ethics in Cambridge, Massachusetts: MIT Press. Smit, I., Lasker, L. & Wallach, W., Cognitive, emo- tive and ethical aspects of decision making in humans Clark, A. (1998). Being there: Putting brain, body, and and in artificial intelligence Vol. II. Windsor, Canada: world together again. Cambridge, MA: MIT Press. IIAS. Clark, R. (1993, 1994). Asimov’s laws of robotics: Impli- Wallach, W. (2004). Artificial morality: bounded ratio- cations for information technology. Published in two nality, bounded morality and emotions. In Smit, I., parts, in IEEE Computer 26, 12 (December 1993) pp. Lasker, L. & Wallach, W. (Eds.), Cognitive, emotive 53-61 and 27,1 (January 1994), pp. 57-66. and ethical aspects of decision making and human ac- Damasio, A. (1995). Descartes error. New York: Pan tion Vol. III. Windsor, Canada: IIAS. Macmillan. Williams, B. (1985). Ethics and the limits of philosophy. Danielson, P. (2003). Modeling complex ethical agents. Cambridge, MA: Harvard University Press. Presented at the conference on Computational mod- eling in the social sciences, Univ. of Washington. http://www.ethics.ubc.ca/pad/papers/mcea.pdf. DeMoss, D. (1998). Aristotle, connectionism, and the morally excellent brain. The Paideia Archive. Pro- ceedings of the 20th World Congress of Philosophy. Friedman, B. & Kahn, P. (1992). Human agency and responsible computing. Journal of Systems and Soft- ware 17, 7–14. Friedman, B. & Nissenbaum, H. (1996). Bias in com- puter systems. ACM Transactions on Information Systems 14, 330–347. Gips, J. (1995) Towards the ethical robot. In Ford, K., Glymour, C., & Hayes, P. (eds.) Android epistemol- ogy. MIT Press, Cambridge, MA, pp. 243-252.