Alevel --- Ruleutilnotes
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Paradox of Happiness Ben Eggleston
1 Paradox of Happiness Ben Eggleston The paradox of happiness is the puzzling but apparently inescapable fact that regarding happiness as the sole ultimately valuable end or objective, and acting accordingly, often results in less happiness than results from regarding other goods as ultimately valuable (and acting accordingly). That is, in many circumstances, happiness is more effectively achieved when other objectives are regarded as worth pursuing for their own sakes than when happiness alone is regarded as worth pursuing for its own sake (see happiness; hedonism; intrinsic value). These other objectives might be regarded as ultimately valuable instead of happiness, or merely in addition to happiness; but they must be valued for their own sakes, and not merely as means to the achievement of happiness. These other objectives may include loving family relationships and friendships, meaningful professional relationships, immersion in rewarding work, the exercise of skills and abilities, accomplishments and triumphs, participation in religion or a large cause or movement, and contributions to one’s culture or nation. The paradox of happiness can be understood as applying to people individually or in groups. With respect to people individually, the paradox of happiness is the fact that any given individual is likely to be less happy if happiness is her sole ultimate objective than she would be if other goods were among her ultimate objectives. With respect to groups of people, the paradox of happiness is the fact that any given group of people, such as a community or a society, is likely to be less happy, collectively, if happiness is its sole collective ultimate objective than it would be if other goods were among its collective ultimate objectives. -
John Stuart Mill's Sanction Utilitarianism
JOHN STUART MILL’S SANCTION UTILITARIANISM: A PHILOSOPHICAL AND HISTORICAL INTERPRETATION A Dissertation by DAVID EUGENE WRIGHT Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY Chair of Committee, Linda Radzik Committee Members, Clare Palmer Scott Austin R.J.Q. Adams Head of Department, Gary Varner May 2014 Major Subject: Philosophy Copyright 2014 David Eugene Wright ABSTRACT This dissertation argues for a particular interpretation of John Stuart Mill’s utilitarianism, namely that Mill is best read as a sanction utilitarian. In general, scholars commonly interpret Mill as some type of act or rule utilitarian. In making their case for these interpretations, it is also common for scholars to use large portions of Mill’s Utilitarianism as the chief source of insight into his moral theory. By contrast, I argue that Utilitarianism is best read as an ecumenical text where Mill explains and defends the general tenets of utilitarianism rather than setting out his own preferred theory. The exception to this ecumenical approach to the text comes in the fifth chapter on justice which, I argue on textual and historical grounds, outlines the central features of Mill’s utilitarianism. With this understanding of Utilitarianism in place, many of the passages commonly cited in favor of the previous interpretations are rendered less plausible, and interpretations emphasizing Mill’s other writings are strengthened. Using this methodology, I critique four of the most prominent act or rule utilitarian interpretations of Mill’s moral theory. I then provide an interpretation of Mill’s theory of moral obligation and utilitarianism. -
Satisficing Consequentialism Author(S): Michael Slote and Philip Pettit Source: Proceedings of the Aristotelian Society, Supplementary Volumes, Vol
Satisficing Consequentialism Author(s): Michael Slote and Philip Pettit Source: Proceedings of the Aristotelian Society, Supplementary Volumes, Vol. 58 (1984), pp. 139-163+165-176 Published by: Blackwell Publishing on behalf of The Aristotelian Society Stable URL: http://www.jstor.org/stable/4106846 Accessed: 15/10/2008 09:26 Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at http://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in the JSTOR archive only for your personal, non-commercial use. Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained at http://www.jstor.org/action/showPublisher?publisherCode=black. Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission. JSTOR is a not-for-profit organization founded in 1995 to build trusted digital archives for scholarship. We work with the scholarly community to preserve their work and the materials they rely upon, and to build a common research platform that promotes the discovery and use of these resources. For more information about JSTOR, please contact [email protected]. The Aristotelian Society and Blackwell Publishing are collaborating with JSTOR to digitize, preserve and extend access to Proceedings of the Aristotelian Society, Supplementary Volumes. -
Consequentialism and Respect: Two Strategies for Justifying Act Utilitarianism
Utilitas (2020), 32,1–18 doi:10.1017/S0953820819000086 ARTICLE Consequentialism and Respect: Two Strategies for Justifying Act Utilitarianism Ben Eggleston* University of Kansas *Corresponding author. Email: [email protected] (First published online 23 April 2019) Abstract Most arguments in support of act utilitarianism are elaborations of one of two basic strat- egies. One is the consequentialist strategy. This strategy relies on the consequentialist premise that an act is right if and only if it produces the best possible consequences and the welfarist premise that the value of a state of affairs is entirely determined by its overall amount of well-being. The other strategy is based on the idea of treating indivi- duals respectfully and resolving conflicts among individuals in whatever way best con- forms to that idea. Although both of these strategies can be used to argue for the principle of act utilitarianism, they are significantly different from each other, and these differences cause them to have different strengths and weaknesses. It emerges that which argumentative strategy is chosen by a proponent of act utilitarianism has a large impact on which virtues her view has and which objections it is vulnerable to. I. Introduction Act utilitarianism is one of the most heavily debated views in moral philosophy, and the arguments that can be given in support of it are remarkably diverse. But most of them are elaborations of one of two basic strategies. These two strategies are significantly dif- ferent from each other, and these differences cause them to have different strengths and weaknesses. In this article, I describe the two strategies and explore their advantages and disadvantages. -
Advantages & Disadvantages of Consequential Ethics
CONSEQUENTIAL ETHICS: SUMMARY (c) 2019 www.prshockley.org Dr. Paul R Shockley What about Consequences? Significant Types of Utilitarianism: Utilitarianism: The right action is what brings about the greatest good Consequential Ethics (CE): An action is right iff it promotes the 1. Consequentialism = whether an act is morally to the greatest number in the long run. Here are different types of best consequences. The best consequences are those in which right depends only on consequences (not utilitarianism: “happiness” is maximized. Central question: what actions will circumstances, the intrinsic nature of the act, or 1. Hedonistic Utilitarianism: maximize Pleasure & minimalize pain. generate the best consequences? This family of outcome based anything that happens before the act). approaches are varied but two, in particular, heed our attention, 2. Act Utilitarianism: an act should be judged by its results. namely, the utilitarianism of Jeremy Bentham (1748-1832) & John 2. Actual Consequentialism = whether an act is 3. Rule Utilitarianism: an act is right iff it follows the rules that promotes Stuart Mill (1806-1873) & egoism or objectivism of Ayn Rand morally right depends only on the actual the best consequences. Ethical rules are chosen in view of the anticipated (1905-1982). Consequential ethics is also referred to as teleological consequences (not foreseen, foreseeable, intended, results flowing from keeping those rules. ethics hence, Greek word teleos, meaning “having reached one’s or likely consequences). 4. Priority Utilitarianism: maximize the achievement of people’s priorities-it is for each person to decide what constitutes personal end” or “goal directed.” This summary centers on utilitarianism. 3. Direct Consequentialism = whether an act is happiness (R.M. -
The Utilitarian Approach
Confirming Pages CHAPTER 7 The Utilitarian Approach The greatest happiness of the greatest number is the foundation of morals and legislation. Jeremy Bentham, COLLECTED WORKS (1843) 7.1. The Revolution in Ethics The late 18th and 19th centuries witnessed an astonishing series of upheavals: The modern nation-state emerged from the French Revolution and the wreckage of the Napoleonic empire; the revolutions of 1848 showed the transforming power of the ideas of “liberty, equality, and fraternity”; in the New World, America was born, sporting a new kind of consti- tution; and the American Civil War (1861–1865) would finish off slavery in Western civilization. All the while, the Indus- trial Revolution was bringing about a complete restructuring of society. It is not surprising that new ideas about ethics emerged during this era. In particular, Jeremy Bentham (1748–1832) made a powerful argument for a novel conception of morality. Morality, he urged, is not about pleasing God, nor is it about being faithful to abstract rules. Rather, morality is about mak- ing the world as happy as possible. Bentham believed in one ultimate moral principle, namely, the Principle of Utility. This principle requires us, in all circumstances, to produce the most happiness that we can. Bentham was the leader of a group of philosophical radicals whose aim was to reform the laws and institutions of England along utilitarian lines. One of his followers was James Mill, the distinguished Scottish philosopher, historian, and economist. James Mill’s son, John Stuart Mill (1806–1873), would become the leading advocate of utilitarian moral theory. -
6. Railton's “Alienation, Consequentialism, and the Demands of Morality”
6. Railton’s “Alienation, Consequentialism, and the Demands of Morality” Martín Abreu Zavaleta July 14, 2015 Railton is addressing one kind of problem that we raised. In class, we discussed the case of a parent who is offered a choice between saving the life of her own child or saving the life of another child. Saving the life of the other child would produce overall more utility than saving the life of her own, so utilitarian consequentialism claims that she should save the life of the other child. To many people, this seems to be the wrong result: surely, they think, whatever duty she has to save the life of the other child is overridden by her duties and obligations to save the life of her own child. Samuel Scheffler presents us with a less tragic, but related problem for consequentialism. Schef- fler observes that consequentialism demands that we don’t buy new shoes if we can spend that money in ways that will help other people more, or that we don’t watch TV if instead we could be doing more utility maximizing things. These two cases point to one important problem for consequentialism (here, under the guise of utilitarianism). Consequentialism seems to alienate us from the people we love and from things that make our lives richer. By focusing solely on the maximization of utility, it seems to ignore the value that other things bring into our lives: things like friendship, the love for one’s family, entertainment, art appreciation, etc. A life lived in line with consequentialist demands seems to be a very impoverished life, enslaved to the maximization of utility. -
Negative Average Preference Utilitarianism Roger Chao*
Journal of Philosophy of Life Vol.2, No.1 (March 2012):55-66 [Discussion Paper] Negative Average Preference Utilitarianism * Roger Chao Abstract For many philosophers working in the area of Population Ethics, it seems that either they have to confront the Repugnant Conclusion (where they are forced to the conclusion of creating massive amounts of lives barely worth living), or they have to confront the Non-Identity Problem (where no one is seemingly harmed as their existence is dependent on the “harmful” event that took place). To them it seems there is no escape, they either have to face one problem or the other. However, there is a way around this, allowing us to escape the Repugnant Conclusion, by using what I will call Negative Average Preference Utilitarianism (NAPU) – which though similar to anti-frustrationism, has some important differences in practice. Current “positive” forms of utilitarianism have struggled to deal with the Repugnant Conclusion, as their theory actually entails this conclusion; however, it seems that a form of Negative Average Preference Utilitarianism (NAPU) easily escapes this dilemma (it never even arises within it). 1. Introduction For many philosophers working in the area of Population Ethics, it seems that either they have to confront the Repugnant Conclusion, or they have to confront the Non-Identity Problem. To them it seems there is no escape, they either have to face one problem or the other. What I will try to show in this paper however, is that there is a way around this, allowing us to escape the Repugnant Conclusion, by using what I will call Negative Average Preference Utilitarianism (NAPU) – which though similar to anti-frustrationism, has some important differences in practice. -
1 Rule Consequentialism and the Problem of Partial Acceptance 1
1 Rule Consequentialism and the Problem of Partial Acceptance 1. Introduction Often the consequences of the acceptance of a rule depend on how many people accept that rule. The question of how to deal with differing levels of acceptance is one that a wide range of ethical theories must answer. Partial compliance and acceptance worries have stirred recent debate in Rule Utilitarianism (e.g. Ridge 2006, Hooker 2007), but also raise problems for Kantianism (e.g. Korsgaard 1986) and Contractualism (e.g. Watson, 1998). The question seems particularly relevant to Rule Consequentialism, since the number of people following a code of rules typically has a direct influence on the consequences of those rules. In an effort to determine the best version of Rule Utilitarianism1, I will analyze three established types of Rule Utilitarianism before presenting my own suggestion.2 I will first discuss a Fixed Rate suggestion (Hooker, 2000) and Michael Ridge’s Variable-Rate Rule Utilitarianism (Ridge, 2006). I will then turn to Holly Smith’s Optimum Rate Utilitarianism (Smith, 2010) before introducing my own proposal, Maximizing Expectation Rate Rule Utilitarianism. After analyzing these variants of Rule Utilitarianism, I will argue that Optimum Rate and Maximizing Expectation are the two most plausible versions of Rule Utilitarianism, though Maximizing Expectation captures our intuitions about solving partial acceptance problems without the costs of the other three theories discussed. 1 Because much of the debate on this topic has been framed in the context of Utilitarianism rather than Consequentialism, I will discuss Rule Utilitarian formulations. Most arguments here can be easily extended to Rule Consequentialist theories. -
Which Consequentialism? Machine Ethics and Moral Divergence
MIRI MACHINE INTELLIGENCE RESEARCH INSTITUTE Which Consequentialism? Machine Ethics and Moral Divergence Carl Shulman, Henrik Jonsson MIRI Visiting Fellows Nick Tarleton Carnegie Mellon University, MIRI Visiting Fellow Abstract Some researchers in the field of machine ethics have suggested consequentialist or util- itarian theories as organizing principles for Artificial Moral Agents (AMAs) (Wallach, Allen, and Smit 2008) that are ‘full ethical agents’ (Moor 2006), while acknowledging extensive variation among these theories as a serious challenge (Wallach, Allen, and Smit 2008). This paper develops that challenge, beginning with a partial taxonomy ofconse- quentialisms proposed by philosophical ethics. We discuss numerous ‘free variables’ of consequentialism where intuitions conflict about optimal values, and then consider spe- cial problems of human-level AMAs designed to implement a particular ethical theory, by comparison to human proponents of the same explicit principles. In conclusion, we suggest that if machine ethics is to fully succeed, it must draw upon the developing field of moral psychology. Shulman, Carl, Nick Tarleton, and Henrik Jonsson. 2009. “Which Consequentialism? Machine Ethics and Moral Divergence.” In AP-CAP 2009: The Fifth Asia-Pacific Computing and Philosophy Conference, October 1st-2nd, University of Tokyo, Japan, Proceedings, edited by Carson Reynolds and Alvaro Cassinelli, 23–25. AP-CAP 2009. http://ia-cap.org/ap-cap09/proceedings.pdf. This version contains minor changes. Carl Shulman, Henrik Jonsson, Nick Tarleton 1. Free Variables of Consequentialism Suppose that the recommendations of a broadly utilitarian view depend on decisions about ten free binary variables, where we assign a probability of 80% to our favored option for each variable; in this case, if our probabilities are well-calibrated and our errors are not correlated across variables, then we will have only slightly more than a 10% chance of selecting the correct (in some meta-ethical framework) specification. -
What Is the Difference Between Weak Negative and Non-Negative Ethical
What Is the Difference Between Weak Negative and Non-Negative Ethical Views? Simon Knutsson Foundational Research Institute [email protected] June 2016 Abstract Weak negative views in ethics are concerned with both reducing suffering and pro- moting happiness, but are commonly said to give more weight to suffering than to happiness. Such views include weak negative utilitarianism (also called negative- leaning utilitarianism), and other views. In contrast, non-negative views, including traditional utilitarianism, are typically said to give equal weight to happiness and suffering. However, it is not obvious how weak negative and non-negative views differ, and what it means to give happiness and suffering equal weight, or to give suffering more weight. Contents 1 Introduction2 2 Terminology3 3 The views under consideration3 4 Weak negative utilitarianism and consequentialism3 5 The Weak Negative Doctrine4 6 Weak negative axiology4 7 The difference between weak negative and non-negative views assuming hap- piness and suffering are objectively highly measurable5 8 If statements about magnitudes of happiness versus suffering involve value judgements 7 9 Weak negative and non-negative views are equally affected 10 10 When is it accurate to describe non-negative and weak negative views as symmetric vs. asymmetric? 10 Foundational Research Institute Acknowledgments 12 References 12 Appendix: Value for an individual versus value for the world—where does the negativity set in? 12 List of Figures 1 Common illustrations of non-negative -
Superintelligence As a Cause Or Cure for Risks of Astronomical Suffering
Informatica 41 (2017) 389–400 389 Superintelligence as a Cause or Cure for Risks of Astronomical Suffering Kaj Sotala and Lukas Gloor Foundational Research Institute, Berlin, Germany E-mail: [email protected], [email protected] foundational-research.org Keywords: existential risk, suffering risk, superintelligence, AI alignment problem, suffering-focused ethics, population ethics Received: August 31, 2017 Discussions about the possible consequences of creating superintelligence have included the possibility of existential risk, often understood mainly as the risk of human extinction. We argue that suffering risks (s-risks), where an adverse outcome would bring about severe suffering on an astronomical scale, are risks of a comparable severity and probability as risks of extinction. Preventing them is the common interest of many different value systems. Furthermore, we argue that in the same way as superintelligent AI both contributes to existential risk but can also help prevent it, superintelligent AI can be both the cause of suffering risks and a way to prevent them from being realized. Some types of work aimed at making superintelligent AI safe will also help prevent suffering risks, and there may also be a class of safeguards for AI that helps specifically against s-risks. Povzetek: Prispevek analizira prednosti in nevarnosti superinteligence. 1 Introduction Work discussing the possible consequences of creating We term the possibility of such outcomes a suffering superintelligent AI (Yudkowsky 2008, Bostrom 2014, risk: Sotala & Yampolskiy 2015) has discussed Suffering risk (s-risk): One where an adverse superintelligence as a possible existential risk: a risk outcome would bring about severe suffering on "where an adverse outcome would either annihilate an astronomical scale, vastly exceeding all Earth-originating intelligent life or permanently and suffering that has existed on Earth so far.