Survey of Self-Play in Reinforcement Learning

Total Page:16

File Type:pdf, Size:1020Kb

Survey of Self-Play in Reinforcement Learning SURVEY OF SELF-PLAY IN REINFORCEMENT LEARNING APREPRINT Anthony DiGiovanni1 and Ethan Zell2 Departments of Statistics1 & Mathematics2 University of Michigan {adigi,ezell}@umich.edu July 8, 2021 ABSTRACT In reinforcement learning (RL), the term self-play describes a kind of multi-agent learning (MAL) that deploys an algorithm against copies of itself to test compatibility in various stochastic environments. As is typical in MAL, the literature draws heavily from well-established concepts in classical game theory and so this survey quickly reviews some fundamental concepts. In what follows, we present a brief survey of self-play literature, its major themes, criteria, and techniques, and then conclude with an assessment of current shortfalls of the literature as well as suggestions for future directions. 1 Introduction and Background Multi-agent learning (MAL) research aims to study interacting systems of agents in a shared environment, the incentives that govern the agents’ behavior, and the optimal way to behave in such environments. As we will see, the definitions of “optimality,” “environment,” and even “incentive” are crucial to developing useful problem statements and powerful solutions. Several agendas with distinct objectives and evaluation criteria have been developed for MAL [19]. From the perspective of online and reinforcement learning (RL), it is natural to formulate MAL as design of algorithms that can achieve a user’s goals optimally (or minimize regret), conditional on an environment induced by the actions taken by other algorithms. Model classes are constructed for this induced environment, on a spectrum from fully stochastic and equivalent to a Markov decision process (MDP) [4], to fully adversarial and unpredictable [3]. Alternatively, game theoretic approaches typically ask how to converge to an equilibrium, assuming play by other agents that is “rational” in some sense. For the design of an algorithm that multiple users will want to adopt, a guiding principle is self-play compatibility (or simply “compatibility”): achieving some high performance benchmark when copies of the algorithm play each other [16]. Maintaining such performance while also adapting to non-self agents, in line with the RL agenda, arXiv:2107.02850v1 [cs.GT] 6 Jul 2021 presents a challenging research program. To appraise the state of compatibility research, particularly theoretical results, we present a survey of literature on this problem and conjecture some implications of these works. Our survey has revealed a number of limitations, open problems, and themes. First, even within the narrow MAL subproblem of compatibility, different subsets of the literature pursue disparate definitions of “good” performance (detailed in Section 3.2). These range from the relatively simple property of convergence to any Nash equilibrium of the stage game [4, 5], to the more demanding criteria of Pareto efficiency [16] in an equilibrium of the stage or repeated game, and fairness in the senses studied in cooperative bargaining theory [20]. We argue that research targeted at these latter criteria is practically essential, given that multiple users of a MAL algorithm will want it to achieve the highest rewards possible within the constraints of cooperation. Second, finite-time guarantees (particularly regret bounds) of self-play performance are almost nonexistent, and the exceptions [13, 20] are not designed or proven to maintain high rewards against simple classes of non-self players. While several works prove polynomial-time convergence, losses incurred during the time it takes for self-copies to coordinate on an equilibrium may be significant. Survey of Self-Play in Reinforcement Learning APREPRINT Third, most existing algorithms with both provable self-play compatibility, and success against other classes of players, use an approach that can be summarized as: “propose, then adapt.” The proposal is the solution that all self copies eventually coordinate on, such as a Nash equilibrium. At least when the preferences of the other agents are known, satisfying any single property desired for MAL algorithms is not too demanding. The challenge lies in simultaneously maintaining multiple guarantees. Thus, many compatible algorithms identify a target class of non-self players to adapt to, in the event that an initial testing period gives evidence that the other player is not a copy. 2 Research Aims This survey explores and conjectures answers for the following questions: • Which formal models of self-play are most theoretically tractable, and what proof techniques and tools are most commonly used? (For example, how do proofs of convergence in self-play [4, 5] overcome the difficulty of adaptive nonstationarity, which depends on the actions of each copy of the algorithm?) • Because “optimal” self-play does not appear well-defined, what are the standard benchmarks for evaluating an RL algorithm’s performance in self-play, especially in regret definitions? Which computational tradeoffs might exist for higher standards, e.g. achieving Pareto efficient solutions [7, 20] versus any Nash equilibrium [5]? • How do certain algorithms [4, 17, 20] maintain self-play compatibility while also guaranteeing safety against adversarial players, and optimality against stochastic ones? Our purpose is a focused treatment of self-play theory, rather than a broad list of summaries of the many papers in this field. Omission of a given work is therefore not an implicit judgment of the lack of importance of that work. 3 Key Results 3.1 Background: Markov Games Markov games appear as an extension of usual game theory since they incorporate some element of randomness or hidden values to a game theoretic scenario. Since Markov games involve two or more deciding agents, they also extend the notion of an MDP. For our purposes, a Markov game will adhere to several simplifying assumptions. As usual in reinforcement learning, the game’s players interact with the environment by choosing actions from a (finite) set Ai and receiving a (possibly random) reward. Games are also repeated. Moreover, Markov games are often assumed to have infinite or unknown horizon with “enough” time for players to learn the game. For simplicity, we discuss games that progress in discrete steps t 2 N, though it is possible to adapt many of these i techniques for a continuous time analogue. For player i at time t, we will denote Rt to be the reward. Some papers view this reward as deterministic [7], while others view it as the distribution of a random variable. In the case the reward is random, players only receive a sample from the distribution and therefore must make estimates as to what the i th distribution actually is. Often, Rt will depend not only on the i player’s action, but also on the other players’ actions i Q as well. To emphasize this, we may write Rt(a) where a 2 A := i Ai. We will also further restrict ourselves to the case where there are finitely many players, f1; : : : ; ng, and finitely many states, f1; : : : ; dg. For brevity, we will denote these sets by [n] and [d], respectively. Definition 1 A strategy or policy πi for player i is a map that takes as inputs the game stage and player’s possible positions to the set of probability distributions over actions: πi : N × [d] !P(Ai): If the strategy πi is independent of time, we refer to it as a stationary strategy. We will denote the set of all strategies for player i by Si. One could also define a strategy that depends on the positions of all players. th In words, πi tells the i player how to behave given its own state and the stage of the game. Player i will then choose action a 2 Ai with probability (πi)a. Note that this definition includes deterministic strategies since one could specify the probability of a particular action to be 1. Definition 2 We define a strategy profile π as the map which contains the information on the strategies of all players: [n] Y π := π1 × · · · × πn : N × [d] ! P(Ai): i2[n] 2 Survey of Self-Play in Reinforcement Learning APREPRINT In several of the papers we discuss, it is allowed that players have access to different actions [13]. Still, an important special case is called a repeated symmetric game (RSG). Definition 3 A repeated Markov game is called symmetric if, for any strategy profile π and any permutation of the players σ 2 Sn, σ(i) i R (π−i; πi) = R (πσ(−i); πσ(i)) th for all i 2 [n]. By π−i we mean the profile π, but with the i entry omitted. Similarly, by σ(−i) we mean the permutation σ applied to all players except the ith. In both symmetric and general games, the concept of Nash equilibrium (NE) arises naturally. More generally, ∗ Definition 4 Let " ≥ 0. A strategy profile π is called a "-NE if for all strategies πi 2 Si: i ∗ i ∗ E[Rt(πi; π−i)] ≤ E[Rt(π )] + "; for all players i 2 [n]. When " = 0, this is the usual notion of NE. In classical game theory, NE do an excellent job of assessing when players have no incentive to deviate from established behavior. Unfortunately, NE may be sub-optimal. For instance, consider the Stag Hunt game (Table 1). When both players cooperate and hunt the stag, both receive higher reward than all other actions in the game. However, if one player defects and hunts the rabbit instead, then neither succeeds at hunting the stag alone. Here, we can check there are two NE: (Stag; Stag) and (Rabbit; Rabbit). Yet, (Rabbit; Rabbit) results in only a fraction of the maximum reward. Hence, we introduce alternative notions to test for optimality. 3.2 Selected Evaluation Criteria in Multi-Agent Learning The literature on self-play features various solution concepts. We present both the self-play criteria and complementary criteria for other-play.
Recommended publications
  • Journal of Mathematical Economics Implementation of Pareto Efficient Allocations
    Journal of Mathematical Economics 45 (2009) 113–123 Contents lists available at ScienceDirect Journal of Mathematical Economics journal homepage: www.elsevier.com/locate/jmateco Implementation of Pareto efficient allocations Guoqiang Tian a,b,∗ a Department of Economics, Texas A&M University, College Station, TX 77843, USA b School of Economics and Institute for Advanced Research, Shanghai University of Finance and Economics, Shanghai 200433, China. article info abstract Article history: This paper considers Nash implementation and double implementation of Pareto effi- Received 10 October 2005 cient allocations for production economies. We allow production sets and preferences Received in revised form 17 July 2008 are unknown to the planner. We present a well-behaved mechanism that fully imple- Accepted 22 July 2008 ments Pareto efficient allocations in Nash equilibrium. The mechanism then is modified Available online 5 August 2008 to fully doubly implement Pareto efficient allocations in Nash and strong Nash equilibria. The mechanisms constructed in the paper have many nice properties such as feasibility JEL classification: C72 and continuity. In addition, they use finite-dimensional message spaces. Furthermore, the D61 mechanism works not only for three or more agents, but also for two-agent economies. D71 © 2008 Elsevier B.V. All rights reserved. D82 Keywords: Incentive mechanism design Implementation Pareto efficiency Price equilibrium with transfer 1. Introduction 1.1. Motivation This paper considers implementation of Pareto efficient allocations for production economies by presenting well-behaved and simple mechanisms that are continuous, feasible, and use finite-dimensional spaces. Pareto optimality is a highly desir- able property in designing incentive compatible mechanisms. The importance of this property is attributed to what may be regarded as minimal welfare property.
    [Show full text]
  • Arxiv:0803.2996V1 [Q-Fin.GN] 20 Mar 2008 JEL Classification: A10, A12, B0, B40, B50, C69, C9, D5, D1, G1, G10-G14
    The virtues and vices of equilibrium and the future of financial economics J. Doyne Farmer∗ and John Geanakoplosy December 2, 2008 Abstract The use of equilibrium models in economics springs from the desire for parsimonious models of economic phenomena that take human rea- soning into account. This approach has been the cornerstone of modern economic theory. We explain why this is so, extolling the virtues of equilibrium theory; then we present a critique and describe why this approach is inherently limited, and why economics needs to move in new directions if it is to continue to make progress. We stress that this shouldn't be a question of dogma, but should be resolved empir- ically. There are situations where equilibrium models provide useful predictions and there are situations where they can never provide use- ful predictions. There are also many situations where the jury is still out, i.e., where so far they fail to provide a good description of the world, but where proper extensions might change this. Our goal is to convince the skeptics that equilibrium models can be useful, but also to make traditional economists more aware of the limitations of equilib- rium models. We sketch some alternative approaches and discuss why they should play an important role in future research in economics. Key words: equilibrium, rational expectations, efficiency, arbitrage, bounded rationality, power laws, disequilibrium, zero intelligence, mar- ket ecology, agent based modeling arXiv:0803.2996v1 [q-fin.GN] 20 Mar 2008 JEL Classification: A10, A12, B0, B40, B50, C69, C9, D5, D1, G1, G10-G14. ∗Santa Fe Institute, 1399 Hyde Park Rd., Santa Fe NM 87501 and LUISS Guido Carli, Viale Pola 12, 00198, Roma, Italy yJames Tobin Professor of Economics, Yale University, New Haven CT, and Santa Fe Institute 1 Contents 1 Introduction 4 2 What is an equilibrium theory? 5 2.1 Existence of equilibrium and fixed points .
    [Show full text]
  • An Equilibrium-Conserving Taxation Scheme for Income from Capital
    Eur. Phys. J. B (2018) 91: 38 https://doi.org/10.1140/epjb/e2018-80497-x THE EUROPEAN PHYSICAL JOURNAL B Regular Article An equilibrium-conserving taxation scheme for income from capital Jacques Temperea Theory of Quantum and Complex Systems, Universiteit Antwerpen, Universiteitsplein 1, 2610 Antwerpen, Belgium Received 28 August 2017 / Received in final form 23 November 2017 Published online 14 February 2018 c The Author(s) 2018. This article is published with open access at Springerlink.com Abstract. Under conditions of market equilibrium, the distribution of capital income follows a Pareto power law, with an exponent that characterizes the given equilibrium. Here, a simple taxation scheme is proposed such that the post-tax capital income distribution remains an equilibrium distribution, albeit with a different exponent. This taxation scheme is shown to be progressive, and its parameters can be simply derived from (i) the total amount of tax that will be levied, (ii) the threshold selected above which capital income will be taxed and (iii) the total amount of capital income. The latter can be obtained either by using Piketty's estimates of the capital/labor income ratio or by fitting the initial Pareto exponent. Both ways moreover provide a check on the amount of declared income from capital. 1 Introduction distribution of money over the agents involved in additive transactions follows a Boltzmann{Gibbs exponential dis- The distribution of income has been studied for a long tribution. Note that this is a strongly simplified model of time in the economic literature, and has more recently economic activity: it is clear that in reality global money become a topic of investigation for statistical physicists conservation is violated.
    [Show full text]
  • Strong Nash Equilibria and Mixed Strategies
    Strong Nash equilibria and mixed strategies Eleonora Braggiona, Nicola Gattib, Roberto Lucchettia, Tuomas Sandholmc aDipartimento di Matematica, Politecnico di Milano, piazza Leonardo da Vinci 32, 20133 Milano, Italy bDipartimento di Elettronica, Informazione e Bioningegneria, Politecnico di Milano, piazza Leonardo da Vinci 32, 20133 Milano, Italy cComputer Science Department, Carnegie Mellon University, 5000 Forbes Avenue, Pittsburgh, PA 15213, USA Abstract In this paper we consider strong Nash equilibria, in mixed strategies, for finite games. Any strong Nash equilibrium outcome is Pareto efficient for each coalition. First, we analyze the two–player setting. Our main result, in its simplest form, states that if a game has a strong Nash equilibrium with full support (that is, both players randomize among all pure strategies), then the game is strictly competitive. This means that all the outcomes of the game are Pareto efficient and lie on a straight line with negative slope. In order to get our result we use the indifference principle fulfilled by any Nash equilibrium, and the classical KKT conditions (in the vector setting), that are necessary conditions for Pareto efficiency. Our characterization enables us to design a strong–Nash– equilibrium–finding algorithm with complexity in Smoothed–P. So, this problem—that Conitzer and Sandholm [Conitzer, V., Sandholm, T., 2008. New complexity results about Nash equilibria. Games Econ. Behav. 63, 621–641] proved to be computationally hard in the worst case—is generically easy. Hence, although the worst case complexity of finding a strong Nash equilibrium is harder than that of finding a Nash equilibrium, once small perturbations are applied, finding a strong Nash is easier than finding a Nash equilibrium.
    [Show full text]
  • Egalitarianism in Mechanism Design∗
    Egalitarianism in Mechanism Design∗ Geoffroy de Clippely This Version: January 2012 Abstract This paper examines ways to extend egalitarian notions of fairness to mechanism design. It is shown that the classic properties of con- strained efficiency and monotonicity with respect to feasible options become incompatible, even in quasi-linear settings. An interim egali- tarian criterion is defined and axiomatically characterized. It is applied to find \fair outcomes" in classical examples of mechanism design, such as cost sharing and bilateral trade. Combined with ex-ante utilitari- anism, the criterion characterizes Harsanyi and Selten's (1972) interim Nash product. Two alternative egalitarian criteria are proposed to il- lustrate how incomplete information creates room for debate as to what is socially desirable. ∗The paper beneficiated from insightful comments from Kfir Eliaz, Klaus Nehring, David Perez-Castrillo, Kareen Rozen, and David Wettstein. Financial support from the National Science Foundation (grant SES-0851210) is gratefully acknowledged. yDepartment of Economics, Brown University. Email: [email protected] 1. INTRODUCTION Developments in the theory of mechanism design, since its origin in the sev- enties, have greatly improved our understanding of what is feasible in envi- ronments involving agents that hold private information. Yet little effort has been devoted to the discussion of socially desirable selection criteria, and the computation of incentive compatible mechanisms meeting those criteria in ap- plications. In other words, extending the theory of social choice so as to make it applicable in mechanism design remains a challenging topic to be studied. The present paper makes some progress in that direction, with a focus on the egalitarian principle.
    [Show full text]
  • Arxiv:1701.06410V1 [Q-Fin.EC] 27 Oct 2016 E Od N Phrases
    ECONOMICS CANNOT ISOLATE ITSELF FROM POLITICAL THEORY: A MATHEMATICAL DEMONSTRATION BRENDAN MARKEY-TOWLER ABSTRACT. The purpose of this paper is to provide a confession of sorts from an economist to political science and philosophy. A confession of the weaknesses of the political position of the economist. It is intended as a guide for political scientists and philosophers to the ostensible policy criteria of economics, and an illustration of an argument that demonstrates logico-mathematically, therefore incontrovertibly, that any policy statement by an economist contains, or is, a political statement. It develops an inescapable compulsion that the absolute primacy and priority of political theory and philosophy in the development of policy criteria must be recognised. Economic policy cannot be divorced from politics as a matter of mathematical fact, and rather, as Amartya Sen has done, it ought embrace political theory and philosophy. 1. THE PLACE AND IMPORTANCE OF PARETO OPTIMALITY IN ECONOMICS Economics, having pretensions to being a “science”, makes distinctions between “positive” statements about how the economy functions and “normative” statements about how it should function. It is a core attribute, inherited largely from its intellectual heritage in British empiricism and Viennese logical positivism (McCloskey, 1983) that normative statements are to be avoided where possible, and ought contain little by way of political presupposition as possible where they cannot. Political ideology is the realm of the politician and the demagogue. To that end, the most basic policy decision criterion of Pareto optimality is offered. This cri- terion is weaker than the extremely strong Hicks-Kaldor criterion. The Hicks-Kaldor criterion presupposes a consequentialist, specifically utilitarian philosophy and states any policy should be adopted which yields net positive utility for society, any compensation for lost utility on the part arXiv:1701.06410v1 [q-fin.EC] 27 Oct 2016 of one to be arranged by the polity, not the economist, out of the gains to the other.
    [Show full text]
  • Pareto Efficiency by MEGAN MARTORANA
    RF Fall Winter 07v42-sig3-INT 2/27/07 8:45 AM Page 8 JARGONALERT Pareto Efficiency BY MEGAN MARTORANA magine you and a friend are walking down the street competitive equilibrium is typically included among them. and a $100 bill magically appears. You would likely A major drawback of Pareto efficiency, some ethicists claim, I share the money evenly, each taking $50, deeming this is that it does not suggest which of the Pareto efficient the fairest division. According to Pareto efficiency, however, outcomes is best. any allocation of the $100 would be optimal — including the Furthermore, the concept does not require an equitable distribution you would likely prefer: keeping all $100 for distribution of wealth, nor does it necessarily suggest taking yourself. remedial steps to correct for existing inequality. If the Pareto efficiency says that an allocation is efficient if an incomes of the wealthy increase while the incomes of every- action makes some individual better off and no individual one else remain stable, such a change is Pareto efficient. worse off. The concept was developed by Vilfredo Pareto, an Martin Feldstein, an economist at Harvard University and Italian economist and sociologist known for his application president of the National Bureau of Economic Research, of mathematics to economic analysis, and particularly for his explains that some see this as unfair. Such critics, while con- Manual of Political Economy (1906). ceding that the outcome is Pareto Pareto used this work to develop efficient, might complain: “I don’t his theory of pure economics, have fewer material goods, but I analyze “ophelimity,” his own have the extra pain of living in a term indicating the power of more unequal world.” In short, giving satisfaction, and intro- they are concerned about not only duce indifference curves.
    [Show full text]
  • Agent Based Models
    1 1 BUT INDIVIDUALS ARE NOT GAS MOLECULES A. Chakraborti TOPICS TO BE COVERED IN THIS CHAPTER: • Agent based models: going beyond the simple statistical mechanics of colliding particles • Explaining the hidden hand of economy: self-organization in a collection of interacting "selfish" agents • Bio-box on A Smith • Minority game and its variants (evolutionary, adaptive etc) • Box on contributions of: WB Arthur, YC Zhang, D Challet, M Marsili et al • Agent-based models for explaining the power law for price fluctuations 1.1 Agent based models: going beyond the simple statistical mechanics of colliding particles One might actually wonder how the mathematical theories and laws which try to explain the physical world of electrons, protons, atoms and molecules could be applied to understand the complex social structure and economic behavior of human beings. Specially so, since human beings are complex and widely varing in nature, properties or characteristics, whereas e.g., the electrons are identical (we do not have to electrons of different types) and indistinguishable. Each electron has a charge of 1.60218 × 10−19C, mass of 9.10938 × 10−31kg, and is a spin-1/2 particle. Moreover, such properties of Econophysics. Sinha, Chatterjee, Chakraborti and Chakrabarti Copyright c 2008 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim ISBN: 3-527-XXXXX-X 2 1 BUT INDIVIDUALS ARE NOT GAS MOLECULES electrons are universal (identical for all electrons), and this amount of informa- tion is sufficient to explain many physical phenomena concerning electrons, once you know the inter actions. But is such little information sufficient to in- fer the complex behavior of human beings? Is it possible to quantify the nature of the interactions between human beings? The answers to these questions are on the negative.
    [Show full text]
  • The Turning Point of Macroeconomics?
    ETH Zürich Invitation for The Turning Point Macro-Econophysics of 30, November 2010 Macroeconomics? Hiroshi Yoshikawa University of Tokyo “Macroeconomics was born as a distinct field The 2008-09 Financial Crisis in the 1940‟s, as a part of the intellectual (2005=100) 130 response to the Great Depression. The term 120 then referred to the body of knowledge and Exports expertise that economic disaster. My thesis 110 in this lecture is that macroeconomics in this 100 original sense has succeeded: Its central Industrial Production problem of depression prevention has been 90 solved, for all practical purposes, and has in 80 fact been solved for many decades. ” 70 Robert Lucas, Nobel Laureate 60 2006 2007 2008 2009 Presidential Address to the 2003 American Economic Association Source: METI, Cabinet Office Neoclassical Economics “ Most macroeconomics of the past 30 years was spectacularly Price Mechanism useless at best, and positively or Markets harmful at worst.” Paul Krugman、June 2009 Pareto Efficiency at Lionel Robbins Lectures in London School of Economics Adam Smith (1776) Invisible Hand “The most interesting recent developments The Adjustment of Price in macroeconomic theory seem to me describable as the reincorporation of Equates aggregative problems such as inflation and the business cycle within the general Demand for and Supply of Apples framework of „microeconomic‟ theory. If Price of Apple these developments succeed, the term „macroeconomic‟ will simply disappear from S use and the modifier „micro‟ will become superfluous.” P* Robert Lucas (1987) D Quantity of Apples Q* Behind Demand and Supply Curves …. Equilibrium As a Point The Marginal Utilities of All the Consumers (j=1,…,J) out of Consuming Apples are Equal to the Price.
    [Show full text]
  • The Fair and Laissez-Faire Markets: from a Neoliberal Laissez- Faire Baseline to a Fair Market
    Pursuit - The Journal of Undergraduate Research at The University of Tennessee Volume 5 Issue 1 Article 8 June 2014 The Fair and Laissez-Faire Markets: From a Neoliberal Laissez- Faire Baseline to a Fair Market Eric L. Dixon University of Tennessee, Knoxville, [email protected] Follow this and additional works at: https://trace.tennessee.edu/pursuit Part of the Economic Policy Commons, Economic Theory Commons, Ethics and Political Philosophy Commons, Inequality and Stratification Commons, Political Economy Commons, and the Public Policy Commons Recommended Citation Dixon, Eric L. (2014) "The Fair and Laissez-Faire Markets: From a Neoliberal Laissez-Faire Baseline to a Fair Market," Pursuit - The Journal of Undergraduate Research at The University of Tennessee: Vol. 5 : Iss. 1 , Article 8. Available at: https://trace.tennessee.edu/pursuit/vol5/iss1/8 This Article is brought to you for free and open access by Volunteer, Open Access, Library Journals (VOL Journals), published in partnership with The University of Tennessee (UT) University Libraries. This article has been accepted for inclusion in Pursuit - The Journal of Undergraduate Research at The University of Tennessee by an authorized editor. For more information, please visit https://trace.tennessee.edu/pursuit. The Fair and Laissez-Faire Markets: From a Neoliberal Laissez-Faire Baseline to a Fair Market Cover Page Footnote I would like to thank David Reidy, Jon Garthoff, and especially Jon Shefner for their continued assistance of my studies in political economy and political philosophy. I would also like to thank my friends and family for their unconditional support. This article is available in Pursuit - The Journal of Undergraduate Research at The University of Tennessee: https://trace.tennessee.edu/pursuit/vol5/iss1/8 Pursuit: The Journal of Undergraduate Research at the University of Tennessee Copyright © The University of Tennessee PURSUIT trace.tennessee.edu/pursuit Fair & Laissez-Faire Markets: From a laissez-faire baseline to a fair market conception ERIC L.
    [Show full text]
  • The Two Faces of Emergence in Economics
    Swarthmore College Works Economics Faculty Works Economics 2007 The Two Faces Of Emergence In Economics Mark Kuperberg Swarthmore College, [email protected] Follow this and additional works at: https://works.swarthmore.edu/fac-economics Part of the Economics Commons Let us know how access to these works benefits ouy Recommended Citation Mark Kuperberg. (2007). "The Two Faces Of Emergence In Economics". Soundings. Volume 90, Issue 1/2. 49-63. https://works.swarthmore.edu/fac-economics/20 This work is brought to you for free by Swarthmore College Libraries' Works. It has been accepted for inclusion in Economics Faculty Works by an authorized administrator of Works. For more information, please contact [email protected]. THE TWO FACES OF EMERGENCE IN ECONOMICS Author(s): Mark Kuperberg Source: Soundings: An Interdisciplinary Journal, Vol. 90, No. 1/2, Emergence Theory (Spring/Summer 2007), pp. 49-63 Published by: Penn State University Press Stable URL: http://www.jstor.org/stable/41179190 Accessed: 04-10-2017 18:47 UTC REFERENCES Linked references are available on JSTOR for this article: http://www.jstor.org/stable/41179190?seq=1&cid=pdf-reference#references_tab_contents You may need to log in to JSTOR to access the linked references. JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about
    [Show full text]
  • The Concept of "Efficiency" in Economics
    WWS-597 REINHARDT THE CONCEPT OF "EFFICIENCY" IN ECONOMICS The concept of “efficiency” as used in economics is multi-faceted, as is shown in the chart below. First, a distinction is made between (a) efficiency in the production of goods and services and (b) (b) efficiency in the distribution of services from producers to end users. FIGURE 1 -- FACETS OF “ECONOMIC EFFICIENCY” IN AN ECONOMY OVERALL PARETO EFICIENCY IN THE ECONOMY EFFICIENCY IN Distribution of PRODUCTION available output among members of society in a Pareto efficient manner, given an initial Full use of Efficient (cost Production of Efficient (cost income distribution available minimizing) the right minimizing) distribution (whatever it may resources input combination channels from have been). in the combinations for of outputs producers to end users economy a given output desired by (transportation, society wholesale, retail, or advertising, etc). Maximum output for a given set of inputs Efficiency in production requires a) That available resources are fully used (which means among other things that en economic with involuntary unemployment is ipso facto inefficient) b) Real resources are used so as to maximize the total social value of the output to be had from any bundle of real resources or, which is the flip side, that any level of output with a given value to society be produced by the combination of real-resource inputs which minimizes the opportunity costs of those real resources. c) Producers collectively produce the right quantity and combination of out puts. d) that goods and services are carried from their producers to end users through cost-minimizing distribution channels.
    [Show full text]