Learning Dynamics and Reinforcement in Stochastic Games By
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
CSC304 Lecture 4 Game Theory
CSC304 Lecture 4 Game Theory (Cost sharing & congestion games, Potential function, Braess’ paradox) CSC304 - Nisarg Shah 1 Recap • Nash equilibria (NE) ➢ No agent wants to change their strategy ➢ Guaranteed to exist if mixed strategies are allowed ➢ Could be multiple • Pure NE through best-response diagrams • Mixed NE through the indifference principle CSC304 - Nisarg Shah 2 Worst and Best Nash Equilibria • What can we say after we identify all Nash equilibria? ➢ Compute how “good” they are in the best/worst case • How do we measure “social good”? ➢ Game with only rewards? Higher total reward of players = more social good ➢ Game with only penalties? Lower total penalty to players = more social good ➢ Game with rewards and penalties? No clear consensus… CSC304 - Nisarg Shah 3 Price of Anarchy and Stability • Price of Anarchy (PoA) • Price of Stability (PoS) “Worst NE vs optimum” “Best NE vs optimum” Max total reward Max total reward Min total reward in any NE Max total reward in any NE or or Max total cost in any NE Min total cost in any NE Min total cost Min total cost PoA ≥ PoS ≥ 1 CSC304 - Nisarg Shah 4 Revisiting Stag-Hunt Hunter 2 Stag Hare Hunter 1 Stag (4 , 4) (0 , 2) Hare (2 , 0) (1 , 1) • Max total reward = 4 + 4 = 8 • Three equilibria ➢ (Stag, Stag) : Total reward = 8 ➢ (Hare, Hare) : Total reward = 2 1 2 1 2 ➢ ( Τ3 Stag – Τ3 Hare, Τ3 Stag – Τ3 Hare) 1 1 1 1 o Total reward = ∗ ∗ 8 + 1 − ∗ ∗ 2 ∈ (2,8) 3 3 3 3 • Price of stability? Price of anarchy? CSC304 - Nisarg Shah 5 Revisiting Prisoner’s Dilemma John Stay Silent Betray Sam -
Potential Games. Congestion Games. Price of Anarchy and Price of Stability
8803 Connections between Learning, Game Theory, and Optimization Maria-Florina Balcan Lecture 13: October 5, 2010 Reading: Algorithmic Game Theory book, Chapters 17, 18 and 19. Price of Anarchy and Price of Staility We assume a (finite) game with n players, where player i's set of possible strategies is Si. We let s = (s1; : : : ; sn) denote the (joint) vector of strategies selected by players in the space S = S1 × · · · × Sn of joint actions. The game assigns utilities ui : S ! R or costs ui : S ! R to any player i at any joint action s 2 S: any player maximizes his utility ui(s) or minimizes his cost ci(s). As we recall from the introductory lectures, any finite game has a mixed Nash equilibrium (NE), but a finite game may or may not have pure Nash equilibria. Today we focus on games with pure NE. Some NE are \better" than others, which we formalize via a social objective function f : S ! R. Two classic social objectives are: P sum social welfare f(s) = i ui(s) measures social welfare { we make sure that the av- erage satisfaction of the population is high maxmin social utility f(s) = mini ui(s) measures the satisfaction of the most unsatisfied player A social objective function quantifies the efficiency of each strategy profile. We can now measure how efficient a Nash equilibrium is in a specific game. Since a game may have many NE we have at least two natural measures, corresponding to the best and the worst NE. We first define the best possible solution in a game Definition 1. -
Lecture Notes
GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies .................................. -
Potential Games and Competition in the Supply of Natural Resources By
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by ASU Digital Repository Potential Games and Competition in the Supply of Natural Resources by Robert H. Mamada A Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Approved March 2017 by the Graduate Supervisory Committee: Carlos Castillo-Chavez, Co-Chair Charles Perrings, Co-Chair Adam Lampert ARIZONA STATE UNIVERSITY May 2017 ABSTRACT This dissertation discusses the Cournot competition and competitions in the exploita- tion of common pool resources and its extension to the tragedy of the commons. I address these models by using potential games and inquire how these models reflect the real competitions for provisions of environmental resources. The Cournot models are dependent upon how many firms there are so that the resultant Cournot-Nash equilibrium is dependent upon the number of firms in oligopoly. But many studies do not take into account how the resultant Cournot-Nash equilibrium is sensitive to the change of the number of firms. Potential games can find out the outcome when the number of firms changes in addition to providing the \traditional" Cournot-Nash equilibrium when the number of firms is fixed. Hence, I use potential games to fill the gaps that exist in the studies of competitions in oligopoly and common pool resources and extend our knowledge in these topics. In specific, one of the rational conclusions from the Cournot model is that a firm’s best policy is to split into separate firms. In real life, we usually witness the other way around; i.e., several firms attempt to merge and enjoy the monopoly profit by restricting the amount of output and raising the price. -
Stochastic Game Theory Applications for Power Management in Cognitive Networks
STOCHASTIC GAME THEORY APPLICATIONS FOR POWER MANAGEMENT IN COGNITIVE NETWORKS A thesis submitted to Kent State University in partial fulfillment of the requirements for the degree of Master of Digital Science by Sham Fung Feb 2014 Thesis written by Sham Fung M.D.S, Kent State University, 2014 Approved by Advisor , Director, School of Digital Science ii TABLE OF CONTENTS LIST OF FIGURES . vi LIST OF TABLES . vii Acknowledgments . viii Dedication . ix 1 BACKGROUND OF COGNITIVE NETWORK . 1 1.1 Motivation and Requirements . 1 1.2 Definition of Cognitive Networks . 2 1.3 Key Elements of Cognitive Networks . 3 1.3.1 Learning and Reasoning . 3 1.3.2 Cognitive Network as a Multi-agent System . 4 2 POWER ALLOCATION IN COGNITIVE NETWORKS . 5 2.1 Overview . 5 2.2 Two Communication Paradigms . 6 2.3 Power Allocation Scheme . 8 2.3.1 Interference Constraints for Primary Users . 8 2.3.2 QoS Constraints for Secondary Users . 9 2.4 Related Works on Power Management in Cognitive Networks . 10 iii 3 GAME THEORY|PRELIMINARIES AND RELATED WORKS . 12 3.1 Non-cooperative Game . 14 3.1.1 Nash Equilibrium . 14 3.1.2 Other Equilibriums . 16 3.2 Cooperative Game . 16 3.2.1 Bargaining Game . 17 3.2.2 Coalition Game . 17 3.2.3 Solution Concept . 18 3.3 Stochastic Game . 19 3.4 Other Types of Games . 20 4 GAME THEORETICAL APPROACH ON POWER ALLOCATION . 23 4.1 Two Fundamental Types of Utility Functions . 23 4.1.1 QoS-Based Game . 23 4.1.2 Linear Pricing Game . -
Online Learning and Game Theory
Online Learning and Game Theory Your guide: Avrim Blum Carnegie Mellon University [Algorithmic Economics Summer School 2012] Itinerary • Stop 1: Minimizing regret and combining advice. – Randomized Wtd Majority / Multiplicative Weights alg – Connections to minimax-optimality in zero-sum games • Stop 2: Bandits and Pricing – Online learning from limited feedback (bandit algs) – Use for online pricing/auction problems • Stop 3: Internal regret & Correlated equil – Internal regret minimization – Connections to correlated equilib. in general-sum games • Stop 4: Guiding dynamics to higher-quality equilibria – Helpful “nudging” of simple dynamics in potential games Stop 1: Minimizing regret and combining expert advice Consider the following setting… Each morning, you need to pick one of N possible routes to drive to work. Robots But traffic is different each day. R Us 32 min Not clear a priori which will be best. When you get there you find out how long your route took. (And maybe others too or maybe not.) Is there a strategy for picking routes so that in the long run, whatever the sequence of traffic patterns has been, you’ve done nearly as well as the best fixed route in hindsight? (In expectation, over internal randomness in the algorithm) Yes. “No-regret” algorithms for repeated decisions A bit more generally: Algorithm has N options. World chooses cost vector. Can view as matrix like this (maybe infinite # cols) World – life - fate Algorithm At each time step, algorithm picks row, life picks column. Alg pays cost for action chosen. Alg gets column as feedback (or just its own cost in the “bandit” model). Need to assume some bound on max cost. -
Deterministic and Stochastic Prisoner's Dilemma Games: Experiments in Interdependent Security
NBER TECHNICAL WORKING PAPER SERIES DETERMINISTIC AND STOCHASTIC PRISONER'S DILEMMA GAMES: EXPERIMENTS IN INTERDEPENDENT SECURITY Howard Kunreuther Gabriel Silvasi Eric T. Bradlow Dylan Small Technical Working Paper 341 http://www.nber.org/papers/t0341 NATIONAL BUREAU OF ECONOMIC RESEARCH 1050 Massachusetts Avenue Cambridge, MA 02138 August 2007 We appreciate helpful discussions in designing the experiments and comments on earlier drafts of this paper by Colin Camerer, Vince Crawford, Rachel Croson, Robyn Dawes, Aureo DePaula, Geoff Heal, Charles Holt, David Krantz, Jack Ochs, Al Roth and Christian Schade. We also benefited from helpful comments from participants at the Workshop on Interdependent Security at the University of Pennsylvania (May 31-June 1 2006) and at the Workshop on Innovation and Coordination at Humboldt University (December 18-20, 2006). We thank George Abraham and Usmann Hassan for their help in organizing the data from the experiments. Support from NSF Grant CMS-0527598 and the Wharton Risk Management and Decision Processes Center is gratefully acknowledged. The views expressed herein are those of the author(s) and do not necessarily reflect the views of the National Bureau of Economic Research. © 2007 by Howard Kunreuther, Gabriel Silvasi, Eric T. Bradlow, and Dylan Small. All rights reserved. Short sections of text, not to exceed two paragraphs, may be quoted without explicit permission provided that full credit, including © notice, is given to the source. Deterministic and Stochastic Prisoner's Dilemma Games: Experiments in Interdependent Security Howard Kunreuther, Gabriel Silvasi, Eric T. Bradlow, and Dylan Small NBER Technical Working Paper No. 341 August 2007 JEL No. C11,C12,C22,C23,C73,C91 ABSTRACT This paper examines experiments on interdependent security prisoner's dilemma games with repeated play. -
Potential Games with Continuous Player Sets
Potential Games with Continuous Player Sets William H. Sandholm* Department of Economics University of Wisconsin 1180 Observatory Drive Madison, WI 53706 [email protected] www.ssc.wisc.edu/~whs First Version: April 30, 1997 This Version: May 20, 2000 * This paper is based on Chapter 4 of my doctoral dissertation (Sandholm [29]). I thank an anonymous referee and Associate Editor, Josef Hofbauer, and seminar audiences at Caltech, Chicago, Columbia, Harvard (Kennedy), Michigan, Northwestern, Rochester, Washington University, and Wisconsin for their comments. I especially thank Eddie Dekel and Jeroen Swinkels for their advice and encouragement. Financial support from a State Farm Companies Foundation Dissertation Fellowship is gratefully acknowledged. Abstract We study potential games with continuous player sets, a class of games characterized by an externality symmetry condition. Examples of these games include random matching games with common payoffs and congestion games. We offer a simple description of equilibria which are locally stable under a broad class of evolutionary dynamics, and prove that behavior converges to Nash equilibrium from all initial conditions. W e consider a subclass of potential games in which evolution leads to efficient play. Finally, we show that the games studied here are the limits of convergent sequences of the finite player potential games studied by Monderer and Shapley [22]. JEL Classification Numbers: C72, C73, D62, R41. 1. Introduction Nash equilibrium is the cornerstone of non-cooperative game theory, providing a necessary condition for stable behavior among rational agents. Still, to justify the prediction of Nash equilibrium play, one must explain how players arrive at a Nash equilibrium; if equilibrium is not reached, the fact that it is self-sustaining becomes moot. -
Stochastic Games with Hidden States∗
Stochastic Games with Hidden States¤ Yuichi Yamamoto† First Draft: March 29, 2014 This Version: January 14, 2015 Abstract This paper studies infinite-horizon stochastic games in which players ob- serve noisy public information about a hidden state each period. We find that if the game is connected, the limit feasible payoff set exists and is invariant to the initial prior about the state. Building on this invariance result, we pro- vide a recursive characterization of the equilibrium payoff set and establish the folk theorem. We also show that connectedness can be replaced with an even weaker condition, called asymptotic connectedness. Asymptotic con- nectedness is satisfied for generic signal distributions, if the state evolution is irreducible. Journal of Economic Literature Classification Numbers: C72, C73. Keywords: stochastic game, hidden state, connectedness, stochastic self- generation, folk theorem. ¤The author thanks Naoki Aizawa, Drew Fudenberg, Johannes Horner,¨ Atsushi Iwasaki, Michi- hiro Kandori, George Mailath, Takeaki Sunada, and Masatoshi Tsumagari for helpful conversa- tions, and seminar participants at various places. †Department of Economics, University of Pennsylvania. Email: [email protected] 1 1 Introduction When agents have a long-run relationship, underlying economic conditions may change over time. A leading example is a repeated Bertrand competition with stochastic demand shocks. Rotemberg and Saloner (1986) explore optimal col- lusive pricing when random demand shocks are i.i.d. each period. Haltiwanger and Harrington (1991), Kandori (1991), and Bagwell and Staiger (1997) further extend the analysis to the case in which demand fluctuations are cyclic or persis- tent. One of the crucial assumptions of these papers is that demand shocks are publicly observable before firms make their decisions in each period. -
Noisy Directional Learning and the Logit Equilibrium*
CSE: AR SJOE 011 Scand. J. of Economics 106(*), *–*, 2004 DOI: 10.1111/j.1467-9442.2004.000376.x Noisy Directional Learning and the Logit Equilibrium* Simon P. Anderson University of Virginia, Charlottesville, VA 22903-3328, USA [email protected] Jacob K. Goeree California Institute of Technology, Pasadena, CA 91125, USA [email protected] Charles A. Holt University of Virginia, Charlottesville, VA 22903-3328, USA [email protected] PROOF Abstract We specify a dynamic model in which agents adjust their decisions toward higher payoffs, subject to normal error. This process generates a probability distribution of players’ decisions that evolves over time according to the Fokker–Planck equation. The dynamic process is stable for all potential games, a class of payoff structures that includes several widely studied games. In equilibrium, the distributions that determine expected payoffs correspond to the distributions that arise from the logit function applied to those expected payoffs. This ‘‘logit equilibrium’’ forms a stochastic generalization of the Nash equilibrium and provides a possible explanation of anomalous laboratory data. ECTED Keywords: Bounded rationality; noisy directional learning; Fokker–Planck equation; potential games; logit equilibrium JEL classification: C62; C73 I. Introduction Small errors and shocks may have offsetting effects in some economic con- texts, in which case there is not much to be gained from an explicit analysis of stochasticNCORR elements. In other contexts, a small amount of randomness can * We gratefully acknowledge financial support from the National Science Foundation (SBR- 9818683U and SBR-0094800), the Alfred P. Sloan Foundation and the Dutch National Science Foundation (NWO-VICI 453.03.606). -
570: Minimax Sample Complexity for Turn-Based Stochastic Game
Minimax Sample Complexity for Turn-based Stochastic Game Qiwen Cui1 Lin F. Yang2 1School of Mathematical Sciences, Peking University 2Electrical and Computer Engineering Department, University of California, Los Angeles Abstract guarantees are rather rare due to complex interaction be- tween agents that makes the problem considerably harder than single agent reinforcement learning. This is also known The empirical success of multi-agent reinforce- as non-stationarity in MARL, which means when multi- ment learning is encouraging, while few theoret- ple agents alter their strategies based on samples collected ical guarantees have been revealed. In this work, from previous strategy, the system becomes non-stationary we prove that the plug-in solver approach, proba- for each agent and the improvement can not be guaranteed. bly the most natural reinforcement learning algo- One fundamental question in MBRL is that how to design rithm, achieves minimax sample complexity for efficient algorithms to overcome non-stationarity. turn-based stochastic game (TBSG). Specifically, we perform planning in an empirical TBSG by Two-players turn-based stochastic game (TBSG) is a two- utilizing a ‘simulator’ that allows sampling from agents generalization of Markov decision process (MDP), arbitrary state-action pair. We show that the em- where two agents choose actions in turn and one agent wants pirical Nash equilibrium strategy is an approxi- to maximize the total reward while the other wants to min- mate Nash equilibrium strategy in the true TBSG imize it. As a zero-sum game, TBSG is known to have and give both problem-dependent and problem- Nash equilibrium strategy [Shapley, 1953], which means independent bound. -
Session Types and Game Semantics a Synchronous Side and an Asynchronous Side
Two Sides of the Same Coin: Session Types and Game Semantics A Synchronous Side and an Asynchronous Side SIMON CASTELLAN, Imperial College London, United Kingdom NOBUKO YOSHIDA, Imperial College London, United Kingdom Game semantics and session types are two formalisations of the same concept: message-passing open programs following certain protocols. Game semantics represents protocols as games, and programs as strategies; while session types specify protocols, and well-typed π-calculus processes model programs. Giving faithful models of the π-calculus and giving a precise description of strategies as a programming language are two difficult 27 problems. In this paper, we show how these two problems can be tackled at the same time by building an accurate game semantics model of the session π-calculus. Our main contribution is to fill a semantic gap between the synchrony of the (session) π-calculus and the asynchrony of game semantics, by developing an event-structure based game semantics for synchronous concurrent computation. This model supports the first truly concurrent fully abstract (for barbed congruence) interpretation of the synchronous (session) π-calculus. We further strengthen this correspondence, establishing finite definability of asynchronous strategies by the internal session π-calculus. As an application of these results, we propose a faithful encoding of synchronous strategies into asynchronous strategies by call-return protocols, which induces automatically an encoding at the level of processes. Our results bring session types and game semantics into the same picture, proposing the session calculus as a programming language for strategies, and strategies as a very accurate model of the session calculus.