Rationalizability in General Situations"
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Lecture 4 Rationalizability & Nash Equilibrium Road
Lecture 4 Rationalizability & Nash Equilibrium 14.12 Game Theory Muhamet Yildiz Road Map 1. Strategies – completed 2. Quiz 3. Dominance 4. Dominant-strategy equilibrium 5. Rationalizability 6. Nash Equilibrium 1 Strategy A strategy of a player is a complete contingent-plan, determining which action he will take at each information set he is to move (including the information sets that will not be reached according to this strategy). Matching pennies with perfect information 2’s Strategies: HH = Head if 1 plays Head, 1 Head if 1 plays Tail; HT = Head if 1 plays Head, Head Tail Tail if 1 plays Tail; 2 TH = Tail if 1 plays Head, 2 Head if 1 plays Tail; head tail head tail TT = Tail if 1 plays Head, Tail if 1 plays Tail. (-1,1) (1,-1) (1,-1) (-1,1) 2 Matching pennies with perfect information 2 1 HH HT TH TT Head Tail Matching pennies with Imperfect information 1 2 1 Head Tail Head Tail 2 Head (-1,1) (1,-1) head tail head tail Tail (1,-1) (-1,1) (-1,1) (1,-1) (1,-1) (-1,1) 3 A game with nature Left (5, 0) 1 Head 1/2 Right (2, 2) Nature (3, 3) 1/2 Left Tail 2 Right (0, -5) Mixed Strategy Definition: A mixed strategy of a player is a probability distribution over the set of his strategies. Pure strategies: Si = {si1,si2,…,sik} σ → A mixed strategy: i: S [0,1] s.t. σ σ σ i(si1) + i(si2) + … + i(sik) = 1. If the other players play s-i =(s1,…, si-1,si+1,…,sn), then σ the expected utility of playing i is σ σ σ i(si1)ui(si1,s-i) + i(si2)ui(si2,s-i) + … + i(sik)ui(sik,s-i). -
Learning and Equilibrium
Learning and Equilibrium Drew Fudenberg1 and David K. Levine2 1Department of Economics, Harvard University, Cambridge, Massachusetts; email: [email protected] 2Department of Economics, Washington University of St. Louis, St. Louis, Missouri; email: [email protected] Annu. Rev. Econ. 2009. 1:385–419 Key Words First published online as a Review in Advance on nonequilibrium dynamics, bounded rationality, Nash equilibrium, June 11, 2009 self-confirming equilibrium The Annual Review of Economics is online at by 140.247.212.190 on 09/04/09. For personal use only. econ.annualreviews.org Abstract This article’s doi: The theory of learning in games explores how, which, and what 10.1146/annurev.economics.050708.142930 kind of equilibria might arise as a consequence of a long-run non- Annu. Rev. Econ. 2009.1:385-420. Downloaded from arjournals.annualreviews.org Copyright © 2009 by Annual Reviews. equilibrium process of learning, adaptation, and/or imitation. If All rights reserved agents’ strategies are completely observed at the end of each round 1941-1383/09/0904-0385$20.00 (and agents are randomly matched with a series of anonymous opponents), fairly simple rules perform well in terms of the agent’s worst-case payoffs, and also guarantee that any steady state of the system must correspond to an equilibrium. If players do not ob- serve the strategies chosen by their opponents (as in extensive-form games), then learning is consistent with steady states that are not Nash equilibria because players can maintain incorrect beliefs about off-path play. Beliefs can also be incorrect because of cogni- tive limitations and systematic inferential errors. -
Game Theory Lecture Notes
Game Theory: Penn State Math 486 Lecture Notes Version 2.1.1 Christopher Griffin « 2010-2021 Licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License With Major Contributions By: James Fan George Kesidis and Other Contributions By: Arlan Stutler Sarthak Shah Contents List of Figuresv Preface xi 1. Using These Notes xi 2. An Overview of Game Theory xi Chapter 1. Probability Theory and Games Against the House1 1. Probability1 2. Random Variables and Expected Values6 3. Conditional Probability8 4. The Monty Hall Problem 11 Chapter 2. Game Trees and Extensive Form 15 1. Graphs and Trees 15 2. Game Trees with Complete Information and No Chance 18 3. Game Trees with Incomplete Information 22 4. Games of Chance 24 5. Pay-off Functions and Equilibria 26 Chapter 3. Normal and Strategic Form Games and Matrices 37 1. Normal and Strategic Form 37 2. Strategic Form Games 38 3. Review of Basic Matrix Properties 40 4. Special Matrices and Vectors 42 5. Strategy Vectors and Matrix Games 43 Chapter 4. Saddle Points, Mixed Strategies and the Minimax Theorem 45 1. Saddle Points 45 2. Zero-Sum Games without Saddle Points 48 3. Mixed Strategies 50 4. Mixed Strategies in Matrix Games 53 5. Dominated Strategies and Nash Equilibria 54 6. The Minimax Theorem 59 7. Finding Nash Equilibria in Simple Games 64 8. A Note on Nash Equilibria in General 66 Chapter 5. An Introduction to Optimization and the Karush-Kuhn-Tucker Conditions 69 1. A General Maximization Formulation 70 2. Some Geometry for Optimization 72 3. -
Lecture Notes
GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies .................................. -
The Monty Hall Problem in the Game Theory Class
The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice. -
Evolutionary Stable Strategy Application of Nash Equilibrium in Biology
GENERAL ARTICLE Evolutionary Stable Strategy Application of Nash Equilibrium in Biology Jayanti Ray-Mukherjee and Shomen Mukherjee Every behaviourally responsive animal (including us) make decisions. These can be simple behavioural decisions such as where to feed, what to feed, how long to feed, decisions related to finding, choosing and competing for mates, or simply maintaining ones territory. All these are conflict situations between competing individuals, hence can be best understood Jayanti Ray-Mukherjee is using a game theory approach. Using some examples of clas- a faculty member in the School of Liberal Studies sical games, we show how evolutionary game theory can help at Azim Premji University, understand behavioural decisions of animals. Game theory Bengaluru. Jayanti is an (along with its cousin, optimality theory) continues to provide experimental ecologist who a strong conceptual and theoretical framework to ecologists studies mechanisms of species coexistence among for understanding the mechanisms by which species coexist. plants. Her research interests also inlcude plant Most of you, at some point, might have seen two cats fighting. It invasion ecology and is often accompanied with the cats facing each other with puffed habitat restoration. up fur, arched back, ears back, tail twitching, with snarls, growls, Shomen Mukherjee is a and howls aimed at each other. But, if you notice closely, they faculty member in the often try to avoid physical contact, and spend most of their time School of Liberal Studies in the above-mentioned behavioural displays. Biologists refer to at Azim Premji University, this as a ‘limited war’ or conventional (ritualistic) strategy (not Bengaluru. He uses field experiments to study causing serious injury), as opposed to dangerous (escalated) animal behaviour and strategy (Figure 1) [1]. -
Game Theory]: Basics of Game Theory
Artificial Intelligence Methods for Social Good M2-1 [Game Theory]: Basics of Game Theory 08-537 (9-unit) and 08-737 (12-unit) Instructor: Fei Fang [email protected] Wean Hall 4126 1 5/8/2018 Quiz 1: Recap: Optimization Problem Given coordinates of 푛 residential areas in a city (assuming 2-D plane), denoted as 푥1, … , 푥푛, the government wants to find a location that minimizes the sum of (Euclidean) distances to all residential areas to build a hospital. The optimization problem can be written as A: min |푥푖 − 푥| 푥 푖 B: min 푥푖 − 푥 푥 푖 2 2 C: min 푥푖 − 푥 푥 푖 D: none of above 2 Fei Fang 5/8/2018 From Games to Game Theory The study of mathematical models of conflict and cooperation between intelligent decision makers Used in economics, political science etc 3/72 Fei Fang 5/8/2018 Outline Basic Concepts in Games Basic Solution Concepts Compute Nash Equilibrium Compute Strong Stackelberg Equilibrium 4 Fei Fang 5/8/2018 Learning Objectives Understand the concept of Game, Player, Action, Strategy, Payoff, Expected utility, Best response Dominant Strategy, Maxmin Strategy, Minmax Strategy Nash Equilibrium Stackelberg Equilibrium, Strong Stackelberg Equilibrium Describe Minimax Theory Formulate the following problem as an optimization problem Find NE in zero-sum games (LP) Find SSE in two-player general-sum games (multiple LP and MILP) Know how to find the method/algorithm/solver/package you can use for solving the games Compute NE/SSE by hand or by calling a solver for small games 5 Fei Fang 5/8/2018 Let’s Play! Classical Games -
Incomplete Information I. Bayesian Games
Bayesian Games Debraj Ray, November 2006 Unlike the previous notes, the material here is perfectly standard and can be found in the usual textbooks: see, e.g., Fudenberg-Tirole. For the examples in these notes (except for the very last section), I draw heavily on Martin Osborne’s excellent recent text, An Introduction to Game Theory, Oxford University Press. Obviously, incomplete information games — in which one or more players are privy to infor- mation that others don’t have — has enormous applicability: credit markets / auctions / regulation of firms / insurance / bargaining /lemons / public goods provision / signaling / . the list goes on and on. 1. A Definition A Bayesian game consists of 1. A set of players N. 2. A set of states Ω, and a common prior µ on Ω. 3. For each player i a set of actions Ai and a set of signals or types Ti. (Can make actions sets depend on type realizations.) 4. For each player i, a mapping τi :Ω 7→ Ti. 5. For each player i, a vN-M payoff function fi : A × Ω 7→ R, where A is the product of the Ai’s. Remarks A. Typically, the mapping τi tells us player i’s type. The easiest case is just when Ω is the product of the Ti’s and τi is just the appropriate projection mapping. B. The prior on states need not be common, but one view is that there is no loss of generality in assuming it (simply put all differences in the prior into differences in signals and redefine the state space and si-mappings accordingly). -
Strong Stackelberg Reasoning in Symmetric Games: an Experimental
Strong Stackelberg reasoning in symmetric games: An experimental ANGOR UNIVERSITY replication and extension Pulford, B.D.; Colman, A.M.; Lawrence, C.L. PeerJ DOI: 10.7717/peerj.263 PRIFYSGOL BANGOR / B Published: 25/02/2014 Publisher's PDF, also known as Version of record Cyswllt i'r cyhoeddiad / Link to publication Dyfyniad o'r fersiwn a gyhoeddwyd / Citation for published version (APA): Pulford, B. D., Colman, A. M., & Lawrence, C. L. (2014). Strong Stackelberg reasoning in symmetric games: An experimental replication and extension. PeerJ, 263. https://doi.org/10.7717/peerj.263 Hawliau Cyffredinol / General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal ? Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. 23. Sep. 2021 Strong Stackelberg reasoning in symmetric games: An experimental replication and extension Briony D. Pulford1, Andrew M. Colman1 and Catherine L. Lawrence2 1 School of Psychology, University of Leicester, Leicester, UK 2 School of Psychology, Bangor University, Bangor, UK ABSTRACT In common interest games in which players are motivated to coordinate their strate- gies to achieve a jointly optimal outcome, orthodox game theory provides no general reason or justification for choosing the required strategies. -
Collusion Constrained Equilibrium
Theoretical Economics 13 (2018), 307–340 1555-7561/20180307 Collusion constrained equilibrium Rohan Dutta Department of Economics, McGill University David K. Levine Department of Economics, European University Institute and Department of Economics, Washington University in Saint Louis Salvatore Modica Department of Economics, Università di Palermo We study collusion within groups in noncooperative games. The primitives are the preferences of the players, their assignment to nonoverlapping groups, and the goals of the groups. Our notion of collusion is that a group coordinates the play of its members among different incentive compatible plans to best achieve its goals. Unfortunately, equilibria that meet this requirement need not exist. We instead introduce the weaker notion of collusion constrained equilibrium. This al- lows groups to put positive probability on alternatives that are suboptimal for the group in certain razor’s edge cases where the set of incentive compatible plans changes discontinuously. These collusion constrained equilibria exist and are a subset of the correlated equilibria of the underlying game. We examine four per- turbations of the underlying game. In each case,we show that equilibria in which groups choose the best alternative exist and that limits of these equilibria lead to collusion constrained equilibria. We also show that for a sufficiently broad class of perturbations, every collusion constrained equilibrium arises as such a limit. We give an application to a voter participation game that shows how collusion constraints may be socially costly. Keywords. Collusion, organization, group. JEL classification. C72, D70. 1. Introduction As the literature on collective action (for example, Olson 1965) emphasizes, groups often behave collusively while the preferences of individual group members limit the possi- Rohan Dutta: [email protected] David K. -
Economics 201B Economic Theory (Spring 2021) Strategic Games
Economics 201B Economic Theory (Spring 2021) Strategic Games Topics: terminology and notations (OR 1.7), games and solutions (OR 1.1-1.3), rationality and bounded rationality (OR 1.4-1.6), formalities (OR 2.1), best-response (OR 2.2), Nash equilibrium (OR 2.2), 2 2 examples × (OR 2.3), existence of Nash equilibrium (OR 2.4), mixed strategy Nash equilibrium (OR 3.1, 3.2), strictly competitive games (OR 2.5), evolution- ary stability (OR 3.4), rationalizability (OR 4.1), dominance (OR 4.2, 4.3), trembling hand perfection (OR 12.5). Terminology and notations (OR 1.7) Sets For R, ∈ ≥ ⇐⇒ ≥ for all . and ⇐⇒ ≥ for all and some . ⇐⇒ for all . Preferences is a binary relation on some set of alternatives R. % ⊆ From % we derive two other relations on : — strict performance relation and not  ⇐⇒ % % — indifference relation and ∼ ⇐⇒ % % Utility representation % is said to be — complete if , or . ∀ ∈ % % — transitive if , and then . ∀ ∈ % % % % can be presented by a utility function only if it is complete and transitive (rational). A function : R is a utility function representing if → % ∀ ∈ () () % ⇐⇒ ≥ % is said to be — continuous (preferences cannot jump...) if for any sequence of pairs () with ,and and , . { }∞=1 % → → % — (strictly) quasi-concave if for any the upper counter set ∈ { ∈ : is (strictly) convex. % } These guarantee the existence of continuous well-behaved utility function representation. Profiles Let be a the set of players. — () or simply () is a profile - a collection of values of some variable,∈ one for each player. — () or simply is the list of elements of the profile = ∈ { } − () for all players except . ∈ — ( ) is a list and an element ,whichistheprofile () . -
Interim Correlated Rationalizability
Interim Correlated Rationalizability The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Dekel, Eddie, Drew Fudenberg, and Stephen Morris. 2007. Interim correlated rationalizability. Theoretical Economics 2, no. 1: 15-40. Published Version http://econtheory.org/ojs/index.php/te/article/view/20070015 Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:3196333 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA Theoretical Economics 2 (2007), 15–40 1555-7561/20070015 Interim correlated rationalizability EDDIE DEKEL Department of Economics, Northwestern University, and School of Economics, Tel Aviv University DREW FUDENBERG Department of Economics, Harvard University STEPHEN MORRIS Department of Economics, Princeton University This paper proposes the solution concept of interim correlated rationalizability, and shows that all types that have the same hierarchies of beliefs have the same set of interim-correlated-rationalizable outcomes. This solution concept charac- terizes common certainty of rationality in the universal type space. KEYWORDS. Rationalizability, incomplete information, common certainty, com- mon knowledge, universal type space. JEL CLASSIFICATION. C70, C72. 1. INTRODUCTION Harsanyi (1967–68) proposes solving games of incomplete