Stochastic Games

Total Page:16

File Type:pdf, Size:1020Kb

Stochastic Games PERSPECTIVE PERSPECTIVE Stochastic games Eilon Solana,1 and Nicolas Vieilleb,1 aDepartment of Statistics and Operations Research, School of Mathematical Sciences, Tel Aviv University, Tel Aviv 6997800, Israel; and bDepartment of Economics and Decision Sciences, HEC Paris, 78 351 Jouy-en-Josas, France Edited by Jose A. Scheinkman, Columbia University, New York, NY, and approved September 22, 2015 (received for review July 9, 2015) In 1953, Lloyd Shapley contributed his paper “Stochastic games” to PNAS. In this paper, he defined the model of stochastic games, which were the first general dynamic model of a game to be defined, and proved that it admits a stationary equilibrium. In this Perspective, we summarize the historical context and the impact of Shapley’s contribution. game theory | stochastic games The 1950s were the decade in which game available actions. The collection of actions the first t stages, as well as the actions that the theory was shaped. After von Neumann and that the players choose, together with the players played in the first t − 1stages.A Morgenstern’s Theory of Games and Economic current state, determine the stage payoff that strategy of a player is a prescription how Behavior (1) was published in 1944, a group of each player receives, as well as a probability to play the game; that is, a function that assigns young and bright researchers started working distribution according to which the new state to every finite history an action to play should on game theory, each of whom published pa- that the play will visit. that history occur. A behavior strategy of a pers that opened new areas. In 1950, John Stochastic games extend the model of stra- player is a function that assigns to every finite Nash published two papers, one on the con- tegic-form games, which is attributable to von history a lottery over the set of available actions. cept of Nash equilibrium (2) and the other on Neumann (8), to dynamic situations in which A zero-sum game is defined to have a the bargaining problem (3). In 1953, Lloyd the environment changes in response to the “value” v if (i) player 1 has a strategy (which Shapley published two papers, one on stochas- players’ choices. They also extend the model is then said to be “optimal”), which ensures tic games (4) and the other on the Shapley of Markov decision problems, which was de- that his expected overall payoff over time value for coalitional games (5). In 1957, veloped by several researchers at the RAND does not fall below v,nomatterwhatisthe Harold Kuhn (6) provided a formulation Corporation in 1949–1952, to competitive sit- strategy followed by player 2, and (ii)ifthe for extensive-form games, and Duncan Luce uations with more than one decision maker. symmetric property holds when exchanging and Howard Raiffa published their book The complexity of stochastic games stems Games and Decisions: Introduction and Critical the roles of the two players. Shapley proved fromthefactthatthechoicesmadebythe the existence of a value. Because the param- Survey (7). These researchers and others, like players have two, sometimes contradictory, eters that define the game are independent Kenneth Arrow, Richard Bellman, and David effects. First, together with the current state, of time, the situation that the players face if Blackwell, interacted, worked and played to- ’ the players actions determine the immediate today the play is in a certain state is the same gether, exchanged ideas, and developed the payoff that each player receives. Second, the situation they would face tomorrow if to- theory of dynamic games as we know it today. current state and the players’ actions influence While in graduate school at Princeton, morrow the play is in that state. In particular, the choice of the new state, which determines one expects to have optimal strategies that Shapley befriended Martin Shubik and John the potential of future payoffs. In particular, Nash. The former became a well-known are stationary Markov, that is, they depend when choosing his actions, each player has to only on the current state of the game. Shapley economist, and the latter defined the concept balance these forces, a decision that may often of Nash equilibrium, the most prevalent proved that indeed such optimal strategies be difficult. Although this dichotomy is also solution concept in game theory and eco- exist, and characterized the value as the present in one-player sequential decision nomictheorytodate.In2013,Shapleywas unique fixed point of a nonlinear functional problems, the presence of additional players awarded the Nobel Prize in economics, 8 y operator—a two-player version of the dynamic who maximize their own goals adds com- after Nash was awarded the prize. programming principle. plexity to the analysis of the situation. The existence of stationary Markov opti- ’ 1. Stochastic Games Shapley s paper in PNAS introduced the mal strategies implies that, to play well, a Stochastic games model dynamic interactions model of stochastic games with positive player needs to know only the current state. in which the environment changes in response stopping probabilities, assuming a finite set In particular, the value of the game does not ’ ’ “ to players behavior. In Shapley swords, In a of states, and a finite set of possible actions change if players receive partial information stochastic game the play proceeds by steps for each player in each state. In this model, ’ from position to position, according to tran- thecurrentstateandtheplayers actions Author contributions: E.S. and N.V. wrote the paper. sition probabilities controlled jointly by the determine the probability that the game will The authors declare no conflict of interest. two players” (4). terminate once the current stage is over. His This article is a PNAS Direct Submission. A stochastic game is played by a set of paper deals with two-player zero-sum games, This article is part of the special series of PNAS 100th Anniversary players. In each stage of the game, the play is so that the gain of player 1 is always equal to articles to commemorate exceptional research published in PNAS ’ over the last century. See the companion article, “Stochastic in a given state (or position, in Shapley slan- the loss of player 2. ” t Games on page 1095 in issue 10 of volume 39. guage), taken from a set of states, and every A history of length in a stochastic game is 1To whom correspondence may be addressed. Email: eilons@post. player chooses an action from a set of the sequence of states that the game visited in tau.ac.il or [email protected]. www.pnas.org/cgi/doi/10.1073/pnas.1513508112 PNAS | November 10, 2015 | vol. 112 | no. 45 | 13743–13746 Downloaded by guest on October 1, 2021 on each other’s actions, and/or if they forget initial wealth, repeatedly play a zero-sum game on algebraic properties of the value of the previously visited states. that affects their wealth, until one player is discounted game, as a function of the dis- Inthefollowingparagraphs,webriefly ruined and declared loser. Everett (18) studied count factor, proven by Bewley and Kohlberg describe some of the directions in which the the so-called “recursive games,” in which ter- (22). One can view the monetary policy of theory has developed following Shapley’s mination probabilities are endogenous, and a the central bank, that of changing the interest groundbreaking paper. payoff is only received upon termination of the rate from time to time, as an implementation game, as a function of the terminating state. of this type of strategy. 2. Discounted Stochastic Games Gillette (19) introduced a deceptively simple- The theory of the non–zero-sum case has ’ Shapley s model is equivalent to one in which looking game, inspired by playground prac- flourished in the last two decades but has not players discount their future payoffs accord- tice, whose analysis resisted all attempts until been settled yet. A collection of strategies, one ing to a discount factor that depends on the Blackwell and Ferguson (20) provided its for each player, is a uniform «-equilibrium if current state and on the players’ actions. The solution. bydeviatingnoplayercanprofitmorethan«, game is called “discounted” if all stopping In a different but related framework, provided the game is sufficiently long or the probabilities equal the same constant, and one Aumann, Maschler, and Stearns (ref. 21, discount factor sufficiently close to 1, and if minusthisconstantiscalledthe“discount reprinted in 1995; see pp. 130–139) conceptu- the payoff vector is almost independent of the factor.” Models of discounted stochastic games alized a notion of a uniform solution to duration of the game. Existence of uniform are prevalent in economics, where the discount infinite-horizon games, in which payoffs are «-equilibria has been established in few cases: factor has a clear economic interpretation. received in each round. This approach views Vieille (25, 26) for two-player games, Solan A large literature followed, relaxing both the the infinite-horizon game as an idealized (27) for three-player games in which the state zero-sum and the finiteness assumptions of model for long-run interactions, in which the of the game changes at most once, and Solan Shapley’spaper.Innon–zero-sum games, a duration of the game is sufficiently long and/or and Vieille (28), Flesch et al. (29), and Simon collection of strategies, one for each player, is a the discount factor is sufficiently close to 1, and (30) in other classes of stochastic games. The “(Nash) equilibrium” if no player can profit by both are not perfectly known. It aims at pro- question in its most generality remains open. deviating from his strategy, assuming all other viding solutions whose optimality properties A weaker concept than Nash equilibrium players follow their prescribed strategies.
Recommended publications
  • Lecture Notes
    GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies ..................................
    [Show full text]
  • Collusion Constrained Equilibrium
    Theoretical Economics 13 (2018), 307–340 1555-7561/20180307 Collusion constrained equilibrium Rohan Dutta Department of Economics, McGill University David K. Levine Department of Economics, European University Institute and Department of Economics, Washington University in Saint Louis Salvatore Modica Department of Economics, Università di Palermo We study collusion within groups in noncooperative games. The primitives are the preferences of the players, their assignment to nonoverlapping groups, and the goals of the groups. Our notion of collusion is that a group coordinates the play of its members among different incentive compatible plans to best achieve its goals. Unfortunately, equilibria that meet this requirement need not exist. We instead introduce the weaker notion of collusion constrained equilibrium. This al- lows groups to put positive probability on alternatives that are suboptimal for the group in certain razor’s edge cases where the set of incentive compatible plans changes discontinuously. These collusion constrained equilibria exist and are a subset of the correlated equilibria of the underlying game. We examine four per- turbations of the underlying game. In each case,we show that equilibria in which groups choose the best alternative exist and that limits of these equilibria lead to collusion constrained equilibria. We also show that for a sufficiently broad class of perturbations, every collusion constrained equilibrium arises as such a limit. We give an application to a voter participation game that shows how collusion constraints may be socially costly. Keywords. Collusion, organization, group. JEL classification. C72, D70. 1. Introduction As the literature on collective action (for example, Olson 1965) emphasizes, groups often behave collusively while the preferences of individual group members limit the possi- Rohan Dutta: [email protected] David K.
    [Show full text]
  • Correlated Equilibria and Communication in Games Françoise Forges
    Correlated equilibria and communication in games Françoise Forges To cite this version: Françoise Forges. Correlated equilibria and communication in games. Computational Complex- ity. Theory, Techniques, and Applications, pp.295-704, 2012, 10.1007/978-1-4614-1800-9_45. hal- 01519895 HAL Id: hal-01519895 https://hal.archives-ouvertes.fr/hal-01519895 Submitted on 9 May 2017 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Correlated Equilibrium and Communication in Games Françoise Forges, CEREMADE, Université Paris-Dauphine Article Outline Glossary I. De…nition of the Subject and its Importance II. Introduction III. Correlated Equilibrium: De…nition and Basic Properties IV. Correlated Equilibrium and Communication V. Correlated Equilibrium in Bayesian Games VI. Related Topics and Future Directions VII. Bibliography Acknowledgements The author wishes to thank Elchanan Ben-Porath, Frédéric Koessler, R. Vijay Krishna, Ehud Lehrer, Bob Nau, Indra Ray, Jérôme Renault, Eilon Solan, Sylvain Sorin, Bernhard von Stengel, Tristan Tomala, Amparo Ur- bano, Yannick Viossat and, especially, Olivier Gossner and Péter Vida, for useful comments and suggestions. Glossary Bayesian game: an interactive decision problem consisting of a set of n players, a set of types for every player, a probability distribution which ac- counts for the players’ beliefs over each others’ types, a set of actions for every player and a von Neumann-Morgenstern utility function de…ned over n-tuples of types and actions for every player.
    [Show full text]
  • Characterizing Solution Concepts in Terms of Common Knowledge Of
    Characterizing Solution Concepts in Terms of Common Knowledge of Rationality Joseph Y. Halpern∗ Computer Science Department Cornell University, U.S.A. e-mail: [email protected] Yoram Moses† Department of Electrical Engineering Technion—Israel Institute of Technology 32000 Haifa, Israel email: [email protected] March 15, 2018 Abstract Characterizations of Nash equilibrium, correlated equilibrium, and rationaliz- ability in terms of common knowledge of rationality are well known (Aumann 1987; arXiv:1605.01236v1 [cs.GT] 4 May 2016 Brandenburger and Dekel 1987). Analogous characterizations of sequential equi- librium, (trembling hand) perfect equilibrium, and quasi-perfect equilibrium in n-player games are obtained here, using results of Halpern (2009, 2013). ∗Supported in part by NSF under grants CTC-0208535, ITR-0325453, and IIS-0534064, by ONR un- der grant N00014-02-1-0455, by the DoD Multidisciplinary University Research Initiative (MURI) pro- gram administered by the ONR under grants N00014-01-1-0795 and N00014-04-1-0725, and by AFOSR under grants F49620-02-1-0101 and FA9550-05-1-0055. †The Israel Pollak academic chair at the Technion; work supported in part by Israel Science Foun- dation under grant 1520/11. 1 Introduction Arguably, the major goal of epistemic game theory is to characterize solution concepts epistemically. Characterizations of the solution concepts that are most commonly used in strategic-form games, namely, Nash equilibrium, correlated equilibrium, and rational- izability, in terms of common knowledge of rationality are well known (Aumann 1987; Brandenburger and Dekel 1987). We show how to get analogous characterizations of sequential equilibrium (Kreps and Wilson 1982), (trembling hand) perfect equilibrium (Selten 1975), and quasi-perfect equilibrium (van Damme 1984) for arbitrary n-player games, using results of Halpern (2009, 2013).
    [Show full text]
  • Correlated Equilibria 1 the Chicken-Dare Game 2 Correlated
    MS&E 334: Computation of Equilibria Lecture 6 - 05/12/2009 Correlated Equilibria Lecturer: Amin Saberi Scribe: Alex Shkolnik 1 The Chicken-Dare Game The chicken-dare game can be throught of as two drivers racing towards an intersection. A player can chose to dare (d) and pass through the intersection or chicken out (c) and stop. The game results in a draw when both players chicken out and the worst possible outcome if they both dare. A player wins when he dares while the other chickens out. The game has one possible payoff matrix given by d c d 0; 0 4; 1 c 1; 4 3; 3 with two pure strategy Nash equilibria (d; c) and (c; d) and one mixed equilibrium where each player mixes the pure strategies with probability 1=2 each. Now suppose that prior to playing the game the players performed the following experiment. The players draw a ball labeled with a strategy, either (c) or (d) from a bag containing three balls labelled c; c; d. The players then agree to follow the strategy suggested by the ball. It can be verified that there is no incentive to deviate from such an agreement since the suggested strategy is best in expectation. This experiment is equivalent to having the following strategy profile chosen for the players by some third party, a correlation device. d c d 0 1=3 c 1=3 1=3 This matrix above is not of rank one and so is not a Nash profile. And, the social welfare in this scenario is 16=3 which is greater than that of any Nash equilibrium.
    [Show full text]
  • Correlated Equilibrium Is a Distribution D Over Action Profiles a Such That for Every ∗ Player I, and Every Action Ai
    NETS 412: Algorithmic Game Theory February 9, 2017 Lecture 8 Lecturer: Aaron Roth Scribe: Aaron Roth Correlated Equilibria Consider the following two player traffic light game that will be familiar to those of you who can drive: STOP GO STOP (0,0) (0,1) GO (1,0) (-100,-100) This game has two pure strategy Nash equilibria: (GO,STOP), and (STOP,GO) { but these are clearly not ideal because there is one player who never gets any utility. There is also a mixed strategy Nash equilibrium: Suppose player 1 plays (p; 1 − p). If the equilibrium is to be fully mixed, player 2 must be indifferent between his two actions { i.e.: 0 = p − 100(1 − p) , 101p = 100 , p = 100=101 So in the mixed strategy Nash equilibrium, both players play STOP with probability p = 100=101, and play GO with probability (1 − p) = 1=101. This is even worse! Now both players get payoff 0 in expectation (rather than just one of them), and risk a horrific negative utility. The four possible action profiles have roughly the following probabilities under this equilibrium: STOP GO STOP 98% <1% GO <1% ≈ 0.01% A far better outcome would be the following, which is fair, has social welfare 1, and doesn't risk death: STOP GO STOP 0% 50% GO 50% 0% But there is a problem: there is no set of mixed strategies that creates this distribution over action profiles. Therefore, fundamentally, this can never result from Nash equilibrium play. The reason however is not that this play is not rational { it is! The issue is that we have defined Nash equilibria as profiles of mixed strategies, that require that players randomize independently, without any communication.
    [Show full text]
  • Maximum Entropy Correlated Equilibria
    Maximum Entropy Correlated Equilibria Luis E. Ortiz Robert E. Schapire Sham M. Kakade ECE Deptartment Dept. of Computer Science Toyota Technological Institute Univ. of Puerto Rico at Mayaguez¨ Princeton University Chicago, IL 60637 Mayaguez,¨ PR 00681 Princeton, NJ 08540 [email protected] [email protected] [email protected] Abstract cept in game theory and in mathematical economics, where game theory is applied to model competitive economies We study maximum entropy correlated equilib- [Arrow and Debreu, 1954]. An equilibrium is a point of ria (Maxent CE) in multi-player games. After strategic stance between the players in a game that charac- motivating and deriving some interesting impor- terizes and serves as a prediction to a possible outcome of tant properties of Maxent CE, we provide two the game. gradient-based algorithms that are guaranteed to The general problem of equilibrium computation is funda- converge to it. The proposed algorithms have mental in computer science [Papadimitriou, 2001]. In the strong connections to algorithms for statistical last decade, there have been many great advances in our estimation (e.g., iterative scaling), and permit a understanding of the general prospects and limitations of distributed learning-dynamics interpretation. We computing equilibria in games. Just late last year, there also briefly discuss possible connections of this was a breakthrough on what was considered a major open work, and more generally of the Maximum En- problem in the field: computing equilibria was shown tropy Principle in statistics, to the work on learn- to be hard, even for two-player games [Chen and Deng, ing in games and the problem of equilibrium se- 2005b,a, Daskalakis and Papadimitriou, 2005, Daskalakis lection.
    [Show full text]
  • Online Learning, Regret Minimization, Minimax Optimality, and Correlated
    High level Last time we discussed notion of Nash equilibrium. Static concept: set of prob. Distributions (p,q,…) such that Online Learning, Regret Minimization, nobody has any incentive to deviate. Minimax Optimality, and Correlated But doesn’t talk about how system would get there. Troubling that even finding one can be hard in large games. Equilibrium What if agents adapt (learn) in ways that are well-motivated in terms of their own rewards? What can we say about the system then? Avrim Blum High level Consider the following setting… Today: Each morning, you need to pick Online learning guarantees that are achievable when acting one of N possible routes to drive in a changing and unpredictable environment. to work. Robots R Us What happens when two players in a zero-sum game both But traffic is different each day. 32 min use such strategies? Not clear a priori which will be best. When you get there you find out how Approach minimax optimality. long your route took. (And maybe Gives alternative proof of minimax theorem. others too or maybe not.) What happens in a general-sum game? Is there a strategy for picking routes so that in the Approaches (the set of) correlated equilibria. long run, whatever the sequence of traffic patterns has been, you’ve done nearly as well as the best fixed route in hindsight? (In expectation, over internal randomness in the algorithm) Yes. “No-regret” algorithms for repeated decisions “No-regret” algorithms for repeated decisions A bit more generally: Algorithm has N options. World chooses cost vector.
    [Show full text]
  • Correlated Equilibrium
    Announcements Ø Minbiao’s office hour will be changed to Thursday 1-2 pm, starting from next week, at Rice Hall 442 1 CS6501: Topics in Learning and Game Theory (Fall 2019) Introduction to Game Theory (II) Instructor: Haifeng Xu Outline Ø Correlated and Coarse Correlated Equilibrium Ø Zero-Sum Games Ø GANs and Equilibrium Analysis 3 Recap: Normal-Form Games Ø � players, denoted by set � = {1, ⋯ , �} Ø Player � takes action �* ∈ �* Ø An outcome is the action profile � = (�., ⋯ , �/) • As a convention, �1* = (�., ⋯ , �*1., �*2., ⋯ , �/) denotes all actions excluding �* / ØPlayer � receives payoff �*(�) for any outcome � ∈ Π*5.�* • �* � = �*(�*, �1*) depends on other players’ actions Ø �* , �* *∈[/] are public knowledge ∗ ∗ ∗ A mixed strategy profile � = (�., ⋯ , �/) is a Nash equilibrium ∗ ∗ (NE) if for any �, �* is a best response to �1*. 4 NE Is Not the Only Solution Concept ØNE rests on two key assumptions 1. Players move simultaneously (so they cannot see others’ strategies before the move) 2. Players take actions independently ØLast lecture: sequential move results in different player behaviors • The corresponding game is called Stackelberg game and its equilibrium is called Strong Stackelberg equilibrium Today: we study what happens if players do not take actions independently but instead are “coordinated” by a central mediator Ø This results in the study of correlated equilibrium 5 An Illustrative Example B STOP GO A STOP (-3, -2) (-3, 0) GO (0, -2) (-100, -100) The Traffic Light Game Well, we did not see many crushes in reality…
    [Show full text]
  • Rationality and Common Knowledge Herbert Gintis
    Rationality and Common Knowledge Herbert Gintis May 30, 2010 1 Introduction Interactive epistemology is the study of the distributionof knowledge among ratio- nal agents, using modal logicin the traditionof Hintikka(1962) and Kripke (1963), and agent rationality based on the rational actor model of economic theory, in the tradition of Von Neumann and Morgenstern (1944) and Savage (1954). Epistemic game theory, which is interactive epistemology adjoined to classical game theory (Aumann, 1987, 1995), has demonstrated an intimate relationship between ratio- nality and correlated equilibrium (Aumann 1987, Brandenburger and Dekel 1987), and has permitted the rigorous specification of the conditions under which rational agents will play a Nash equilibrium (Aumann and Brandenburger 1995). A central finding in this research is that rational agents use the strategies sug- gested by game-theoretic equilibrium concepts when there is a commonality of knowledge in the form of common probability distributions over the stochastic variables that arise in the play of the game (so-called common priors), and com- mon knowledge of key aspects of the strategic interaction. We say an event E is common knowledge for agents i D 1;:::;n if the each agent i knows E, each i knows that each agent j knows E, each agent k knows that each j knows that each i knows E, and so on (Lewis 1969, Aumann 1976). Specifying when individuals share knowledge and when they know that they share knowledge is among the most challenging of epistemological issues. Con- temporary psychological research has shown that these issues cannot be resolved by analyzing the general features of high-level cognitive functioning alone, but in fact concern the particular organization of the human brain.
    [Show full text]
  • Correlated Equilibrium in a Nutshell
    The University of Manchester Research Correlated equilibrium in a nutshell DOI: 10.1007/s11238-017-9609-9 Document Version Accepted author manuscript Link to publication record in Manchester Research Explorer Citation for published version (APA): Amir, R., Belkov, S., & Evstigneev, I. V. (2017). Correlated equilibrium in a nutshell. Theory and Decision, 83(4), 1- 12. https://doi.org/10.1007/s11238-017-9609-9 Published in: Theory and Decision Citing this paper Please note that where the full-text provided on Manchester Research Explorer is the Author Accepted Manuscript or Proof version this may differ from the final Published version. If citing, it is advised that you check and use the publisher's definitive version. General rights Copyright and moral rights for the publications made accessible in the Research Explorer are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. Takedown policy If you believe that this document breaches copyright please refer to the University of Manchester’s Takedown Procedures [http://man.ac.uk/04Y6Bo] or contact [email protected] providing relevant details, so we can investigate your claim. Download date:25. Sep. 2021 Noname manuscript No. (will be inserted by the editor) Correlated Equilibrium in a Nutshell Rabah Amir · Sergei Belkov · Igor V. Evstigneev Received: date / Accepted: date Abstract We analyze the concept of correlated equilibrium in the framework of two-player two-strategy games. This simple framework makes it possible to clearly demonstrate the characteristic features of this concept.
    [Show full text]
  • Yale Lecture on Correlated Equilibrium and Incomplete Information
    Yale lecture on Correlated Equilibrium and Incomplete Information Stephen Morris March 2012 Introduction Game Theoretic Predictions are very sensitive to beliefs/expectations and higher order beliefs/expectations or (equivalently) information structure (Higher order) beliefs/expectations are rarely observed What predictions can we make and analysis can we do if we do not observe (higher order) beliefs/expectations? Robust Predictions Agenda: Basic Question Fix "payo¤ relevant environment" = action sets, payo¤-relevant variables ("states"), payo¤ functions, distribution over states = incomplete information game without higher order beliefs about states Assume payo¤ relevant environment is observed by the analyst/econometrician Analyze what could happen for all possible higher order beliefs (maintaining common prior assumption and equilibrium assumptions) Make set valued predictions about joint distribution of actions and states Robust Predictions Agenda: Minimal Information More generally, the analyst might know something about players’higher order beliefs but cannot rule out the possibility that they know more What predictions can he then make? Narrower set of predictions (more incentive constraints) More knowledge of agents’information thus allows tighter identi…cation of underlying game Partial Identi…cation (not this lecture) Incomplete Information Correlated Equilibrium Robust predictions = set of outcomes consistent with equilibrium given any additional information the players may observe set of outcomes that could arise if a mediator who knew the payo¤ state could privately make action recommendations set of incomplete information correlated equilibrium outcomes very permissive de…nition of incomplete information correlated equilibrium: we dub it "Bayes correlated equilibrium" This Talk 1 De…nition of Bayes correlated equilibrium 2 "Epistemic result": Bayes correlated equilibria equal set of distributions that could arise if players observed additional information 3 One player, two action, two state example (c.f.
    [Show full text]