Interim Correlated Rationalizability in Infinite Games

Total Page:16

File Type:pdf, Size:1020Kb

Interim Correlated Rationalizability in Infinite Games INTERIM CORRELATED RATIONALIZABILITY IN INFINITE GAMES JONATHAN WEINSTEIN AND MUHAMET YILDIZ Abstract. In a Bayesian game, assume that the type space is a complete, separable metric space, the action space is a compact metric space, and the payoff functions are continuous. We show that the iterative and fixed-point definitions of interim correlated rationalizability (ICR) coincide, and ICR is non-empty-valued and upper hemicontinuous. This extends the finite-game results of Dekel, Fudenberg and Morris (2007), who introduced ICR. Our result applies, for instance, to discounted infinite-horizon dynamic games. JEL Numbers: C72, C73. 1. Introduction Interim correlated rationalizability (henceforth ICR) has emerged as the main notion of rationalizability for Bayesian games. Among other reasons, it has the following two desirable properties: Firstly, it is upper hemicontinuous in types, i.e., one cannot obtain a substantially new ICR solution by perturbing a type. Secondly, two distinct definitions of ICR coincide: The fixed-point definition states that ICR is the weakest solution concept with a specific best-response property, and this is the definition that gives ICR its epistemological meaning, as a characterization of actions that are consistent with common knowledge of rationality. Under an alternate definition, ICR is computed by iteratively eliminating the actions that are never a best response (type by type), and this iterative definition is often more amenable for analysis. The above properties were originally proven for games with finite action spaces and finite sets of payoff parameters (by Dekel, Fudenberg, and Morris (2006, 2007), who introduced ICR). However, in many important games these sets are quite large. For example, in the infinitely repeated prisoners’dilemma game, the set of outcomes is uncountable. Hence, the In earlier versions of the paper, the results were confined to compactly metrizable type spaces. We thank an anonymous referee for detailed comments and for providing the main arguments that extended our proofs to the type spaces that are completely metrizable and separable. 1 2 JONATHANWEINSTEINANDMUHAMETYILDIZ set of all possible payoff functions (with payoffs in [0, 1]) is the unit cube with uncountably many dimensions. Therefore, to analyze the infinitely repeated prisoners’ dilemma game under payoff uncertainty without restricting the payoffs, we would need to consider a space of underlying payoff parameters larger than the continuum. The action space is also un- countably infinite, because the action space consists of all mappings that specify whether one cooperates or defects at each of infinitely many histories. In this note, we establish the aforementioned two properties of ICR in greater general- ity so that it can be applied to commonly-used games with large action and type spaces. Specifically, we establish that ICR is upper hemicontinuous and its iterative and fixed-point definitions coincide in games where (i) the spaces of payoff parameters and types are com- plete, separable metric spaces, (ii) the space of actions is a compact metric space, and (iii) the belief and payoff functions are continuous. An immediate consequence of our result is that ICR is upper hemicontinuous when types are endowed with the product topology of belief hierarchies. That is, one cannot obtain a substantially distinct ICR solution for a type by perturbing the type’s higher-order beliefs (possibly by considering another type space). In that sense, ICR predictions are robust to higher-order uncertainty. Interestingly, ICR is unique in that regard: any strict refine- ment of ICR, such as Bayesian Nash equilibrium, fails to be upper hemicontinuous as a correspondence from belief hierarchies (Weinstein-Yildiz, 2007). When the action space is a compact metric space, the set of all continuous payoff func- tions is a complete separable metric space under the uniform metric. Then, the universal type space is also complete and separable metric space. Therefore, we obtain an upper- hemicontinuity result for ICR on universal type space under the same conditions for individ- ual best response: the action space is a complete separable metric space and it is common knowledge that the utility functions are continuous. In Section 6, we apply our results to infinite-horizon games with finitely many moves at every stage, where the relevant notion of continuity is continuity at infinity. As examples, we show that ICR is upper hemicontinuous in discounted repeated games with uniformly bounded stage-game payoffs, even if discount factors are unknown and stage payoffs are unknown, history-dependent or even stochastic. We applied our results here to general infinite-horizon games, including repeated games, in Weinstein and Yildiz (2013). INTERIMCORRELATEDRATIONALIZABILITY 3 Our note relates to the existing literature as follows. The usual solution concepts, includ- ing Nash equilibrium and rationalizability under complete information (Bernheim (1984) and Pearce (1984)), are upper hemicontinuous with respect to the parameters of the game. However, as we mentioned above, most incomplete-information versions of these concepts are not upper hemicontinuous with respect to higher-order beliefs, and we are not aware of any positive result on this matter other than the result of Dekel, Fudenberg and Morris that we extend here. On the other hand, there is a sizeable literature on the equivalence of the iterative and fixed-point definitions of rationalizability under complete information. The equivalence holds when the action spaces are compact metric spaces and the payoff functions are continuous (Tan and Werlang, 1988), and we extend this result to Bayesian games (using ICR as the notion of rationalizability).1 Under complete information, Lipman (1994) shows that the equivalence may fail in games with countably infinite, discrete action spaces, because reaching the fixed point may require iterations through a transfinite ordinal; Arieli (2010) discusses the same issue in continuous games, in which the payoff functions are continuous and the strategy spaces are complete separable metric spaces, finding that the iterations may need to reach the first uncountable ordinal (but not further). These results show that our continuity and compactness assumptions are not superfluous for the equivalence result.2 2. Basic Definitions Given any topological space X, we will write (X) for the space of Borel probability measures on X and endow it with the weak topology. We will also endow all product spaces with their product topology. We write f : X Y to mean that f is a correspondence from X to Y , i.e. an arbitrary subset of X Y with the convention that f(x) denotes y :(x, y) f . f 2 g We consider the defining subset of X Y to be the graph of the correspondence and denote this G(f) for clarity. 1The equivalence was also discussed, for a more narrow class of games, in Bernheim (1984), and in Dufwenberg-Stegeman (2002). This extension to Bayesian games was also covered by Ely-Peski (2006), using interim independent rationalizability as the solution concept. 2Also related is Chen-Long-Luo (2007) who form an iterative definition which is independent of the order of elimination, by allowing dominance by previously eliminated strategies; standard concepts may be order- dependent without the presence of regularity assumptions, as noted by Dufwenberg-Stegeman (2002). 4 JONATHANWEINSTEINANDMUHAMETYILDIZ The correspondence f is said to have the closed-graph property if G (f) is closed. In general, the closed-graph property is weaker than upperhemicontinuity, but the two concepts coincide when Y is compact. When Y is compact, a correspondence f : X Y is said to be upper hemicontinuous (UHC) if f has the closed-graph property. Consider a Bayesian game (N, A, u, , T, ) in normal form where N is a set of players, • 3 A = i N Ai is a compact metric space of action profiles, • 2 is a complete, separable metric space of payoff parameters, • ui : A [0, 1] is a continuous payoff function for each player i N, • ! 2 T = i N Ti is a complete, separable metric space of type profiles, and • 2 ti ( T i) is the interim belief of type ti on the payoff functions and the other • 2 players’types. The map ti t is assumed to be continuous. 7! i In the definition of a Bayesian game above, we made some topological assumptions that we will maintain throughout. These assumptions are satisfied in many applications, including discounted dynamic games, as discussed in Section 6. For each player i, define the best-response correspondence BRi : ( A i) Ai by setting BRi () = arg max ui (, ai, a i) d (, a i) ai Ai 2 Z for each ( A i), where arg max is the set of maximizers. The best-response 2 correspondence is extended to ( T i A i) by BRi () = BRi marg A i , where margX takes the marginal with respect to X. Since Ai is compact and ui is continuous and bounded, BRi is always non-empty. Moreover: Fact 1. BRi is UHC. This fact is a version of Berge’s Maximum Theorem suitable for our purposes; see the appendix for a straightforward proof. 3 For any family (Xi)i N of sets and any family of functions fi : Xi Yi, i N, we write X = i N Xi 2 ! 2 2 and X i = j N i Xj and define functions f : X Y and f i : X i Y i by setting f i (x i) = 2 nf g ! ! (fj (xj))j N i . 2 nf g INTERIMCORRELATEDRATIONALIZABILITY 5 We next present the two definitions of interim correlated rationalizability (ICR). The definitions may differ in general infinite games, but will coincide under the present conditions. The first definition is given by iterated elimination of strictly dominated actions that are never a weak best response, as follows: m Define the family of correspondences Si : Ti Ai, i N, m N, iteratively, by setting 0 2 m 2 S (ti) = Ai and for each m > 0 and ti Ti, let ai S (ti) if and only if there exists i 2 2 i ( T i A i) such that 2 (2.1) ai BRi () , 2 (2.2) ti = marg T i µ, m 1 (2.3) G S i = 1.
Recommended publications
  • Repeated Games
    6.254 : Game Theory with Engineering Applications Lecture 15: Repeated Games Asu Ozdaglar MIT April 1, 2010 1 Game Theory: Lecture 15 Introduction Outline Repeated Games (perfect monitoring) The problem of cooperation Finitely-repeated prisoner's dilemma Infinitely-repeated games and cooperation Folk Theorems Reference: Fudenberg and Tirole, Section 5.1. 2 Game Theory: Lecture 15 Introduction Prisoners' Dilemma How to sustain cooperation in the society? Recall the prisoners' dilemma, which is the canonical game for understanding incentives for defecting instead of cooperating. Cooperate Defect Cooperate 1, 1 −1, 2 Defect 2, −1 0, 0 Recall that the strategy profile (D, D) is the unique NE. In fact, D strictly dominates C and thus (D, D) is the dominant equilibrium. In society, we have many situations of this form, but we often observe some amount of cooperation. Why? 3 Game Theory: Lecture 15 Introduction Repeated Games In many strategic situations, players interact repeatedly over time. Perhaps repetition of the same game might foster cooperation. By repeated games, we refer to a situation in which the same stage game (strategic form game) is played at each date for some duration of T periods. Such games are also sometimes called \supergames". We will assume that overall payoff is the sum of discounted payoffs at each stage. Future payoffs are discounted and are thus less valuable (e.g., money and the future is less valuable than money now because of positive interest rates; consumption in the future is less valuable than consumption now because of time preference). We will see in this lecture how repeated play of the same strategic game introduces new (desirable) equilibria by allowing players to condition their actions on the way their opponents played in the previous periods.
    [Show full text]
  • Lecture 4 Rationalizability & Nash Equilibrium Road
    Lecture 4 Rationalizability & Nash Equilibrium 14.12 Game Theory Muhamet Yildiz Road Map 1. Strategies – completed 2. Quiz 3. Dominance 4. Dominant-strategy equilibrium 5. Rationalizability 6. Nash Equilibrium 1 Strategy A strategy of a player is a complete contingent-plan, determining which action he will take at each information set he is to move (including the information sets that will not be reached according to this strategy). Matching pennies with perfect information 2’s Strategies: HH = Head if 1 plays Head, 1 Head if 1 plays Tail; HT = Head if 1 plays Head, Head Tail Tail if 1 plays Tail; 2 TH = Tail if 1 plays Head, 2 Head if 1 plays Tail; head tail head tail TT = Tail if 1 plays Head, Tail if 1 plays Tail. (-1,1) (1,-1) (1,-1) (-1,1) 2 Matching pennies with perfect information 2 1 HH HT TH TT Head Tail Matching pennies with Imperfect information 1 2 1 Head Tail Head Tail 2 Head (-1,1) (1,-1) head tail head tail Tail (1,-1) (-1,1) (-1,1) (1,-1) (1,-1) (-1,1) 3 A game with nature Left (5, 0) 1 Head 1/2 Right (2, 2) Nature (3, 3) 1/2 Left Tail 2 Right (0, -5) Mixed Strategy Definition: A mixed strategy of a player is a probability distribution over the set of his strategies. Pure strategies: Si = {si1,si2,…,sik} σ → A mixed strategy: i: S [0,1] s.t. σ σ σ i(si1) + i(si2) + … + i(sik) = 1. If the other players play s-i =(s1,…, si-1,si+1,…,sn), then σ the expected utility of playing i is σ σ σ i(si1)ui(si1,s-i) + i(si2)ui(si2,s-i) + … + i(sik)ui(sik,s-i).
    [Show full text]
  • Learning and Equilibrium
    Learning and Equilibrium Drew Fudenberg1 and David K. Levine2 1Department of Economics, Harvard University, Cambridge, Massachusetts; email: [email protected] 2Department of Economics, Washington University of St. Louis, St. Louis, Missouri; email: [email protected] Annu. Rev. Econ. 2009. 1:385–419 Key Words First published online as a Review in Advance on nonequilibrium dynamics, bounded rationality, Nash equilibrium, June 11, 2009 self-confirming equilibrium The Annual Review of Economics is online at by 140.247.212.190 on 09/04/09. For personal use only. econ.annualreviews.org Abstract This article’s doi: The theory of learning in games explores how, which, and what 10.1146/annurev.economics.050708.142930 kind of equilibria might arise as a consequence of a long-run non- Annu. Rev. Econ. 2009.1:385-420. Downloaded from arjournals.annualreviews.org Copyright © 2009 by Annual Reviews. equilibrium process of learning, adaptation, and/or imitation. If All rights reserved agents’ strategies are completely observed at the end of each round 1941-1383/09/0904-0385$20.00 (and agents are randomly matched with a series of anonymous opponents), fairly simple rules perform well in terms of the agent’s worst-case payoffs, and also guarantee that any steady state of the system must correspond to an equilibrium. If players do not ob- serve the strategies chosen by their opponents (as in extensive-form games), then learning is consistent with steady states that are not Nash equilibria because players can maintain incorrect beliefs about off-path play. Beliefs can also be incorrect because of cogni- tive limitations and systematic inferential errors.
    [Show full text]
  • Lecture Notes
    GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies ..................................
    [Show full text]
  • The Monty Hall Problem in the Game Theory Class
    The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice.
    [Show full text]
  • Finitely Repeated Games
    Repeated games 1: Finite repetition Universidad Carlos III de Madrid 1 Finitely repeated games • A finitely repeated game is a dynamic game in which a simultaneous game (the stage game) is played finitely many times, and the result of each stage is observed before the next one is played. • Example: Play the prisoners’ dilemma several times. The stage game is the simultaneous prisoners’ dilemma game. 2 Results • If the stage game (the simultaneous game) has only one NE the repeated game has only one SPNE: In the SPNE players’ play the strategies in the NE in each stage. • If the stage game has 2 or more NE, one can find a SPNE where, at some stage, players play a strategy that is not part of a NE of the stage game. 3 The prisoners’ dilemma repeated twice • Two players play the same simultaneous game twice, at ! = 1 and at ! = 2. • After the first time the game is played (after ! = 1) the result is observed before playing the second time. • The payoff in the repeated game is the sum of the payoffs in each stage (! = 1, ! = 2) • Which is the SPNE? Player 2 D C D 1 , 1 5 , 0 Player 1 C 0 , 5 4 , 4 4 The prisoners’ dilemma repeated twice Information sets? Strategies? 1 .1 5 for each player 2" for each player D C E.g.: (C, D, D, C, C) Subgames? 2.1 5 D C D C .2 1.3 1.5 1 1.4 D C D C D C D C 2.2 2.3 2 .4 2.5 D C D C D C D C D C D C D C D C 1+1 1+5 1+0 1+4 5+1 5+5 5+0 5+4 0+1 0+5 0+0 0+4 4+1 4+5 4+0 4+4 1+1 1+0 1+5 1+4 0+1 0+0 0+5 0+4 5+1 5+0 5+5 5+4 4+1 4+0 4+5 4+4 The prisoners’ dilemma repeated twice Let’s find the NE in the subgames.
    [Show full text]
  • Incomplete Information I. Bayesian Games
    Bayesian Games Debraj Ray, November 2006 Unlike the previous notes, the material here is perfectly standard and can be found in the usual textbooks: see, e.g., Fudenberg-Tirole. For the examples in these notes (except for the very last section), I draw heavily on Martin Osborne’s excellent recent text, An Introduction to Game Theory, Oxford University Press. Obviously, incomplete information games — in which one or more players are privy to infor- mation that others don’t have — has enormous applicability: credit markets / auctions / regulation of firms / insurance / bargaining /lemons / public goods provision / signaling / . the list goes on and on. 1. A Definition A Bayesian game consists of 1. A set of players N. 2. A set of states Ω, and a common prior µ on Ω. 3. For each player i a set of actions Ai and a set of signals or types Ti. (Can make actions sets depend on type realizations.) 4. For each player i, a mapping τi :Ω 7→ Ti. 5. For each player i, a vN-M payoff function fi : A × Ω 7→ R, where A is the product of the Ai’s. Remarks A. Typically, the mapping τi tells us player i’s type. The easiest case is just when Ω is the product of the Ti’s and τi is just the appropriate projection mapping. B. The prior on states need not be common, but one view is that there is no loss of generality in assuming it (simply put all differences in the prior into differences in signals and redefine the state space and si-mappings accordingly).
    [Show full text]
  • Norms, Repeated Games, and the Role of Law
    Norms, Repeated Games, and the Role of Law Paul G. Mahoneyt & Chris William Sanchiricot TABLE OF CONTENTS Introduction ............................................................................................ 1283 I. Repeated Games, Norms, and the Third-Party Enforcement P rob lem ........................................................................................... 12 88 II. B eyond T it-for-Tat .......................................................................... 1291 A. Tit-for-Tat for More Than Two ................................................ 1291 B. The Trouble with Tit-for-Tat, However Defined ...................... 1292 1. Tw o-Player Tit-for-Tat ....................................................... 1293 2. M any-Player Tit-for-Tat ..................................................... 1294 III. An Improved Model of Third-Party Enforcement: "D ef-for-D ev". ................................................................................ 1295 A . D ef-for-D ev's Sim plicity .......................................................... 1297 B. Def-for-Dev's Credible Enforceability ..................................... 1297 C. Other Attractive Properties of Def-for-Dev .............................. 1298 IV. The Self-Contradictory Nature of Self-Enforcement ....................... 1299 A. The Counterfactual Problem ..................................................... 1300 B. Implications for the Self-Enforceability of Norms ................... 1301 C. Game-Theoretic Workarounds ................................................
    [Show full text]
  • Economics 201B Economic Theory (Spring 2021) Strategic Games
    Economics 201B Economic Theory (Spring 2021) Strategic Games Topics: terminology and notations (OR 1.7), games and solutions (OR 1.1-1.3), rationality and bounded rationality (OR 1.4-1.6), formalities (OR 2.1), best-response (OR 2.2), Nash equilibrium (OR 2.2), 2 2 examples × (OR 2.3), existence of Nash equilibrium (OR 2.4), mixed strategy Nash equilibrium (OR 3.1, 3.2), strictly competitive games (OR 2.5), evolution- ary stability (OR 3.4), rationalizability (OR 4.1), dominance (OR 4.2, 4.3), trembling hand perfection (OR 12.5). Terminology and notations (OR 1.7) Sets For R, ∈ ≥ ⇐⇒ ≥ for all . and ⇐⇒ ≥ for all and some . ⇐⇒ for all . Preferences is a binary relation on some set of alternatives R. % ⊆ From % we derive two other relations on : — strict performance relation and not  ⇐⇒ % % — indifference relation and ∼ ⇐⇒ % % Utility representation % is said to be — complete if , or . ∀ ∈ % % — transitive if , and then . ∀ ∈ % % % % can be presented by a utility function only if it is complete and transitive (rational). A function : R is a utility function representing if → % ∀ ∈ () () % ⇐⇒ ≥ % is said to be — continuous (preferences cannot jump...) if for any sequence of pairs () with ,and and , . { }∞=1 % → → % — (strictly) quasi-concave if for any the upper counter set ∈ { ∈ : is (strictly) convex. % } These guarantee the existence of continuous well-behaved utility function representation. Profiles Let be a the set of players. — () or simply () is a profile - a collection of values of some variable,∈ one for each player. — () or simply is the list of elements of the profile = ∈ { } − () for all players except . ∈ — ( ) is a list and an element ,whichistheprofile () .
    [Show full text]
  • Interim Correlated Rationalizability
    Interim Correlated Rationalizability The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters Citation Dekel, Eddie, Drew Fudenberg, and Stephen Morris. 2007. Interim correlated rationalizability. Theoretical Economics 2, no. 1: 15-40. Published Version http://econtheory.org/ojs/index.php/te/article/view/20070015 Citable link http://nrs.harvard.edu/urn-3:HUL.InstRepos:3196333 Terms of Use This article was downloaded from Harvard University’s DASH repository, and is made available under the terms and conditions applicable to Other Posted Material, as set forth at http:// nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of- use#LAA Theoretical Economics 2 (2007), 15–40 1555-7561/20070015 Interim correlated rationalizability EDDIE DEKEL Department of Economics, Northwestern University, and School of Economics, Tel Aviv University DREW FUDENBERG Department of Economics, Harvard University STEPHEN MORRIS Department of Economics, Princeton University This paper proposes the solution concept of interim correlated rationalizability, and shows that all types that have the same hierarchies of beliefs have the same set of interim-correlated-rationalizable outcomes. This solution concept charac- terizes common certainty of rationality in the universal type space. KEYWORDS. Rationalizability, incomplete information, common certainty, com- mon knowledge, universal type space. JEL CLASSIFICATION. C70, C72. 1. INTRODUCTION Harsanyi (1967–68) proposes solving games of incomplete
    [Show full text]
  • (501B) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann
    Dirk Bergemann Department of Economics Yale University Microeconomic Theory (501b) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann 1. (Market for Lemons) Here I ask that you work out some of the details in perhaps the most famous of all information economics models. By contrast to discussion in class, we give a complete formulation of the game. A seller is privately informed of the value v of the good that she sells to a buyer. The buyer's prior belief on v is uniformly distributed on [x; y] with 3 0 < x < y: The good is worth 2 v to the buyer. (a) Suppose the buyer proposes a price p and the seller either accepts or rejects p: If she accepts, the seller gets payoff p−v; and the buyer gets 3 2 v − p: If she rejects, the seller gets v; and the buyer gets nothing. Find the optimal offer that the buyer can make as a function of x and y: (b) Show that if the buyer and the seller are symmetrically informed (i.e. either both know v or neither party knows v), then trade takes place with probability 1. (c) Consider a simultaneous acceptance game in the model with private information as in part a, where a price p is announced and then the buyer and the seller simultaneously accept or reject trade at price p: The payoffs are as in part a. Find the p that maximizes the probability of trade. [SOLUTION] (a) We look for the subgame perfect equilibrium, thus we solve by back- ward induction.
    [Show full text]
  • Tacit Collusion in Oligopoly"
    Tacit Collusion in Oligopoly Edward J. Green,y Robert C. Marshall,z and Leslie M. Marxx February 12, 2013 Abstract We examine the economics literature on tacit collusion in oligopoly markets and take steps toward clarifying the relation between econo- mists’analysis of tacit collusion and those in the legal literature. We provide examples of when the economic environment might be such that collusive pro…ts can be achieved without communication and, thus, when tacit coordination is su¢ cient to elevate pro…ts versus when com- munication would be required. The authors thank the Human Capital Foundation (http://www.hcfoundation.ru/en/), and especially Andrey Vavilov, for …nancial support. The paper bene…tted from discussions with Roger Blair, Louis Kaplow, Bill Kovacic, Vijay Krishna, Steven Schulenberg, and Danny Sokol. We thank Gustavo Gudiño for valuable research assistance. [email protected], Department of Economics, Penn State University [email protected], Department of Economics, Penn State University [email protected], Fuqua School of Business, Duke University 1 Introduction In this chapter, we examine the economics literature on tacit collusion in oligopoly markets and take steps toward clarifying the relation between tacit collusion in the economics and legal literature. Economists distinguish between tacit and explicit collusion. Lawyers, using a slightly di¤erent vocabulary, distinguish between tacit coordination, tacit agreement, and explicit collusion. In hopes of facilitating clearer communication between economists and lawyers, in this chapter, we attempt to provide a coherent resolution of the vernaculars used in the economics and legal literature regarding collusion.1 Perhaps the easiest place to begin is to de…ne explicit collusion.
    [Show full text]