(501B) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann

Total Page:16

File Type:pdf, Size:1020Kb

(501B) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann Dirk Bergemann Department of Economics Yale University Microeconomic Theory (501b) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann 1. (Market for Lemons) Here I ask that you work out some of the details in perhaps the most famous of all information economics models. By contrast to discussion in class, we give a complete formulation of the game. A seller is privately informed of the value v of the good that she sells to a buyer. The buyer's prior belief on v is uniformly distributed on [x; y] with 3 0 < x < y: The good is worth 2 v to the buyer. (a) Suppose the buyer proposes a price p and the seller either accepts or rejects p: If she accepts, the seller gets payoff p−v; and the buyer gets 3 2 v − p: If she rejects, the seller gets v; and the buyer gets nothing. Find the optimal offer that the buyer can make as a function of x and y: (b) Show that if the buyer and the seller are symmetrically informed (i.e. either both know v or neither party knows v), then trade takes place with probability 1. (c) Consider a simultaneous acceptance game in the model with private information as in part a, where a price p is announced and then the buyer and the seller simultaneously accept or reject trade at price p: The payoffs are as in part a. Find the p that maximizes the probability of trade. [SOLUTION] (a) We look for the subgame perfect equilibrium, thus we solve by back- ward induction. From the next point it should be clear why there are other Nash equilibrium which are not subgame perfect. If the buyer proposes price p, then the seller accepts if and only if v ≥ p (we assume that, if indifferent the seller accepts). Thus, the payoff of the buyer conditional on offering price p is: 3 fv ≥ pg · [ v − pjv ≥ p]: P E 2 That is, the expected utility of the buyer is the probability of the seller accepting the offer times the expected utility conditional on 1 the seller accepting the offer. Expanding the terms, we get: 3 (p − x) 3 fv ≥ pg · [ v − pjv ≥ p] = ( [ vjv ≤ p] − p) P E 2 y − x E 2 (p − x) 3 x + p = ( − p) y − x 2 2 p − x = (3(x + p) − 4p) 4(y − x) p − x = (3x − p): 4(y − x) One should note that the previous equality holds only for p 2 [x; y]. Obviously for p ≤ x we have that Pfv ≥ pg = 0 and for p ≥ y we have that Pfv ≥ pg = 1. Taking into account the previous observation, we can differentiate with respect to p and get the F.O.C., since the objective function is strictly concave in p we know that the maximum will be achieved at a corner (that is, p 2 fx; yg), if the solution to the F.O.C. yields a p 62 [x; y]. Taking the derivative, we get: @ 3 1 fv ≥ pg · [ v − pjv ≥ p] = (4x − 2p) @pP E 2 4(y − x) @ 3 ) fv ≥ pg · [ v − pjv ≥ p] = 0 () p = 2x @pP E 2 Thus, we have that the buyer proposes the following price depending on the distribution of v. ( 2x 2x ≤ y p∗ = y 2x > y (b) If the buyer knows the value v, then the unique subgame perfect Nash equilibrium is that the buyer offers v and the seller accepts. There are two things to note. First, this is a complete information game, thus the right solution concept is subgame perfect Nash equilibrium. Note that there are other Nash equilibrium which are not sub-game perfect. For example, the buyer offers 0 and the seller rejects all offers. If neither agent know the valuation v, then the game can also be treated as a complete information game. In this case, the buyer offers E[v] = (y + x)=2, and the sellet accepts. Note that without asymmetric information, the game can be treated as complete infor- mation game, in which the value of buyer and seller are 3E[v]=2 and E[v] respectively. (c) From part (a) it is clear that the seller accepts if and only if v ≥ p. On the other hand, the buyer accepts if and only if he gets non-negative 2 profits by accepting. Thus, the buyer accepts if and only if 3 p − x fv ≥ pg · [ v − pjv ≥ p] = (3x − p) ≥ 0: P E 2 4(y − x) To maximize the probability of trade, we just need to find the max- imum p such that the buyer gets non-negative profits. Clearly, we can just equate the previous inequality to 0 to get the price that maximizes trade. That is, p∗ − x (3x − p∗) = 0 4(y − x) The quadratic equation has two solutions, as the buyer also gets 0 profits when there is 0 probability of trade. Thus, we consider the other solution. Thus, the price that maximizes the probability of trade is: ( 3x 3x ≤ y p∗ = ; [y; 3x] 3x ≥ y where we note that whenever 3x ≥ y we have a continuum of prices that ensure trade with probability 1. 2. (Akerlof's Lemon in a General Equilibrium Model) Suppose there is a continuum of buyers with a characteristic vb 2 [0; 1] distributed ac- cording to a distribution function G(vb). There is also a continuum of sellers with mass one, all with the same characteristic vs 2 (0; 1). Each seller owns an object of value θ 2 [0; 1] which is distributed according to a distribution function F (θ) on the unit interval, independent across sell- ers. The value θ is known to the seller and unknown to the buyer. The valuations for the object are us = vsθ and ub = vbθ: (a) Determine the efficient volume of trade and the efficient mix of sellers (in terms of types of θ) and buyer (in terms of types of vb). (b) Suppose for the moment that the characteristic of each object is observable and hence that information is symmetric across sellers and buyers. What would be the equilibrium price for the object with value θ be and would trade be efficient. (c) Under asymmetric information, compute the demand and supply functions. Explain what the demand is not necessarily downward sloping (despite the absence of income effects.) 3 (d) Solve for the competitive equilibrium for f and g uniform on [0; 1]. Verify that trade is suboptimal. (e) Argue that, in general, there can exist multiple equilibria. Then as- sume that it is the case. Show that a higher price equilibrium Pareto dominates a lower price one. Compare with the first fundamental theorem of welfare economics. (f) Show that a minimum quality standard s0 > 0, if enforceable, may improve welfare. [SOLUTION] (a) If you are the benevolent social planner, you will want trade to occur only when ub ≥ us; that is, when vb ≥ vs: Thus, the probability of trade is Pr (Trade) = Pr (vb > vs) = 1−G (vs) : In other words, you are only going to match high-value buyers to some sellers. To which sellers? To the ones that allow you (the planner) to maximize gains from trade θ (vb − vs) : Therefore, you will choose the highest-valuation sellers (the best objects) to sell. The set of matched sellers must have the same measure of the set of buyers. Let θ∗ be the lowest seller that gets matched. Then, ∗ 1 − F (θ ) = 1 − G (vs) ∗ −1 θ = F G (vs) : So the efficient mix of sellers and buyers (in terms of maximizing social ∗ ∗ welfare) involves buyers with vb ≥ vs and sellers with θ ≥ θ , with θ as defined above. (b) Let p(θ) be the price of an object of value θ (since θ is observable we can condition on it). In equilibrium, prices p(θ) must be such that market clears. We expect p(θ) to be strictly increasing in θ. Suppose not, say there are two sellers with v2 > v1 and p(v2) < p(v1), then clearly good 2 will \dominate" good 1 (2 has a higher quality which is valued by consumers and a lower price). Also note that, the higher vb buyers have higher valuation for quality and hence are willing to pay more for quality. Therefore, we expect high θ goods, which have higher prices, to be matched with high vb buyers. By the market clearing condition we require that for all items sold in equilibrium G(vb) = F (θ) which in turn implies the matching between sellers and buyers −1 vb(θ) = G (F (θ)) (1) −1 That is, a good with quality θ is matched with a buyer with vb = G (F (θ)). From the participation conditions (i.e., matched buyer and seller both will- ing to trade), we can immediately see that, if θ is sold in equilibrium, p(θ) 4 should satisfy −1 vsθ ≤ p(θ) ≤ G (F (θ))θ (2) But not all functions satisfying (2) will be equilibrium prices1 because, to support the matching, buyers have to be willing to buy the θ corresponding to them and not other. Type-vb buyers, if in equilibrium buying, are going to choose a good so that max vbθ − p(θ) θ From the FOC we have, 0 p (θ) = vb (3) Combining conditions (1) and (3) we get that, in equilibrium, p(θ) should satisfy the following condition, p0(θ) = G−1(F (θ)): (4) Can we further characterize p(θ)? Yes, we can put additional restrictions on the last unit sold. Let θ be the last unit sold, that is, for θ < θ @p(θ) such that (2) and (4) hold.
Recommended publications
  • Game Theory 2: Extensive-Form Games and Subgame Perfection
    Game Theory 2: Extensive-Form Games and Subgame Perfection 1 / 26 Dynamics in Games How should we think of strategic interactions that occur in sequence? Who moves when? And what can they do at different points in time? How do people react to different histories? 2 / 26 Modeling Games with Dynamics Players Player function I Who moves when Terminal histories I Possible paths through the game Preferences over terminal histories 3 / 26 Strategies A strategy is a complete contingent plan Player i's strategy specifies her action choice at each point at which she could be called on to make a choice 4 / 26 An Example: International Crises Two countries (A and B) are competing over a piece of land that B occupies Country A decides whether to make a demand If Country A makes a demand, B can either acquiesce or fight a war If A does not make a demand, B keeps land (game ends) A's best outcome is Demand followed by Acquiesce, worst outcome is Demand and War B's best outcome is No Demand and worst outcome is Demand and War 5 / 26 An Example: International Crises A can choose: Demand (D) or No Demand (ND) B can choose: Fight a war (W ) or Acquiesce (A) Preferences uA(D; A) = 3 > uA(ND; A) = uA(ND; W ) = 2 > uA(D; W ) = 1 uB(ND; A) = uB(ND; W ) = 3 > uB(D; A) = 2 > uB(D; W ) = 1 How can we represent this scenario as a game (in strategic form)? 6 / 26 International Crisis Game: NE Country B WA D 1; 1 3X; 2X Country A ND 2X; 3X 2; 3X I Is there something funny here? I Is there something funny here? I Specifically, (ND; W )? I Is there something funny here?
    [Show full text]
  • Equilibrium Refinements
    Equilibrium Refinements Mihai Manea MIT Sequential Equilibrium I In many games information is imperfect and the only subgame is the original game. subgame perfect equilibrium = Nash equilibrium I Play starting at an information set can be analyzed as a separate subgame if we specify players’ beliefs about at which node they are. I Based on the beliefs, we can test whether continuation strategies form a Nash equilibrium. I Sequential equilibrium (Kreps and Wilson 1982): way to derive plausible beliefs at every information set. Mihai Manea (MIT) Equilibrium Refinements April 13, 2016 2 / 38 An Example with Incomplete Information Spence’s (1973) job market signaling game I The worker knows her ability (productivity) and chooses a level of education. I Education is more costly for low ability types. I Firm observes the worker’s education, but not her ability. I The firm decides what wage to offer her. In the spirit of subgame perfection, the optimal wage should depend on the firm’s beliefs about the worker’s ability given the observed education. An equilibrium needs to specify contingent actions and beliefs. Beliefs should follow Bayes’ rule on the equilibrium path. What about off-path beliefs? Mihai Manea (MIT) Equilibrium Refinements April 13, 2016 3 / 38 An Example with Imperfect Information Courtesy of The MIT Press. Used with permission. Figure: (L; A) is a subgame perfect equilibrium. Is it plausible that 2 plays A? Mihai Manea (MIT) Equilibrium Refinements April 13, 2016 4 / 38 Assessments and Sequential Rationality Focus on extensive-form games of perfect recall with finitely many nodes. An assessment is a pair (σ; µ) I σ: (behavior) strategy profile I µ = (µ(h) 2 ∆(h))h2H: system of beliefs ui(σjh; µ(h)): i’s payoff when play begins at a node in h randomly selected according to µ(h), and subsequent play specified by σ.
    [Show full text]
  • Lecture Notes
    GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies ..................................
    [Show full text]
  • The Monty Hall Problem in the Game Theory Class
    The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice.
    [Show full text]
  • Incomplete Information I. Bayesian Games
    Bayesian Games Debraj Ray, November 2006 Unlike the previous notes, the material here is perfectly standard and can be found in the usual textbooks: see, e.g., Fudenberg-Tirole. For the examples in these notes (except for the very last section), I draw heavily on Martin Osborne’s excellent recent text, An Introduction to Game Theory, Oxford University Press. Obviously, incomplete information games — in which one or more players are privy to infor- mation that others don’t have — has enormous applicability: credit markets / auctions / regulation of firms / insurance / bargaining /lemons / public goods provision / signaling / . the list goes on and on. 1. A Definition A Bayesian game consists of 1. A set of players N. 2. A set of states Ω, and a common prior µ on Ω. 3. For each player i a set of actions Ai and a set of signals or types Ti. (Can make actions sets depend on type realizations.) 4. For each player i, a mapping τi :Ω 7→ Ti. 5. For each player i, a vN-M payoff function fi : A × Ω 7→ R, where A is the product of the Ai’s. Remarks A. Typically, the mapping τi tells us player i’s type. The easiest case is just when Ω is the product of the Ti’s and τi is just the appropriate projection mapping. B. The prior on states need not be common, but one view is that there is no loss of generality in assuming it (simply put all differences in the prior into differences in signals and redefine the state space and si-mappings accordingly).
    [Show full text]
  • Subgame Perfect (-)Equilibrium in Perfect Information Games
    Subgame perfect (-)equilibrium in perfect information games J´anosFlesch∗ April 15, 2016 Abstract We discuss recent results on the existence and characterization of subgame perfect ({)equilibrium in perfect information games. The game. We consider games with perfect information and deterministic transitions. Such games can be given by a directed tree.1 In this tree, each node is associated with a player, who controls this node. The outgoing arcs at this node represent the actions available to this player at this node. We assume that each node has at least one successor, rather than having terminal nodes.2 Play of the game starts at the root. At any node z that play visits, the player who controls z has to choose one of the actions at z, which brings play to a next node. This induces an infinite path in the tree from the root, which we call a play. Depending on this play, each player receives a payoff. Note that these payoffs are fairly general. This setup encompasses the case when the actions induce instantaneous rewards which are then aggregated into a payoff, possibly by taking the total discounted sum or the long-term average. It also includes payoff functions considered in the literature of computer science (reachability games, etc.). Subgame-perfect (-)equilibrium. We focus on pure strategies for the players. A central solution concept in such games is subgame-perfect equilibrium, which is a strategy profile that induces a Nash equilibrium in every subgame, i.e. when starting at any node in the tree, no player has an incentive to deviate individually from his continuation strategy.
    [Show full text]
  • Subgame-Perfect Equilibria in Mean-Payoff Games
    Subgame-perfect Equilibria in Mean-payoff Games Léonard Brice ! Université Gustave Eiffel, France Jean-François Raskin ! Université Libre de Bruxelles, Belgium Marie van den Bogaard ! Université Gustave Eiffel, France Abstract In this paper, we provide an effective characterization of all the subgame-perfect equilibria in infinite duration games played on finite graphs with mean-payoff objectives. To this end, we introduce the notion of requirement, and the notion of negotiation function. We establish that the plays that are supported by SPEs are exactly those that are consistent with the least fixed point of the negotiation function. Finally, we show that the negotiation function is piecewise linear, and can be analyzed using the linear algebraic tool box. As a corollary, we prove the decidability of the SPE constrained existence problem, whose status was left open in the literature. 2012 ACM Subject Classification Software and its engineering: Formal methods; Theory of compu- tation: Logic and verification; Theory of computation: Solution concepts in game theory. Keywords and phrases Games on graphs, subgame-perfect equilibria, mean-payoff objectives. Digital Object Identifier 10.4230/LIPIcs... 1 Introduction The notion of Nash equilibrium (NE) is one of the most important and most studied solution concepts in game theory. A profile of strategies is an NE when no rational player has an incentive to change their strategy unilaterally, i.e. while the other players keep their strategies. Thus an NE models a stable situation. Unfortunately, it is well known that, in sequential games, NEs suffer from the problem of non-credible threats, see e.g. [18]. In those games, some NE only exists when some players do not play rationally in subgames and so use non-credible threats to force the NE.
    [Show full text]
  • 14.12 Game Theory Lecture Notes∗ Lectures 7-9
    14.12 Game Theory Lecture Notes∗ Lectures 7-9 Muhamet Yildiz Intheselecturesweanalyzedynamicgames(withcompleteinformation).Wefirst analyze the perfect information games, where each information set is singleton, and develop the notion of backwards induction. Then, considering more general dynamic games, we will introduce the concept of the subgame perfection. We explain these concepts on economic problems, most of which can be found in Gibbons. 1 Backwards induction The concept of backwards induction corresponds to the assumption that it is common knowledge that each player will act rationally at each node where he moves — even if his rationality would imply that such a node will not be reached.1 Mechanically, it is computed as follows. Consider a finite horizon perfect information game. Consider any node that comes just before terminal nodes, that is, after each move stemming from this node, the game ends. If the player who moves at this node acts rationally, he will choose the best move for himself. Hence, we select one of the moves that give this player the highest payoff. Assigning the payoff vector associated with this move to the node at hand, we delete all the moves stemming from this node so that we have a shorter game, where our node is a terminal node. Repeat this procedure until we reach the origin. ∗These notes do not include all the topics that will be covered in the class. See the slides for a more complete picture. 1 More precisely: at each node i the player is certain that all the players will act rationally at all nodes j that follow node i; and at each node i the player is certain that at each node j that follows node i the player who moves at j will be certain that all the players will act rationally at all nodes k that follow node j,...ad infinitum.
    [Show full text]
  • Lecture Notes
    Chapter 12 Repeated Games In real life, most games are played within a larger context, and actions in a given situation affect not only the present situation but also the future situations that may arise. When a player acts in a given situation, he takes into account not only the implications of his actions for the current situation but also their implications for the future. If the players arepatient andthe current actionshavesignificant implications for the future, then the considerations about the future may take over. This may lead to a rich set of behavior that may seem to be irrational when one considers the current situation alone. Such ideas are captured in the repeated games, in which a "stage game" is played repeatedly. The stage game is repeated regardless of what has been played in the previous games. This chapter explores the basic ideas in the theory of repeated games and applies them in a variety of economic problems. As it turns out, it is important whether the game is repeated finitely or infinitely many times. 12.1 Finitely-repeated games Let = 0 1 be the set of all possible dates. Consider a game in which at each { } players play a "stage game" , knowing what each player has played in the past. ∈ Assume that the payoff of each player in this larger game is the sum of the payoffsthat he obtains in the stage games. Denote the larger game by . Note that a player simply cares about the sum of his payoffs at the stage games. Most importantly, at the beginning of each repetition each player recalls what each player has 199 200 CHAPTER 12.
    [Show full text]
  • Collusion and Cartels
    Collusion and Cartels Allan Collard-Wexler Duke November 4, 2016 Defining Collusion: A secret agreement between two or more parties for a fraudulent, illegal, or deceitful purpose. These are “cooperative” outcomes in the sense that firms agree to act together. Why do firms want to achieve a cooperative outcome? They earn higher profits! Some famous recent examples of collusion (we’ll see more when we cover antitrust issues): Lysine (mid-1990’s) From Wikipedia: The lysine price-fixing conspiracy was an organized effort during the mid-1990s to raise the price of the animal feed additive lysine. It involved five companies that had commercialized high-tech fermentation technologies, including American company Archer Daniels Midland (ADM), Japanese companies Ajinomoto and Kyowa Hakko Kogyo, and Korean companies Sewon America Inc. and Cheil Jedang Ltd. A criminal investigation resulted in fines and three-year prison sentences for three executives of ADM who colluded with the other companies to fix prices. The foreign companies settled with the United States Department of Justice Antitrust Division in September through December 1996. Each firm and four executives from the Asian firms pled guilty as part of a plea bargain to aid in further investigation against ADM. The cartel had been able to raise lysine prices 70% within their first nine months of cooperation. The investigation yielded $105 million in criminal fines, a record antitrust penalty at the time, including a $70 million fine against ADM. ADM was fined an additional $30 million for its participation in a separate conspiracy in the citric acid market and paid a total fine of $100 million.
    [Show full text]
  • The Effective Minimax Value of Asynchronously Repeated Games*
    Int J Game Theory (2003) 32: 431–442 DOI 10.1007/s001820300161 The effective minimax value of asynchronously repeated games* Kiho Yoon Department of Economics, Korea University, Anam-dong, Sungbuk-gu, Seoul, Korea 136-701 (E-mail: [email protected]) Received: October 2001 Abstract. We study the effect of asynchronous choice structure on the possi- bility of cooperation in repeated strategic situations. We model the strategic situations as asynchronously repeated games, and define two notions of effec- tive minimax value. We show that the order of players’ moves generally affects the effective minimax value of the asynchronously repeated game in significant ways, but the order of moves becomes irrelevant when the stage game satisfies the non-equivalent utilities (NEU) condition. We then prove the Folk Theorem that a payoff vector can be supported as a subgame perfect equilibrium out- come with correlation device if and only if it dominates the effective minimax value. These results, in particular, imply both Lagunoff and Matsui’s (1997) result and Yoon (2001)’s result on asynchronously repeated games. Key words: Effective minimax value, folk theorem, asynchronously repeated games 1. Introduction Asynchronous choice structure in repeated strategic situations may affect the possibility of cooperation in significant ways. When players in a repeated game make asynchronous choices, that is, when players cannot always change their actions simultaneously in each period, it is quite plausible that they can coordinate on some particular actions via short-run commitment or inertia to render unfavorable outcomes infeasible. Indeed, Lagunoff and Matsui (1997) showed that, when a pure coordination game is repeated in a way that only *I thank three anonymous referees as well as an associate editor for many helpful comments and suggestions.
    [Show full text]
  • Contemporaneous Perfect Epsilon-Equilibria
    Games and Economic Behavior 53 (2005) 126–140 www.elsevier.com/locate/geb Contemporaneous perfect epsilon-equilibria George J. Mailath a, Andrew Postlewaite a,∗, Larry Samuelson b a University of Pennsylvania b University of Wisconsin Received 8 January 2003 Available online 18 July 2005 Abstract We examine contemporaneous perfect ε-equilibria, in which a player’s actions after every history, evaluated at the point of deviation from the equilibrium, must be within ε of a best response. This concept implies, but is stronger than, Radner’s ex ante perfect ε-equilibrium. A strategy profile is a contemporaneous perfect ε-equilibrium of a game if it is a subgame perfect equilibrium in a perturbed game with nearly the same payoffs, with the converse holding for pure equilibria. 2005 Elsevier Inc. All rights reserved. JEL classification: C70; C72; C73 Keywords: Epsilon equilibrium; Ex ante payoff; Multistage game; Subgame perfect equilibrium 1. Introduction Analyzing a game begins with the construction of a model specifying the strategies of the players and the resulting payoffs. For many games, one cannot be positive that the specified payoffs are precisely correct. For the model to be useful, one must hope that its equilibria are close to those of the real game whenever the payoff misspecification is small. To ensure that an equilibrium of the model is close to a Nash equilibrium of every possible game with nearly the same payoffs, the appropriate solution concept in the model * Corresponding author. E-mail addresses: [email protected] (G.J. Mailath), [email protected] (A. Postlewaite), [email protected] (L.
    [Show full text]