ARTIFICIAL INTELLIGENCE Decision Making: Opponent Based

Total Page:16

File Type:pdf, Size:1020Kb

ARTIFICIAL INTELLIGENCE Decision Making: Opponent Based INFOB2KI 2019-2020 Utrecht University The Netherlands ARTIFICIAL INTELLIGENCE Decision making: opponent based Lecturer: Silja Renooij These slides are part of the INFOB2KI Course Notes available from www.cs.uu.nl/docs/vakken/b2ki/schema.html Game theory 2 Outline . Game theory: rules for defeating opponents – Perfect information games • Deterministic turn‐taking games – Mini‐max algorithm, Alpha‐beta pruning – Best response, Nash equilibrium – Imperfect information games • Simultaneous‐move games: Mixed strategy – Incomplete information games • Prisoner’s dilemma 3 Game Theory Developed to explain the optimal strategy in two‐person (nowadays n ≥ 2) interactions. Initially, von Neumann and Morgenstern – Zero‐sum games (your win == opponents loss) . John Nash – Nonzero‐sum games . Harsanyi, Selten – Incomplete information (Bayesian games) 4 Zero-sum games Better term: constant‐sum game Examples of zero‐sum games: . 2‐player; payoffs: 1 for win, ‐1 for loss, 0 for draw total payoff is 0, regardless of the outcome . 2‐player; payoffs: 1 for win, 0 for loss, ½ for draw total payoff is 1, regadless of outcome . 3‐player; payoffs: distribute 3 points over players, depending on performance Example of non‐zero‐sum game: . 2‐player; payoffs: 3 for win, 0 for loss, 1 for draw total payoff is either 3 or 2, depending on the outcome. 5 Game types Complete information games: . Perfect information games upon making a move, players know full history of the game, all moves by all players, all payoffs, etc . Imperfect information games players know all outcomes/payoffs, types of other players & their strategies, but are unaware of (or unsure about) possible actions of other playes ‐ simultaneous moves: what action will others choose? ‐ (temporarily) shielded attributes: who has which cards? Complete information games can be deterministic or involve chance. 6 Game types Incomplete information games: Uncertainty about game being played: factors outside the rules of the game, not known to one or more players, may affect the outcome of the game • E.g players may not know other players "type", their strategies, payoffs or preferences Incomplete information games can be deterministic or involve chance. 7 Complete information games Deterministic Chance Perfect Chess, checkers, Backgammon, information go, othello, monopoly Tic‐tac‐toe Imperfect Battleships, Bridge, poker, information Minesweeper scrabble NB(!) textbook says randomness is the difference between perfect and imperfect. Other sources state imperfect == incomplete …. be aware of this! 8 Deterministic Two-player, turn-taking Perfect information 9 Game tree • alternates between two players. • represents all possibilities from perspective of one player (‘me’) from current root. me: you: me: you: Zero‐sum or Non‐zero sum? (my rewards) 10 Minimax . Compute value of perfect play for deterministic, perfect information games – Traverse game tree in DFS‐like manner – ‘bubble up’ values of evaluation function: maximise if my turn, minimise for opponent’s turn . Serves for selecting next move: choose move to position with highest minimax value = best achievable payoff against best play . May also serve for finding optimal strategy = from start to finish my best move, for every move of opponent. 11 Minimax: example NB book uses circles and squares instead of triangles! 12 Minimax algorithm 13 Properties of minimax . Complete? Yes (if tree is finite) . Optimal? Yes (against an optimal opponent) . Time complexity? O(bm) for branching factor b and max depth m (in general, worst case, we cannot improve on this, so O(bm) suggested in textbook AI4G must be a typo) . Space complexity? O(bm) (if depth‐first exploration; can be reduced to O(m) with backtracking variant of DFS which generates one successor at a time, rather than all ) For chess, with b ≈ 35, m ≈100 for "reasonable" games exact solution completely infeasible 14 Solution: pruning m n . Idea: If m is better for me than n, we will never actually get to n in play (MAX will avoid it) prune that leaf/subtree . let α be the best (= highest) value (to MAX) found so far on current path . define β similarly for MIN: best (= lowest) value found so far 15 Alpha-Beta (α-β) pruning Minimax, augmented with upper‐ and lowerbounds • Init: for all non‐leaf nodes set α = −∞ lowerbound on achievable score β= ∞ upperbound on achievable score • Upwards: update α in MAX move; update β in MIN move α = −∞, β = ∞ α = −∞, β = ∞ 16 α-β Pruning function MIN-VALUE is similarly extended: if then return v β ← MIN(β, v) 17 α-β pruning example α= 3 β= ∞ α= −∞ β= 3 α = best score till now (lowerbound ) updated in own (MAX) move β= upperbound on achievable score updated in opponents (MIN) move 18 α-β pruning example α= 3 β= ∞ α= −∞ α= −∞ β= 3 β= 2 19 α-β pruning example α= 3 β= ∞ α= −∞ α= −∞ α= −∞ β= 3 β= 2 β= 14 Prune or continue? 20 α-β pruning example α= 3 β= ∞ α= −∞ α= −∞ α= −∞ β= 3 β= 2 β= 5 Prune or continue? 21 α-β pruning example α= 3 β= ∞ α= −∞ β= 2 α= −∞ α= −∞ β= 3 β= 2 22 Properties of α-β . Pruning does not affect final result! . Good move ordering improves effectiveness of pruning . With “perfect ordering”, time complexity = O(bm/2) effective branching factor is √b allows search depth to double for same cost . A simple example of the value of reasoning about which computations are relevant (a form of meta‐reasoning) 23 Practical feasibility: resource limits Suppose we have 100 secs and explore 104 nodes/sec 106 nodes can be explored per move What if we have too little time to reach terminal states (=utility function)? Standard approach combines: . cutoff test: e.g., depth limit (perhaps add quiescence search: disregard positions that are unlikely to exhibit wild swings in value in near future) . evaluation function = estimated desirability of position 24 Evaluation functions Evaluation function (cf heuristic with A*): . Returns estimate of expected utility of the game from a given position . Must agree with utility function on terminal nodes . Must not take too long For chess, typically linear weighted sum of features Eval(s) = w1 f1(s) + w2 f2(s) + … + wn fn(s) e.g., w1 = 9 with f1(s) = (# white queens) – (# black queens), etc. 25 Cutting off search MinimaxCutoff is identical to MinimaxValue except: 1. Terminal? is replaced by Cutoff? 2. Utility is replaced by Eval Does it work in practice? bm = 106, b=35 m=4 4‐ply lookahead is a hopeless chess player! – 4‐ply ≈ human novice – 8‐ply ≈ typical PC, human master – 12‐ply ≈ Deep Blue, Kasparov 26 Deterministic games in practice . Checkers: Chinook ended 40‐year‐reign of human world champion Marion Tinsley in 1994. Used a pre‐computed endgame database defining perfect play for all positions involving 8 or fewer pieces on the board, a total of 444 billion positions. Chess: Deep Blue defeated human world champion Garry Kasparov in a six‐game match in 1997. Deep Blue searches 200 million positions per second, uses very sophisticated evaluation, and undisclosed methods for extending some lines of search up to 40 ply. Othello: human champions refuse to compete against computers, who are too good. Go: human champions refuse to compete against computers, who are too bad. In Go, b > 300, so most programs use pattern knowledge bases to suggest plausible moves. 2016/2017: Alpha Go beats world’s number 1 and 2 using deep learning. 27 Strategies and equilibria 28 Big Monkey, Little Monkey example . Monkeys usually eat ground‐level fruit . Occasionally they climb a tree to shake loose a coconut (1 per tree) . A Coconut yields 10 Calories . Big Monkey expends 2 Calories climbing up the tree, shaking and climbing down. Little Monkey expends 0 Calories on this exercise. 29 BM and LM utilities . If only BM climbs the tree LM eats some before BM gets down BM gets 6 C, LM gets 4 C . If only LM climbs the tree BM eats almost all before LM gets down BM gets 9 C, LM gets 1 C . If both climb the tree BM is first to hog coconut BM gets 7 C, LM gets 3 C How should the monkeys each act so as to maximize their own calorie gain? 30 BM and LM: strategies Strategies are determined prior to ‘playing the game’ Assume BM will be allowed to move first. BM has two (single action) strategies: – wait (w), or – climb (c) LM has four strategies: – If BM waits, then wait; if BM climbs then wait (xw) – If BM waits, then wait; if BM climbs then climb (xx) – If BM waits, then climb; if BM climbs then wait (x¬x) – If BM waits, then climb; if BM climbs then climb (xc) 31 BM and LM: BM moves first Big monkey w c Little monkey w c w c 0,0 9,1 6-2,4 7-2,3 (BM,LM) What should Big Monkey do? If BM waits, will outcome be at least that of climbing, regardless of what LM does? No: 0 vs 4, 9 vs 5 …. What if we believe LM will act rationally? 32 BM and LM: BM moves first Big monkey w c Little monkey w c w c 0,0 9,1 6-2,4 7-2,3 (BM,LM) What should Big Monkey do? • If BM waits, LM will climb BM gets 9 • If BM climbs, LM will wait BM gets 4 BM should wait (w) What about Little Monkey? Opposite of BM (x¬x) (eventhough we’ll never get to the right side of the game tree unless BM errs) 33 BM and LM: BM moves first The game‐tree representation of a game is called extensive form1, as opposed to normal form2 : LM xc xw xx x¬x c 5,3 4,4 5,3 4,4 BM w 9,1 0,0 0,0 9,1 What should BM do? What about Little Monkey? wait (w) Opposite of BM (x¬x) 1 game tree that explicitly shows the players´ moves and resulting payoffs 2 table showing payoffs of outcomes of simultaneous ‘decisions’ (strategies) 34 Dominant strategies Consider a player’s strategies s1 and s2.
Recommended publications
  • Lecture 4 Rationalizability & Nash Equilibrium Road
    Lecture 4 Rationalizability & Nash Equilibrium 14.12 Game Theory Muhamet Yildiz Road Map 1. Strategies – completed 2. Quiz 3. Dominance 4. Dominant-strategy equilibrium 5. Rationalizability 6. Nash Equilibrium 1 Strategy A strategy of a player is a complete contingent-plan, determining which action he will take at each information set he is to move (including the information sets that will not be reached according to this strategy). Matching pennies with perfect information 2’s Strategies: HH = Head if 1 plays Head, 1 Head if 1 plays Tail; HT = Head if 1 plays Head, Head Tail Tail if 1 plays Tail; 2 TH = Tail if 1 plays Head, 2 Head if 1 plays Tail; head tail head tail TT = Tail if 1 plays Head, Tail if 1 plays Tail. (-1,1) (1,-1) (1,-1) (-1,1) 2 Matching pennies with perfect information 2 1 HH HT TH TT Head Tail Matching pennies with Imperfect information 1 2 1 Head Tail Head Tail 2 Head (-1,1) (1,-1) head tail head tail Tail (1,-1) (-1,1) (-1,1) (1,-1) (1,-1) (-1,1) 3 A game with nature Left (5, 0) 1 Head 1/2 Right (2, 2) Nature (3, 3) 1/2 Left Tail 2 Right (0, -5) Mixed Strategy Definition: A mixed strategy of a player is a probability distribution over the set of his strategies. Pure strategies: Si = {si1,si2,…,sik} σ → A mixed strategy: i: S [0,1] s.t. σ σ σ i(si1) + i(si2) + … + i(sik) = 1. If the other players play s-i =(s1,…, si-1,si+1,…,sn), then σ the expected utility of playing i is σ σ σ i(si1)ui(si1,s-i) + i(si2)ui(si2,s-i) + … + i(sik)ui(sik,s-i).
    [Show full text]
  • Minimax TD-Learning with Neural Nets in a Markov Game
    Minimax TD-Learning with Neural Nets in a Markov Game Fredrik A. Dahl and Ole Martin Halck Norwegian Defence Research Establishment (FFI) P.O. Box 25, NO-2027 Kjeller, Norway {Fredrik-A.Dahl, Ole-Martin.Halck}@ffi.no Abstract. A minimax version of temporal difference learning (minimax TD- learning) is given, similar to minimax Q-learning. The algorithm is used to train a neural net to play Campaign, a two-player zero-sum game with imperfect information of the Markov game class. Two different evaluation criteria for evaluating game-playing agents are used, and their relation to game theory is shown. Also practical aspects of linear programming and fictitious play used for solving matrix games are discussed. 1 Introduction An important challenge to artificial intelligence (AI) in general, and machine learning in particular, is the development of agents that handle uncertainty in a rational way. This is particularly true when the uncertainty is connected with the behavior of other agents. Game theory is the branch of mathematics that deals with these problems, and indeed games have always been an important arena for testing and developing AI. However, almost all of this effort has gone into deterministic games like chess, go, othello and checkers. Although these are complex problem domains, uncertainty is not their major challenge. With the successful application of temporal difference learning, as defined by Sutton [1], to the dice game of backgammon by Tesauro [2], random games were included as a standard testing ground. But even backgammon features perfect information, which implies that both players always have the same information about the state of the game.
    [Show full text]
  • SEQUENTIAL GAMES with PERFECT INFORMATION Example
    SEQUENTIAL GAMES WITH PERFECT INFORMATION Example 4.9 (page 105) Consider the sequential game given in Figure 4.9. We want to apply backward induction to the tree. 0 Vertex B is owned by player two, P2. The payoffs for P2 are 1 and 3, with 3 > 1, so the player picks R . Thus, the payoffs at B become (0, 3). 00 Next, vertex C is also owned by P2 with payoffs 1 and 0. Since 1 > 0, P2 picks L , and the payoffs are (4, 1). Player one, P1, owns A; the choice of L gives a payoff of 0 and R gives a payoff of 4; 4 > 0, so P1 chooses R. The final payoffs are (4, 1). 0 00 We claim that this strategy profile, { R } for P1 and { R ,L } is a Nash equilibrium. Notice that the 0 00 strategy profile gives a choice at each vertex. For the strategy { R ,L } fixed for P2, P1 has a maximal payoff by choosing { R }, ( 0 00 0 00 π1(R, { R ,L }) = 4 π1(R, { R ,L }) = 4 ≥ 0 00 π1(L, { R ,L }) = 0. 0 00 In the same way, for the strategy { R } fixed for P1, P2 has a maximal payoff by choosing { R ,L }, ( 00 0 00 π2(R, {∗,L }) = 1 π2(R, { R ,L }) = 1 ≥ 00 π2(R, {∗,R }) = 0, where ∗ means choose either L0 or R0. Since no change of choice by a player can increase that players own payoff, the strategy profile is called a Nash equilibrium. Notice that the above strategy profile is also a Nash equilibrium on each branch of the game tree, mainly starting at either B or starting at C.
    [Show full text]
  • Economics 201B Economic Theory (Spring 2021) Strategic Games
    Economics 201B Economic Theory (Spring 2021) Strategic Games Topics: terminology and notations (OR 1.7), games and solutions (OR 1.1-1.3), rationality and bounded rationality (OR 1.4-1.6), formalities (OR 2.1), best-response (OR 2.2), Nash equilibrium (OR 2.2), 2 2 examples × (OR 2.3), existence of Nash equilibrium (OR 2.4), mixed strategy Nash equilibrium (OR 3.1, 3.2), strictly competitive games (OR 2.5), evolution- ary stability (OR 3.4), rationalizability (OR 4.1), dominance (OR 4.2, 4.3), trembling hand perfection (OR 12.5). Terminology and notations (OR 1.7) Sets For R, ∈ ≥ ⇐⇒ ≥ for all . and ⇐⇒ ≥ for all and some . ⇐⇒ for all . Preferences is a binary relation on some set of alternatives R. % ⊆ From % we derive two other relations on : — strict performance relation and not  ⇐⇒ % % — indifference relation and ∼ ⇐⇒ % % Utility representation % is said to be — complete if , or . ∀ ∈ % % — transitive if , and then . ∀ ∈ % % % % can be presented by a utility function only if it is complete and transitive (rational). A function : R is a utility function representing if → % ∀ ∈ () () % ⇐⇒ ≥ % is said to be — continuous (preferences cannot jump...) if for any sequence of pairs () with ,and and , . { }∞=1 % → → % — (strictly) quasi-concave if for any the upper counter set ∈ { ∈ : is (strictly) convex. % } These guarantee the existence of continuous well-behaved utility function representation. Profiles Let be a the set of players. — () or simply () is a profile - a collection of values of some variable,∈ one for each player. — () or simply is the list of elements of the profile = ∈ { } − () for all players except . ∈ — ( ) is a list and an element ,whichistheprofile () .
    [Show full text]
  • Pure and Bayes-Nash Price of Anarchy for Generalized Second Price Auction
    Pure and Bayes-Nash Price of Anarchy for Generalized Second Price Auction Renato Paes Leme Eva´ Tardos Department of Computer Science Department of Computer Science Cornell University, Ithaca, NY Cornell University, Ithaca, NY [email protected] [email protected] Abstract—The Generalized Second Price Auction has for advertisements and slots higher on the page are been the main mechanism used by search companies more valuable (clicked on by more users). The bids to auction positions for advertisements on search pages. are used to determine both the assignment of bidders In this paper we study the social welfare of the Nash equilibria of this game in various models. In the full to slots, and the fees charged. In the simplest model, information setting, socially optimal Nash equilibria are the bidders are assigned to slots in order of bids, and known to exist (i.e., the Price of Stability is 1). This paper the fee for each click is the bid occupying the next is the first to prove bounds on the price of anarchy, and slot. This auction is called the Generalized Second Price to give any bounds in the Bayesian setting. Auction (GSP). More generally, positions and payments Our main result is to show that the price of anarchy is small assuming that all bidders play un-dominated in the Generalized Second Price Auction depend also on strategies. In the full information setting we prove a bound the click-through rates associated with the bidders, the of 1.618 for the price of anarchy for pure Nash equilibria, probability that the advertisement will get clicked on by and a bound of 4 for mixed Nash equilibria.
    [Show full text]
  • Extensive Form Games with Perfect Information
    Notes Strategy and Politics: Extensive Form Games with Perfect Information Matt Golder Pennsylvania State University Sequential Decision-Making Notes The model of a strategic (normal form) game suppresses the sequential structure of decision-making. When applying the model to situations in which players move sequentially, we assume that each player chooses her plan of action once and for all. She is committed to this plan, which she cannot modify as events unfold. The model of an extensive form game, by contrast, describes the sequential structure of decision-making explicitly, allowing us to study situations in which each player is free to change her mind as events unfold. Perfect Information Notes Perfect information describes a situation in which players are always fully informed about all of the previous actions taken by all players. This assumption is used in all of the following lecture notes that use \perfect information" in the title. Later we will also study more general cases where players may only be imperfectly informed about previous actions when choosing an action. Extensive Form Games Notes To describe an extensive form game with perfect information we need to specify the set of players and their preferences, just as for a strategic game. In addition, we also need to specify the order of the players' moves and the actions each player may take at each point (or decision node). We do so by specifying the set of all sequences of actions that can possibly occur, together with the player who moves at each point in each sequence. We refer to each possible sequence of actions (a1; a2; : : : ak) as a terminal history and to the function that denotes the player who moves at each point in each terminal history as the player function.
    [Show full text]
  • Bayesian Games Professors Greenwald 2018-01-31
    Bayesian Games Professors Greenwald 2018-01-31 We describe incomplete-information, or Bayesian, normal-form games (formally; no examples), and corresponding equilibrium concepts. 1 A Bayesian Model of Interaction A Bayesian, or incomplete information, game is a generalization of a complete-information game. Recall that in a complete-information game, the game is assumed to be common knowledge. In a Bayesian game, in addition to what is common knowledge, players may have private information. This private information is captured by the notion of an epistemic type, which describes a player’s knowledge. The Bayesian-game formalism makes two simplifying assumptions: • Any information that is privy to any of the players pertains only to utilities. In all realizations of a Bayesian game, the number of players and their actions are identical. • Players maintain beliefs about the game (i.e., about utilities) in the form of a probability distribution over types. Prior to receiving any private information, this probability distribution is common knowledge: i.e., it is a common prior. After receiving private information, players conditional on this information to update their beliefs. As a consequence of the com- mon prior assumption, any differences in beliefs can be attributed entirely to differences in information. Rational players are again assumed to maximize their utility. But further, they are assumed to update their beliefs when they obtain new information via Bayes’ rule. Thus, in a Bayesian game, in addition to players, actions, and utilities, there is a type space T = ∏i2[n] Ti, where Ti is the type space of player i.There is also a common prior F, which is a probability distribution over type profiles.
    [Show full text]
  • Games with Hidden Information
    Games with Hidden Information R&N Chapter 6 R&N Section 17.6 • Assumptions so far: – Two-player game : Player A and B. – Perfect information : Both players see all the states and decisions. Each decision is made sequentially . – Zero-sum : Player’s A gain is exactly equal to player B’s loss. • We are going to eliminate these constraints. We will eliminate first the assumption of “perfect information” leading to far more realistic models. – Some more game-theoretic definitions Matrix games – Minimax results for perfect information games – Minimax results for hidden information games 1 Player A 1 L R Player B 2 3 L R L R Player A 4 +2 +5 +2 L Extensive form of game: Represent the -1 +4 game by a tree A pure strategy for a player 1 defines the move that the L R player would make for every possible state that the player 2 3 would see. L R L R 4 +2 +5 +2 L -1 +4 2 Pure strategies for A: 1 Strategy I: (1 L,4 L) Strategy II: (1 L,4 R) L R Strategy III: (1 R,4 L) Strategy IV: (1 R,4 R) 2 3 Pure strategies for B: L R L R Strategy I: (2 L,3 L) Strategy II: (2 L,3 R) 4 +2 +5 +2 Strategy III: (2 R,3 L) L R Strategy IV: (2 R,3 R) -1 +4 In general: If N states and B moves, how many pure strategies exist? Matrix form of games Pure strategies for A: Pure strategies for B: Strategy I: (1 L,4 L) Strategy I: (2 L,3 L) Strategy II: (1 L,4 R) Strategy II: (2 L,3 R) Strategy III: (1 R,4 L) Strategy III: (2 R,3 L) 1 Strategy IV: (1 R,4 R) Strategy IV: (2 R,3 R) R L I II III IV 2 3 L R I -1 -1 +2 +2 L R 4 II +4 +4 +2 +2 +2 +5 +1 L R III +5 +1 +5 +1 IV +5 +1 +5 +1 -1 +4 3 Pure strategies for Player B Player A’s payoff I II III IV if game is played I -1 -1 +2 +2 with strategy I by II +4 +4 +2 +2 Player A and strategy III by III +5 +1 +5 +1 Player B for Player A for Pure strategies IV +5 +1 +5 +1 • Matrix normal form of games: The table contains the payoffs for all the possible combinations of pure strategies for Player A and Player B • The table characterizes the game completely, there is no need for any additional information about rules, etc.
    [Show full text]
  • Outline for Static Games of Complete Information I
    Outline for Static Games of Complete Information I. Definition of a game II. Examples III. Definition of Nash equilibrium IV. Examples, continued V. Iterated elimination of dominated strategies VI. Mixed strategies VII. Existence theorem on Nash equilibria VIII. The Hotelling model and extensions Copyright © 2004 by Lawrence M. Ausubel Definition: An n-player, static game of complete information consists of an n-tuple of strategy sets and an n-tuple of payoff functions, denoted by G = {S1, … , Sn; u1, … , un} Si, the strategy set of player i, is the set of all permissible moves for player i. We write si ∈ Si for one of player i’s strategies. ui, the payoff function of player i, is the utility, profit, etc. for player i, and depends on the strategies chosen by all the players: ui(s1, … , sn). Example: Prisoners’ Dilemma Prisoner II Remain Confess Silent Remain –1 , –1 –5 , 0 Prisoner Silent I Confess 0, –5 –4 ,–4 Example: Battle of the Sexes F Boxing Ballet Boxing 2, 1 0, 0 M Ballet 0, 0 1, 2 Definition: A Nash equilibrium of G (in pure strategies) consists of a strategy for every player with the property that no player can improve her payoff by unilaterally deviating: (s1*, … , sn*) with the property that, for every player i: ui (s1*, … , si–1*, si*, si+1*, … , sn*) ≥ ui (s1*, … , si–1*, si, si+1*, … , sn*) for all si ∈ Si. Equivalently, a Nash equilibrium is a mutual best response. That is, for every player i, si* is a solution to: ***** susssssiiiiin∈arg max{ (111 , ... ,−+ , , , ... , )} sSii∈ Example: Prisoners’ Dilemma Prisoner II
    [Show full text]
  • Efficiency and Welfare in Economies with Incomplete Information∗
    Efficiency and Welfare in Economies with Incomplete Information∗ George-Marios Angeletos Alessandro Pavan MIT, NBER, and FRB of Minneapolis Northwestern University September 2005. Abstract This paper examines a class of economies with externalities, strategic complementarity or substitutability, and incomplete information. We first characterize efficient allocations and com- pare them to equilibrium. We show how the optimal degree of coordination and the efficient use of information depend on the primitives of the environment, and how this relates to the welfare losses due to volatility and heterogeneity. We next examine the social value of informa- tion in equilibrium. When the equilibrium is efficient, welfare increases with the transparency of information if and only if agents’ actions are strategic complements. When the equilibrium is inefficient, additional effects are introduced by the interaction of externalities and informa- tion. We conclude with a few applications, including investment complementarities, inefficient fluctuations, and market competition. Keywords: Social value of information, coordination, higher-order beliefs, externalities, transparency. ∗An earlier version of this paper was entitled “Social Value of Information and Coordination.” For useful com- ments we thank Daron Acemoglu, Gadi Barlevi, Olivier Blanchard, Marco Bassetto, Christian Hellwig, Kiminori Matsuyama, Stephen Morris, Thomas Sargent, and especially Ivàn Werning. Email Addresses: [email protected], [email protected]. 1Introduction Is coordination socially desirable? Do markets use available information efficiently? Is the private collection of information good from a social perspective? What are the welfare effects of the public information disseminated by prices, market experts, or the media? Should central banks and policy makers disclose the information they collect and the forecasts they make about the economy in a transparent and timely fashion, or is there room for “constructive ambiguity”? These questions are non-trivial.
    [Show full text]
  • Introduction to Game Theory: a Discovery Approach
    Linfield University DigitalCommons@Linfield Linfield Authors Book Gallery Faculty Scholarship & Creative Works 2020 Introduction to Game Theory: A Discovery Approach Jennifer Firkins Nordstrom Linfield College Follow this and additional works at: https://digitalcommons.linfield.edu/linfauth Part of the Logic and Foundations Commons, Set Theory Commons, and the Statistics and Probability Commons Recommended Citation Nordstrom, Jennifer Firkins, "Introduction to Game Theory: A Discovery Approach" (2020). Linfield Authors Book Gallery. 83. https://digitalcommons.linfield.edu/linfauth/83 This Book is protected by copyright and/or related rights. It is brought to you for free via open access, courtesy of DigitalCommons@Linfield, with permission from the rights-holder(s). Your use of this Book must comply with the Terms of Use for material posted in DigitalCommons@Linfield, or with other stated terms (such as a Creative Commons license) indicated in the record and/or on the work itself. For more information, or if you have questions about permitted uses, please contact [email protected]. Introduction to Game Theory a Discovery Approach Jennifer Firkins Nordstrom Linfield College McMinnville, OR January 4, 2020 Version 4, January 2020 ©2020 Jennifer Firkins Nordstrom This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. Preface Many colleges and universities are offering courses in quantitative reasoning for all students. One model for a quantitative reasoning course is to provide students with a single cohesive topic. Ideally, such a topic can pique the curios- ity of students with wide ranging academic interests and limited mathematical background. This text is intended for use in such a course.
    [Show full text]
  • Extensive-Form Games with Perfect Information
    Extensive-Form Games with Perfect Information Jeffrey Ely April 22, 2015 Jeffrey Ely Extensive-Form Games with Perfect Information Extensive-Form Games With Perfect Information Recall the alternating offers bargaining game. It is an extensive-form game. (Described in terms of sequences of moves.) In this game, each time a player moves, he knows everything that has happened in the past. He can compute exactly how the game will end given any profile of continuation strategies Such a game is said to have perfect information. Jeffrey Ely Extensive-Form Games with Perfect Information Simple Extensive-Form Game With Perfect Information 1b ¨ H L ¨ H R ¨¨ HH 2r¨¨ HH2r AB @ C @ D @ @ r @r r @r 1, 0 2, 3 4, 1 −1, 0 Figure: An extensive game Jeffrey Ely Extensive-Form Games with Perfect Information Game Trees Nodes H Initial node. Terminal Nodes Z ⊂ H. Player correspondence N : H ⇒ f1, ... , ng (N(h) indicates who moves at h. They move simultaneously) Actions Ai (h) 6= Æ at each h where i 2 N(h). I Action profiles can be identified with branches in the tree. Payoffs ui : Z ! R. Jeffrey Ely Extensive-Form Games with Perfect Information Strategies A strategy for player i is a function si specifying si (h) 2 Ai (h) for each h such that i 2 N(h). Jeffrey Ely Extensive-Form Games with Perfect Information Strategic Form Given any profile of strategies s, there is a unique terminal node z(s) that will be reached. We define ui (s) = ui (z(s)). We have thus obtained a strategic-form representation of the game.
    [Show full text]