Combinatorial Games

Total Page:16

File Type:pdf, Size:1020Kb

Combinatorial Games Midterm 1 Stat155 Game Theory Lecture 11: Midterm 1 Review “This is an open book exam: you can use any printed or written material, but you cannot use a laptop, tablet, or phone (or any device that can communicate). There are three questions, each consisting of three parts. Peter Bartlett Each part of each question carries equal weight. Answer each question in the space provided.” September 29, 2016 1 / 35 2 / 35 Topics Definitions: Combinatorial games Combinatorial games Positions, moves, terminal positions, impartial/partisan, progressively A combinatorial game has: bounded Two players, Player I and Player II. Progressively bounded impartial and partisan games The sets N and P A set of positions X . Theorem: Someone can win For each player, a set of legal moves between positions, that is, a set Examples: Subtraction, Chomp, Nim, Rims, Staircase Nim, Hex of ordered pairs, (current position, next position): Zero sum games Payoff matrices, pure and mixed strategies, safety strategies MI , MII X X . Von Neumann’s minimax theorem ⊂ × Solving two player zero-sum games Saddle points Equalizing strategies Players alternately choose moves. Solving2 2 games × Play continues until some player cannot move. Dominated strategies 2 n and m 2 games Normal play: the player that cannot move loses the game. × × Principle of indifference Symmetry: Invariance under permutations 3 / 35 4 / 35 Definitions: Combinatorial games Impartial combinatorial games and winning strategies Terminology: An impartial game has the same set of legal moves for both players: MI = MII . Theorem A partisan game has different sets of legal moves for the players. In a progressively bounded impartial A terminal position for a player has no legal move to another position. combinatorial game under normal x is terminal for player I if there is no y X with( x, y) MI . play, X = N P. ∈ ∈ ∪ A combinatorial game is progressively bounded if, for every starting That is, from any initial position, one position x X , there is finite bound on the number of moves before of the players has a winning strategy. 0 ∈ the game ends. (That is, if B(x) denotes the maximum number of Proof: induction on number of moves before the game ends, then B(x) < .) ∞ moves until the end. A winning strategy for a player from position x: a mapping from non-terminal positions to legal moves that is guaranteed to result in a win for that player from that position. 5 / 35 6 / 35 Key Ideas: Progressively bounded impartial games Example: Nim k piles of chips. Remove some (positive) number of chips from some pile. A player wins when they take the last chip. Bouton’s Theorem P: Every move leads to N. A Nim position( x ,..., x ) is in P iff N: Some move leads to P (hence cannot contain terminal positions). 1 k the Nim-sum of its components is0. The Nim-sum of x = (x ,..., x ) is written x x x . 1 k 1 ⊕ 2 ⊕ · · · ⊕ k The binary representation of the Nim-sum is the bitwise sum, in modulo-two arithmetic, of the binary representations of the components of x. Proof: Show that Z := (x ,..., x ): x x = 0 is P: { 1 k 1 ⊕ · · · ⊕ k } Z must lead to Z c ; from Z c , there is a move into Z. 7 / 35 8 / 35 Partisan Games Topics Combinatorial games Positions, moves, terminal positions, impartial/partisan, progressively Recall: bounded Progressively bounded impartial and partisan games An impartial game has the same set of legal moves for both players: The sets N and P MI = MII . Theorem: Someone can win A partisan game has different sets of legal moves for the players. Examples: Subtraction, Chomp, Nim, Rims, Staircase Nim, Hex Zero sum games Theorem Payoff matrices, pure and mixed strategies, safety strategies Consider a progressively bounded partisan Von Neumann’s minimax theorem combinatorial game under normal play, with no ties Solving two player zero-sum games allowed. From any initial position, one of the players Saddle points Equalizing strategies has a winning strategy. Solving2 2 games × Dominated strategies 2 n and m 2 games × × Principle of indifference Symmetry: Invariance under permutations 9 / 35 10 / 35 Two-player zero-sum games Two-player zero-sum games Definitions Player I has m actions, 1, 2,..., m. Definitions Player II has n actions, 1, 2,..., n. A mixed strategy is a probability distribution over actions. m n The payoff matrix A R × represents the payoff to Player I: ∈ A mixed strategy for Player I is a vector a11 a12 a1n ··· x1 a a a 21 22 2n x m A = . ···. 2 m . x = . ∆m := x R : xi 0, xi = 1 . ∈ ∈ ≥ . ( i=1 ) am1 am2 amn X ··· xm If Player I chooses i and Player II chooses j, the payoff to Player I is a and the payoff to Player II is a . ij − ij The sum of the payoff to Player I and the payoff to Player II is0. 11 / 35 12 / 35 Two-player zero-sum games Two-player zero-sum games The expected payoff to Player I when Player I plays mixed strategy A mixed strategy for Player II is a vector x ∆m and Player II plays mixed strategy y ∆n is ∈ ∈ m n y1 n EI x EJ y aIJ = xi aij yj y2 ∼ ∼ n i=1 j=1 y = . ∆n := y R : yi 0, yi = 1 . X X . ∈ ( ∈ ≥ ) Xi=1 = x>Ay yn a a a y 11 12 ··· 1n 1 a21 a22 a2n y2 A pure strategy is a mixed strategy where one entry is1 and the = x1 x2 xm . ···. . ··· . .. others0. (This is a canonical basis vector ei .) a a a y m1 m2 ··· mn n 13 / 35 14 / 35 Two-player zero-sum games Two-player zero-sum games Von Neumann’s Minimax Theorem A safety strategy for Player I is an x ∆ that satisfies m n ∗ m For any two-person zero-sum game with payoff matrix A R × , ∈ ∈ min x∗>Ay = max min x>Ay. max min x>Ay = min max x>Ay. y ∆n x ∆m y ∆n x ∆m y ∆n y ∆n x ∆m ∈ ∈ ∈ ∈ ∈ ∈ ∈ This mixed strategy maximizes the worst case expected gain for Player I. We call the optimal expected payoff the value of the game, Similarly, a safety strategy for Player II is a y ∗ ∆n that satisfies V := max min x>Ay = min max x>Ay. x ∆m y ∆n y ∆n x ∆m ∈ ∈ ∈ ∈ ∈ max x>Ay ∗ = min max x>Ay. x ∆m y ∆n x ∆m LHS: Player I plays x ∆m first, then Player II responds with y ∆n. ∈ ∈ ∈ ∈ ∈ RHS: Player II plays y ∆ first, then Player I responds with x ∆ . ∈ n ∈ m Notice that we should always prefer to play last: This mixed strategy minimizes the worst case expected loss for Player II. max min x>Ay min max x>Ay. x ∆m y ∆n ≤ y ∆n x ∆m ∈ ∈ ∈ ∈ The astonishing part is that it does not help. 15 / 35 16 / 35 Two-player zero-sum games Topics Combinatorial games Von Neumann’s Minimax Theorem Positions, moves, terminal positions, impartial/partisan, progressively m n For any two-person zero-sum game with payoff matrix A R × , ∈ bounded Progressively bounded impartial and partisan games max min x>Ay = min max x>Ay. x ∆m y ∆n y ∆n x ∆m Zero sum games ∈ ∈ ∈ ∈ Payoff matrices, pure and mixed strategies, safety strategies Von Neumann’s minimax theorem Solving two player zero-sum games Safety strategies are optimal strategies: Saddle points For safety strategies x∗ for Player I and y ∗ for Player II, Equalizing strategies Solving2 2 games × min x∗>Ay = max x>Ay ∗ = x∗>Ay ∗ = V . Dominated strategies y ∆n x ∆m 2 n and m 2 games ∈ ∈ × × Principle of indifference Symmetry: Invariance under permutations 17 / 35 18 / 35 Saddle points Saddle points Definition A pair( i ∗, j∗) 1,..., m 1,..., n is a saddle point for a payoff m∈n { } × { } matrix A R × if Theorem ∈ m n If( i ∗, j∗) is a saddle point for a payoff matrix A R × , then max aij = ai j = min ai j . ∈ i ∗ ∗ ∗ j ∗ ei ∗ is an optimal strategy for Player I, ej∗ is an optimal strategy for Player II, and If Player I plays i and Player II plays j , neither player has an ∗ ∗ the value of the game is ai ∗j∗ . incentive to change. Think of these as locally optimal strategies for the players. They are also globally optimal strategies. 19 / 35 20 / 35 2 2 games Dominated pure strategies × How to solve a2 2 game × 1 Check for a saddle point. (Is the max of row mins= min of column maxes?) 2 If there are no saddle points, find equalizing strategies. Definition Equalizing strategies satisfy: A pure strategy ei for Player I is dominated by ei 0 in payoff matrix A if, for all j 1,..., n , ∈ { } x1a11 + (1 x1)a21 = x1a12 + (1 x1)a22, a a . − − ij ≤ i 0j y a + (1 y )a = y a + (1 y )a . 1 11 − 1 12 1 21 − 1 22 Solving gives a21 a22 x1 = − , a21 a22 + a12 a11 − a a − y = 12 − 22 . 1 a a + a a 12 − 22 21 − 11 21 / 35 22 / 35 Solving2 n games Principle of indifference × Payoff matrix Theorem m n Suppose a game with payoff matrix A R × has value V . If x ∆m 2315 ∈ ∈ and y ∆ are optimal strategies for Players I and II, then 4160 n ∈ m n for all j, x a V , for all i, a y V , l lj ≥ il l ≤ The maximum occurs at the Xl=1 Xl=1 intersection of the lines m n corresponding to columns 2 and 3.
Recommended publications
  • Alpha-Beta Pruning
    Alpha-Beta Pruning Carl Felstiner May 9, 2019 Abstract This paper serves as an introduction to the ways computers are built to play games. We implement the basic minimax algorithm and expand on it by finding ways to reduce the portion of the game tree that must be generated to find the best move. We tested our algorithms on ordinary Tic-Tac-Toe, Hex, and 3-D Tic-Tac-Toe. With our algorithms, we were able to find the best opening move in Tic-Tac-Toe by only generating 0.34% of the nodes in the game tree. We also explored some mathematical features of Hex and provided proofs of them. 1 Introduction Building computers to play board games has been a focus for mathematicians ever since computers were invented. The first computer to beat a human opponent in chess was built in 1956, and towards the late 1960s, computers were already beating chess players of low-medium skill.[1] Now, it is generally recognized that computers can beat even the most accomplished grandmaster. Computers build a tree of different possibilities for the game and then work backwards to find the move that will give the computer the best outcome. Al- though computers can evaluate board positions very quickly, in a game like chess where there are over 10120 possible board configurations it is impossible for a computer to search through the entire tree. The challenge for the computer is then to find ways to avoid searching in certain parts of the tree. Humans also look ahead a certain number of moves when playing a game, but experienced players already know of certain theories and strategies that tell them which parts of the tree to look.
    [Show full text]
  • Game Theory with Translucent Players
    Game Theory with Translucent Players Joseph Y. Halpern and Rafael Pass∗ Cornell University Department of Computer Science Ithaca, NY, 14853, U.S.A. e-mail: fhalpern,[email protected] November 27, 2012 Abstract A traditional assumption in game theory is that players are opaque to one another—if a player changes strategies, then this change in strategies does not affect the choice of other players’ strategies. In many situations this is an unrealistic assumption. We develop a framework for reasoning about games where the players may be translucent to one another; in particular, a player may believe that if she were to change strategies, then the other player would also change strategies. Translucent players may achieve significantly more efficient outcomes than opaque ones. Our main result is a characterization of strategies consistent with appro- priate analogues of common belief of rationality. Common Counterfactual Belief of Rationality (CCBR) holds if (1) everyone is rational, (2) everyone counterfactually believes that everyone else is rational (i.e., all players i be- lieve that everyone else would still be rational even if i were to switch strate- gies), (3) everyone counterfactually believes that everyone else is rational, and counterfactually believes that everyone else is rational, and so on. CCBR characterizes the set of strategies surviving iterated removal of minimax dom- inated strategies: a strategy σi is minimax dominated for i if there exists a 0 0 0 strategy σ for i such that min 0 u (σ ; µ ) > max u (σ ; µ ). i µ−i i i −i µ−i i i −i ∗Halpern is supported in part by NSF grants IIS-0812045, IIS-0911036, and CCF-1214844, by AFOSR grant FA9550-08-1-0266, and by ARO grant W911NF-09-1-0281.
    [Show full text]
  • The Effective Minimax Value of Asynchronously Repeated Games*
    Int J Game Theory (2003) 32: 431–442 DOI 10.1007/s001820300161 The effective minimax value of asynchronously repeated games* Kiho Yoon Department of Economics, Korea University, Anam-dong, Sungbuk-gu, Seoul, Korea 136-701 (E-mail: [email protected]) Received: October 2001 Abstract. We study the effect of asynchronous choice structure on the possi- bility of cooperation in repeated strategic situations. We model the strategic situations as asynchronously repeated games, and define two notions of effec- tive minimax value. We show that the order of players’ moves generally affects the effective minimax value of the asynchronously repeated game in significant ways, but the order of moves becomes irrelevant when the stage game satisfies the non-equivalent utilities (NEU) condition. We then prove the Folk Theorem that a payoff vector can be supported as a subgame perfect equilibrium out- come with correlation device if and only if it dominates the effective minimax value. These results, in particular, imply both Lagunoff and Matsui’s (1997) result and Yoon (2001)’s result on asynchronously repeated games. Key words: Effective minimax value, folk theorem, asynchronously repeated games 1. Introduction Asynchronous choice structure in repeated strategic situations may affect the possibility of cooperation in significant ways. When players in a repeated game make asynchronous choices, that is, when players cannot always change their actions simultaneously in each period, it is quite plausible that they can coordinate on some particular actions via short-run commitment or inertia to render unfavorable outcomes infeasible. Indeed, Lagunoff and Matsui (1997) showed that, when a pure coordination game is repeated in a way that only *I thank three anonymous referees as well as an associate editor for many helpful comments and suggestions.
    [Show full text]
  • 570: Minimax Sample Complexity for Turn-Based Stochastic Game
    Minimax Sample Complexity for Turn-based Stochastic Game Qiwen Cui1 Lin F. Yang2 1School of Mathematical Sciences, Peking University 2Electrical and Computer Engineering Department, University of California, Los Angeles Abstract guarantees are rather rare due to complex interaction be- tween agents that makes the problem considerably harder than single agent reinforcement learning. This is also known The empirical success of multi-agent reinforce- as non-stationarity in MARL, which means when multi- ment learning is encouraging, while few theoret- ple agents alter their strategies based on samples collected ical guarantees have been revealed. In this work, from previous strategy, the system becomes non-stationary we prove that the plug-in solver approach, proba- for each agent and the improvement can not be guaranteed. bly the most natural reinforcement learning algo- One fundamental question in MBRL is that how to design rithm, achieves minimax sample complexity for efficient algorithms to overcome non-stationarity. turn-based stochastic game (TBSG). Specifically, we perform planning in an empirical TBSG by Two-players turn-based stochastic game (TBSG) is a two- utilizing a ‘simulator’ that allows sampling from agents generalization of Markov decision process (MDP), arbitrary state-action pair. We show that the em- where two agents choose actions in turn and one agent wants pirical Nash equilibrium strategy is an approxi- to maximize the total reward while the other wants to min- mate Nash equilibrium strategy in the true TBSG imize it. As a zero-sum game, TBSG is known to have and give both problem-dependent and problem- Nash equilibrium strategy [Shapley, 1953], which means independent bound.
    [Show full text]
  • Games with Hidden Information
    Games with Hidden Information R&N Chapter 6 R&N Section 17.6 • Assumptions so far: – Two-player game : Player A and B. – Perfect information : Both players see all the states and decisions. Each decision is made sequentially . – Zero-sum : Player’s A gain is exactly equal to player B’s loss. • We are going to eliminate these constraints. We will eliminate first the assumption of “perfect information” leading to far more realistic models. – Some more game-theoretic definitions Matrix games – Minimax results for perfect information games – Minimax results for hidden information games 1 Player A 1 L R Player B 2 3 L R L R Player A 4 +2 +5 +2 L Extensive form of game: Represent the -1 +4 game by a tree A pure strategy for a player 1 defines the move that the L R player would make for every possible state that the player 2 3 would see. L R L R 4 +2 +5 +2 L -1 +4 2 Pure strategies for A: 1 Strategy I: (1 L,4 L) Strategy II: (1 L,4 R) L R Strategy III: (1 R,4 L) Strategy IV: (1 R,4 R) 2 3 Pure strategies for B: L R L R Strategy I: (2 L,3 L) Strategy II: (2 L,3 R) 4 +2 +5 +2 Strategy III: (2 R,3 L) L R Strategy IV: (2 R,3 R) -1 +4 In general: If N states and B moves, how many pure strategies exist? Matrix form of games Pure strategies for A: Pure strategies for B: Strategy I: (1 L,4 L) Strategy I: (2 L,3 L) Strategy II: (1 L,4 R) Strategy II: (2 L,3 R) Strategy III: (1 R,4 L) Strategy III: (2 R,3 L) 1 Strategy IV: (1 R,4 R) Strategy IV: (2 R,3 R) R L I II III IV 2 3 L R I -1 -1 +2 +2 L R 4 II +4 +4 +2 +2 +2 +5 +1 L R III +5 +1 +5 +1 IV +5 +1 +5 +1 -1 +4 3 Pure strategies for Player B Player A’s payoff I II III IV if game is played I -1 -1 +2 +2 with strategy I by II +4 +4 +2 +2 Player A and strategy III by III +5 +1 +5 +1 Player B for Player A for Pure strategies IV +5 +1 +5 +1 • Matrix normal form of games: The table contains the payoffs for all the possible combinations of pure strategies for Player A and Player B • The table characterizes the game completely, there is no need for any additional information about rules, etc.
    [Show full text]
  • Recursion & the Minimax Algorithm Winning Tic-Tac-Toe How to Pass The
    Recursion & the Minimax Algorithm Winning Tic-tac-toe Key to Acing Computer Science Anticipate the implications of your move. If you understand everything, ace your computer science course. Avoid moves which could enable your opponent to win. Otherwise, study & program something you don't understand, Attempt moves which would then see force your opponent to lose, so you win. Key to Acing Computer Science CompSci 4 Recursion & Minimax 27.1 CompSci 4 Recursion & Minimax 27.2 How to pass the buck: recursion! How could this approach work? Imagine this: What actually happens is that your friends doing most of You have all sorts of friends who are willing to help you do your work have friends of their own. Everyone your work with two caveats: eventually does a small part of the work and piece it o They won't do all of your work back together to do the work in its entirity. o They will only do the type of work you have to do In the last example, 100 total people would be working Example: on the homework, each person doing only one problem How would you do your homework consisting of 100 calculus each and passing on the rest. problems? Important: the person to do the last homework problem Easy – do the first problem yourself and ask your friend to do does it entirely without help. the rest CompSci 4 Recursion & Minimax 27.3 CompSci 4 Recursion & Minimax 27.4 The Parts of Recursion Example: Finding the Maximum ÿ Base Case – this is the simplest form of the problem ÿ Imagine you are given a stack of 1000 unsorted papers which can be solved directly.
    [Show full text]
  • Maximin Equilibrium∗
    Maximin equilibrium∗ Mehmet ISMAILy March, 2014. This version: June, 2014 Abstract We introduce a new theory of games which extends von Neumann's theory of zero-sum games to nonzero-sum games by incorporating common knowledge of individual and collective rationality of the play- ers. Maximin equilibrium, extending Nash's value approach, is based on the evaluation of the strategic uncertainty of the whole game. We show that maximin equilibrium is invariant under strictly increasing transformations of the payoffs. Notably, every finite game possesses a maximin equilibrium in pure strategies. Considering the games in von Neumann-Morgenstern mixed extension, we demonstrate that the maximin equilibrium value is precisely the maximin (minimax) value and it coincides with the maximin strategies in two-player zero-sum games. We also show that for every Nash equilibrium that is not a maximin equilibrium there exists a maximin equilibrium that Pareto dominates it. In addition, a maximin equilibrium is never Pareto dominated by a Nash equilibrium. Finally, we discuss maximin equi- librium predictions in several games including the traveler's dilemma. JEL-Classification: C72 ∗I thank Jean-Jacques Herings for his feedback. I am particularly indebted to Ronald Peeters for his continuous comments and suggestions about the material in this paper. I am also thankful to the participants of the MLSE seminar at Maastricht University. Of course, any mistake is mine. yMaastricht University. E-mail: [email protected]. 1 Introduction In their ground-breaking book, von Neumann and Morgenstern (1944, p. 555) describe the maximin strategy1 solution for two-player games as follows: \There exists precisely one solution.
    [Show full text]
  • Maximin Equilibrium
    Munich Personal RePEc Archive Maximin equilibrium Ismail, Mehmet Maastricht University October 2014 Online at https://mpra.ub.uni-muenchen.de/97322/ MPRA Paper No. 97322, posted 04 Dec 2019 13:32 UTC Mehmet Ismail Maximin equilibrium RM/14/037 Maximin equilibrium∗ Mehmet ISMAIL† First version March, 2014. This version: October, 2014 Abstract We introduce a new concept which extends von Neumann and Mor- genstern’s maximin strategy solution by incorporating ‘individual ra- tionality’ of the players. Maximin equilibrium, extending Nash’s value approach, is based on the evaluation of the strategic uncertainty of the whole game. We show that maximin equilibrium is invariant under strictly increasing transformations of the payoffs. Notably, every finite game possesses a maximin equilibrium in pure strategies. Considering the games in von Neumann-Morgenstern mixed extension, we demon- strate that the maximin equilibrium value is precisely the maximin (minimax) value and it coincides with the maximin strategies in two- person zerosum games. We also show that for every Nash equilibrium that is not a maximin equilibrium there exists a maximin equilibrium that Pareto dominates it. Hence, a strong Nash equilibrium is always a maximin equilibrium. In addition, a maximin equilibrium is never Pareto dominated by a Nash equilibrium. Finally, we discuss max- imin equilibrium predictions in several games including the traveler’s dilemma. JEL-Classification: C72 Keywords: Non-cooperative games, maximin strategy, zerosum games. ∗I thank Jean-Jacques Herings
    [Show full text]
  • Foundations of AI 5
    Foundations of AI 5. Board Games Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard and Luc De Raedt SA -1 Contents Board Games Minimax Search Alpha-Beta Search Games with an Element of Chance State of the Art 05/2 Why Board Games? Board games are one of the oldest branches of AI (Shannon und Turing 1950). Board games present a very abstract and pure form of competition between two opponents and clearly require a form on “intelligence”. The states of a game are easy to represent. The possible actions of the players are well defined. Realization of the game as a search problem The world states are fully accessible It is nonetheless a contingency problem, because the characteristics of the opponent are not known in advance. 05/3 Problems Board games are not only difficult because they are contingency problems , but also because the search trees can become astronomically large . Examples : • Chess : On average 35 possible actions from every position, 100 possible moves 35 100 nodes in the search tree (with “only” ca. 10 40 legal chess positions). • Go : On average 200 possible actions with ca. 300 moves 200 300 nodes. Good game programs have the properties that they • delete irrelevant branches of the game tree , • use good evaluation functions for in-between states , and • look ahead as many moves as possible . 05/4 Terminology of Two-Person Board Games Players are MAX and MIN, where MAX begins. Initial position (e.g., board arrangement) Operators (= legal moves) Termination test , determines when the game is over.
    [Show full text]
  • Minimax Theory
    Minimax Theory 1 Introduction When solving a statistical learning problem, there are often many procedures to choose from. This leads to the following question: how can we tell if one statistical learning procedure is better than another? One answer is provided by minimax theory which is a set of techniques for finding the minimum, worst case behavior of a procedure. 2 Definitions and Notation Let P be a set of distributions and let X1;:::;Xn be a sample from some distribution P 2 P. Let θ(P ) be some function of P . For example, θ(P ) could be the mean of P , the variance of P or the density of P . Let θb = θb(X1;:::;Xn) denote an estimator. Given a metric d, the minimax risk is Rn ≡ Rn(P) = inf sup EP [d(θ;b θ(P ))] (1) θb P 2P where the infimum is over all estimators. The sample complexity is n o n(, P) = min n : Rn(P) ≤ : (2) Example 1 Suppose that P = fN(θ; 1) : θ 2 Rg where N(θ; 1) denotes a Gaussian with mean θ and variance 1. Consider estimating θ with the metric d(a; b) = (a − b)2. The minimax risk is 2 Rn = inf sup EP [(θb− θ) ]: (3) θb P 2P In this example, θ is a scalar. Example 2 Let (X1;Y1);:::; (Xn;Yn) be a sample from a distribution P . Let m(x) = R EP (Y jX = x) = y dP (yjX = x) be the regression function. In this case, we might use R 2 the metric d(m1; m2) = (m1(x) − m2(x)) dx in which case the minimax risk is Z 2 Rn = inf sup EP (mb (x) − m(x)) : (4) mb P 2P In this example, θ is a function.
    [Show full text]
  • John Von Neumann's Work in the Theory of Games and Mathematical Economics
    JOHN VON NEUMANN'S WORK IN THE THEORY OF GAMES AND MATHEMATICAL ECONOMICS H. W. KUHN AND A. W. TUCKER Of the many areas of mathematics shaped by his genius, none shows more clearly the influence of John von Neumann than the Theory of Games. This modern approach to problems of competition and cooperation was given a broad foundation in his superlative paper of 1928 [A].1 In scope and youthful vigor this work can be compared only to his papers of the same period on the axioms of set theory and the mathematical foundations of quantum mechanics. A decade later, when the Austrian economist Oskar Morgenstern came to Princeton, von Neumann's interest in the theory was reawakened. The result of their active and intensive collaboration during the early years of World War II was the treatise Theory of games and economic behavior [D], in which the basic structure of the 1928 paper is elaborated and extended. Together, the paper and treatise contain a remarkably complete outline of the subject as we know it today, and every writer in the field draws in some measure upon concepts which were there united into a coherent theory. The crucial innovation of von Neumann, which was to be both the keystone of his Theory of Games and the central theme of his later research in the area, was the assertion and proof of the Minimax Theorem. Ideas of pure and randomized strategies had been intro­ duced earlier, especially by Êmile Borel [3]. However, these efforts were restricted either to individual examples or, at best, to zero-sum two-person games with skew-symmetric payoff matrices.
    [Show full text]
  • 3B Extensive-Form Games.Pptx
    Introduction to Game Theory 3b. Extensive-Form Games Dana Nau University of Maryland Nau: Game Theory 1 The Sharing Game Suppose agents 1 and 2 are two children Someone offers them two cookies, but only if they can agree how to share them Agent 1 chooses one of the following options: Agent 1 gets 2 cookies, agent 2 gets 0 cookies They each get 1 cookie Agent 1 gets 0 cookies, agent 2 gets 2 cookies Agent 2 chooses to accept or reject the split: 1 Accept => 2-0 0-2 they each get their cookies(s) 1-1 Otherwise, neither gets any 2 2 2 no yes no yes no yes (0,0) (2,0) (0,0) (1,1) (0,0) (2,0) Nau: Game Theory 2 Extensive Form The sharing game is a game in extensive form A game representation that makes the temporal structure explicit Doesn’t assume agents act simultaneously Extensive form can be converted to normal form, so previous results carry over But there are additional results that depend on the temporal structure In a perfect-information game, the extensive form is a game tree: Nonterminal node = place where an agent chooses an action Edge = an available action or move 1 Terminal node = a final outcome 2-0 0-2 1-1 • At each terminal node h, each 2 2 2 agent i has a utility ui(h) no yes no yes no yes (0,0) (2,0) (0,0) (1,1) (0,0) (0,2) Nau: Game Theory 3 Pure Strategies Pure strategy for agent i in a perfect-information game: specifies which action to take at every node where it’s i’s choice Sharing game: Agent 1 has 3 pure strategies: S1 = {2-0, 1-1, 0-2} Agent 2 has 8 pure strategies: S2 = {(yes, yes, yes), (yes, yes, no), (yes, no, yes), (yes, no, no), (no, yes, yes), (no, yes, no), 1 2-0 0-2 (no, no, yes), (no, no, no)} 1-1 2 2 2 no yes no yes no yes (0,0) (2,0) (0,0) (1,1) (0,0) (0,2) Nau: Game Theory 4 1 Extensive form vs.
    [Show full text]