Game Theory Lecture Notes

Total Page:16

File Type:pdf, Size:1020Kb

Game Theory Lecture Notes Game Theory: Penn State Math 486 Lecture Notes Version 2.1.1 Christopher Griffin « 2010-2021 Licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License With Major Contributions By: James Fan George Kesidis and Other Contributions By: Arlan Stutler Sarthak Shah Contents List of Figuresv Preface xi 1. Using These Notes xi 2. An Overview of Game Theory xi Chapter 1. Probability Theory and Games Against the House1 1. Probability1 2. Random Variables and Expected Values6 3. Conditional Probability8 4. The Monty Hall Problem 11 Chapter 2. Game Trees and Extensive Form 15 1. Graphs and Trees 15 2. Game Trees with Complete Information and No Chance 18 3. Game Trees with Incomplete Information 22 4. Games of Chance 24 5. Pay-off Functions and Equilibria 26 Chapter 3. Normal and Strategic Form Games and Matrices 37 1. Normal and Strategic Form 37 2. Strategic Form Games 38 3. Review of Basic Matrix Properties 40 4. Special Matrices and Vectors 42 5. Strategy Vectors and Matrix Games 43 Chapter 4. Saddle Points, Mixed Strategies and the Minimax Theorem 45 1. Saddle Points 45 2. Zero-Sum Games without Saddle Points 48 3. Mixed Strategies 50 4. Mixed Strategies in Matrix Games 53 5. Dominated Strategies and Nash Equilibria 54 6. The Minimax Theorem 59 7. Finding Nash Equilibria in Simple Games 64 8. A Note on Nash Equilibria in General 66 Chapter 5. An Introduction to Optimization and the Karush-Kuhn-Tucker Conditions 69 1. A General Maximization Formulation 70 2. Some Geometry for Optimization 72 3. Gradients, Constraints and Optimization 76 4. Convex Sets and Combinations 78 iii iv CONTENTS 5. Convex and Concave Functions 79 6. Karush-Kuhn-Tucker Conditions 80 7. Relating Back to Game Theory 83 Chapter 6. Zero-Sum Matrix Games with Linear Programming 85 1. Linear Programs 85 2. Intuition on the Solution of Linear Programs 86 3. A Linear Program for Zero-Sum Game Players 90 4. Matrix Notation, Slack and Surplus Variables for Linear Programming 93 5. Solving Linear Programs by Computer 95 6. Duality and Optimality Conditions for Zero-Sum Game Linear Programs 98 Chapter 7. Quadratic Programs and General Sum Games 105 1. Introduction to Quadratic Programming 105 2. Solving QP's by Computer 106 3. General Sum Games and Quadratic Programming 106 Chapter 8. Nash's Bargaining Problem and Cooperative Games 115 1. Payoff Regions in Two Player Games 115 2. Collaboration and Multi-criteria Optimization 119 3. Nash's Bargaining Axioms 122 4. Nash's Bargaining Theorem 123 Chapter 9. A Short Introduction to N-Player Cooperative Games 131 1. Motivating Cooperative Games 131 2. Basic Results on Coalition Games 132 3. Division of Payoff to the Coalition 133 4. The Core 134 5. Shapley Values 136 Appendix A. Utility Theory 139 1. Decision Making Under Certainty 139 2. Advanced Decision Making under Uncertainty 146 Bibliography 151 List of Figures 0.1 There are several sub-disciplines within Game Theory. Each one has its own unique sets of problems and applications. We will study Classical Game Theory, which focuses on questions like, \What is my best decision in a given economic scenario, where a reward function provides a way for me to understand how my decision will impact my result." We may also investigate Combinatorial Game Theory, which is interested in games like Chess or Go. If there's time, we'll study Evolutionary Game Theory, which is interesting in its own right. xiii 1.1 An (American) roulette wheel is shown above. A French roulette wheel lacks the 00 pocket. This image was obtained from http://www.math.uah.edu/stat/ games/Roulette.html.7 1.2 Example card counting strategies. This table is adapted from https: //en.wikipedia.org/wiki/Card_counting. 10 1.3 You are sitting at a Black Jack Table. The dealer holds a king and something. You hold a 7 and a King. Do you hit? 10 1.4 The Monty Hall Problem is a multi-stage decision problem whose solution relies on conditional probability. The stages of decision making are shown in the diagram. We assume that the prizes are randomly assigned to the doors. We can't see this step{so we've adorned this decision with a square box. We'll discuss these boxes more when we talk about game trees. You the player must first choose a door. Lastly, you must decide whether or not to switch doors having been shown a door that is incorrect. 12 2.1 Digraphs on 3 Vertices: There are 64 = 26 distinct graphs on three vertices. The increased number of edges graphs is caused by the fact that the edges are now directed. 16 2.2 Two Paths: We illustrate two paths in a digraph on three vertices. 16 2.3 Directed Tree: We illustrate a directed tree. Every directed tree has a unique vertex called the root. The root is connected by a directed path to every other vertex in the directed tree. 17 2.4 Sub Tree: We illustrate a sub-tree. This tree is the collection of all nodes that are descended from a vertex u. 18 2.5 Rock-Paper-Scissors with Perfect Information: Player 1 moves first and holds up a symbol for either rock, paper or scissors. This is illustrated by the three edges leaving the root node, which is assigned to Player 1. Player 2 then holds up a symbol for either rock, paper or scissors. Payoffs are assigned to Player 1 and 2 at terminal nodes. The index of the payoff vector corresponds to the players. 20 v vi LIST OF FIGURES 2.6 New Guinea is located in the south pacific and was a major region of contention during World War II. The northern half was controlled by Japan through 1943, while the southern half was controlled by the Allies. (Image created from Wikipedia (http://en.wikipedia.org/wiki/File:LocationNewGuinea.svg), originally sourced from http://commons.wikimedia.org/wiki/File: LocationPapuaNewGuinea.svg. 20 2.7 The game tree for the Battle of the Bismark Sea. The Japanese could choose to sail either north or south of New Britain. The Americans (Allies) could choose to concentrate their search efforts on either the northern or southern routes. Given this game tree, the Americans would always choose to search the North if they knew the Japanese had chosen to sail on the north side of New Britain; alternatively, they would search the south route, if they knew the Japanese had taken that. Assuming the Americans have perfect intelligence, the Japanese would always choose to sail the northern route as in this instance they would expose themselves to only 2 days of bombing as opposed to 3 with the southern route. 21 2.8 Simple tic-tac-toe: Players in this case try to get two in a row. 22 2.9 The game tree for the Battle of the Bismark Sea with incomplete information. Obviously Kenney could not have known a priori which path the Japanese would choose to sail. He could have reasoned (as they might) that there best plan was to sail north, but he wouldn't really know. We can capture this fact by showing that when Kenney chooses his move, he cannot distinguish between the two intermediate nodes that belong to the Allies. 24 2.10 Poker: The root node of the game tree is controlled by Nature. At this node, a single random card is dealt to Player 1. Player 1 can then decide whether to end the game by folding (and thus receiving a payoff or not) or continuing the game by raising. At this point, Player 2 can then decide whether to call or fold, thus potentially receiving a payoff. 26 2.11 Reduced Red Black Poker: We are told that Player 1 receives a red card. The resulting game tree is substantially simpler. Because the information set on Player 2 controlled nodes indicated a lack of knowledge of Player 1's card, we can see that this sub-game is now a complete information game. 27 2.12 A unique path through the game tree of the Battle of the Bismark Sea. Since each player determines a priori the unique edge he/she will select when confronted with a specific information set, a path through the tree can be determined from these selections. 28 2.13 The probability space constructed from fixed player strategies in a game of chance. The strategy space is constructed from the unique choices determined by the strategy of the players and the independent random events that are determined by the chance moves. 30 2.14 The probability space constructed from fixed player strategies in a game of chance. The strategy space is constructed from the unique choices determined by the strategy of the players and the independent random events that are determined by the chance moves. Note in this example that constructing the LIST OF FIGURES vii probabilities of the various events requires multiplying the probabilities of the chance moves in each path. 31 2.15 Game tree paths derived from the Simple Poker Game as a result of the strategy (Fold, Fold). The probability of each of these paths is 1=2. 32 2.16 The game tree for the Battle of the Bismark Sea. If the Japanese sail north, the best move for the Allies is to search north. If the Japanese sail south, then the best move for the Allies is to search south. The Japanese, observing the payoffs, note that given these best strategies for the Allies, there best course of action is to sail North.
Recommended publications
  • A History of Von Neumann and Morgenstern's Theory of Games, 12Pt Escolha Sob Incerteza Ax
    Marcos Thiago Graciani Axiomatic choice under uncertainty: a history of von Neumann and Morgenstern’s Theory of Games Escolha sob incerteza axiomática: uma história do Theory of Games de von Neumann e Morgenstern São Paulo 2019 Prof. Dr. Vahan Agopyan Reitor da Universidade de São Paulo Prof. Dr. Fábio Frezatti Diretor da Faculdade de Economia, Administração e Contabilidade Prof. Dr. José Carlos de Souza Santos Chefe do Departamento de Economia Prof. Dr. Ariaster Baumgratz Chimeli Coordenador do Programa de Pós-Graduação em Economia Marcos Thiago Graciani Axiomatic choice under uncertainty: a history of von Neumann and Morgenstern’s Theory of Games Escolha sob incerteza axiomática: uma história do Theory of Games de von Neumann e Morgenstern Dissertação de Mestrado apresentada ao Departamento de Economia da Faculdade de Economia, Administração e Contabilidade da Universidade de São Paulo (FEA-USP) como requisito parcial à obtenção do título de Mestre em Ciências. Área de Concentração: Teoria Econômica. Universidade de São Paulo — USP Faculdade de Economia, Administração e Contabilidade Programa de Pós-Graduação Orientador: Pedro Garcia Duarte Versão Corrigida (A versão original está disponível na biblioteca da Faculdade de Economia, Administração e Contabilidade.) São Paulo 2019 FICHA CATALOGRÁFICA Elaborada pela Seção de Processamento Técnico do SBD/FEA com dados inseridos pelo autor. Graciani, Marcos Thiago. Axiomatic choice under uncertainty: a history of von Neumann and Morgenstern’s Theory of Games / Marcos Thiago Graciani. – São Paulo, 2019. 133p. Dissertação (Mestrado) – Universidade de São Paulo, 2019. Orientador: Pedro Garcia Duarte. 1. História da teoria dos jogos 2. Axiomática 3. Incerteza 4. Von Neumann e Morgenstern I.
    [Show full text]
  • Frequently Asked Questions in Mathematics
    Frequently Asked Questions in Mathematics The Sci.Math FAQ Team. Editor: Alex L´opez-Ortiz e-mail: [email protected] Contents 1 Introduction 4 1.1 Why a list of Frequently Asked Questions? . 4 1.2 Frequently Asked Questions in Mathematics? . 4 2 Fundamentals 5 2.1 Algebraic structures . 5 2.1.1 Monoids and Groups . 6 2.1.2 Rings . 7 2.1.3 Fields . 7 2.1.4 Ordering . 8 2.2 What are numbers? . 9 2.2.1 Introduction . 9 2.2.2 Construction of the Number System . 9 2.2.3 Construction of N ............................... 10 2.2.4 Construction of Z ................................ 10 2.2.5 Construction of Q ............................... 11 2.2.6 Construction of R ............................... 11 2.2.7 Construction of C ............................... 12 2.2.8 Rounding things up . 12 2.2.9 What’s next? . 12 3 Number Theory 14 3.1 Fermat’s Last Theorem . 14 3.1.1 History of Fermat’s Last Theorem . 14 3.1.2 What is the current status of FLT? . 14 3.1.3 Related Conjectures . 15 3.1.4 Did Fermat prove this theorem? . 16 3.2 Prime Numbers . 17 3.2.1 Largest known Mersenne prime . 17 3.2.2 Largest known prime . 17 3.2.3 Largest known twin primes . 18 3.2.4 Largest Fermat number with known factorization . 18 3.2.5 Algorithms to factor integer numbers . 18 3.2.6 Primality Testing . 19 3.2.7 List of record numbers . 20 3.2.8 What is the current status on Mersenne primes? .
    [Show full text]
  • Game Theoretic Interaction and Decision: a Quantum Analysis
    games Article Game Theoretic Interaction and Decision: A Quantum Analysis Ulrich Faigle 1 and Michel Grabisch 2,* 1 Mathematisches Institut, Universität zu Köln, Weyertal 80, 50931 Köln, Germany; [email protected] 2 Paris School of Economics, University of Paris I, 106-112, Bd. de l’Hôpital, 75013 Paris, France * Correspondence: [email protected]; Tel.: +33-144-07-8744 Received: 12 July 2017; Accepted: 21 October 2017; Published: 6 November 2017 Abstract: An interaction system has a finite set of agents that interact pairwise, depending on the current state of the system. Symmetric decomposition of the matrix of interaction coefficients yields the representation of states by self-adjoint matrices and hence a spectral representation. As a result, cooperation systems, decision systems and quantum systems all become visible as manifestations of special interaction systems. The treatment of the theory is purely mathematical and does not require any special knowledge of physics. It is shown how standard notions in cooperative game theory arise naturally in this context. In particular, states of general interaction systems are seen to arise as linear superpositions of pure quantum states and Fourier transformation to become meaningful. Moreover, quantum games fall into this framework. Finally, a theory of Markov evolution of interaction states is presented that generalizes classical homogeneous Markov chains to the present context. Keywords: cooperative game; decision system; evolution; Fourier transform; interaction system; measurement; quantum game 1. Introduction In an interaction system, economic (or general physical) agents interact pairwise, but do not necessarily cooperate towards a common goal. However, this model arises as a natural generalization of the model of cooperative TU games, for which already Owen [1] introduced the co-value as an assessment of the pairwise interaction of two cooperating players1.
    [Show full text]
  • The Monty Hall Problem in the Game Theory Class
    The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice.
    [Show full text]
  • Introduction to Probability
    Massachusetts Institute of Technology Course Notes 10 6.042J/18.062J, Fall ’02: Mathematics for Computer Science November 4 Professor Albert Meyer and Dr. Radhika Nagpal revised November 6, 2002, 572 minutes Introduction to Probability 1 Probability Probability will be the topic for the rest of the term. Probability is one of the most important sub- jects in Mathematics and Computer Science. Most upper level Computer Science courses require probability in some form, especially in analysis of algorithms and data structures, but also in infor- mation theory, cryptography, control and systems theory, network design, artificial intelligence, and game theory. Probability also plays a key role in fields such as Physics, Biology, Economics and Medicine. There is a close relationship between Counting/Combinatorics and Probability. In many cases, the probability of an event is simply the fraction of possible outcomes that make up the event. So many of the rules we developed for finding the cardinality of finite sets carry over to Proba- bility Theory. For example, we’ll apply an Inclusion-Exclusion principle for probabilities in some examples below. In principle, probability boils down to a few simple rules, but it remains a tricky subject because these rules often lead unintuitive conclusions. Using “common sense” reasoning about probabilis- tic questions is notoriously unreliable, as we’ll illustrate with many real-life examples. This reading is longer than usual . To keep things in bounds, several sections with illustrative examples that do not introduce new concepts are marked “[Optional].” You should read these sections selectively, choosing those where you’re unsure about some idea and think another ex- ample would be helpful.
    [Show full text]
  • Strong Stackelberg Reasoning in Symmetric Games: an Experimental
    Strong Stackelberg reasoning in symmetric games: An experimental ANGOR UNIVERSITY replication and extension Pulford, B.D.; Colman, A.M.; Lawrence, C.L. PeerJ DOI: 10.7717/peerj.263 PRIFYSGOL BANGOR / B Published: 25/02/2014 Publisher's PDF, also known as Version of record Cyswllt i'r cyhoeddiad / Link to publication Dyfyniad o'r fersiwn a gyhoeddwyd / Citation for published version (APA): Pulford, B. D., Colman, A. M., & Lawrence, C. L. (2014). Strong Stackelberg reasoning in symmetric games: An experimental replication and extension. PeerJ, 263. https://doi.org/10.7717/peerj.263 Hawliau Cyffredinol / General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal ? Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. 23. Sep. 2021 Strong Stackelberg reasoning in symmetric games: An experimental replication and extension Briony D. Pulford1, Andrew M. Colman1 and Catherine L. Lawrence2 1 School of Psychology, University of Leicester, Leicester, UK 2 School of Psychology, Bangor University, Bangor, UK ABSTRACT In common interest games in which players are motivated to coordinate their strate- gies to achieve a jointly optimal outcome, orthodox game theory provides no general reason or justification for choosing the required strategies.
    [Show full text]
  • Game, Set, and Match Carl W
    Game, Set, and Match Carl W. Lee September 2016 Note: Some of the text below comes from Martin Gardner's articles in Scientific American and some from Mathematical Circles by Fomin, Genkin, and Itenberg. 1 1 Fifteen This is a two player game. Take a set of nine cards, numbered one (ace) through nine. They are spread out on the table, face up. Players alternately select a card to add their hands, which they keep face up in front of them. The goal is to achieve a subset of three cards in your hand so that the values of these three cards sum to exactly fifteen. (The ace counts as 1.) This is an example of a game that is isomorphic to (the \same as") a well-known game: Tic-Tac-Toe. Number the cells of a tic-tac-toe board with the integers from 1 to 9 arranged in the form of a magic square. 8 1 6 3 5 7 4 9 2 Then winning combinations correspond exactly to triples of numbers that sum to 15. This is a combinatorial two person game (but one that permits ties). There is no hidden information (e.g., cards that one player has that the other cannot see) and no random elements (e.g., rolling of dice). 2 2 Ones and Twos Ten 1's and ten 2's are written on a blackboard. In one turn, a player may erase (or cross out) any two numbers. If the two numbers erased are identical, they are replaced with a single 2.
    [Show full text]
  • Introduction to Probability 1 Monty Hall
    Massachusetts Institute of Technology Course Notes, Week 12 6.042J/18.062J, Fall ’05: Mathematics for Computer Science November 21 Prof. Albert R. Meyer and Prof. Ronitt Rubinfeld revised November 29, 2005, 1285 minutes Introduction to Probability Probability is the last topic in this course and perhaps the most important. Many algo­ rithms rely on randomization. Investigating their correctness and performance requires probability theory. Moreover, many aspects of computer systems, such as memory man­ agement, branch prediction, packet routing, and load balancing are designed around probabilistic assumptions and analyses. Probability also comes up in information theory, cryptography, artificial intelligence, and game theory. Beyond these engineering applica­ tions, an understanding of probability gives insight into many everyday issues, such as polling, DNA testing, risk assessment, investing, and gambling. So probability is good stuff. 1 Monty Hall In the September 9, 1990 issue of Parade magazine, the columnist Marilyn vos Savant responded to this letter: Suppose you’re on a game show, and you’re given the choice of three doors. Behind one door is a car, behind the others, goats. You pick a door, say number 1, and the host, who knows what’s behind the doors, opens another door, say number 3, which has a goat. He says to you, ”Do you want to pick door number 2?” Is it to your advantage to switch your choice of doors? Craig. F. Whitaker Columbia, MD The letter roughly describes a situation faced by contestants on the 1970’s game show Let’s Make a Deal, hosted by Monty Hall and Carol Merrill.
    [Show full text]
  • Economics 201B Economic Theory (Spring 2021) Strategic Games
    Economics 201B Economic Theory (Spring 2021) Strategic Games Topics: terminology and notations (OR 1.7), games and solutions (OR 1.1-1.3), rationality and bounded rationality (OR 1.4-1.6), formalities (OR 2.1), best-response (OR 2.2), Nash equilibrium (OR 2.2), 2 2 examples × (OR 2.3), existence of Nash equilibrium (OR 2.4), mixed strategy Nash equilibrium (OR 3.1, 3.2), strictly competitive games (OR 2.5), evolution- ary stability (OR 3.4), rationalizability (OR 4.1), dominance (OR 4.2, 4.3), trembling hand perfection (OR 12.5). Terminology and notations (OR 1.7) Sets For R, ∈ ≥ ⇐⇒ ≥ for all . and ⇐⇒ ≥ for all and some . ⇐⇒ for all . Preferences is a binary relation on some set of alternatives R. % ⊆ From % we derive two other relations on : — strict performance relation and not  ⇐⇒ % % — indifference relation and ∼ ⇐⇒ % % Utility representation % is said to be — complete if , or . ∀ ∈ % % — transitive if , and then . ∀ ∈ % % % % can be presented by a utility function only if it is complete and transitive (rational). A function : R is a utility function representing if → % ∀ ∈ () () % ⇐⇒ ≥ % is said to be — continuous (preferences cannot jump...) if for any sequence of pairs () with ,and and , . { }∞=1 % → → % — (strictly) quasi-concave if for any the upper counter set ∈ { ∈ : is (strictly) convex. % } These guarantee the existence of continuous well-behaved utility function representation. Profiles Let be a the set of players. — () or simply () is a profile - a collection of values of some variable,∈ one for each player. — () or simply is the list of elements of the profile = ∈ { } − () for all players except . ∈ — ( ) is a list and an element ,whichistheprofile () .
    [Show full text]
  • ECONS 424 – STRATEGY and GAME THEORY MIDTERM EXAM #2 – Answer Key
    ECONS 424 – STRATEGY AND GAME THEORY MIDTERM EXAM #2 – Answer key Exercise #1. Hawk-Dove game. Consider the following payoff matrix representing the Hawk-Dove game. Intuitively, Players 1 and 2 compete for a resource, each of them choosing to display an aggressive posture (hawk) or a passive attitude (dove). Assume that payoff > 0 denotes the value that both players assign to the resource, and > 0 is the cost of fighting, which only occurs if they are both aggressive by playing hawk in the top left-hand cell of the matrix. Player 2 Hawk Dove Hawk , , 0 2 2 − − Player 1 Dove 0, , 2 2 a) Show that if < , the game is strategically equivalent to a Prisoner’s Dilemma game. b) The Hawk-Dove game commonly assumes that the value of the resource is less than the cost of a fight, i.e., > > 0. Find the set of pure strategy Nash equilibria. Answer: Part (a) • When Player 2 (in columns) chooses Hawk (in the left-hand column), Player 1 (in rows) receives a positive payoff of by paying Hawk, which is higher than his payoff of zero from playing Dove. − Therefore, for Player2 1 Hawk is a best response to Player 2 playing Hawk. Similarly, when Player 2 chooses Dove (in the right-hand column), Player 1 receives a payoff of by playing Hawk, which is higher than his payoff from choosing Dove, ; entailing that Hawk is Player 1’s best response to Player 2 choosing Dove. Therefore, Player 1 chooses2 Hawk as his best response to all of Player 2’s strategies, implying that Hawk is a strictly dominant strategy for Player 1.
    [Show full text]
  • The Monty Hall Problem
    The Monty Hall Problem Richard D. Gill Mathematical Institute, University of Leiden, Netherlands http://www.math.leidenuniv.nl/∼gill 17 March, 2011 Abstract A range of solutions to the Monty Hall problem is developed, with the aim of connecting popular (informal) solutions and the formal mathematical solutions of in- troductory text-books. Under Riemann's slogan of replacing calculations with ideas, we discover bridges between popular solutions and rigorous (formal) mathematical solutions by using various combinations of symmetry, symmetrization, independence, and Bayes' rule, as well as insights from game theory. The result is a collection of intuitive and informal logical arguments which can be converted step by step into formal mathematics. The Monty Hall problem can be used simultaneously to develop probabilistic intuition and to give a deeper understanding of the paradox, not just to provide a routine exercise in computation of conditional probabilities. Simple insights from game theory can be used to prove probability inequalities. Acknowledgement: this text is the result of an evolution through texts written for three internet encyclopedias (wikipedia.org, citizendium.org, statprob.com) and each time has benefitted from the contributions of many other editors. I thank them all, for all the original ideas in this paper. That leaves only the mistakes, for which I bear responsibility. 1 Introduction Imagine you are a guest in a TV game show. The host, a certain Mr. Monty Hall, shows you three large doors and tells you that behind one of the doors there is a car while behind the other two there are goats. You want to win the car.
    [Show full text]
  • Turing Learning with Nash Memory
    Turing Learning with Nash Memory Machine Learning, Game Theory and Robotics Shuai Wang Master of Science Thesis University of Amsterdam Local Supervisor: Dr. Leen Torenvliet (University of Amsterdam, NL) Co-Supervisor: Dr. Frans Oliehoek (University of Liverpool, UK) External Advisor: Dr. Roderich Groß (University of Sheffield, UK) ABSTRACT Turing Learning is a method for the reverse engineering of agent behaviors. This approach was inspired by the Turing test where a machine can pass if its behaviour is indistinguishable from that of a human. Nash memory is a memory mechanism for coevolution. It guarantees monotonicity in convergence. This thesis explores the integration of such memory mechanism with Turing Learning for faster learning of agent behaviors. We employ the Enki robot simulation platform and learns the aggregation behavior of epuck robots. Our experiments indicate that using Nash memory can reduce the computation time by 35.4% and results in faster convergence for the aggregation game. In addition, we present TuringLearner, the first Turing Learning platform. Keywords: Turing Learning, Nash Memory, Game Theory, Multi-agent System. COMMITTEE MEMBER: DR.KATRIN SCHULZ (CHAIR) PROF. DR.FRANK VAN HARMELEN DR.PETER VAN EMDE BOAS DR.PETER BLOEM YFKE DULEK MSC MASTER THESIS,UNIVERSITY OF AMSTERDAM Amsterdam, August 2017 Contents 1 Introduction ....................................................7 2 Background ....................................................9 2.1 Introduction to Game Theory9 2.1.1 Players and Games................................................9 2.1.2 Strategies and Equilibria........................................... 10 2.2 Computing Nash Equilibria 11 2.2.1 Introduction to Linear Programming.................................. 11 2.2.2 Game Solving with LP............................................. 12 2.2.3 Asymmetric Games............................................... 13 2.2.4 Bimatrix Games.................................................
    [Show full text]