General Game Playing with Imperfect Information
Total Page:16
File Type:pdf, Size:1020Kb
Load more
										Recommended publications
									
								- 
												
												Lecture 4 Rationalizability & Nash Equilibrium Road
Lecture 4 Rationalizability & Nash Equilibrium 14.12 Game Theory Muhamet Yildiz Road Map 1. Strategies – completed 2. Quiz 3. Dominance 4. Dominant-strategy equilibrium 5. Rationalizability 6. Nash Equilibrium 1 Strategy A strategy of a player is a complete contingent-plan, determining which action he will take at each information set he is to move (including the information sets that will not be reached according to this strategy). Matching pennies with perfect information 2’s Strategies: HH = Head if 1 plays Head, 1 Head if 1 plays Tail; HT = Head if 1 plays Head, Head Tail Tail if 1 plays Tail; 2 TH = Tail if 1 plays Head, 2 Head if 1 plays Tail; head tail head tail TT = Tail if 1 plays Head, Tail if 1 plays Tail. (-1,1) (1,-1) (1,-1) (-1,1) 2 Matching pennies with perfect information 2 1 HH HT TH TT Head Tail Matching pennies with Imperfect information 1 2 1 Head Tail Head Tail 2 Head (-1,1) (1,-1) head tail head tail Tail (1,-1) (-1,1) (-1,1) (1,-1) (1,-1) (-1,1) 3 A game with nature Left (5, 0) 1 Head 1/2 Right (2, 2) Nature (3, 3) 1/2 Left Tail 2 Right (0, -5) Mixed Strategy Definition: A mixed strategy of a player is a probability distribution over the set of his strategies. Pure strategies: Si = {si1,si2,…,sik} σ → A mixed strategy: i: S [0,1] s.t. σ σ σ i(si1) + i(si2) + … + i(sik) = 1. If the other players play s-i =(s1,…, si-1,si+1,…,sn), then σ the expected utility of playing i is σ σ σ i(si1)ui(si1,s-i) + i(si2)ui(si2,s-i) + … + i(sik)ui(sik,s-i). - 
												
												Game Theory Lecture Notes
Game Theory: Penn State Math 486 Lecture Notes Version 2.1.1 Christopher Griffin « 2010-2021 Licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License With Major Contributions By: James Fan George Kesidis and Other Contributions By: Arlan Stutler Sarthak Shah Contents List of Figuresv Preface xi 1. Using These Notes xi 2. An Overview of Game Theory xi Chapter 1. Probability Theory and Games Against the House1 1. Probability1 2. Random Variables and Expected Values6 3. Conditional Probability8 4. The Monty Hall Problem 11 Chapter 2. Game Trees and Extensive Form 15 1. Graphs and Trees 15 2. Game Trees with Complete Information and No Chance 18 3. Game Trees with Incomplete Information 22 4. Games of Chance 24 5. Pay-off Functions and Equilibria 26 Chapter 3. Normal and Strategic Form Games and Matrices 37 1. Normal and Strategic Form 37 2. Strategic Form Games 38 3. Review of Basic Matrix Properties 40 4. Special Matrices and Vectors 42 5. Strategy Vectors and Matrix Games 43 Chapter 4. Saddle Points, Mixed Strategies and the Minimax Theorem 45 1. Saddle Points 45 2. Zero-Sum Games without Saddle Points 48 3. Mixed Strategies 50 4. Mixed Strategies in Matrix Games 53 5. Dominated Strategies and Nash Equilibria 54 6. The Minimax Theorem 59 7. Finding Nash Equilibria in Simple Games 64 8. A Note on Nash Equilibria in General 66 Chapter 5. An Introduction to Optimization and the Karush-Kuhn-Tucker Conditions 69 1. A General Maximization Formulation 70 2. Some Geometry for Optimization 72 3. - 
												
												Frequently Asked Questions in Mathematics
Frequently Asked Questions in Mathematics The Sci.Math FAQ Team. Editor: Alex L´opez-Ortiz e-mail: [email protected] Contents 1 Introduction 4 1.1 Why a list of Frequently Asked Questions? . 4 1.2 Frequently Asked Questions in Mathematics? . 4 2 Fundamentals 5 2.1 Algebraic structures . 5 2.1.1 Monoids and Groups . 6 2.1.2 Rings . 7 2.1.3 Fields . 7 2.1.4 Ordering . 8 2.2 What are numbers? . 9 2.2.1 Introduction . 9 2.2.2 Construction of the Number System . 9 2.2.3 Construction of N ............................... 10 2.2.4 Construction of Z ................................ 10 2.2.5 Construction of Q ............................... 11 2.2.6 Construction of R ............................... 11 2.2.7 Construction of C ............................... 12 2.2.8 Rounding things up . 12 2.2.9 What’s next? . 12 3 Number Theory 14 3.1 Fermat’s Last Theorem . 14 3.1.1 History of Fermat’s Last Theorem . 14 3.1.2 What is the current status of FLT? . 14 3.1.3 Related Conjectures . 15 3.1.4 Did Fermat prove this theorem? . 16 3.2 Prime Numbers . 17 3.2.1 Largest known Mersenne prime . 17 3.2.2 Largest known prime . 17 3.2.3 Largest known twin primes . 18 3.2.4 Largest Fermat number with known factorization . 18 3.2.5 Algorithms to factor integer numbers . 18 3.2.6 Primality Testing . 19 3.2.7 List of record numbers . 20 3.2.8 What is the current status on Mersenne primes? . - 
												
												SEQUENTIAL GAMES with PERFECT INFORMATION Example
SEQUENTIAL GAMES WITH PERFECT INFORMATION Example 4.9 (page 105) Consider the sequential game given in Figure 4.9. We want to apply backward induction to the tree. 0 Vertex B is owned by player two, P2. The payoffs for P2 are 1 and 3, with 3 > 1, so the player picks R . Thus, the payoffs at B become (0, 3). 00 Next, vertex C is also owned by P2 with payoffs 1 and 0. Since 1 > 0, P2 picks L , and the payoffs are (4, 1). Player one, P1, owns A; the choice of L gives a payoff of 0 and R gives a payoff of 4; 4 > 0, so P1 chooses R. The final payoffs are (4, 1). 0 00 We claim that this strategy profile, { R } for P1 and { R ,L } is a Nash equilibrium. Notice that the 0 00 strategy profile gives a choice at each vertex. For the strategy { R ,L } fixed for P2, P1 has a maximal payoff by choosing { R }, ( 0 00 0 00 π1(R, { R ,L }) = 4 π1(R, { R ,L }) = 4 ≥ 0 00 π1(L, { R ,L }) = 0. 0 00 In the same way, for the strategy { R } fixed for P1, P2 has a maximal payoff by choosing { R ,L }, ( 00 0 00 π2(R, {∗,L }) = 1 π2(R, { R ,L }) = 1 ≥ 00 π2(R, {∗,R }) = 0, where ∗ means choose either L0 or R0. Since no change of choice by a player can increase that players own payoff, the strategy profile is called a Nash equilibrium. Notice that the above strategy profile is also a Nash equilibrium on each branch of the game tree, mainly starting at either B or starting at C. - 
												
												The Monty Hall Problem in the Game Theory Class
The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice. - 
												
												Game, Set, and Match Carl W
Game, Set, and Match Carl W. Lee September 2016 Note: Some of the text below comes from Martin Gardner's articles in Scientific American and some from Mathematical Circles by Fomin, Genkin, and Itenberg. 1 1 Fifteen This is a two player game. Take a set of nine cards, numbered one (ace) through nine. They are spread out on the table, face up. Players alternately select a card to add their hands, which they keep face up in front of them. The goal is to achieve a subset of three cards in your hand so that the values of these three cards sum to exactly fifteen. (The ace counts as 1.) This is an example of a game that is isomorphic to (the \same as") a well-known game: Tic-Tac-Toe. Number the cells of a tic-tac-toe board with the integers from 1 to 9 arranged in the form of a magic square. 8 1 6 3 5 7 4 9 2 Then winning combinations correspond exactly to triples of numbers that sum to 15. This is a combinatorial two person game (but one that permits ties). There is no hidden information (e.g., cards that one player has that the other cannot see) and no random elements (e.g., rolling of dice). 2 2 Ones and Twos Ten 1's and ten 2's are written on a blackboard. In one turn, a player may erase (or cross out) any two numbers. If the two numbers erased are identical, they are replaced with a single 2. - 
												
												Introduction to Probability 1 Monty Hall
Massachusetts Institute of Technology Course Notes, Week 12 6.042J/18.062J, Fall ’05: Mathematics for Computer Science November 21 Prof. Albert R. Meyer and Prof. Ronitt Rubinfeld revised November 29, 2005, 1285 minutes Introduction to Probability Probability is the last topic in this course and perhaps the most important. Many algo rithms rely on randomization. Investigating their correctness and performance requires probability theory. Moreover, many aspects of computer systems, such as memory man agement, branch prediction, packet routing, and load balancing are designed around probabilistic assumptions and analyses. Probability also comes up in information theory, cryptography, artificial intelligence, and game theory. Beyond these engineering applica tions, an understanding of probability gives insight into many everyday issues, such as polling, DNA testing, risk assessment, investing, and gambling. So probability is good stuff. 1 Monty Hall In the September 9, 1990 issue of Parade magazine, the columnist Marilyn vos Savant responded to this letter: Suppose you’re on a game show, and you’re given the choice of three doors. Behind one door is a car, behind the others, goats. You pick a door, say number 1, and the host, who knows what’s behind the doors, opens another door, say number 3, which has a goat. He says to you, ”Do you want to pick door number 2?” Is it to your advantage to switch your choice of doors? Craig. F. Whitaker Columbia, MD The letter roughly describes a situation faced by contestants on the 1970’s game show Let’s Make a Deal, hosted by Monty Hall and Carol Merrill. - 
												
												Economics 201B Economic Theory (Spring 2021) Strategic Games
Economics 201B Economic Theory (Spring 2021) Strategic Games Topics: terminology and notations (OR 1.7), games and solutions (OR 1.1-1.3), rationality and bounded rationality (OR 1.4-1.6), formalities (OR 2.1), best-response (OR 2.2), Nash equilibrium (OR 2.2), 2 2 examples × (OR 2.3), existence of Nash equilibrium (OR 2.4), mixed strategy Nash equilibrium (OR 3.1, 3.2), strictly competitive games (OR 2.5), evolution- ary stability (OR 3.4), rationalizability (OR 4.1), dominance (OR 4.2, 4.3), trembling hand perfection (OR 12.5). Terminology and notations (OR 1.7) Sets For R, ∈ ≥ ⇐⇒ ≥ for all . and ⇐⇒ ≥ for all and some . ⇐⇒ for all . Preferences is a binary relation on some set of alternatives R. % ⊆ From % we derive two other relations on : — strict performance relation and not  ⇐⇒ % % — indifference relation and ∼ ⇐⇒ % % Utility representation % is said to be — complete if , or . ∀ ∈ % % — transitive if , and then . ∀ ∈ % % % % can be presented by a utility function only if it is complete and transitive (rational). A function : R is a utility function representing if → % ∀ ∈ () () % ⇐⇒ ≥ % is said to be — continuous (preferences cannot jump...) if for any sequence of pairs () with ,and and , . { }∞=1 % → → % — (strictly) quasi-concave if for any the upper counter set ∈ { ∈ : is (strictly) convex. % } These guarantee the existence of continuous well-behaved utility function representation. Profiles Let be a the set of players. — () or simply () is a profile - a collection of values of some variable,∈ one for each player. — () or simply is the list of elements of the profile = ∈ { } − () for all players except . ∈ — ( ) is a list and an element ,whichistheprofile () . - 
												
												Ai12-General-Game-Playing-Pre-Handout
Introduction GDL GGP Alpha Zero Conclusion References Artificial Intelligence 12. General Game Playing One AI to Play All Games and Win Them All Jana Koehler Alvaro´ Torralba Summer Term 2019 Thanks to Dr. Peter Kissmann for slide sources Koehler and Torralba Artificial Intelligence Chapter 12: GGP 1/53 Introduction GDL GGP Alpha Zero Conclusion References Agenda 1 Introduction 2 The Game Description Language (GDL) 3 Playing General Games 4 Learning Evaluation Functions: Alpha Zero 5 Conclusion Koehler and Torralba Artificial Intelligence Chapter 12: GGP 2/53 Introduction GDL GGP Alpha Zero Conclusion References Deep Blue Versus Garry Kasparov (1997) Koehler and Torralba Artificial Intelligence Chapter 12: GGP 4/53 Introduction GDL GGP Alpha Zero Conclusion References Games That Deep Blue Can Play 1 Chess Koehler and Torralba Artificial Intelligence Chapter 12: GGP 5/53 Introduction GDL GGP Alpha Zero Conclusion References Chinook Versus Marion Tinsley (1992) Koehler and Torralba Artificial Intelligence Chapter 12: GGP 6/53 Introduction GDL GGP Alpha Zero Conclusion References Games That Chinook Can Play 1 Checkers Koehler and Torralba Artificial Intelligence Chapter 12: GGP 7/53 Introduction GDL GGP Alpha Zero Conclusion References Games That a General Game Player Can Play 1 Chess 2 Checkers 3 Chinese Checkers 4 Connect Four 5 Tic-Tac-Toe 6 ... Koehler and Torralba Artificial Intelligence Chapter 12: GGP 8/53 Introduction GDL GGP Alpha Zero Conclusion References Games That a General Game Player Can Play (Ctd.) 5 ... 18 Zhadu 6 Othello 19 Pancakes 7 Nine Men's Morris 20 Quarto 8 15-Puzzle 21 Knight's Tour 9 Nim 22 n-Queens 10 Sudoku 23 Blob Wars 11 Pentago 24 Bomberman (simplified) 12 Blocker 25 Catch a Mouse 13 Breakthrough 26 Chomp 14 Lights Out 27 Gomoku 15 Amazons 28 Hex 16 Knightazons 29 Cubicup 17 Blocksworld 30 .. - 
												
												Pure and Bayes-Nash Price of Anarchy for Generalized Second Price Auction
Pure and Bayes-Nash Price of Anarchy for Generalized Second Price Auction Renato Paes Leme Eva´ Tardos Department of Computer Science Department of Computer Science Cornell University, Ithaca, NY Cornell University, Ithaca, NY [email protected] [email protected] Abstract—The Generalized Second Price Auction has for advertisements and slots higher on the page are been the main mechanism used by search companies more valuable (clicked on by more users). The bids to auction positions for advertisements on search pages. are used to determine both the assignment of bidders In this paper we study the social welfare of the Nash equilibria of this game in various models. In the full to slots, and the fees charged. In the simplest model, information setting, socially optimal Nash equilibria are the bidders are assigned to slots in order of bids, and known to exist (i.e., the Price of Stability is 1). This paper the fee for each click is the bid occupying the next is the first to prove bounds on the price of anarchy, and slot. This auction is called the Generalized Second Price to give any bounds in the Bayesian setting. Auction (GSP). More generally, positions and payments Our main result is to show that the price of anarchy is small assuming that all bidders play un-dominated in the Generalized Second Price Auction depend also on strategies. In the full information setting we prove a bound the click-through rates associated with the bidders, the of 1.618 for the price of anarchy for pure Nash equilibria, probability that the advertisement will get clicked on by and a bound of 4 for mixed Nash equilibria. - 
												
												Towards General Game-Playing Robots: Models, Architecture and Game Controller
Towards General Game-Playing Robots: Models, Architecture and Game Controller David Rajaratnam and Michael Thielscher The University of New South Wales, Sydney, NSW 2052, Australia fdaver,[email protected] Abstract. General Game Playing aims at AI systems that can understand the rules of new games and learn to play them effectively without human interven- tion. Our paper takes the first step towards general game-playing robots, which extend this capability to AI systems that play games in the real world. We develop a formal model for general games in physical environments and provide a sys- tems architecture that allows the embedding of existing general game players as the “brain” and suitable robotic systems as the “body” of a general game-playing robot. We also report on an initial robot prototype that can understand the rules of arbitrary games and learns to play them in a fixed physical game environment. 1 Introduction General game playing is the attempt to create a new generation of AI systems that can understand the rules of new games and then learn to play these games without human intervention [5]. Unlike specialised systems such as the chess program Deep Blue, a general game player cannot rely on algorithms that have been designed in advance for specific games. Rather, it requires a form of general intelligence that enables the player to autonomously adapt to new and possibly radically different problems. General game- playing systems therefore are a quintessential example of a new generation of software that end users can customise for their own specific tasks and special needs. - 
												
												Extensive Form Games with Perfect Information
Notes Strategy and Politics: Extensive Form Games with Perfect Information Matt Golder Pennsylvania State University Sequential Decision-Making Notes The model of a strategic (normal form) game suppresses the sequential structure of decision-making. When applying the model to situations in which players move sequentially, we assume that each player chooses her plan of action once and for all. She is committed to this plan, which she cannot modify as events unfold. The model of an extensive form game, by contrast, describes the sequential structure of decision-making explicitly, allowing us to study situations in which each player is free to change her mind as events unfold. Perfect Information Notes Perfect information describes a situation in which players are always fully informed about all of the previous actions taken by all players. This assumption is used in all of the following lecture notes that use \perfect information" in the title. Later we will also study more general cases where players may only be imperfectly informed about previous actions when choosing an action. Extensive Form Games Notes To describe an extensive form game with perfect information we need to specify the set of players and their preferences, just as for a strategic game. In addition, we also need to specify the order of the players' moves and the actions each player may take at each point (or decision node). We do so by specifying the set of all sequences of actions that can possibly occur, together with the player who moves at each point in each sequence. We refer to each possible sequence of actions (a1; a2; : : : ak) as a terminal history and to the function that denotes the player who moves at each point in each terminal history as the player function.