Gradient Play in N-Cluster Games with Zero-Order Information

Total Page:16

File Type:pdf, Size:1020Kb

Gradient Play in N-Cluster Games with Zero-Order Information Gradient Play in n-Cluster Games with Zero-Order Information Tatiana Tatarenko, Jan Zimmermann, Jurgen¨ Adamy Abstract— We study a distributed approach for seeking a models, each agent intends to find a strategy to achieve a Nash equilibrium in n-cluster games with strictly monotone Nash equilibrium in the resulting n-cluster game, which is mappings. Each player within each cluster has access to the a stable state that minimizes the cluster’s cost functions in current value of her own smooth local cost function estimated by a zero-order oracle at some query point. We assume the response to the actions of the agents from other clusters. agents to be able to communicate with their neighbors in Continuous time algorithms for the distributed Nash equi- the same cluster over some undirected graph. The goal of libria seeking problem in multi-cluster games were proposed the agents in the cluster is to minimize their collective cost. in [17]–[19]. The paper [17] solves an unconstrained multi- This cost depends, however, on actions of agents from other cluster game by using gradient-based algorithms, whereas the clusters. Thus, a game between the clusters is to be solved. We present a distributed gradient play algorithm for determining works [18] and [19] propose a gradient-free algorithm, based a Nash equilibrium in this game. The algorithm takes into on zero-order information, for seeking Nash and generalized account the communication settings and zero-order information Nash equilibria respectively. In discrete time domain, the under consideration. We prove almost sure convergence of this work [5] presents a leader-follower based algorithm, which algorithm to a Nash equilibrium given appropriate estimations can solve unconstrained multi-cluster games in linear time. of the local cost functions’ gradients. The authors in [20] extend this result to the case of leaderless I. INTRODUCTION architecture. Both papers [5], [20] prove linear convergence in games with strongly monotone mappings and first-order Distributed optimization and game theory provide power- information, meaning that agents can calculate gradients ful frameworks to deal with optimization problems arising of their cost functions and use this information to update in multi-agent systems. In generic distributed optimization their states. In contrast to that, the work [16] deals with problems, the cost functions of agents are distributed across a gradient-free approach to the cluster games. However, the network, meaning that each agent has only partial infor- the gradient estimations are constructed in such a way that mation about the whole optimization problem which is to only convergence to a neighborhood of the equilibrium can be solved. Game theoretic problems arise in such networks be guaranteed. Moreover, these estimations are obtained by when the agents do not cooperate with each other and the using two query points, for which an extra coordination cost functions of these non-cooperative agents are coupled by between the agents is required. the decisions of all agents in the system. The applications Motivated by relevancy of n-cluster game models in of game theoretic and distributed optimization approaches many engineering applications, we present a discrete time include, for example, electricity markets, power systems, distributed procedure to seek Nash equilibria in n-cluster flow control problems and communication networks [6], [11], games with zero-order information. We consider settings, [12]. where agents can communicate with their direct neighbors On the other hand, cooperation and competition coex- within the corresponding cluster over some undirected graph. ists in many practical situations, such as cloud computing, However, in many practical situations the agents do not hierarchical optimization in Smart Grid, and adversarial know the functional form of their objectives and can only networks [3], [4], [8]. A body of recent work has been access the current values of their objective functions at some arXiv:2107.12648v1 [cs.GT] 27 Jul 2021 devoted to analysis of non-cooperative games and distributed query point. Such situations arise, for example, in electricity optimization problems in terms of a single model called n- markets with unknown price functions [15]. In such cases, cluster games [5], [16]–[20]. In such n-cluster games, each the information structure is referred to as zero-order oracle. cluster corresponds to a player whose goal is to minimize Our work focuses on zero-order oracle information settings her own cost function. However, the clusters in this game and, thus, assumes agents to have no access to the analytical are not the actual decision-makers as the optimization of the form of their cost functions and gradients. The agents instead cluster’s objective is controlled by the agents belonging to construct their local query points and get the corresponding the corresponding cluster. Each of such agents has her own cost values from the oracle. Based on these values, the local cost function, which is available only to this agent, agents estimate their local gradients to be able to follow but depends on the joint actions of agents in all clusters. the step in the gradient play procedure. We formulate the The cluster’s objective, in turn, is the sum of the local cost sufficient conditions and provide some concrete example on functions of the agents within the cluster. Therefore, in such how to estimate the gradients to guarantee the almost sure convergence of the resulting algorithm to Nash equilibria in The authors are with Control Methods and Robotics Lab at the TU Darmstadt, Germany. n-cluster games with strictly monotone game mappings. To the best of our knowledge, we present the first algorithm solving n-cluster games with zero-order oracle and the clusters. Instead, we consider the following zero-order infor- corresponding one-point gradient estimations. mation structure in the system: No agent has access to the The paper is organized as follows. In SectionII we for- analytical form of any cost function, including its own. Each mulated the n-cluster game with undirected communication agent can only observe the value of its local cost function topology in each cluster and zero-order oracle information. given any joint action of all agents in the system. Formally, Section III introduces the gradient play algorithm which is given a joint action x Ω, each agent j [ni], i [n] j 2 2 2 based on the one-point gradient estimations. The convergence receives the value Ji (x) from a zero-order oracle. Especially, result is presented in Section III as well. SectionIV provides no agent has or receives any information about the gradient. an example of query points and gradient estimations which Let us denote the game between the clusters introduced guarantee convergence of the algorithm discussed in Sec- above by Γ(n; J ; Ω ; ). We make the following f ig f ig fGig tion III. SectionV presents some simulation results. Finally, assumptions regarding the game Γ: SectionVI concludes the paper. Assumption 1. The n-cluster game under consideration is Notations. The set 1; : : : ; n is denoted by [n]. For any strictly convex. Namely, for all , the set is convex, f gn @f(x) i [n] Ωi function f : K R, K R , if(x) = is the 2 ! ⊆ r @xi the cost function Ji(xi; x i) is continuously differentiable in partial derivative taken in respect to the ith coordinate of the for each fixed . Moreover,− the game mapping, which n xi x i vector variable x R . We consider real normed space E, is defined as − which is the space2 of real vectors, i.e. E = n. We use (u; v) R T to denote the inner product in E. We use to denote the F(x) , [ 1J1(x1; x 1);:::; nJn(xn; x n)] (1) k·k r − r − Euclidean norm induced by the standard dot product in E. is strictly monotone on Ω. Any mapping g : E E is said to be strictly monotone on ! j Q E, if (g(u) g(v); u v) > 0 for any u; v Q, where Assumption 2. Each function iJi (xi; x i) is Lipschitz ⊆ − − 2 continuous on Ω. r − u = v. We use Br(p) to denote the ball of the radius r 0 6 ≥ and the center p E and to denote the unit sphere with i S Assumption 3. The action sets Ωj, j [ni], i [n], are the center in 0 E2. We use v to denote the projection 2 2 2 PΩ f g compact. Moreover, for each i there exists a so called safety of v E to a set Ω E. The mathematical expectation 2 2 ⊆ ball Br(p) Ωi with ri > 0 and pi Ωi . of a random value ξ is denoted by E ξ . We use the big-O ⊆ 2 f g The assumptions above are standard in the literature on notation, that is, the function f(x): R R is O(g(x)) as ! f(x) both game-theoretic and zero-order optimization [1]. Finally, x a, f(x) = O(g(x)) as x a, if limx a jg(x)j K for some! positive constant K. ! ! j j ≤ we make the following assumption on the communication graph, which guarantees sufficient information ”mixing” in II. NASH EQUILIBRIUM SEEKING the network within each cluster. A. Problem Formulation Assumption 4. The underlying undirected communication We consider a non-cooperative game between n clusters. graph i([ni]; i) is connected for all i = 1; : : : ; n. The j G A i n n associated non-negative mixing matrix Wi = [wkj] R × Each cluster i [n] itself consists of ni agents. Let Ji 2 j 1 2 defines the weights on the undirected arcs such that wi > 0 and Ωi R denote respectively the cost function and kj ⊆ if and only if (k; j) and Pni wi = 1, k [n ].
Recommended publications
  • Game Playing
    Game Playing Philipp Koehn 29 September 2015 Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 Outline 1 ● Games ● Perfect play – minimax decisions – α–β pruning ● Resource limits and approximate evaluation ● Games of chance ● Games of imperfect information Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 2 games Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 Games vs. Search Problems 3 ● “Unpredictable” opponent ⇒ solution is a strategy specifying a move for every possible opponent reply ● Time limits ⇒ unlikely to find goal, must approximate ● Plan of attack: – computer considers possible lines of play (Babbage, 1846) – algorithm for perfect play (Zermelo, 1912; Von Neumann, 1944) – finite horizon, approximate evaluation (Zuse, 1945; Wiener, 1948; Shannon, 1950) – first Chess program (Turing, 1951) – machine learning to improve evaluation accuracy (Samuel, 1952–57) – pruning to allow deeper search (McCarthy, 1956) Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 Types of Games 4 deterministic chance perfect Chess Backgammon information Checkers Monopoly Go Othello imperfect battleships Bridge information Blind Tic Tac Toe Poker Scrabble Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 Game Tree (2-player, Deterministic, Turns) 5 Philipp Koehn Artificial Intelligence: Game Playing 29 September 2015 Simple Game Tree 6 ● 2 player game ● Each player has one move ● You move first ● Goal: optimize your payoff (utility) Start Your move Opponent
    [Show full text]
  • Tic-Tac-Toe and Machine Learning David Holmstedt Davho304 729G43
    Tic-Tac-Toe and machine learning David Holmstedt Davho304 729G43 Table of Contents Introduction ............................................................................................................................... 1 What is tic-tac-toe.................................................................................................................. 1 Tic-tac-toe Strategies ............................................................................................................. 1 Search-Algorithms .................................................................................................................. 1 Machine learning ................................................................................................................... 2 Weights .............................................................................................................................. 2 Training data ...................................................................................................................... 3 Bringing them together ...................................................................................................... 3 Implementation ......................................................................................................................... 4 Overview ................................................................................................................................ 4 UI .......................................................................................................................................
    [Show full text]
  • Graph Ramsey Games
    TECHNICAL R EPORT INSTITUT FUR¨ INFORMATIONSSYSTEME ABTEILUNG DATENBANKEN UND ARTIFICIAL INTELLIGENCE Graph Ramsey games DBAI-TR-99-34 Wolfgang Slany arXiv:cs/9911004v1 [cs.CC] 10 Nov 1999 DBAI TECHNICAL REPORT Institut f¨ur Informationssysteme NOVEMBER 5, 1999 Abteilung Datenbanken und Artificial Intelligence Technische Universit¨at Wien Favoritenstr. 9 A-1040 Vienna, Austria Tel: +43-1-58801-18403 Fax: +43-1-58801-18492 [email protected] http://www.dbai.tuwien.ac.at/ DBAI TECHNICAL REPORT DBAI-TR-99-34, NOVEMBER 5, 1999 Graph Ramsey games Wolfgang Slany1 Abstract. We consider combinatorial avoidance and achievement games based on graph Ramsey theory: The players take turns in coloring still uncolored edges of a graph G, each player being assigned a distinct color, choosing one edge per move. In avoidance games, completing a monochromatic subgraph isomorphic to another graph A leads to immedi- ate defeat or is forbidden and the first player that cannot move loses. In the avoidance+ variants, both players are free to choose more than one edge per move. In achievement games, the first player that completes a monochromatic subgraph isomorphic to A wins. Erd˝os & Selfridge [16] were the first to identify some tractable subcases of these games, followed by a large number of further studies. We complete these investigations by settling the complexity of all unrestricted cases: We prove that general graph Ramsey avoidance, avoidance+, and achievement games and several variants thereof are PSPACE-complete. We ultra-strongly solve some nontrivial instances of graph Ramsey avoidance games that are based on symmetric binary Ramsey numbers and provide strong evidence that all other cases based on symmetric binary Ramsey numbers are effectively intractable.
    [Show full text]
  • From Incan Gold to Dominion: Unraveling Optimal Strategies in Unsolved Games
    From Incan Gold to Dominion: Unraveling Optimal Strategies in Unsolved Games A Senior Comprehensive Paper presented to the Faculty of Carleton College Department of Mathematics Tommy Occhipinti, Advisor by Grace Jaffe Tyler Mahony Lucinda Robinson March 5, 2014 Acknowledgments We would like to thank our advisor, Tommy Occhipinti, for constant technical help, brainstorming, and moral support. We would also like to thank Tristan Occhipinti for providing the software that allowed us to run our simulations, Miles Ott for helping us with statistics, Mike Tie for continual technical support, Andrew Gainer-Dewar for providing the LaTeX templates, and Harold Jaffe for proofreading. Finally, we would like to express our gratitude to the entire math department for everything they have done for us over the past four years. iii Abstract While some games have inherent optimal strategies, strategies that will win no matter how the opponent plays, perhaps more interesting are those that do not possess an objective best strategy. In this paper, we examine three such games: the iterated prisoner’s dilemma, Incan Gold, and Dominion. Through computer simulations, we attempted to develop strategies for each game that could win most of the time, though not necessarily all of the time. We strived to create strategies that had a large breadth of success; that is, facing most common strategies of each game, ours would emerge on top. We begin with an analysis of the iterated prisoner’s dilemma, running an Axelrod-style tournament to determine the balance of characteristics of winning strategies. Next, we turn our attention to Incan Gold, where we examine the ramifications of different styles of decision making, hoping to enhance an already powerful strategy.
    [Show full text]
  • Math-GAMES IO1 EN.Pdf
    Math-GAMES Compendium GAMES AND MATHEMATICS IN EDUCATION FOR ADULTS COMPENDIUMS, GUIDELINES AND COURSES FOR NUMERACY LEARNING METHODS BASED ON GAMES ENGLISH ERASMUS+ PROJECT NO.: 2015-1-DE02-KA204-002260 2015 - 2018 www.math-games.eu ISBN 978-3-89697-800-4 1 The complete output of the project Math GAMES consists of the here present Compendium and a Guidebook, a Teacher Training Course and Seminar and an Evaluation Report, mostly translated into nine European languages. You can download all from the website www.math-games.eu ©2018 Erasmus+ Math-GAMES Project No. 2015-1-DE02-KA204-002260 Disclaimer: "The European Commission support for the production of this publication does not constitute an endorsement of the contents which reflects the views only of the authors, and the Commission cannot be held responsible for any use which may be made of the information contained therein." This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. ISBN 978-3-89697-800-4 2 PRELIMINARY REMARKS CONTRIBUTION FOR THE PREPARATION OF THIS COMPENDIUM The Guidebook is the outcome of the collaborative work of all the Partners for the development of the European Erasmus+ Math-GAMES Project, namely the following: 1. Volkshochschule Schrobenhausen e. V., Co-ordinating Organization, Germany (Roland Schneidt, Christl Schneidt, Heinrich Hausknecht, Benno Bickel, Renate Ament, Inge Spielberger, Jill Franz, Siegfried Franz), reponsible for the elaboration of the games 1.1 to 1.8 and 10.1. to 10.3 2. KRUG Art Movement, Kardzhali, Bulgaria (Radost Nikolaeva-Cohen, Galina Dimova, Deyana Kostova, Ivana Gacheva, Emil Robert), reponsible for the elaboration of the games 2.1 to 2.3 3.
    [Show full text]
  • The Pgsolver Collection of Parity Game Solvers
    Report The PGSolver Collection of Parity Game Solvers Version 3 Oliver Friedmann Martin Lange Institut f¨urInformatik Ludwig-Maximilians-Universit¨atM¨unchen, Germany March 5, 2014 1 1 5 5 8 5 1 7 2 0 Contents 1 Introduction 3 1.1 The Importance of Parity Games . .3 1.2 Aim and Content of PGSolver .......................3 1.3 Structure of this Report . .4 2 Solving Parity Games 6 2.1 Technicalities . .6 2.1.1 Parity Games . .6 2.1.2 Dominions . .8 2.1.3 Attractors and Attractor Strategies . .8 2.2 Local vs. Global Solvers . .8 2.3 Verifying Strategies . 10 2.4 Universal Optimisations . 10 2.4.1 SCC Decomposition . 10 2.4.2 Detection of Special Cases . 11 2.4.3 Priority Compression . 12 2.4.4 Priority Propagation . 13 2.5 The Generic Solver . 13 2.6 Implemented Algorithms . 14 2.6.1 The Recursive Algorithm . 14 2.6.2 The Local Model Checking Algorithm . 15 2.6.3 The Strategy Improvement Algorithm . 16 2.6.4 The Optimal Strategy Improvement Method . 16 2.6.5 The Strategy Improvement by Reduction to Discounted Payoff Games . 17 2.6.6 Probabilistic Strategy Improvement . 17 2.6.7 Probabilistic Strategy Improvement 2 . 17 2.6.8 Local Strategy Improvement . 18 2.6.9 The Small Progress Measures Algorithm . 18 2.6.10 The Small Progress Measures Reduction to SAT . 19 2.6.11 The Direct Reduction to SAT . 20 2.6.12 The Dominion Decomposition Algorithm . 20 2.6.13 The Big-Step Algorithm . 20 2.7 Implemented Heuristics .
    [Show full text]
  • Introduction to Computing at SIO: Notes for Fall Class, 2018
    Introduction to Computing at SIO: Notes for Fall class, 2018 Peter Shearer Scripps Institution of Oceanography University of California, San Diego September 24, 2018 ii Contents 1 Introduction 1 1.1 Scientific Computing at SIO . .2 1.1.1 Hardware . .2 1.1.2 Software . .2 1.2 Why you should learn a \real" language . .3 1.3 Learning to program . .4 1.4 FORTRAN vs. C . .5 1.5 Python . .6 2 UNIX introduction 7 2.1 Getting started . .7 2.2 Basic commands . .9 2.3 Files and editing . 12 2.4 Basic commands, continued . 14 2.4.1 Wildcards . 15 2.4.2 The .bashrc and .bash profile files . 16 2.5 Scripts . 18 2.6 File transfer and compression . 21 2.6.1 FTP command . 21 2.6.2 File compression . 22 2.6.3 Using the tar command . 23 2.6.4 Remote logins and job control . 24 2.7 Miscellaneous commands . 26 2.7.1 Common sense . 28 2.8 Advanced UNIX . 28 2.8.1 Some sed and awk examples . 30 2.9 Example of UNIX script to process data . 31 2.10 Common UNIX command summary . 35 3 A GMT Tutorial 37 3.0.1 Setting default values . 52 4 LaTeX 53 4.1 A simple example . 54 4.2 Example with equations . 55 4.3 Changing the default parameters . 57 4.3.1 Font size . 57 4.3.2 Font attributes . 57 4.3.3 Line spacing . 58 iii iv CONTENTS 4.4 Including graphics . 59 4.5 Want to know more? .
    [Show full text]
  • Mechanical Tic-Tac-Toe Board
    Mechanical Tic-Tac-Toe Board A Major Qualifying Project Report submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE in partial fulfilment of the requirements for the Degree of Bachelor of Science. By: Shane Bell Timothy Bill Abigail McAdams Taylor Teed Spring 2018 Submitted to: Professor Robert Daniello Mechanical Engineering Department 1 Abstract The goal of this MQP was to design, test and build a fully mechanical computer capable of playing tic-tac-toe against a human being. There is no question that modern computers could solve this problem more efficiently. However, our team aims to prove that old school technology still has a place in society today. Our design includes numerous methods of mechanical motion that are found in many designs today such as an escapement, gears, racks and pinions, and hydraulics. The machine was built almost entirely in the Higgins machine shop, except for a couple parts that were either cut with a water-jet or purchased. Our design uses an indexing module to detect position and data stored on a physical punch card to produce the best strategic answer. If the user makes the first move, the computer will never lose, only win or tie. 2 Acknowledgements Our group would like that thank our advisor Robert Daniello for his guidance and support throughout the duration of our project, Thomas Kouttron and Michael Cooke for their positive attitudes and constant aid in manufacturing, and Dane Kouttron for allowing access to machines not available within WPI facilities. We would also like to thank James Loiselle, Ian Anderson, Karl Ehlers, and Cam Collins for their good hearted aid, teaching our team the ins and outs of computer aided manufacturing and helping to figure out machining problems that we could not.
    [Show full text]
  • Learning Based Game Playing Agents for Abstract Board Games with Increasing State-Space Complexity
    ABSTRACT ANALYSIS OF Q- LEARNING BASED GAME PLAYING AGENTS FOR ABSTRACT BOARD GAMES WITH INCREASING STATE-SPACE COMPLEXITY by Indrima Upadhyay This thesis investigates Q-learning agents in a Reinforcement Learning framework for abstract board games. The two key contributions are: exploring the training of Q-learning agents, and a methodology to evaluate different agents playing abstract games. We focus on - Tic-Tac-Toe, Nine-Men’s Morris, and Mancala, noting that each of these games is solved. To train our Q-agent, we test the impact of a teaching agent (Q-learning twin, deterministic Min-Max, and a non-deterministic Min-Max) based on the number of training epochs needed until an agent converges. We find that a deterministic Min-Max is the best teaching agent, but a Q- learning twin allows us to converge without needing a pre-existing agent. In training, we include a methodology to create weaker agents for entertainment purposes. To evaluate a range of competitive agents, we provide a methodology and conduct a round- robin tournament. We find that a deterministic Min-Max agent scores maximum points for all three games, the Q-learning based agent places second for both Tic-Tac-Toe and Nine Men’s Morris, and a non-deterministic Min-Max places second for Mancala. From these two contributions we summarize our conclusions and provide discussion on these results and provide insight on how this space might be further investigated. ANALYSIS OF Q- LEARNING BASED GAME PLAYING AGENTS FOR ABSTRACT BOARD GAMES WITH INCREASING STATE-SPACE COMPLEXITY A Thesis Submitted to the Faculty of Miami University in partial fulfillment of the requirements for the degree of Master of Science by Indrima Upadhyay Miami University Oxford, Ohio 2021 Advisor: Dr.
    [Show full text]
  • Measuring the Solution Strength of Learning Agents in Adversarial Perfect Information Games
    Measuring the Solution Strength of Learning Agents in Adversarial Perfect Information Games Zaheen Farraz Ahmad, Nathan Sturtevant, Michael Bowling Department of Computing Science University of Alberta, Amii Edmonton, AB fzfahmad, nathanst, [email protected] Abstract were all evaluated against humans and demonstrated strate- gic capabilities that surpassed even top-level professional Self-play reinforcement learning has given rise to capable human players. game-playing agents in a number of complex domains such A commonly used method to evaluate the performance of as Go and Chess. These players were evaluated against other agents in two-player games is to have the agents play against state-of-the-art agents and professional human players and have demonstrated competence surpassing these opponents. other agents with theoretical guarantees of performance, or But does strong competition performance also mean the humans in a number of one-on-one games and then mea- agents can (weakly or strongly) solve the game? Or even suring the proportions of wins and losses. The agent which approximately solve the game? No existing work has con- wins the most games against the others is said to be the more sidered this question. We propose aligning our evaluation of capable agent. However, this metric only provides a loose self-play agents with metrics of strong/weakly solving strate- ranking among the agents — it does not provide a quantita- gies to provide a measure of an agent’s strength. Using small tive measure of the actual strength of the agents or how well games, we establish methodology on measuring the strength they generalize to their respective domains.
    [Show full text]
  • COMBINATORIAL GAME THEORY in CLOBBER Jeroen Claessen
    COMBINATORIAL GAME THEORY IN CLOBBER Jeroen Claessen Master Thesis DKE 11-06 Thesis submitted in partial fulfilment of the requirements for the degree of Master of Science of Artificial Intelligence at the Faculty of Humanities and Sciences of Maastricht University Thesis committee: Dr. M.H.M. Winands Dr. ir. J.W.H.M. Uiterwijk J.A.M. Nijssen, M.Sc. Maastricht University Department of Knowledge Engineering Maastricht, The Netherlands July 2011 Preface This master thesis was written at the Department of Knowledge Engineering at Maastricht University. In this thesis I investigate the application of combinatorial game theory in the game Clobber, and its incorporation in a Monte-Carlo Tree Search based Clobber player, with the aim of improving its perfor- mance. I would like to thank the following people for making this thesis possible. Most importantly, I would like to express my gratitude to my supervisor Dr. Mark Winands for his guidance during the past months. He has supplied me with many ideas and has taken the time to read and correct this thesis every week. Further, I would like to thank the other committee members for assessing this thesis. Finally, I would like to thank my family and friends for supporting me while I was writing this thesis. Jeroen Claessen Limbricht, June 2011 Abstract The past couple of decades classic board games have received much attention in the field of Artificial Intelligence. Recently, increasingly more research has been done on modern board games, because of the more complicated nature of these games. They are often non-deterministic games with imperfect information, that can be played by more than two players, which leads to new challenges.
    [Show full text]
  • Solving Difficult Game Positions
    Solving Difficult Game Positions Solving Difficult Game Positions PROEFSCHRIFT ter verkrijging van de graad van doctor aan de Universiteit Maastricht, op gezag van de Rector Magnificus, Prof. mr. G.P.M.F. Mols, volgens het besluit van het College van Decanen, in het openbaar te verdedigen op woensdag 15 december 2010 om 14.00 uur door Jahn-Takeshi Saito Promotor: Prof. dr. G. Weiss Copromotor: Dr. M.H.M. Winands Dr. ir. J.W.H.M. Uiterwijk Leden van de beoordelingscommissie: Prof. dr. ir. R.L.M. Peeters (voorzitter) Prof. dr. T. Cazenave (Universit´eParis-Dauphine) Prof. dr. M. Gyssens (Universiteit Hasselt / Universiteit Maastricht) Prof. dr. ir. J.C. Scholtes Prof. dr. C. Witteveen (Technische Universiteit Delft) The research has been funded by the Netherlands Organisation for Scientific Research (NWO), in the framework of the project Go for Go, grant number 612.066.409. Dissertation Series No. 2010-49 The research reported in this thesis has been carried out under the auspices of SIKS, the Dutch Research School for Information and Knowledge Systems. ISBN: 978-90-8559-164-1 c 2010 J.-T. Saito All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronically, mechanically, photo- copying, recording or otherwise, without prior permission of the author. Preface After receiving my Master's degree in Computational Linguistics and Artificial Intel- ligence in Osnabruck,¨ I faced the pleasant choice between becoming a Ph.D. student in Osnabruck¨ or Maastricht. The first option would have led me further into the field of Computational Linguistics.
    [Show full text]