Optimal Search in Discrete Locations: Extensions and New Findings

Total Page:16

File Type:pdf, Size:1020Kb

Optimal Search in Discrete Locations: Extensions and New Findings Optimal Search in Discrete Locations: Extensions and New Findings Jake Clarkson, M.Sci.(Hons.), M.Res Submitted for the degree of Doctor of Philosophy at Lancaster University. July 2020 Abstract A hidden target needs to be found by a searcher in many real-life situations, some of which involve large costs and significant consequences with failure. Therefore, efficient search methods are paramount. In our search model, the target lies in one of several discrete locations according to some hiding distribution, and the searcher's goal is to discover the target in minimum expected time by making successive searches of individual locations. In Part I of the thesis, the searcher knows the hiding distribution. Here, if there is only one way to search each location, the solution to the search problem, discovered in the 1960s, is simple; search next any location with a maximal probability per unit time of detecting the target. An equivalent solution is derived by viewing the search problem as a multi-armed bandit and following a Gittins index policy. Motivated by modern search technology, we introduce two modes|fast and slow| to search each location. The fast mode takes less time, but the slow mode is more likely to find the target. An optimal policy is difficult to obtain in general, because it requires an optimal sequence of search modes for each location, in addition to a set of sequence-dependent Gittins indices for choosing between locations. For each mode, we identify a sufficient condition for a location to use only that search mode in an optimal policy. For locations meeting neither sufficient condition, an optimal choice of search mode is extremely complicated, depending both on the hiding distribution and the search parameters of the other locations. We propose several heuristic policies motivated by our analysis, and demonstrate their near-optimal performance in an I Abstract II extensive numerical study. In Part II of the thesis, the searcher has only one search mode per location, but does not know the hiding distribution, which is chosen by an intelligent hider who aims to maximise the expected time until the target is discovered. Such a search game, modelled via two-person, zero-sum game theory, is relevant if the target is a bomb, intruder, or, of increasing importance due to advances in technology, a computer hacker. By Part I, if the hiding distribution is known, an optimal counter strategy for the searcher is any corresponding Gittins index policy. To develop an optimal search strategy in the search game, the searcher must account for the hider's motivation to choose an optimal hiding distribution, and consider the set of corresponding Gittins index policies. However, the searcher must choose carefully from this set of Gittins index policies to ensure the same expected time to discover the target regardless of where it is hidden by the hider. As a result, finding an optimal search strategy, or even proving one exists, is difficult. We extend several results for special cases from the literature to the fully-general search game; in particular, we show an optimal search strategy exists and may take a simple form. Using a novel test, we investigate the frequency of the optimality of a particular hiding strategy that gives the searcher no preference over any location at the beginning of the search. Acknowledgements I'll begin by thanking my supervisors, without whom this PhD would not have been possible. First, Kevin. It seems a long time since I mistakenly thought you were called Keith (coincidently, the name of your twin) after my STOR-i welcome day! Since then we've spent many hours discussing challenging problems in your office, but enough about the owners of our respective football clubs; I think we made more progress talking about search problems. Thanks for your invaluable guidance, encouragement and wisdom regarding not only the PhD but also my future career. You were also an excellent tour guide in the Monterey Peninsula, teaching me a lot about, inter alia, wildlife (sorry no whale watching) and Motown. I also encountered some fantastic new fish-based dishes there; who couldn't get swept up in your love of cioppino and, in particular, clam chowder? Monterey brings me on to Kyle, who enriched each NPS visit by being a most-welcoming host. Yet, your influence extended far beyond Monterey trips, thank you for your continual help and meticulous feedback throughout the three years - your input was crucial. In particular, you've helped my writing; now I tend to find that I use less unnecessary words and phrases and, therefore, can be more `to the point', so to speak, without losing any information that I wish to convey to the reader I'm more succinct. I hope to work more with both of you in the future. Trips to NPS and the PhD itself would not have been possible without the financial support of ESPRC, of which I am grateful, and, of course, STOR-i itself, a wonderfully supportive, friendly and vibrant atmosphere in which to do research. I'd like to thank everyone involved at STOR-i, staff and students alike, but in particular Jon Tawn and III Acknowledgements IV the admin team, Kim Wilson, Jen Bull and Wendy Shimmin, who make everything run so smoothly whilst all being so friendly and approachable. I'd especially like to thank them for dealing with my requests regarding lumbar support and ever-changing thesis hand in date. I've had so many enjoyable experiences at STOR-i: dressing up for the Christmas meal, dancing until cramp hits at the STOR-i ball, away days (the good and the `interesting'). I've made lots of great friends who I'd like to thank for enduring my endless tales of buses, bargains and dreams, humouring my latest fad, and supporting me through tougher times. Since I celebrated it recently, in reverse-chronological order of the date of their golden birthday (Paperless Post, 2020), I'd like, in particular, to thank: Graham: no-one else takes such an interest in my dreams (or my kitchen uten- sils...); I raise my eyebrows at others! Sean: at his peak, a fantastic food critic and user of satire, and otherwise still pretty funny. Lucy: our humour never failed to brighten my day, most of it Terribly Silly, some slightly crude, laughing until we're both crazy like a fool! Sam: the most superb, astounding, majestic good sir, someone Hoo's always great for a Chuckle no matter Watt the situation; you think of everything! Thank you also for introducing me to the concept of a `golden birthday'; I'm sure your unbelievable memory for dates will correct me if the order of this list is wrong. Emily: always there for support whenever I needed it, and always up for a good old natter! Never lose your Twinkling Sparkle! Kathryn: a regularly-supportive ear over peppermint tea, always helping put the pieces into place ... Luke: a fount of wise words and stimulating conversation, not least about who should open and how many spinners to play in the next test. Euan: a natural quipster, who kept me from going nuts by turning the seemingly Acknowledgements V mundane into a zesty mix of trpamolining dogs, fetching emails and much more. Next come the only four people I've met who come close to making me look sane. Jack: the man with the best-smelling arms in Devon. Bike-safety gear aside, a fellow seeker of good value; I got the best value with your friendship. David: a bin-diving, graffiti-dishing, ‘fit’-assessing, Spaniard-in-Belgium-spotting, sign-slapping, `vegetarian' Tasmanian devil. Whilst your antics never failed to en- lighten my day, in some tough times you've been a knight in shining armour. Harry: a frequent partner in crime, always egging eachother on in our latest craze. A refreshingly-free spirit with relentless enthusiasm and tr`esBon life advice when required. Er ist alles in Butter (as long as it's grass fed). Harjit: I've mentioned the best Baker, now time for the worst baker, the confident, insightful, whimsical and practical Harjit Hullait. Thanks for all the moments that can only be described as `CH' (and C doesn't stand for conventional); social events were always better when you hadn't food in the oven. And that just leaves . Rob, my LIC1 for 8 (eight) glorious years, where do I start? From the Brummie hipster who I Met on an unquestionably-hot day in October 2011, you've changed a fair bit. You've learnt about the washing machine, how to use pots and pans (as opposed to baking trays and more baking trays), and that AM doesn't end when you go to bed. On a serious note, your excellent sense of humour, willingness to play `guessing games' and continuing interest in my Cricket Captain saves has significantly enriched my time at Lancaster; we've laughed uncontrollably an uncountable number of times, mainly at things an outsider would find most baffling. Away from STOR-i, I'd like to thank my home friends (who've been bored my tales far more than anyone else) Aidan, Imogen, James and Robyn for many jokes, mainly directed (amicably) at eachother. I'd also like to thank Andrew for his Terrific Sass. Finally, I'd like to thank my family, especially my parents, grandparents and sister. 1live-in-chum Acknowledgements VI I don't think they were expecting me to have 23 years of education, but they've been a pillar of strength throughout all of it. I consider myself extremely lucky to have them; they've picked me up when I'm down, and we've regularly laughed with (and often at) eachother in better times.
Recommended publications
  • Rendezvous Search with Revealed Information: Applications to the Line
    J. Appl. Prob. 44, 1–15 (2007) Printed in England © Applied Probability Trust 2007 RENDEZVOUS SEARCH WITH REVEALED INFORMATION: APPLICATIONS TO THE LINE STEVE ALPERN,∗ London School of Economics Abstract The symmetric rendezvous problem on a network Q asks how two players, forced to use the same mixed strategy, can minimize their expected meeting time, starting from a known initial distribution on the nodes of Q. This minimum is called the (symmetric) ‘rendezvous value’of Q. Traditionally, the players are assumed to receive no information while playing the game. We consider the effect on rendezvous times of giving the players some information about past actions and chance moves, enabling each of them to apply Bayesian updates to improve his knowledge of the other’s whereabouts. This technique can be used to give lower bounds on the rendezvous times of the original game (without any revealed information). We consider the case in which they are placed a known distance apart on the line graph Q (known as ‘symmetric rendezvous on the line’). Our approach is to concentrate on a general analysis of the effect of revelations, rather than compute the best bounds possible with our technique. Keywords: Rendezvous; search; common interest game; team theory 2000 Mathematics Subject Classification: Primary 90B40 1. Introduction In this paper we consider how the revelation of information can affect an important unsolved problem in rendezvous theory: the symmetric rendezvous problem on the line or line graph (whose nodes are the integers, Z) posed in [1], dubbed . (In fact, given the recent solution by Weber [14] of the symmetric rendezvous problem on the complete graph K3,is arguably the main unsolved problem.) The problem begins with a chance move α, called the initial configuration.
    [Show full text]
  • Adversarial Search (Game)
    Adversarial Search CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2019 Soleymani “Artificial Intelligence: A Modern Approach”, 3rd Edition, Chapter 5 Most slides have been adopted from Klein and Abdeel, CS188, UC Berkeley. Outline } Game as a search problem } Minimax algorithm } �-� Pruning: ignoring a portion of the search tree } Time limit problem } Cut off & Evaluation function 2 Games as search problems } Games } Adversarial search problems (goals are in conflict) } Competitive multi-agent environments } Games in AI are a specialized kind of games (in the game theory) 3 Adversarial Games 4 Types of Games } Many different kinds of games! } Axes: } Deterministic or stochastic? } One, two, or more players? } Zero sum? } Perfect information (can you see the state)? } Want algorithms for calculating a strategy (policy) which recommends a move from each state 5 Zero-Sum Games } Zero-Sum Games } General Games } Agents have opposite utilities } Agents have independent utilities (values on outcomes) (values on outcomes) } Lets us think of a single value that one } Cooperation, indifference, competition, maximizes and the other minimizes and more are all possible } Adversarial, pure competition } More later on non-zero-sum games 6 Primary assumptions } We start with these games: } Two -player } Tu r n taking } agents act alternately } Zero-sum } agents’ goals are in conflict: sum of utility values at the end of the game is zero or constant } Perfect information } fully observable Examples: Tic-tac-toe,
    [Show full text]
  • Best-First Minimax Search Richard E
    Artificial Intelligence ELSEVIER Artificial Intelligence 84 ( 1996) 299-337 Best-first minimax search Richard E. Korf *, David Maxwell Chickering Computer Science Department, University of California, Los Angeles, CA 90024, USA Received September 1994; revised May 1995 Abstract We describe a very simple selective search algorithm for two-player games, called best-first minimax. It always expands next the node at the end of the expected line of play, which determines the minimax value of the root. It uses the same information as alpha-beta minimax, and takes roughly the same time per node generation. We present an implementation of the algorithm that reduces its space complexity from exponential to linear in the search depth, but at significant time cost. Our actual implementation saves the subtree generated for one move that is still relevant after the player and opponent move, pruning subtrees below moves not chosen by either player. We also show how to efficiently generate a class of incremental random game trees. On uniform random game trees, best-first minimax outperforms alpha-beta, when both algorithms are given the same amount of computation. On random trees with random branching factors, best-first outperforms alpha-beta for shallow depths, but eventually loses at greater depths. We obtain similar results in the game of Othello. Finally, we present a hybrid best-first extension algorithm that combines alpha-beta with best-first minimax, and performs significantly better than alpha-beta in both domains, even at greater depths. In Othello, it beats alpha-beta in two out of three games. 1. Introduction and overview The best chess machines, such as Deep-Blue [lo], are competitive with the best humans, but generate billions of positions per move.
    [Show full text]
  • Exploding Offers with Experimental Consumer Goods∗
    Exploding Offers with Experimental Consumer Goods∗ Alexander L. Brown Ajalavat Viriyavipart Department of Economics Department of Economics Texas A&M University Texas A&M University Xiaoyuan Wang Department of Economics Texas A&M University August 29, 2014 Abstract Recent theoretical research indicates that search deterrence strategies are generally opti- mal for sellers in consumer goods markets. Yet search deterrence is not always employed in such markets. To understand this incongruity, we develop an experimental market where profit-maximizing strategy dictates sellers should exercise one form of search deterrence, ex- ploding offers. We find that buyers over-reject exploding offers relative to optimal. Sellers underutilize exploding offers relative to optimal play, even conditional on buyer over-rejection. This tendency dissipates when sellers make offers to computerized buyers, suggesting their persistent behavior with human buyers may be due to a preference rather than a miscalculation. Keywords: exploding offer, search deterrence, experimental economics, quantal response equi- librium JEL: C91 D21 L10 M31 ∗We thank the Texas A&M Humanities and Social Science Enhancement of Research Capacity Program for providing generous financial support for our research. We have benefited from comments by Gary Charness, Catherine Eckel, Silvana Krasteva and seminar participants of the 2013 North American Economic Science Association and Southern Economic Association Meetings. Daniel Stephenson and J. Forrest Williams provided invaluable help in conducting the experimental sessions. 1 Introduction One of the first applications of microeconomic theory is to the consumer goods market: basic supply and demand models can often explain transactions in centralized markets quite well. When markets are decentralized, often additional modeling is required to explain the interactions of consumers and producers.
    [Show full text]
  • Adversarial Search (Game Playing)
    Artificial Intelligence Adversarial Search (Game Playing) Chapter 5 Adapted from materials by Tim Finin, Marie desJardins, and Charles R. Dyer Outline • Game playing – State of the art and resources – Framework • Game trees – Minimax – Alpha-beta pruning – Adding randomness State of the art • How good are computer game players? – Chess: • Deep Blue beat Gary Kasparov in 1997 • Garry Kasparav vs. Deep Junior (Feb 2003): tie! • Kasparov vs. X3D Fritz (November 2003): tie! – Checkers: Chinook (an AI program with a very large endgame database) is the world champion. Checkers has been solved exactly – it’s a draw! – Go: Computer players are decent, at best – Bridge: “Expert” computer players exist (but no world champions yet!) • Good place to learn more: http://www.cs.ualberta.ca/~games/ Chinook • Chinook is the World Man-Machine Checkers Champion, developed by researchers at the University of Alberta. • It earned this title by competing in human tournaments, winning the right to play for the (human) world championship, and eventually defeating the best players in the world. • Visit http://www.cs.ualberta.ca/~chinook/ to play a version of Chinook over the Internet. • The developers have fully analyzed the game of checkers and have the complete game tree for it. – Perfect play on both sides results in a tie. • “One Jump Ahead: Challenging Human Supremacy in Checkers” Jonathan Schaeffer, University of Alberta (496 pages, Springer. $34.95, 1998). Ratings of human and computer chess champions Typical case • 2-person game • Players alternate moves • Zero-sum: one player’s loss is the other’s gain • Perfect information: both players have access to complete information about the state of the game.
    [Show full text]
  • Learning Near-Optimal Search in a Minimal Explore/Exploit Task
    Explore/exploit tradeoff strategies with accumulating resources 1 Explore/exploit tradeoff strategies in a resource accumulation search task [revised/resubmitted to Cognitive Science] Ke Sang1, Peter M. Todd1, Robert L. Goldstone1, and Thomas T. Hills2 1 Cognitive Science Program and Department of Psychological and Brain Sciences, Indiana University Bloomington 2 Department of Psychology, University of Warwick REVISION 7/5/2018 Contact information: Peter M. Todd Indiana University, Cognitive Science Program 1101 E. 10th Street, Bloomington, IN 47405 USA phone: +1 812 855-3914 email: [email protected] Explore/exploit tradeoff strategies with accumulating resources 2 Abstract How, and how well, do people switch between exploration and exploitation to search for and accumulate resources? We study the decision processes underlying such exploration/exploitation tradeoffs by using a novel card selection task. With experience, participants learn to switch appropriately between exploration and exploitation and approach optimal performance. We model participants’ behavior on this task with random, threshold, and sampling strategies, and find that a linear decreasing threshold rule best fits participants’ results. Further evidence that participants use decreasing threshold-based strategies comes from reaction time differences between exploration and exploitation; however, participants themselves report non-decreasing thresholds. Decreasing threshold strategies that “front-load” exploration and switch quickly to exploitation are particularly effective
    [Show full text]
  • CS 561: Artificial Intelligence
    CS 561: Artificial Intelligence Instructor: Sofus A. Macskassy, [email protected] TAs: Nadeesha Ranashinghe ([email protected]) William Yeoh ([email protected]) Harris Chiu ([email protected]) Lectures: MW 5:00-6:20pm, OHE 122 / DEN Office hours: By appointment Class page: http://www-rcf.usc.edu/~macskass/CS561-Spring2010/ This class will use http://www.uscden.net/ and class webpage - Up to date information - Lecture notes - Relevant dates, links, etc. Course material: [AIMA] Artificial Intelligence: A Modern Approach, by Stuart Russell and Peter Norvig. (2nd ed) This time: Outline (Adversarial Search – AIMA Ch. 6] Game playing Perfect play ◦ The minimax algorithm ◦ alpha-beta pruning Resource limitations Elements of chance Imperfect information CS561 - Lecture 7 - Macskassy - Spring 2010 2 What kind of games? Abstraction: To describe a game we must capture every relevant aspect of the game. Such as: ◦ Chess ◦ Tic-tac-toe ◦ … Accessible environments: Such games are characterized by perfect information Search: game-playing then consists of a search through possible game positions Unpredictable opponent: introduces uncertainty thus game-playing must deal with contingency problems CS561 - Lecture 7 - Macskassy - Spring 2010 3 Searching for the next move Complexity: many games have a huge search space ◦ Chess: b = 35, m=100 nodes = 35 100 if each node takes about 1 ns to explore then each move will take about 10 50 millennia to calculate. Resource (e.g., time, memory) limit: optimal solution not feasible/possible, thus must approximate 1. Pruning: makes the search more efficient by discarding portions of the search tree that cannot improve quality result. 2. Evaluation functions: heuristics to evaluate utility of a state without exhaustive search.
    [Show full text]
  • Optimal Symmetric Rendezvous Search on Three Locations
    Optimal Symmetric Rendezvous Search on Three Locations Richard Weber Centre for Mathematical Sciences, Wilberforce Road, Cambridge CB2 0WB email: [email protected] www.statslab.cam.ac.uk/~rrw1 In the symmetric rendezvous search game played on n locations two players are initially placed at two distinct locations. The game is played in discrete steps, at each of which each player can either stay where he is or move to a different location. The players share no common labelling of the locations. We wish to find a strategy such that, if both players follow it independently, then the expected number of steps until they are in the same location is minimized. Informal versions of the rendezvous problem have received considerable attention in the popular press. The problem was proposed by Steve Alpern in 1976 and it has proved notoriously difficult to analyse. In this paper we prove a 20 year old conjecture that the following strategy is optimal for the game on three locations: in each block of two steps, stay where you, or search the other two locations in random order, doing these with probabilities 1=3 and 2=3 respectively. This is now the first nontrivial symmetric rendezvous search game to be fully solved. Key words: rendezvous search; search games; semidefinite programming MSC2000 Subject Classification: Primary: 90B40; Secondary: 49N75, 90C22, 91A12, 93A14 OR/MS subject classification: Primary: Games/group decisions, cooperative; Secondary: Analysis of algorithms; Programming, quadratic 1. Symmetric rendezvous search on three locations In the symmetric rendezvous search game played on n locations two players are initially placed at two distinct locations.
    [Show full text]
  • Algorithm Design and Analysis in Wireless Networks Lin Chen
    Algorithm design and analysis in wireless networks Lin Chen To cite this version: Lin Chen. Algorithm design and analysis in wireless networks. Data Structures and Algorithms [cs.DS]. Paris-Sud XI, 2017. tel-01688436 HAL Id: tel-01688436 https://hal.archives-ouvertes.fr/tel-01688436 Submitted on 24 Jan 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Algorithm Design and Analysis in Wireless Networks Habilitation Thesis Lin CHEN Laboratoire de Recherche en Informatique (UMR 8623) Université Paris-Sud Defended on July 11, 2017 Committee M. Tamer BASAR Reviewer M. Pierre FRAIGNIAUD Reviewer M. Bruno GAUJAL Reviewer M. Eitan ALTMAN Examiner M. Joffroy BEAUQUIER Examiner Mme. Johanne COHEN Examiner M. Mérouane DEBBAH Examiner M. Fabio MARTIGNON Examiner Abstract Algorithms are perhaps the most fundamental and fascinating elements in computer science as a whole. Networks and networked systems are no exception. This habilitation thesis summarizes my research during the last eight years on some algorithmic problems of both fundamental and practical importance in modern networks and networked systems, more specifically, wireless net- works. Generically, wireless networks have a number of common features which form a common ground on which algorithms for wireless networks are designed.
    [Show full text]
  • THE THEORY of SEARCH GAMES and RENDEZVOUS INTERNATIONAL SERIES in OPERATIONS RESEARCH & MANAGEMENT SCIENCE Frederick S
    THE THEORY OF SEARCH GAMES AND RENDEZVOUS INTERNATIONAL SERIES IN OPERATIONS RESEARCH & MANAGEMENT SCIENCE Frederick S. Hillier, Series Editor Stanford University Weyant, J./ ENERGY AND ENVIRONMENTAL POLICY MODELING Shanthikumar, J.G. & Sumita, U. / APPLIED PROBABILITY AND STOCHASTIC PROCESSES Liu, B. & Esogbue, A.O. / DECISION CRITERIA AND OPTIMAL INVENTORY PROCESSES Gal, T., Stewart, T.J., Hanne, T. / MULTICRITERIA DECISION MAKING: Advances in MCDM Models, Algorithms, Theory, and Applications Fox, B.L. / STRATEGIES FOR QUASI-MONTE CARLO Hall, R.W. / HANDBOOK OF TRANSPORTATION SCIENCE Grassman, W.K. / COMPUTATIONAL PROBABILITY Pomerol, J-C. & Barba-Romero, S. / MULTICRITERION DECISION IN MANAGEMENT Axsäter, S./ INVENTORY CONTROL Wolkowicz, H., Saigal, R., & Vandenberghe, L. / HANDBOOK OF SEMI-DEFINITE PROGRAMMING: Theory, Algorithms, and Applications Hobbs, B.F. & Meier, P. / ENERGY DECISIONS AND THE ENVIRONMENT: A Guide to the Use of Multicriteria Methods Dar-El, E. / HUMAN LEARNING: From Learning Curves to Learning Organizations Armstrong, J.S. / PRINCIPLES OF FORECASTING: A Handbook for Researchers and Practitioners Balsamo, S., Personé, V., & Onvural, R./ ANALYSIS OF QUEUEING NETWORKS WITH BLOCKING Bouyssou, D. et al. / EVALUATION AND DECISION MODELS: A Critical Perspective Hanne, T. / INTELLIGENT STRATEGIES FOR META MULTIPLE CRITERIA DECISION MAKING Saaty, T. & Vargas, L. / MODELS, METHODS, CONCEPTS and APPLICATIONS OF THE ANALYTIC HIERARCHY PROCESS Chatterjee, K. & Samuelson, W. / GAME THEORY AND BUSINESS APPLICATIONS Hobbs, B. et al. / THE NEXT GENERATION OF ELECTRIC POWER UNIT COMMITMENT MODELS Vanderbei, R.J. / LINEAR PROGRAMMING: Foundations and Extensions, 2nd Ed. Kimms, A. / MATHEMATICAL PROGRAMMING AND FINANCIAL OBJECTIVES FOR SCHEDULING PROJECTS Baptiste, P., Le Pape, C. & Nuijten, W. / CONSTRAINT-BASED SCHEDULING Feinberg, E. & Shwartz, A.
    [Show full text]
  • Download Book
    Lecture Notes in Computer Science 3279 Commenced Publication in 1973 Founding and Former Series Editors: Gerhard Goos, Juris Hartmanis, and Jan van Leeuwen Editorial Board David Hutchison Lancaster University, UK Takeo Kanade Carnegie Mellon University, Pittsburgh, PA, USA Josef Kittler University of Surrey, Guildford, UK Jon M. Kleinberg Cornell University, Ithaca, NY, USA Friedemann Mattern ETH Zurich, Switzerland John C. Mitchell Stanford University, CA, USA Moni Naor Weizmann Institute of Science, Rehovot, Israel Oscar Nierstrasz University of Bern, Switzerland C. Pandu Rangan Indian Institute of Technology, Madras, India Bernhard Steffen University of Dortmund, Germany Madhu Sudan Massachusetts Institute of Technology, MA, USA Demetri Terzopoulos New York University, NY, USA Doug Tygar University of California, Berkeley, CA, USA Moshe Y. Vardi Rice University, Houston, TX, USA Gerhard Weikum Max-Planck Institute of Computer Science, Saarbruecken, Germany Geoffrey M. Voelker Scott Shenker (Eds.) Peer-to-Peer Systems III Third International Workshop, IPTPS 2004 La Jolla, CA, USA, February 26-27, 2004 Revised Selected Papers 13 Volume Editors Geoffrey M. Voelker University of California, San Diego Department of Computer Science and Engineering 9500 Gilman Dr., MC 0114, La Jolla, CA 92093-0114, USA E-mail: [email protected] Scott Shenker University of California, Berkeley Computer Science Division, EECS Department 683 Soda Hall, 1776, Berkeley, CA 94720, USA E-mail: [email protected] Library of Congress Control Number: Applied for CR Subject Classification (1998): C.2.4, C.2, H.3, H.4, D.4, F.2.2, E.1, D.2 ISSN 0302-9743 ISBN 3-540-24252-X Springer Berlin Heidelberg New York This work is subject to copyright.
    [Show full text]
  • Game Playing (Adversarial Search) CS 580 (001) - Spring 2018
    Lecture 5: Game Playing (Adversarial Search) CS 580 (001) - Spring 2018 Amarda Shehu Department of Computer Science George Mason University, Fairfax, VA, USA February 21, 2018 Amarda Shehu (580) 1 1 Outline of Today's Class 2 Games vs Search Problems 3 Perfect Play Minimax Decision Alpha-Beta Pruning 4 Resource Limits and Approximate Evaluation Expectiminimax 5 Games of Imperfect Information 6 Game Playing Summary Amarda Shehu (580) Outline of Today's Class 2 Game Playing { Adversarial Search Search in a multi-agent, competitive environment ! Adversarial Search/Game Playing Mathematical game theory treats any multi-agent environment as a game, with possibly co-operative behaviors (study of economies) Most games studied in AI: deterministic, turn-taking, two-player, zero-sum games of perfect information Amarda Shehu (580) Games vs Search Problems 3 Game Playing - Adversarial Search Most games studied in AI: deterministic, turn-taking, two-player, zero-sum games of perfect information zero-sum: utilities of the two players sum to 0 (no win-win) deterministic: precise rules with known outcomes perfect information: fully observable Amarda Shehu (580) Games vs Search Problems 4 Game Playing - Adversarial Search Most games studied in AI: deterministic, turn-taking, two-player, zero-sum games of perfect information zero-sum: utilities of the two players sum to 0 (no win-win) deterministic: precise rules with known outcomes perfect information: fully observable Search algorithms designed for such games make use of interesting general techniques (meta-heuristics) such as evaluation functions, search pruning, and more. However, games are to AI what grand prix racing is to automobile design.
    [Show full text]