Computing Game-Theoretic Solutions and Applications to Security∗

Total Page:16

File Type:pdf, Size:1020Kb

Computing Game-Theoretic Solutions and Applications to Security∗ Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence Computing Game-Theoretic Solutions and Applications to Security∗ Vincent Conitzer Departments of Computer Science and Economics Duke University Durham, NC 27708, USA [email protected] Abstract theory in the design of their agents, enabling them to reason strategically. For this to work, it is essential to design effi- The multiagent systems community has adopted game cient algorithms for computing game-theoretic solutions— theory as a framework for the design of systems of mul- algorithms that take a game as input, and produce a way to tiple self-interested agents. For this to be effective, effi- cient algorithms must be designed to compute the solu- play as output. tions that game theory prescribes. In this paper, I sum- The rest of this paper is organized as follows. We first dis- marize some of the state of the art on this topic, focus- cuss various ways to represent games. We then discuss vari- ing particularly on how this line of work has contributed ous solution concepts—notions of what it means to solve the to several highly visible deployed security applications, game—and whether and how these solutions can be com- developed at the University of Southern California. puted efficiently, with a focus on two-player normal-form games. Finally, we discuss exciting recent developments in the deployment of these techniques in high-value security Introduction applications. How should an agent choose its actions optimally? The stan- dard decision-theoretic framework is to calculate a utility for Representing Games every possible outcome, as well as, for each possible course of action, a conditional probability distribution over the pos- We first discuss how to represent games. This is important sible outcomes. Then, the agent should choose an action that first of all from a conceptual viewpoint: we cannot have gen- maximizes expected utility. However, this framework is not eral tools for analyzing games precisely without making the straightforward to apply when there are other rational agents games themselves precise. There are several standard repre- in the environment, who have potentially different objec- sentation schemes from the game theory literature that al- tives. In this case, the agent needs to determine a probability low us to think clearly about games. Additionally, the rep- distribution over the other agents’ actions. To do so, it is resentation scheme of course has a major impact on scala- natural to assume that those agents are themselves trying to bility. From this perspective, a good representation scheme optimize their actions for their own utility functions. But this allows us to compactly represent natural games, in the same quickly leads one into circularities: if the optimal action for sense that Bayesian networks allow us to compactly rep- agent 1 depends on agent 2’s action, but the optimal action resent many natural probability distributions; and it allows for agent 2 depends on agent 1’s action, how can we ever for efficient algorithms. While the standard representation come to a conclusion? This is precisely the type of prob- schemes from game theory are sufficient in some cases, in lem that game theory aims to address. It allows agents to other cases the representation size blows up exponentially. form beliefs over each other’s actions and to act rationally In this section, we review standard representation on those beliefs in a consistent manner. The word “game” schemes from game theory, as well as, briefly, representa- in game theory is used to refer to any strategic situation, in- tion schemes that have been introduced in the AI commu- cluding games in the common sense of the word, such as nity to represent families of natural games more compactly. board and card games, but also important domains that are Discussion of compact representation schemes for security not so playful, such as the security games that we will dis- games is postponed towards the end of the paper. cuss towards the end of this paper. Social scientists and even biologists use game theory to Normal-form games model behaviors observed in the world. In contrast, AI re- Perhaps the best-known representation scheme for games is searchers tend to be interested in constructively using game the normal form (also known as the strategic form, or, in the ∗This paper was invited as a “What’s Hot” paper to the case of two players, the bimatrix form). Here, each player i AAAI’12 Sub-Area Spotlights track. has a set of available pure strategies Si, of which she must Copyright c 2012, Association for the Advancement of Artificial choose one. Each player i also has a utility function ui : Intelligence (www.aaai.org). All rights reserved. S1 × ::: × Sn ! R which maps outcomes (consisting of a 2106 pure strategy for each player) to utilities. Two-player games are often written down as follows: R P S R 0,0 -1,1 1,-1 P 1,-1 0,0 -1,1 S -1,1 1,-1 0,0 This is the familiar game of rock-paper-scissors. Player 1 is the “row” player and selects a row, and player two is the “column” player and selects a column. The numbers in each entry are the utilities of the row and column players, respec- tively, for the corresponding outcome. Rock-paper-scissors is an example of a zero-sum game: in each entry, the num- Figure 1: An extremely simple poker game. bers sum to zero, so whatever one player gains, the other loses. (taking the expectation over Nature’s randomness). This re- sults in the following normal-form game: Extensive-form games CC CF FC FF Extensive-form games allow us to directly model the tempo- BB 0,0 0,0 1,-1 1,-1 ral and informational structure of a game. They are a gener- BC .5,-.5 1.5,-1.5 0,0 1,-1 alization of the game trees familiar to most AI researchers. CB -.5,.5 -.5,.5 1,-1 1,-1 They allow chance nodes (often called “moves by Nature” in CC 0,0 1,-1 0,0 1,-1 game theory), non-zero-sum outcomes, as well as imperfect information. In a perfect-information game, such as chess, While any extensive-form game can be converted to nor- checkers, or tic-tac-toe, nothing about the current state of mal form in this manner, the representation size can blow up the game is ever hidden from a player. In contrast, in (for exponentially; we will return to this problem later. example) most card games, no single player has full knowl- edge of the state of the game, because she does not know the Bayesian games cards of the other players. This is represented using infor- In recreational games, it is generally common knowledge mation sets: an information set is a set of nodes in the game how much each player values each outcome, because this is tree that all have the same acting player, such that the player specified by the rules of the game. In contrast, when mod- cannot distinguish these nodes from each other. eling real-world strategic situations, generally each player For example, consider the following simple (perhaps the has some uncertainty over how much the other players value simplest possible interesting) poker game. We have a two(!)- the different outcomes. This type of uncertainty is natu- card deck, with a King and a Jack. The former is a winning rally modeled by Bayesian games. In a Bayesian game, each card, the latter a losing card. Both players put $1 into the pot player draws a type from a distribution before playing the at the outset. Player 1 draws a card; player 2 does not get game. This type determines how much the player values a card. Player 1 must then choose to bet (put an additional each outcome.1 Each player knows her own type, but not $1 into the pot) or check (do nothing). Player 2 must then those of the other players. Again, Bayesian games can be choose to call (match whatever additional amount player 2 converted to normal form, at the cost of an exponential in- put into the pot, possibly $0), or fold (in which case player 1 crease in size. wins the pot). If player 2 calls, player 1 must show her card; if it is a King, she wins the pot, and if it is a Jack, player 2 Stochastic games wins the pot. Stochastic games are simply a multiplayer generalization of Figure 1 shows the extensive form of this game. The Markov decision processes. In each state, all players choose dashed lines between nodes for player 2 connect nodes in- an action, and the profile of actions selected determines the side the same information set, i.e., between which player 2 immediate rewards to all players as well as the transition cannot tell the difference (because he does not know which probabilities to other states. A stochastic game with only one card player 1 has). state is a repeated game (where the players play the same It is possible to convert extensive-form games to normal- normal-form game over and over). form games, as follows. A pure strategy in an extensive-form game consists of a plan that tells the player which action Compact representation schemes to take at every information set that she has in the game. One downside of the normal form is that its size is ex- (A player must choose the same action at every node inside ponential in the number of players. Several representation the same information set, because after all she cannot dis- schemes have been proposed to be able to represent “natu- tinguish them.) One pure strategy for player 1 is “Bet on a ral” games with many players compactly.
Recommended publications
  • Lecture Notes
    GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies ..................................
    [Show full text]
  • The Monty Hall Problem in the Game Theory Class
    The Monty Hall Problem in the Game Theory Class Sasha Gnedin∗ October 29, 2018 1 Introduction Suppose you’re on a game show, and you’re given the choice of three doors: Behind one door is a car; behind the others, goats. You pick a door, say No. 1, and the host, who knows what’s behind the doors, opens another door, say No. 3, which has a goat. He then says to you, “Do you want to pick door No. 2?” Is it to your advantage to switch your choice? With these famous words the Parade Magazine columnist vos Savant opened an exciting chapter in mathematical didactics. The puzzle, which has be- come known as the Monty Hall Problem (MHP), has broken the records of popularity among the probability paradoxes. The book by Rosenhouse [9] and the Wikipedia entry on the MHP present the history and variations of the problem. arXiv:1107.0326v1 [math.HO] 1 Jul 2011 In the basic version of the MHP the rules of the game specify that the host must always reveal one of the unchosen doors to show that there is no prize there. Two remaining unrevealed doors may hide the prize, creating the illusion of symmetry and suggesting that the action does not matter. How- ever, the symmetry is fallacious, and switching is a better action, doubling the probability of winning. There are two main explanations of the paradox. One of them, simplistic, amounts to just counting the mutually exclusive cases: either you win with ∗[email protected] 1 switching or with holding the first choice.
    [Show full text]
  • Incomplete Information I. Bayesian Games
    Bayesian Games Debraj Ray, November 2006 Unlike the previous notes, the material here is perfectly standard and can be found in the usual textbooks: see, e.g., Fudenberg-Tirole. For the examples in these notes (except for the very last section), I draw heavily on Martin Osborne’s excellent recent text, An Introduction to Game Theory, Oxford University Press. Obviously, incomplete information games — in which one or more players are privy to infor- mation that others don’t have — has enormous applicability: credit markets / auctions / regulation of firms / insurance / bargaining /lemons / public goods provision / signaling / . the list goes on and on. 1. A Definition A Bayesian game consists of 1. A set of players N. 2. A set of states Ω, and a common prior µ on Ω. 3. For each player i a set of actions Ai and a set of signals or types Ti. (Can make actions sets depend on type realizations.) 4. For each player i, a mapping τi :Ω 7→ Ti. 5. For each player i, a vN-M payoff function fi : A × Ω 7→ R, where A is the product of the Ai’s. Remarks A. Typically, the mapping τi tells us player i’s type. The easiest case is just when Ω is the product of the Ti’s and τi is just the appropriate projection mapping. B. The prior on states need not be common, but one view is that there is no loss of generality in assuming it (simply put all differences in the prior into differences in signals and redefine the state space and si-mappings accordingly).
    [Show full text]
  • (501B) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann
    Dirk Bergemann Department of Economics Yale University Microeconomic Theory (501b) Problem Set 5. Bayesian Games Suggested Solutions by Tibor Heumann 1. (Market for Lemons) Here I ask that you work out some of the details in perhaps the most famous of all information economics models. By contrast to discussion in class, we give a complete formulation of the game. A seller is privately informed of the value v of the good that she sells to a buyer. The buyer's prior belief on v is uniformly distributed on [x; y] with 3 0 < x < y: The good is worth 2 v to the buyer. (a) Suppose the buyer proposes a price p and the seller either accepts or rejects p: If she accepts, the seller gets payoff p−v; and the buyer gets 3 2 v − p: If she rejects, the seller gets v; and the buyer gets nothing. Find the optimal offer that the buyer can make as a function of x and y: (b) Show that if the buyer and the seller are symmetrically informed (i.e. either both know v or neither party knows v), then trade takes place with probability 1. (c) Consider a simultaneous acceptance game in the model with private information as in part a, where a price p is announced and then the buyer and the seller simultaneously accept or reject trade at price p: The payoffs are as in part a. Find the p that maximizes the probability of trade. [SOLUTION] (a) We look for the subgame perfect equilibrium, thus we solve by back- ward induction.
    [Show full text]
  • Bayesian Games Professors Greenwald 2018-01-31
    Bayesian Games Professors Greenwald 2018-01-31 We describe incomplete-information, or Bayesian, normal-form games (formally; no examples), and corresponding equilibrium concepts. 1 A Bayesian Model of Interaction A Bayesian, or incomplete information, game is a generalization of a complete-information game. Recall that in a complete-information game, the game is assumed to be common knowledge. In a Bayesian game, in addition to what is common knowledge, players may have private information. This private information is captured by the notion of an epistemic type, which describes a player’s knowledge. The Bayesian-game formalism makes two simplifying assumptions: • Any information that is privy to any of the players pertains only to utilities. In all realizations of a Bayesian game, the number of players and their actions are identical. • Players maintain beliefs about the game (i.e., about utilities) in the form of a probability distribution over types. Prior to receiving any private information, this probability distribution is common knowledge: i.e., it is a common prior. After receiving private information, players conditional on this information to update their beliefs. As a consequence of the com- mon prior assumption, any differences in beliefs can be attributed entirely to differences in information. Rational players are again assumed to maximize their utility. But further, they are assumed to update their beliefs when they obtain new information via Bayes’ rule. Thus, in a Bayesian game, in addition to players, actions, and utilities, there is a type space T = ∏i2[n] Ti, where Ti is the type space of player i.There is also a common prior F, which is a probability distribution over type profiles.
    [Show full text]
  • Reinterpreting Mixed Strategy Equilibria: a Unification Of
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Research Papers in Economics Reinterpreting Mixed Strategy Equilibria: A Unification of the Classical and Bayesian Views∗ Philip J. Reny Arthur J. Robson Department of Economics Department of Economics University of Chicago Simon Fraser University September, 2003 Abstract We provide a new interpretation of mixed strategy equilibria that incor- porates both von Neumann and Morgenstern’s classical concealment role of mixing as well as the more recent Bayesian view originating with Harsanyi. For any two-person game, G, we consider an incomplete information game, , in which each player’s type is the probability he assigns to the event thatIG his mixed strategy in G is “found out” by his opponent. We show that, generically, any regular equilibrium of G can be approximated by an equilibrium of in which almost every type of each player is strictly opti- mizing. This leadsIG us to interpret i’s equilibrium mixed strategy in G as a combination of deliberate randomization by i together with uncertainty on j’s part about which randomization i will employ. We also show that such randomization is not unusual: For example, i’s randomization is nondegen- erate whenever the support of an equilibrium contains cyclic best replies. ∗We wish to thank Drew Fudenberg, Motty Perry and Hugo Sonnenschein for very helpful comments. We are especially grateful to Bob Aumann for a stimulating discussion of the litera- ture on mixed strategies that prompted us to fit our work into that literature and ultimately led us to the interpretation we provide here, and to Hari Govindan whose detailed suggestions per- mitted a substantial simplification of the proof of our main result.
    [Show full text]
  • Coordination with Local Information
    CORRECTED VERSION OF RECORD; SEE LAST PAGE OF ARTICLE OPERATIONS RESEARCH Vol. 64, No. 3, May–June 2016, pp. 622–637 ISSN 0030-364X (print) ISSN 1526-5463 (online) ó http://dx.doi.org/10.1287/opre.2015.1378 © 2016 INFORMS Coordination with Local Information Munther A. Dahleh Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, [email protected] Alireza Tahbaz-Salehi Columbia Business School, Columbia University, New York, New York 10027, [email protected] John N. Tsitsiklis Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, [email protected] Spyros I. Zoumpoulis Decision Sciences Area, INSEAD, 77300 Fontainebleau, France, [email protected] We study the role of local information channels in enabling coordination among strategic agents. Building on the standard finite-player global games framework, we show that the set of equilibria of a coordination game is highly sensitive to how information is locally shared among different agents. In particular, we show that the coordination game has multiple equilibria if there exists a collection of agents such that (i) they do not share a common signal with any agent outside of that collection and (ii) their information sets form an increasing sequence of nested sets. Our results thus extend the results on the uniqueness and multiplicity of equilibria beyond the well-known cases in which agents have access to purely private or public signals. We then provide a characterization of the set of equilibria as a function of the penetration of local information channels. We show that the set of equilibria shrinks as information becomes more decentralized.
    [Show full text]
  • The Monty Hall Problem As a Bayesian Game
    Preprints (www.preprints.org) | NOT PEER-REVIEWED | Posted: 25 May 2017 doi:10.20944/preprints201705.0187.v1 Peer-reviewed version available at Games 2017, 8, 31; doi:10.3390/g8030031 The Monty Hall Problem as a Bayesian Game Mark Whitmeyer∗ Department of Economics, University of Texas at Austin, Austin, TX 78712, USA Abstract This paper formulates the classic Monty Hall problem as a Bayesian game. Allowing Monty a small amount of freedom in his decisions facilitates a variety of solutions. The solution concept used is the Bayes Nash Equilibrium (BNE), and the set of BNE relies on Monty's motives and incentives. We endow Monty and the contestant with common prior probabilities (p) about the motives of Monty, and show that under certain conditions on p, the unique equilibrium is one where the contestant is indifferent between switching and not switching. This coincides and agrees with the typical responses and explanations by experimental subjects. Finally, we show that our formulation can explain the experimental results in Page (1998) [12]; that more people gradually choose switch as the number of doors in the problem increases. Keywords: Monty Hall; Equiprobability Bias; games of incomplete information; Bayes Nash Equilibrium JEL Classification: C60, C72, D81, D82 ∗Email: [email protected] 1 © 2017 by the author(s). Distributed under a Creative Commons CC BY license. Preprints (www.preprints.org) | NOT PEER-REVIEWED | Posted: 25 May 2017 doi:10.20944/preprints201705.0187.v1 Peer-reviewed version available at Games 2017, 8, 31; doi:10.3390/g8030031 1 Introduction The \Monty Hall" problem is a famous scenario in decision theory.
    [Show full text]
  • Quantum Games: a Review of the History, Current State, and Interpretation
    Quantum games: a review of the history, current state, and interpretation Faisal Shah Khan,∗ Neal Solmeyer & Radhakrishnan Balu,y Travis S. Humblez October 1, 2018 Abstract We review both theoretical and experimental developments in the area of quantum games since the inception of the subject circa 1999. We will also offer a narrative on the controversy that surrounded the subject in its early days, and how this controversy has affected the development of the subject. 1 Introduction Roots of the theory of quantum information lie in the ideas of Wisener and Ingarden [1, 2]. These authors proposed ways to incorporate the uncertainty or entropy of information within the uncertainty inherent in quantum mechanical processes. Uncertainty is a fundamental concept in both physics and the theory of information, hence serving as the natural link between the two disciplines. Uncertainty itself is defined in terms of probability distributions. Every quantum physical object produces a probability distribution when its state is measured. More precisely, a general state of an m-state quantum object is an element of the m m-dimensional complex projective Hilbert space CP , say v = (v1; : : : ; vm) : (1) Upon measurement with respect to the observable states of the quantum object (which are the elements of m an orthogonal basis of CP ), v will produce the probability distribution 1 2 2 Pm 2 jv1j ;:::; jvmj (2) k=1 jvkj arXiv:1803.07919v2 [quant-ph] 27 Sep 2018 over the observable states. Hence, a notion of quantum entropy or uncertainty may be defined that coincides with the corresponding notion in classical information theory [3].
    [Show full text]
  • Interim Correlated Rationalizability in Infinite Games
    INTERIM CORRELATED RATIONALIZABILITY IN INFINITE GAMES JONATHAN WEINSTEIN AND MUHAMET YILDIZ Abstract. In a Bayesian game, assume that the type space is a complete, separable metric space, the action space is a compact metric space, and the payoff functions are continuous. We show that the iterative and fixed-point definitions of interim correlated rationalizability (ICR) coincide, and ICR is non-empty-valued and upper hemicontinuous. This extends the finite-game results of Dekel, Fudenberg and Morris (2007), who introduced ICR. Our result applies, for instance, to discounted infinite-horizon dynamic games. JEL Numbers: C72, C73. 1. Introduction Interim correlated rationalizability (henceforth ICR) has emerged as the main notion of rationalizability for Bayesian games. Among other reasons, it has the following two desirable properties: Firstly, it is upper hemicontinuous in types, i.e., one cannot obtain a substantially new ICR solution by perturbing a type. Secondly, two distinct definitions of ICR coincide: The fixed-point definition states that ICR is the weakest solution concept with a specific best-response property, and this is the definition that gives ICR its epistemological meaning, as a characterization of actions that are consistent with common knowledge of rationality. Under an alternate definition, ICR is computed by iteratively eliminating the actions that are never a best response (type by type), and this iterative definition is often more amenable for analysis. The above properties were originally proven for games with finite action spaces and finite sets of payoff parameters (by Dekel, Fudenberg, and Morris (2006, 2007), who introduced ICR). However, in many important games these sets are quite large.
    [Show full text]
  • PSCI552 - Formal Theory (Graduate) Lecture Notes
    PSCI552 - Formal Theory (Graduate) Lecture Notes Dr. Jason S. Davis Spring 2020 Contents 1 Preliminaries 4 1.1 What is the role of theory in the social sciences? . .4 1.2 So why formalize that theory? And why take this course? . .4 2 Proofs 5 2.1 Introduction . .5 2.2 Logic . .5 2.3 What are we proving? . .6 2.4 Different proof strategies . .7 2.4.1 Direct . .7 2.4.2 Indirect/contradiction . .7 2.4.3 Induction . .8 3 Decision Theory 9 3.1 Preference Relations and Rationality . .9 3.2 Imposing Structure on Preferences . 11 3.3 Utility Functions . 13 3.4 Uncertainty, Expected Utility . 14 3.5 The Many Faces of Decreasing Returns . 15 3.5.1 Miscellaneous stuff . 17 3.6 Optimization . 17 3.7 Comparative Statics . 18 4 Game Theory 22 4.1 Preliminaries . 22 4.2 Static games of complete information . 23 4.3 The Concept of the Solution Concept . 24 4.3.1 Quick side note: Weak Dominance . 27 4.3.2 Quick side note: Weak Dominance . 27 4.4 Evaluating Outcomes/Normative Theory . 27 4.5 Best Response Correspondences . 28 4.6 Nash Equilibrium . 28 4.7 Mixed Strategies . 31 4.8 Extensive Form Games . 36 4.8.1 Introduction . 36 4.8.2 Imperfect Information . 39 4.8.3 Subgame Perfection . 40 4.8.4 Example Questions . 43 4.9 Static Games of Incomplete Information . 45 4.9.1 Learning . 45 4.9.2 Bayesian Nash Equilibrium . 46 4.10 Dynamic Games with Incomplete Information . 50 4.10.1 Perfect Bayesian Nash Equilibrium .
    [Show full text]
  • Reinterpreting Mixed Strategy Equilibria: a Unification of The
    Reinterpreting Mixed Strategy Equilibria: A Uni…cation of the Classical and Bayesian Views¤ Philip J. Reny Arthur J. Robson Department of Economics Department of Economics University of Chicago University of Western Ontario September 2002 Abstract We provide a new interpretation of mixed strategy equilibria that incor- porates both von Neumann and Morgenstern’s classical concealment role of mixing as well as the more recent Bayesian view originating with Harsanyi. For any two-person game, G, we consider an incomplete information game, ; in which each player’s type is the probability he assigns to the event thatIG his mixed strategy in G is “found out” by his opponent. We show that, generically, any regular equilibrium of G can be approximated by an equilibrium of in which almost every type of each player is strictly optimizing. ThisIG leads us to interpret i’s equilibrium mixed strategy in G as a combination of deliberate randomization by i together with uncer- tainty on j’s part about which randomization i will employ. We also show that such randomization is not unusual: For example, i’s randomization is nondegenerate whenever the support of an equilibrium contains cyclic best replies. ¤We wish to thank Drew Fudenberg, Motty Perry and Hugo Sonnenschein for very helpful comments. We are especially grateful to Bob Aumann for a stimulating discussion of the liter- ature on mixed strategies that prompted us to …t our work into that literature and ultimately led us to the interpretation we provide here, and to Hari Govindan whose detailed sugges- tions permitted a substantial simpli…cation of the proof of our main result.
    [Show full text]