<<

Theory with Translucent Players

Joseph Y. Halpern∗ Rafael Pass† Cornell University Cornell University Dept. Computer Science Dept. Computer Science Ithaca, NY 14853, USA Ithaca, NY 14853, USA [email protected] [email protected]

ABSTRACT Keywords A traditional assumption in is that players are opaque Epistemic logic, rationality, counterfactuals to one another—if a player changes strategies, then this change in strategies does not affect the choice of other players’ strategies. In 1. INTRODUCTION many situations this is an unrealistic assumption. We develop a framework for reasoning about where the players may be Two large firms 1 and 2 need to decide whether to cooperate translucent to one another; in particular, a player may believe that (C) or sue (S) the other firm. Suing the other firm always has if she were to change strategies, then the other player would also a small positive reward, but being sued induces a high penalty p; change strategies. Translucent players may achieve significantly more precisely, u(C,C) = (0, 0); u(C,S) = (−p, r); u(S, C) = more efficient outcomes than opaque ones. (r, −p), u(S, S) = (r − p, r − p). In other words, we are consid- Our main result is a characterization of strategies consistent with ering an instance of the Prisoner’s Dilemma. appropriate analogues of common belief of rationality. Common But there is a catch. Before acting, each firms needs to discuss Counterfactual Belief of Rationality (CCBR) holds if (1) every- their decision with its board. Although these discussions are held one is rational, (2) everyone counterfactually believes that everyone behind closed doors, there is always the possibility of the decision else is rational (i.e., all players i believe that everyone else would being “leaked”; as a consequence, the other company may change still be rational even if i were to switch strategies), (3) everyone its course of action. Furthermore, both companies are aware of this counterfactually believes that everyone else is rational, and counter- fact. In other words, the players are translucent to one another. factually believes that everyone else is rational, and so on. CCBR In such a scenario, it may well be rational for both companies to characterizes the set of strategies surviving iterated removal of min- cooperate. For instance, consider the following situation. imax dominated strategies: a σi is minimax dominated for 0 0 0 • Firm i believes that its action is leaked to firm 2 − i with i if there exists a strategy σi for i such that minµ0 ui(σi, µ−i) > −i probability . maxµ−i ui(σi, µ−i). • Firm i believes that if the other firm 2 − i finds out that i is defecting, then 2 − i will also defect. Categories and Subject Descriptors F.4.1 [Mathematical Logic and Formal Languages]: Mathemati- • Finally, p > r (i.e., the penalty for being sued is signifi- cal Logic—modal logic; I.2.11 [Artificial Intelligence]: Distributed cantly higher than the reward of suing the other company). multiagent systems Artificial Intelligence— ; J.4 [Social and Be- Neither firm defects, since defection is noticed by the other firm havioral Sciences]: Economics with probability , which (according to their beliefs) leads to a harsh punishment. Thus, the possibility of the players’ actions be- General Terms ing leaked to the other player allows the players to significantly improve social welfare in equilibrium. (This suggests that it may Economics, Theory be mutually beneficial for two countries to spy on each other!) Even if the Prisoner’s dilemma is not played by corporations but ∗Halpern is supported in part by NSF grants IIS-0812045, IIS- by individuals, each player may believe that if he chooses to de- 0911036, and CCF-1214844, by AFOSR grant FA9550-08-1-0266, fect, his “guilt” over defecting may be visible to the other player. and by ARO grant W911NF-09-1-0281. † (Indeed, facial and bodily cues such as increased pupil size are of- Pass is supported in part by a Alfred P. Sloan Fellowship, ten associated with deception; see e.g., [Ekman and Friesen 1969].) Microsoft New Faculty Fellowship, NSF Award CNS-1217821, NSF CAREER Award CCF-0746990, NSF Award CCF-1214844, Thus, again, the players may choose to cooperate out of fear that if AFOSR YIP Award FA9550-10-1-0093, and DARPA and AFRL they defect, the other player may detect it and act on it. under contract FA8750-11-2- 0211. The views and conclusions Our goal is to capture this type of reasoning formally. We take contained in this document are those of the authors and should not a Bayesian approach: Each player has a (subjective) probability be interpreted as representing the official policies, either expressed distribution (describing the player’s beliefs) over the states of the or implied, of the Defense Advanced Research Projects Agency or world. Traditionally, a player i is said to be rational in a state ω if the US Government the strategy σi that i plays at ω is a to the strategy 1 profile µ−i of the other players induced by i’s beliefs in ω; that is,

TARK 2013, Chennai, India. 1 Copyright 2013 by the authors. Formally, we assume that i has a distribution on states, and at each

216 0 0 ui(σi, µ−i) ≥ ui(σi, µ−i) for all alternative strategies σi for i. In acterizes the set of strategies that survive iterated removal our setting, things are more subtle. Player i may believe that if she of minimax dominated strategies: a strategy σi is minimax 0 0 were to switch strategies from σi to σi, then players other than i dominated for i if there exists a strategy σi for i such that 0 0 might also switch strategies. We capture this using counterfactuals min 0 ui(σ , µ ) > maxµ ui(σi, µ−i); that is, µ−i i −i −i [Lewis 1973; Stalnaker 1968].2 Associated with each state of the 0 0 ui(σi, µ−i) > ui(σi, µ−i) no matter what the strategy pro- world ω, each player i, and f(ω, i, σ0) where player i plays σ0. 0 i i files µ−i and µ−i are. Note that if i changes strategies, then this change in strategies may start a chain reaction, leading to further changes. We can think of We also consider analogues of in our setting, and 0 f(ω, i, σi) as the steady-state of this process: the state show that individually rational strategy profiles that survive iter- 0 that would result if i switched strategies to σ . Let µ 0 be ated removal of minimax dominated strategies characterize such i f(ω,i,σi) the distribution on strategy profiles of −i (the players other than i) equilibria. induced by i’s beliefs at ω about the steady-state outcome of this Note that in our approach, each player i has a belief about how process. We say that i is rational at a state ω where i plays σi the other players’ strategies would change if i were to change strate- 0 and has beliefs µi if ui(σi, µ−i) ≥ ui(σ , µ 0 ) for every gies, but we do not require i to explicitly specify how he would i f(ω,i,σi) 0 respond to other people changing strategies. The latter approach, alternative strategy σi for i. Note that we have required the closest- state function to be deterministic, returning a unique state, rather of having each player pick a “meta-strategy” that takes as input the than a distribution over states. While this may seem incompatible strategy of other players, was explored by Howard [1971] in the with the motivating scenario, it does not seem so implausible in our 1970s. It led to complex formalisms involving infinite hierachies context that, by taking a rich enough representation of states, we of meta-strategies: at the lowest level, each player specifies a strat- can assume that a state contains enough information about players egy in the original game; at level k, each player specifies a “re- to resolve uncertainty about what strategies they would use if one sponse rule” (i.e., a meta-strategy) to other players’ (k − 1)-level player were to switch strategies. response rules. Such hierarchical structures have not proven useful We are interested in considering analogues to in when dealing with applications. Since we do not require players to a setting with translucent players, and providing epistemic charac- specify reaction rules, we avoid the complexities of this approach. terizations of them. To do that, we need some definitions. We say Program equilibria [Tennenholz 2004] and conditional commit- that a player i counterfactually believes ϕ at ω if i believes ϕ holds ments [Kalai et al. 2010] provide a different approach to avoiding even if i were to switch strategies. Common Counterfactual Belief infinite hierarchies. Roughly speaking, each player i simply spec- of Rationality (CCBR) holds if (1) everyone is rational, (2) every- ifies a program Πi; player i’s action is determined by running i’s program on input the (description of) the programs of the other one counterfactually believes that everyone else is rational (i.e., all 0 players i believe that everyone else would still be still rational even players; that is, i action is given by Πi(Π−i). Tennenholtz [2004] if i were to switch strategies), (3) everyone counterfactually be- and Kalai et al. [2010] show that every (correlated) individually lieves that everyone else is rational, and counterfactually believes rational outcome can be sustained in a program equilibrium. Their that everyone else is rational, and so on. model, however, assumes that player’s programs (which should be Our main result is a characterization of strategies consistent with interpreted as their “plan of action”) are commonly known to all CCBR. Roughly speaking, these results can be summarized as fol- players. We dispense with this assumption. It is also not clear how lows: to define common belief of rationality in their model; the study of program equilibria and conditional commitments has considered • If the closest-state function respects “unilateral deviations”— only analogues of Nash equilibrium. when i switches strategies, the strategies and beliefs of play- Counterfactuals have been explored in a game-theoretic setting; ers other than i remain the same—then CCBR characterizes see, for example, [Aumann 1995; Halpern 1999; Samet 1996; Stal- the set of rationalizable strategies. naker 1996; Zambrano 2004]. However, all these papers consid- ered only structures where, in the closest state where i changes • If the closest-state function can be arbitrary, CCBR char- strategies, all other players’ strategies remain the same; thus, these approaches are not applicable in our context. state, a pure strategy profile is played; the distribution on states clearly induces a distribution on strategy profiles for the players other than i, which we denote µ−i. 2. COUNTERFACTUAL STRUCTURES 2 A different, more direct, approach for capturing our original moti- Given a game Γ, let Σ (Γ) denote player i’s pure strategies in Γ vating example would be to consider and analyze an extensive-form i (we occasionally omit the parenthetical Γ if it is clear from context variant G0 of the original normal-form game G that explicitly mod- els the “leakage” of players’ actions in G, allows the player to react or irrelevant). to these leakage signals by choosing a new action in G, which again To reason about the game Γ, we consider a class of Kripke struc- may be leaked and the players may react to, and so on. Doing this tures corresponding to Γ. For simplicity, we here focus on finite is subtle. We would need to model how players respond to receiv- structures. A finite probability structure M appropriate for Γ is a ing leaked information, and to believing that there was a change in tuple (Ω, s, PR1,..., PRn), where Ω is a finite set of states; s plan even if information wasn’t leaked. To make matters worse, it’s associates with each state ω ∈ Ω a pure strategy profile s(ω) in the not clear what it would mean that a player is “intending” to per- form an action a if players can revise what they do as the result of game Γ; and, for each player i, PRi is a probability assignment a leak. Does it mean that a player will do a if no information is that associates with each state ω ∈ Ω a probability distribution leaked to him? What if no information is leaked, but he believes PRi(ω) on Ω, such that that the other side is planning to change their plans in any case? In addition, modeling the game in this way would require a dis- 1. PRi(ω)([[si(ω)]]M ) = 1, where for each strategy σi for tribution over leakage signals to be exogenously given (as part of player i, [[σi]]M = {ω : si(ω) = σi}, where si(ω) denotes the description of the game G0). Moreover, player strategies would player i’s strategy in the strategy profile s(ω); have to be infinite objects, since there is no bound on the sequence of leaks and responses to leaks. In constrast, using counterfactuals, 2. PRi(ω)([[PRi(ω), i]]M ) = 1, where for each probability we can directly reason about the original (finite) game G. measure π and player i, [[π, i]]M = {ω : PRi(ω) = π}.

217 These assumptions say that player i assigns probability 1 to his how he evaluates the state. He thus ascribes probability 0 to ω. In 0 actual strategy and beliefs. f(ω, i, σi) he has also been given the drug, and the drug in par- To deal with counterfactuals, we augment probability structures ticular affects how he evaluates outcomes. Thus, i’s beliefs in the 0 with a “closest-state” function f that associates with each state ω, state f(ω, i, σi) are quite different from his beliefs in all states in 0 c player i, and strategy σi, a state f(ω, i, σi) where player i plays the support of PR 0 (ω). i,σi σ0; if σ0 is already played in ω, then the closest state to ω where σ0 is played is ω itself. Formally, a finite counterfactual structure 2.1 Logics for Counterfactual Games M appropriate for Γ is a tuple (Ω, s, f, PR1,..., PRn), where Let L(Γ) be the language where we start with true and the prim- (Ω, s, PR1,..., PRn) is a probability structure appropriate for Γ 0 itive proposition RAT i and play i(σi) for σi ∈ Σi(Γ), and close and f is a “closest-state” function. We require that if f(ω, i, σi) = ∗ 0 off under the modal operators Bi (player i believes) and Bi (player ω , then i counterfactually believes) for i = 1, . . . , n, CB (common be- 0 0 CB∗ 1. si(ω ) = σ ; lief), and (common counterfactual belief), conjunction, and negation. We think of Biϕ as saying that “i believes ϕ holds with 0 0 ∗ 2. if σi = si(ω), then ω = ω. probability 1” and Bi ϕ as saying “i believes that ϕ holds with probability 1, even if i were to switch strategies”. Given a probability assignment PR for player i, we define i’s i Let L0 be defined exactly like L except that we exclude the counterfactual belief at state ω (“what i believes would happen if “counterfactual” modal operators B∗ and CB∗. We first define he switched to σ0 at ω) as i semantics for L0 using probability structures (without counterfac- c 0 X 00 PR 0 (ω)(ω ) = PRi(ω)(ω ). tuals). We define the notion of a formula ϕ being true at a state ω i,σi {ω00∈Ω:f(ω00,i,σ0 )=ω0} in a probability structure M (written (M, w) |= ϕ) in the standard i way, by induction on the structure of ϕ, as follows: Note that the conditions above imply that each player i knows what c 0 • (M, ω) |= true true strategy he would play if he were to switch; that is, PR 0 (ω)([[σi]]M ) = (so is vacuously true). i,σi 1. • (M, ω) |= play (σi) iff σi = si(ω). Let Supp(π) denote the support of the probability measure π. i c 0 0 0 Note that Supp(PR 0 (ω)) = {f(ω , i, σi): ω ∈ Supp(PRi(ω)}. • (M, ω) |= ¬ϕ iff (M, ω) 6|= ϕ. i,σi Moreover, it is almost immediate from the definition that if PRi(ω) = 0 0 0 c c 0 0 • (M, ω) |= ϕ ∧ ϕ iff (M, ω) |= ϕ and (M, ω) |= ϕ . PRi(ω ), then PR 0 (ω) = PR 0 (ω ) for all strategies σi for i,σi i,σi player i. But it does not in general follow that i knows his counter- • (M, ω) |= Biϕ iff PRi(ω)([[ϕ]]M ) = 1, where [[ϕ]]M = factual beliefs at ω, that is, it may not be the case that for all strate- {ω :(M, ω) |= ϕ}. 0 c c gies σi for player i, PRi,σ0 (ω)([[PRi,σ0 (ω), i]]M ) = 1. Suppose i i • (M, ω) |= RAT iff s (ω) is a best response given player that we think of a state as representing each player’s ex ante view i i i’s beliefs regarding the strategies of other players induced of the game. The fact that player si(ω) = σi should then be inter- by PRi. preted as “i intends to play σi at state ω.” With this view, suppose 0 that ω is a state where si(ω) is a conservative strategy, while σi • Let EBϕ (“everyone believes ϕ”) be an abbreviation of B1ϕ∧ k is a rather reckless strategy. It seems reasonable to expect that i’s ... ∧ Bnϕ; and define EB ϕ for all k inductively, by taking subjective beliefs regarding the likelihood of various outcomes may EB1ϕ to be EBϕ and EBk+1ϕ to be EB(EBkϕ). depend in part on whether he is in a conservative or reckless frame c 0 k of mind. We can think of PR 0 (ω)(ω ) as the probability that i • (M, ω) |= CBϕ iff (M, ω) |= EB ϕ for all k ≥ 1. i,σi 0 ascribes, at state ω, to ω being the outcome of i switching to strat- Semantics for L0 in counterfactual structures is defined in an iden- 0 c 0 egy σi; thus, PR 0 (ω)(ω ) represents i’s evaluation of the like- i,σi tical way, except that we redefine RAT i to take into account the lihood of ω0 when he is in a conservative frame of mind. This may fact that player i’s beliefs about the strategies of players −i may c not be the evaluation that i uses in states in the support PR 0 (ω); change if i changes strategies. i,σi at all these states, i is in a “reckless” frame of mind. Moreover, 0 • (M, ω) |= RAT i iff for every strategy σ for player i, there may not be a unique reckless frame of mind, so that i may not i c X 0 0 have the same beliefs at all the states in the support of PR 0 (ω). i,σi PRi(ω)(ω )ui(si(ω), s−i(ω )) ≥ M is a strongly appropriate counterfactual structure if it is an ω0∈Ω appropriate counterfactual structure and, at every state ω, every X c 0 0 0 PRi,σ0 (ω)(ω )ui(σi, s−i(ω )). player i knows his counterfactual beliefs. As the example above i 0 suggests, strong appropriateness is a nontrivial requirement. As we ω ∈Ω shall see, however, our characterization results hold in both appro- The condition above is equivalent to requiring that priate and strongly appropriate counterfactual structures. X 0 0 Note that even in strongly appropriate counterfactually struc- PRi(ω)(ω )ui(si(ω), s−i(ω )) ≥ 0 c tures, we may not have PRi(f(ω, i, σi)) = PR 0 (ω). We ω0∈Ω i,σi 0 c X 0 0 0 0 do have PRi(f(ω, i, σi)) = PR 0 (ω) in strongly appropriate i,σi PRi(ω)(ω )ui(σi, s−i(f(ω , i, σi))). 0 c 0 counterfactual structures if f(ω, i, σi) is in the support of PR 0 (ω) ω ∈Ω i,σi (which will certainly be the case if ω is in the support of PRi(ω)). Note that, in general, this condition is different from requiring that 0 c To see why we may not want to have PRi(f(ω, i, σi)) = PR 0 (ω) s (ω) i i,σi i is a best reponse given player ’s beliefs regarding the strate- in general, even in strongly appropriate counterfactual structures, gies of other players induced by PRi. consider the example above again. Suppose that, in state ω, al- To give the semantics for L in counterfactual structures, we now ∗ ∗ though i does not realize it, he has been given a drug that affects also need to define the semantics of Bi and CB :

218 ∗ 0 • (M, ω) |= Bi ϕ iff for all strategies σi ∈ Σi(Γ), 3. CHARACTERIZING COMMON COUN- c PR 0 (ω)([[ϕ]]M ) = 1. i,σi TERFACTUAL BELIEF OF RATIONAL- ITY • (M, ω) |= CB∗ϕ (M, ω) |= (EB∗)kϕ k ≥ 1 iff for all . It is well known rationalizability can be characterized in terms ∗ of common belief of common belief of rationality in probability It is easy to see that, like Bi, Bi depends only on i’s beliefs; as 5 0 c structures [?; ?]. In the full version of the paper we show that if we observed above, if PRi(ω) = PRi(ω ), then PRi,σ0 (ω) = i we restrict to counterfactual structures that respect unilateral de- c 0 0 ∗ 0 ∗ PR 0 (ω ) for all σi, so (M, ω) |= Bi ϕ iff (M, ω ) |= Bi ϕ. It i,σi viations—where in the closest state to ω where player i switches ∗ ∗ immediately follows that Bi ϕ ⇒ BiBi ϕ is valid (i.e., true at all strategies, everybody else’s strategy and beliefs remain same—common states in all structures). counterfactual belief of rationality characterizes rationalizable strate- The following abbreviations will be useful in the sequel. Let gies. In a sense (which is made precise in the full version of the pa- RAT be an abbreviation for RAT 1 ∧...∧RAT n, and let play(~σ) per), counterfactual structures respecting unilateral deviations be- be an abbreviation for play 1(σ1) ∧ ... ∧ play n(σn). have just like probability structures (without counterfactuals). We now characterize common counterfactual belief of rational- 2.2 Common Counterfactual Belief of Ratio- ity without putting any restrictions on the counterfactual structures nality (other than them being appropriate, or strongly appropriate). Our We are interested in analyzing strategies being played at states characterization is based on ideas that come from the characteriza- where (1) everyone is rational, (2) everyone counterfactually be- tion of rationalizability. It is well known that rationalizability can lieves that everyone else is rational (i.e., for every player i, i be- be characterized in terms of an iterated deletion procedure, where lieves that everyone else would still be rational even if i were to at each stage, a strategy σ for player i is deleted if there are no switch strategies), (3) everyone counterfactually believes that ev- beliefs that i could have about the undeleted strategies for the play- eryone else is rational, and counterfactually believes that everyone ers other than i that would make σ rational [Pearce 1984]. Thus, else is rational, and so on. For each player i, define the formu- there is a deletion procedure that, when applied repeatedly, results k in only the rationalizable strategies, that is, the strategies that are las SRATi (player i is strongly k-level rational) inductively, by 0 k+1 played in states where there is common belief of rationality, being taking SRATi to be true and SRAT to be an abbreviation of i left undeleted. We now show that there is an analogous way of ∗ k characterizing common counterfactual belief of rationality. RAT i ∧ Bi (∧j6=iSRATj ).

k n k 3.1 Iterated Minimax Domination Let SRAT be an abbreviation of ∧j=1SRATj . Define CCBR (common counterfactual belief of rationality) as The key to our characterization is the notion of minimax domi- follows: nated strategies.

k • (M, ω) |= CCBR iff (M, ω) |= SRAT ϕ for all k ≥ 1. DEFINITION 3.1.: Strategy σi for player i in game Γ is minimax 0 dominated with respect to Σ−i ⊆ Σ−i(Γ) iff there exists a strategy k 0 Note that it is critical in the definition of SRATi that we require σi ∈ Σi(Γ) such that only that player i counterfactually believes that everyone else (i.e., 0 min ui(σi, τ−i) > max ui(σi, τ−i). the players other than i) are rational, and believe that everyone else 0 0 τ−i∈Σ−i τ−i∈Σ−i is rational, and so on. Player i has no reason to believe that his own strategy would be rational if he were to switch strategies; indeed, ∗ Bi RAT i can hold only if every strategy for player i is rational with respect to i’s beliefs. This is why we do not define CCBR as In other words, player i’s strategy σ is minimax dominated with ∗ 3 0 0 CB RAT . respect to Σ−i iff there exists a strategy σ such that the worst-case We also consider the consequence of just common belief of ra- payoff for player i if he uses σ0 is strictly better than his best-case k tionality in our setting. Define W RATi (player i is weakly k-level payoff if he uses σ, given that the other players are restricted to k ∗ 0 rational) just as SRATi , except that Bi is replaced by Bi. An easy using a strategy in Σ−i. k+1 k induction on k shows that W RAT implies W RAT and that In the standard setting, if a strategy σi for player i is dominated k k 4 0 W RAT implies Bi(W RAT ). It follows that we could have by σi then we would expect that a rational player will never player k+1 0 equivalently defined W RATi as σi, because σi is a strictly better choice. As is well known, if σi 0 is dominated by σi, then there are no beliefs that i could have re- n k RAT i ∧ Bi(∧j=1W RATj ). garding the strategies used by the other players according to which σi is a best response [Pearce 1984]. This is no longer the case in Thus, W RAT k+1 is equivalent to RAT ∧ EB(W RAT k). As a our setting. For example, in the standard setting, cooperation is consequence we have the following: dominated by defection in Prisoner’s Dilemma. But in our setting, suppose that player 1 believes that if he cooperates, then the other k PROPOSITION 2.1.: (M, ω) |= CB(RAT ) iff (M, ω) |= W RAT player will cooperate, while if he defects, then the other player will for all k ≥ 0. defect. Then cooperation is not dominated by defection. So when can we guarantee that playing a strategy is irrational in 3Interestingly, Samet [1996] essentially considers an analogue of our setting? This is the case only if the strategy is minimax domi- ∗ 0 CB RAT . This works in his setting since he is considering only nated. If σi is minimax dominated by σi, there are no counterfac- events in the past, not events in the future. tual beliefs that i could have that would justify playing σi. Con- 4We can also show that SRAT k+1 implies SRAT k, but it is not versely, if σi is not minimax dominated by any strategy, then there k ∗ k the case that SRATi implies Bi SRATi , since RAT does not ∗ 5 imply Bi RAT . Available at http://www.cs.cornell.edu/home/halpern/papers/minimax.pdf.

219 k are beliefs and counterfactual beliefs that i could have that would REMARK 3.5.: Note that in the definition of NSD i (Γ), we re- justify playing σi. Specifically, i could believe that the players in move all strategies that are dominated by some strategy in Σi(Γ), k−1 −i are playing the strategy profile that gives i the best possible util- not just those dominated by some strategy in NSD i (Γ). Nev- 0 ity when he plays σi, and that if he switches to another strategy σi, ertheless, the definition would be equivalent even if we had con- k−1 the other players will play the strategy profile that gives i the worst sidered only dominating strategies in NSD i (Γ). For suppose 0 possible utility given that he is playing σi. not. Let k be the smallest integer such that there exists some strat- k−1 Note that we consider only domination by pure strategies. It egy σi ∈ NSD i (Γ) that is minimax dominated by a strategy 0 k−1 k−1 is easy to construct examples of strategies that are not minimax σi ∈/ NSD i (Γ), but there is no strategy in NSD i (Γ) that 0 dominated by any pure strategy, but are minimax dominated by a dominates σi. That is, σi was deleted in some previous itera- 0 q mixed strategy. Our characterization works only if we restrict to tion. Then there exists a sequence of strategies σi , . . . , σi and 0 0 domination by pure strategies. The characterization, just as with indices k0 < k1 < . . . < kq = k − 1 such that σi = σi, the characterization of rationalizability, involves iterated deletion, j kj j σi ∈ NSDi (Γ), and for all 0 ≤ j < q, σi is minimax domi- but now we do not delete dominated strategies in the standard sense, j+1 kj −1 k−2 nated by σi with respect to NSD i (Γ). Since NSD (Γ) ⊆ but minimax dominated strategies. j q NSD (Γ) for j ≤ k−2, an easy induction on j shows that σi min- q−j k−2 k 0 imax dominates σ with respect to NSD for all 0 < j ≤ q. DEFINITION 3.2.: Define NSD j (Γ) inductively: let NSD j (Γ) = q 0 0 In particular, σ minimax dominates σi = σ with respect to k+1 k k−2 Σj and let NSD j (Γ) consist of the strategies in NSD j (Γ) not NSD . k minimax dominated with respect to NSD −j (Γ). Strategy σ sur- vives k rounds of iterated deletion of minimax strategies for player k The following example shows that iteration has bite: there exist i if σ ∈ NSD i (Γ). Strategy σ for player i survives iterated dele- tion of minimax dominated strategies if it survives k rounds of a 2-player game where each player has k actions and k − 1 rounds iterated deletion of strongly dominated for all k, that is, if σ ∈ of iterations are needed. ∞ k NSD (Γ) = ∩kNSD (Γ). i i EXAMPLE 3.6.: Consider a two-player game, where both play- ers announce a value between 1 and k. Both players receive in util- In the deletion procedure above, at each step we remove all ity the smallest of the values announced; additionally, the player strategies that are minimax dominated; that is we perform a “max- who announces the larger value get a reward of p = 0.5.6 That is, imal” deletion at each step. As we now show, the set of strategies u(x, y) = (y + p, y) if x > y, (x, x + p) if y > x, and (x, x) that survives iterated deletion is actually independent of the dele- if x = y. In the first step of the deletion process, 1 is deleted for tion order. 0 m ~ 0 1 m both players; playing 1 can yield a max utility of 1, whereas the Let S ,...,S be sets of strategy profiles. S = (S ,S ,...,S ) mininum utility of any other action is 1.5. Once 1 is deleted, 2 is is a terminating deletion sequence for Γ if, for j = 0, . . . , m − 1, deleted for both players: 2 can yield a max utility of 2, and the min Sj+1 ⊂ Sj (note that we use ⊂ to mean proper subset) and all j+1 utility of any other action (once 1 is deleted) is 2.5. Continuing this players i, Si contains all strategies for player i not minimax j process, we see that only (k, k) survives. dominated with respect to S−i (but may also contain some strate- m gies that are minimax dominated), and Si does not contain any m strategies that are minimax dominated with respect to S−i. A set 3.2 Characterizing Iterated Minimax Domi- T of strategy profiles has ambiguous terminating sets if there exist nation ~ ~0 two terminating deletion sequences S = (T,S1,...,Sm), S = We now show that strategies surviving iterated removal of min- 0 0 0 (T,S1,...,Sm0 ) such that Sm 6= Sm0 ; otherwise, we say that T imax dominated strategies characterize the set of strategies consis- has a unique terminating set. tent with common counterfactual belief of rationality in (strongly) appropriate counterfactual structures. As a first step, we define a PROPOSITION 3.3.: No (finite) set of strategy profiles has am- “minimax” analogue of rationalizability. biguous terminating sets. DEFINITION 3.7.: A strategy profile ~σ in game Γ is minimax Proof: Let T be a set of strategy profiles of least cardinality that has rationalizable if, for each player i, there is a set Z ⊆ Σ (Γ) such ~ i i ambiguous terminating deletion sequences S = (T,S1,...,Sm) that ~0 0 0 0 0 and S = (T,S1,...,Sm0 ), where Sm 6= Sm0 . Let T be the set of strategies that are not minimax dominated with respect to T . • σi ∈ Zi; 0 0 0 0 Clearly T 6= ∅ and, by definition, T ⊆ S1 ∩S1. Since T , S1, and 0 00 0 • for every strategy σi ∈ Zi and strategy σi ∈ Σi(Γ), S1 all have cardinality less than that of T , they must all have unique terminating sets; moreover, the terminating sets must be the same. 0 00 0 max ui(σi, τ−i) ≥ min ui(σi , τ−i). For consider a terminating deletion sequence starting at T . We can τ−i∈Z−i τ−i∈Z−i get a terminating deletion sequence starting at S1 by just appending 0 this sequence to S1 (or taking this sequence itself, if S1 = T ). We 0 can similarly get a terminating deletion sequence starting at S1. Since all these terminating deletion sequences have the same final THEOREM 3.8.: The following are equivalent: element, this must be the unique terminating set. But (S1,...,Sm) ∞ 0 0 (a) ~σ ∈ NSD (Γ); and (S1,...,Sm0 ) are terminating deletion sequences with Sm 6= 0 S 0 , a contradiction. m (b) ~σ is minimax rationalizable in Γ; 6This game can be viewed a a reverse variant of the Traveler’s COROLLARY 3.4.: The set of strategies that survivies interated dilemma [Basu 1994], where the player who announces the smaller deletion of minimax strategies is independent of the deletion order. value gets the reward.

220 (c) there exists a finite counterfactual structure M that is strongly Basu, K. (1994). The traveler’s dilemma: paradoxes of rational- appropriate for Γ and a state ω such that ity in game theory. American Economic Review 84(2), 391– n k 395. (M, ω) |= play(~σ) ∧i=1 SRATi Brandenburger, A. and E. Dekel (1987). Rationalizability and for all k ≥ 0; correlated equilibria. Econometrica 55, 1391–1402. (d) for all players i, there exists a finite counterfactual structure Ekman, P. and W. Friesen (1969). Nonverbal leakage and clues M that is appropriate for Γ and a state ω such that to deception. Psychiatry 32, 88–105. k Halpern, J. Y. (1999). Hypothetical knowledge and counterfac- (M, ω) |= play (σi) ∧ SRAT i i tual reasoning. International Journal of Game Theory 28(3), for all k ≥ 0. 315–330. Howard, N. (1971). Paradoxes of Rationality: Theory of The proof of Theorem 3.8 can be found in the full version of the Metagames and Political Behavior. The MIT Press, Cam- paper. In the full version of the paper, we additionally characterize bridge. analogues of Nash equilibrium in counterfactual structures. These results allow us to more closely relate our model to those of Ten- Kalai, A., E. Kalai, E. Lehrer, and D. Samet (2010). A commit- nenholtz [2004] and Kalai et al. [2010]. ment folk theorem. Games and Economic Behavior 69(1), 127–137. 4. DISCUSSION Lewis, D. K. (1973). Counterfactuals. Cambridge, Mass.: Har- We have introduced a game-theoretic framework for analyzing vard University Press. scenarios where a player may believe that if he were to switch Pearce, D. G. (1984). Rationalizable strategic behavior and the strategies, this intention to switch may be detected by the other problem of perfection. Econometrica 52(4), 1029–1050. players, resulting in them also switching strategies. Our formal Samet, D. (1996). Hypothetical knowledge and games with per- model allows players’ counterfactual beliefs (i.e., their beliefs about fect information. Games and Economic Behavior 17, 230– the state of the world in the event that they switch strategies) to be 251. arbitrary—they may be completely different from the players’ ac- tual beliefs. Stalnaker, R. C. (1968). A semantic analysis of conditional We may also consider a more restricted model where we re- logic. In N. Rescher (Ed.), Studies in Logical Theory, pp. quire that a player i’s counterfactual beliefs regarding other play- 98–112. Oxford University Press. ers’ strategies and beliefs is -close to player i’s actual beliefs in Stalnaker, R. C. (1996). Knowledge, belief and counterfactual total variation distance7—that is, for every state ω ∈ Ω, player reasoning in games. Economics and Philosophy 12, 133– 0 c i, and strategy σi for player i, the projection of PR 0 (ω) onto 163. i,σi strategies and beliefs of players −i is -close to the projection of Tan, T. and S. Werlang (1988). The Bayesian foundation of solu- PRi(ω) onto strategies and beliefs of players −i. tion concepts of games. Journal of Economic Theory 45(45), We refer to counterfactual structures satisfying this property as - 370–391. counterfactual stuctures. Roughly speaking, -counterfactual struc- Tennenholz, M. (2004). Program equilibrium. Games and Eco- tures restrict to scenarios where players are not “too” transparent to nomic Behavior 49(12), 363–373. one another; this captures the situation when a player assigns only some “small” probability to its switch in action being noticed by Zambrano, E. (2004). Counterfactual reasoning and common the other players. knowledge of rationality in normal form. Topics in Theo- As we show in the full paper, 0-counterfactual structures behave retical Economics 4(1). just as counterfactual structures that respect unilateral deviations: common counterfactual belief of rationality in 0-counterfactual struc- tures characterizes rationalizable strategies. The general counter- factual structures investigated in this paper are 1-counterfactual structures (that is, we do not impose any conditions on players’ counterfactual beliefs). We remark that although our characteriza- tion results rely on the fact that we consider 1-counterfactual struc- tures, the motivating example in the introduction (the translucent prisoner’s dilemma game) shows that even considering -counterfactual structures with a small  can result in there being strategies consis- tent with common counterfactual belief of rationality that are not rationalizable. We leave an exploration of common counterfactual belief of rationality in -counterfactual structures for future work.

References Aumann, R. J. (1995). Backwards induction and of rationality. Games and Economic Behavior 8, 6–19. 7Recall that two probability distribution are -close in total vari- ation distance if the probabilities that they assign to any event E differ by at most .

221