AAAI Technical Report SS-12-03 for Security, Sustainability and Health

Rational Irrationality

Dan Ventura Computer Science Department Brigham Young University [email protected]

Abstract not rational). Most people, of course, don’t know how to or don’t want to compute the expected value of the decision We present a game-theoretic account of irrational agent be- to buy a lottery ticket. But even people that do the analysis havior and define conditions under which irrational behav- sometimes will buy lottery tickets. So, the question is, is the ior may be considered quasi-rational. To do so, we use a 2-player, zero-sum strategic game, parameterize the reward lottery really a tax on people that can’t do math, or are they structure and study how the value of the game changes with just doing a different kind of math? this parameter. We argue that for any “underdog” agent, there In this paper we consider a game theoretic setting involv- is a point at which the asymmetry of the game will pro- ing asymmetric reward structure that suggests the latter and voke the agent to act irrationally. This implies that the non- as a result suggests that in asymmetric games of conflict “underdog” player must therefore also act irrationally even it can be quasi-rational for an agent to behave irrationally. though he has no incentive (in the reward structure) for doing And, if this is the case, such an agent’s opponent must also so, which implies, in turn, a meta-level game. then behave (or at least be perceived to be capable of behav- ing) irrationally as well. Introduction The Domination Game Game theory offers an often compelling account of agent be- havior under various forms of cooperation and conflict. One Consider the following payoff matrix for a 2-player, zero- of the basic tenets of game theory is that agents will behave sum strategic game, which we will call Domination: rationally, that is, that agents will act to maximize their re-  1 −1  D = ward in the game. However, it is well-known that humans −1 r do not always act in this way. One natural reason for irra- tional behavior might be that the agent does not understand Player A has two actions (a1, a2) corresponding to a the game, and is, in fact, acting rationally given their (incor- choice of rows in the matrix and player B has two actions rect) understanding of the game. But, what if an agent fully (b1, b2) corresponding to a choice of columns. Solving for understands the game and still acts irrationally? An interest- an equilibrium point yields a value for the game (in terms ing array of security questions can be asked in this context, r−1 of r) of v = r+3 which is obtained by player A playing a particularly in cases of asymmetric conflict such as those mixed with probability p = r+1 of playing action seen in recent cases of terrorism and military conflicts. As r+3 a and by player B playing a mixed strategy with probability an extreme example, why would someone detonate a bomb 1 q = r+1 of playing action b (see Appendix A for details). strapped to their own body, or fly a plane into a building? r+3 1 Can we offer an explanation for an agent that chooses to This game can be understood intuitively as follows. As the play irrationally, knowing that that is what they are choos- value of r increases, B is less and less likely to choose action ing? Is there a kind of rationality to this kind of choice? Can b2 as such a choice could result in a catastrophic payoff. At we determine the to such behavior? the same time, A is less and less likely to choose action a2 A second, day-to-day scenario that exhibits this kind of as this would likely result in a loss (since the chance of get- irrational behavior might be the purchase of lottery tickets. A ting payoff r is increasingly remote). And, in the limit, the tongue-in-cheek acknowledgement of this is the humorous equilibrium strategy is the pure (joint) strategy that results in aphorism that explains that “the lottery is a tax on people the joint action (a1, b1) being chosen with probability 1. If who can’t do math”. When people buy lottery tickets, they both players are rational, their strategies will approach this are (apparently) happy to risk a small amount of money for pure strategy as r approaches ∞. a chance at a huge reward, even though an expected value For this discussion, a useful way to look at the equilibrium analysis will tell them that this is a bad idea (in other words, value v is as the sum of a reward π and a risk ρ. Figure 1 shows a decomposition of v into π and ρ as a function of r, Copyright c 2012, Association for the Advancement of Artificial where π = p(2q − 1) and ρ = (1 − p)(−q − qr + r) (or, Intelligence (www.aaai.org). All rights reserved. alternatively, π = q(2p − 1) and ρ = (1 − q)(−p − pr +

83 r)—see Appendix B for details). The reward component is the result of the opponent playing action 1, while the risk component is the result of the opponent playing action 2. For the joint rational strategy, almost all of the (expected) payoff comes from the reward component, and, as r increases, this contribution continues to increase. In the limit, π = v, and ρ = 0. In other words, in the limit, there is no chance that the opponent will play action 2. The Case for Irrationality For the following, without loss of generality, assume r > 1, so that p > 0.5 and q > 0.5. Let p∗ = [p, 1−p], q∗ = [q, 1− q] with the equilibrium (joint) solution being (p∗, q∗). For s strategy s = [s, 1 − s], let xs be the (row or column) vector Figure 2: Irrational reward and risk for A (B) as a func- of the payoff matrix selected using strategy s and choosing tion of r the action associated with probability s. s s Definition 1. Given strategies s and t, let E [t|s] = t · xs represent the expectation over the possible actions associ- ated with t, given the action associated with probability s for q∗ q∗ p∗ p∗ strategy s. πp∗ + ρp∗ = πq∗ + ρq∗ Definition 2. For strategies s = (s, 1 − s) and t, and as- r2 − 1 2r − 2 = + suming without loss of generality that s > 0.5 we say r2 + 6r + 9 r2 + 6r + 9 s s πt = sE [t|s] is the t-reward against s. It is the expectation r2 + 2r − 3 over the possible choices of t, given the choice of the higher = 2 probability action associated with s. r + 6r + 9 r − 1 Definition 3. For strategies s = (s, 1 − s) and t, and as- = suming without loss of generality that s > 0.5, we say r + 3 s s = v ρt = (1 − s)E [t|1 − s] is the t-risk against s. It is the expec- tation over the possible choices of t, given the choice of the In other words, v can be understood as the sum of the lower probability action associated with s. equilibrium reward and equilibrium risk. We might alterna- In particular, for the Domination game we have (see Ap- tively say that v is the sum of the rational reward and ratio- pendix B for details): nal risk, and this is the classical solution approach in game 2 theory. However, there are alternative solution approaches q∗ p∗ r − 1 π ∗ = π ∗ = that may be considered. p q r2 + 6r + 9 Definition 4. Given a player’s equilibrium strategy s∗ = q∗ p∗ 2r − 2 @ ρ ∗ = ρ ∗ = [s, 1 − s], s = [1 − s, s] is that player’s irrational strategy. p q r2 + 6r + 9 Player A’s (B’s) irrational strategy for the Domination Notice that these quantities can be combined in two natu- game is p@ = [1 − p, p] (q@ = [1 − q, q]), and when this v ral ways, both of which express the value of the game, just strategy is played against B’s (A’s) rational strategy, as in Figure 1: 2 ∗ ∗ 1 − r πq = πp = p@ q@ r2 + 6r + 9 2 ∗ ∗ 2r + 2r − 4 ρq = ρp = p@ q@ r2 + 6r + 9 We will sometimes refer to these quantities as A’s (B’s) irrational payoff or A’s (B’s) irrational risk. Figure 2 shows irrational reward and risk for A against B’s (and B against A’s) rational strategy. Note that the result of playing irra- tionally is that the reward is now inverted – if the opponent plays rationally, its higher probability action will result in a loss for A, while the risk results in a (possibly very large) win. However, the probability of the risk goes to zero as quickly as the risk grows, resulting in a bounded risk whose expectation is only twice v. ∗ Figure 1: Decomposing v into reward π and risk ρ, as a However, since the value πp is less than v (recall that B function of r q@ prefers smaller payoffs), this is the first scenario we’ve seen

84 t t ρs that allows B to actually win (as opposed to just limiting its Definition 8. ψs = # is the risk coefficient of strategy s ρs losses), and it therefore appears to be in B’s best interest to against t. act irrationally. How can this be reconciled with the fact that v represents the equilibrium (best for both) value? The an- For the Domination game, if both players are playing ra- ∗ ∗ swer is because v can now be thought of as also representing tionally (joint strategy (p , q )), an expectation over B’s choices given that A chooses ratio- ∗ ∗ ∗ ∗ ∗ ∗ q p q p p p ∗ 2 nally (because π = π and ρ = ρ , π + ρ = v). q r −1 2 p@ q@ p@ q@ q@ q@ q∗ πp∗ r2+6r+9 r − 1 As r increases, the expectation is skewed more and more φp∗ = # = = 3 2 π ∗ r r + 6r + 9r by the large value while at the same time, the probability p of A actually choosing action a approaches 0. The larger q∗ 2r−2 2 q∗ ρp∗ r2+6r+9 2 − 2r p∗ ψp∗ = # = = 2 the value of r, the farther away the [irrational] risk (ρq@ ) −1 r + 4r + 4 ρp∗ gets from the worst case (r), while the closer the expected ∗ 2 p∗ p r −1 2 ∗ π ∗ 2 1 − r [irrational] reward (π @ ) gets to the best case (−1). So, as p q r +6r+9 q φ ∗ = = = q # −1 r2 + 6r + 9 r increases, B has (relatively) greater and greater incentive πq∗ to act irrationally. In other words, the greater A’s advantage p∗ 2r−2 becomes, the greater incentive B has to act irrationally. p∗ ρq∗ r2+6r+9 2r − 2 ψq∗ = # = = 3 2 Note that if either one of the players plays irrationally ρ ∗ r r + 6r + 9r (while the other plays rationally), the value of the game does q not change and is still the sum of a reward and a risk, For the Domination game, if player A is playing irrationally 2 2 @ ∗ ∗ ∗ ∗ ∗ 1 − r 2r + 2r − 4 while player B is playing rationally (joint strategy (p , q )), πq + ρq = πp + ρp = + p@ p@ q@ q@ r2 + 6r + 9 r2 + 6r + 9 2 ∗ 2 r + 2r − 3 q 1−r 2 = ∗ π @ 2 1 − r r2 + 6r + 9 φq = p = r +6r+9 = p@ # r r3 + 6r2 + 9r r − 1 π @ = p r + 3 ∗ 2 q 2r +2r−4 2 = v ∗ ρ @ 2 −2r − 2r + 4 ψq = p = r +6r+9 = p@ # −1 r2 + 4r + 4 ρp@ As a result, unilateral irrationality does not affect the value v of the game, and these quantities can again be com- On the other hand, if player A is playing rationally while bined to yield two more expressions for v. In other words, v player B is playing irrationally (joint strategy (p∗, q@)), can also be understood as the sum of an irrational expected payoff and an irrational expected risk, given that one of the ∗ 2 players is acting rationally. πp 1−r 2 p∗ q@ r2+6r+9 r − 1 To summarize, v can be interpreted in any of three ways: φq@ = # = = 2 ∗ ∗ −1 r + 6r + 9 as the expected result for the joint strategy (p , q ) (both πq@ ∗ 2 players acting rationally), or as the expected result for the p 2r +2r−4 2 ∗ @ ∗ ρ @ 2 2r + 2r − 4 joint strategy (p , q ) (player 1 acting rationally and player ψp = q = r +6r+9 = 2 acting irrationally) or as the expected result for the joint q@ # r r3 + 6r2 + 9r ρq@ strategy (p@, q∗) (player 1 acting irrationally and player 2 acting rationally). Definition 9. Given a rational (equilibrium) joint strategy The additional joint strategy (p@, q@) may also be ex- (s∗, t∗) for players A and B, where s∗ = [s, 1 − s], say it panded as above; however, the result cannot be used to ex- is quasi-rational for A to play the irrational strategy s@ = press v (see Appendix B for details.) [1 − s, s] against t∗ if the following conditions hold: # Definition 5. πs is the maximal reward for strategy s. Here t∗ t∗ π > π ∗ B maximal refers to the amplitude of the payoff, with the sign 1. s@ s (for player , who wants to minimize pay- determined by the goal of the player using s. off, the inequality is reversed) ∗ φ # 2. φt > θ Definition 6. ρs is the maximal risk for strategy s. Here s@ A maximal refers to the amplitude of the payoff, with the sign t∗ ψ 3. ψs@ < θA s determined by the goal of the player using . The first condition ensures that the irrational reward is In a zero-sum game with players using strategies s and t, better than the rational reward. The second condition ensures # # # # ρs = πt and ρt = πs , and, in the Domination game, that the irrational reward is sufficiently close to the maxi- # # # # # # φ πp∗ = πp@ = r, πq∗ = πq@ = −1, ρp∗ = ρp@ = −1 and mal reward, where “sufficiently close” is defined by θA. The # # third condition ensures that the irrational risk is significantly ρ ∗ = ρ = r. q q@ less than the maximal risk, where “significantly less” is de- t t πs ψ φ Definition 7. φ = # is the reward coefficient of strategy fined by θ . We will refer to θ as A’s reward threshold and s π A A s ψ s against t. to θA as A’s risk threshold. A reckless player will have a

85 p∗ φ low reward threshold and a high risk threshold. A cautious For the second condition we must show φq@ > θB. player will be opposite. Both thresholds low would charac- 2 terize a nickel-and-dime player, while both high would char- ∗ r − 1 φp = acterize a player that pursues high-risk/high-reward scenar- q@ r2 + 6r + 9 t t ios. Note that because ∀ − 1 ≤ π , ρ ≤ 1, it is sufficient 2 s,t s s  φ q φ  φ ψ −3θB − 8θB +1 that −1 ≤ θA, θA ≤ 1. φ − 1 θB −1 > q 2 q  −3θφ − 8θφ +1   −3θφ − 8θφ +1  Definition 10. A rational game is one for which the quasi- B B + 6 B B + 9 θφ −1 θφ −1 rationality conditions do not hold for either player. B B q 9(θφ )2+6θφ 8θφ +1+8θφ +1 B B B B − 1 Definition 11. quasi-rational game φ 2 A is one for which the (θB −1) = q q quasi-rationality conditions hold for exactly one player. φ 2 φ φ φ φ φ 9(θB ) +6θB 8θB +1+8θB +1 −18θB −6 8θB +1 φ 2 + φ + 9 (θB −1) (θB −1) Definition 12. An irrational game is one for which the q φ 2 φ φ φ quasi-rationality conditions hold for both players. 8(θB ) + 6θB 8θB + 1 + 10θB = q 8θφ + 6 8θφ + 1 + 10 Proposition 1. Lots of games are rational, including any B B whose reward structures are not asymmetric. φ φ q φ θB (8θB + 6 8θB + 1 + 10) = q Theorem 1. For any two players A and B, with thresholds φ φ 8θB + 6 8θB + 1 + 10 0 < θφ , θψ , θφ , θψ < 1, there is a game G that is quasi- A A B B = θφ rational. B

p∗ ψ For the third condition we must show ψq@ < θB. Proof. Let G be the Domination game with r > √ ∗ 2 φ φ p 2r + 2r − 4 −3θ − 8θ +1 B B 2 ψq@ = max{ φ , ψ }. We must show that the condi- r3 + 6r2 + 9r θB −1 θB 2 tions of quasi-rationality hold for exactly one player. With-  2   2  2 ψ + 2 ψ − 4 out loss of generality, let that player be B. By Lemma 1, we θB θB < 3 2 have that the quasi-rationality conditions hold for B, and by  2   2   2  Lemma 2, we have that they do not hold for A. ψ + 6 ψ + 9 ψ θB θB θB  ψ   ψ 2  ψ 3 θB θB θB 2 2 + 2 2 − 4 2 =  ψ   ψ 2 ∗ ∗ θB θB Lemma 1. Let (p , q ) be the equilibrium strategy for A 1 + 6 2 + 9 2 and B for the Domination game. Given thresholds 0 < θψ     θφ , θψ < 1, it is quasi-rational for B to play @ against B 2 − θψ 1 + θψ B B √ q 2 B B −3θφ − 8θφ +1 = B B 2  3 ψ   3 ψ  A for r > max{ φ , ψ }. 1 + θ 1 + θ θB −1 θB 2 B 2 B  ψ   ψ  ψ θ 2 − θB 1 + θB = B 2  3 ψ   3 ψ  1 + 2 θB 1 + 2 θB p∗ p∗ Proof. For the first condition we must show π @ < π ∗ . q q θψ (note that since B wants to minimize payoff, we have re- < B · 2 · 1 versed the inequality. 2 ψ = θB 2 ∗ 1 − r πp = q@ r2 + 6r + 9 r2 − 1 < r2 + 6r + 9 Lemma 2. Let (p∗, q∗) be the equilibrium strategy for A p∗ and B for the Domination game. It is not quasi-rational for = π ∗ q A to play p@ against B for r > 1.

φ 2 Note that the inequality holds because θB > 0 implies r > 1. Proof. Since A wants to maximize payoff, it suffices to

86 q∗ q∗ show that πp@ ≤ πp∗ . 2 ∗ 1 − r πq = p@ r2 + 6r + 9 r2 − 1 < r2 + 6r + 9 q∗ = πp∗

Proposition 2. No irrational one-shot, zero-sum strategic game exists.

The Meta-game Insanity Figure 3: Value of the Domination game as a function of Theorem 1 says that for even the most conservative “un- r for various levels of meta-gamesmanship derdog” and the most reckless “bully”, there exists a game for which the “underdog” should play irrationally, but the “bully” should not. Of course, if the “underdog” does play and q1 =qq ˜ + (1 − q˜)(1 − q). With these, we can derive an irrationally, then the “bully” would benefit by playing ir- alternative form of the value of the meta-game Insanity, that rationally, and if the “bully” chooses to play irrationally we will call v1 (see Appendix D for details): then the “underdog” could not afford to. This, in turn, sug- gests a meta-game, which we will call Insanity, in which the v1 = p1q1(1) + (1 − p1)q1(−1) choice of row and column determine the players’ choices + p1(1 − q1)(−1) of whether to play rationally or irrationally for the Domi- +(1 − p1)(1 − q1)(r) nation game. The payoff matrix can be represented as the 2 (expected) value of the game given both player’s choice of = v strategy: It should be obvious that this meta-gamesmanship can be  v −v  I = = vD carried on ad infinitum, and we can identify the meta-level −v rv with the subscript k (we have been discussing the first meta- The upper left represents the expected value of the game for level with k = 1). A little thought should verify that the both players acting rationally and the bottom right for both value of the kth level game is acting irrationally. The bottom left is player A acting irra- v = vk+1 tionally and B acting rationally and the upper right is A act- k ing rationally and B acting irrationally. Note that the form Figure 3 shows how the value of the game changes for vari- of the matrix is exactly the same as that of the Domination ous levels of meta-gamesmanship. While the ultimate value ∗ game, and, indeed, the equilibrium strategies, p˜ = [˜p, 1−p˜] of the game is always the same in the limit (1), the threat ∗ and q˜ = [˜q, 1 − q˜] are the same as those for the Domina- of irrationality can have a significant effect on the value for tion game (see Appendix C): p˜ = p, q˜ = q, and given these, r << ∞. we can compute the equilibrium value of the Insanity game Further, the probability that A will eventually play a1 (Appendix C): v˜ = v2, which is obtained by player A play- given his choice of strategy at the kth level can be defined ∗ ing (meta-)strategy p˜ and player B playing (meta-)strategy recursively as q˜∗. What this means is that player A plays a mixed (meta- )strategy with probability p˜of playing rationally (that is, em- p0 = p ∗ ploying strategy p ) for the Domination game and probabil- pk = ppk−1 + (1 − p)(1 − pk−1), k > 0 ity 1 − p˜ of playing irrationally (strategy p@ for the Dom- ination game) while player B plays a similar mixed (meta- with qk defined similarly. Figure 4 shows how the prob- )strategy over the strategies q∗ and q@. ability p of A playing a1 changes for various levels of We can now ask, given the meta-strategies, how this af- meta-gamemanship. As expected, as k increases, p grows fects the probabilities of the players choosing between their more slowly. Again, while the asymptotic results remain un- two (base) action choices. If p is the probability that A will changed, the threat of irrationality by B affects the behavior of player A. choose action a1 in the base game, we will call p1 the proba- bility that A will choose a1 after first playing the meta-game (and thus selecting whether or not to play the base game ra- Final Comments tionally). If A chooses to play rationally (probability p˜), then The ideas suggested here are underdeveloped and much A will play a1 with probability p. If A chooses to play irra- work remains to be done. Most obviously, experiments must tionally (probability 1 − p˜), then A will play a1 with prob- be designed to place people in situations similar to the ability 1 − p. So, p1 =pp ˜ + (1 − p˜)(1 − p). Likewise, q1 Domination game to see if, in fact, they do behave as pre- will represent the probability that B will choose action b1, dicted (that is, they behave quasi-rationally when the game

87 To find B’s equilibrium strategy, we compute

E[q|p = 1] = q(1) + (1 − q)(−1) = 2q − 1 E[q|p = 0] = q(−1) + (1 − q)(r) = −q − qr + r

and repeat the process to find q∗ = (q, 1 − q):

2q − 1 = −q − qr + r 3q + qr = r + 1 q(r + 3) = r + 1 r + 1 q = r + 3 Figure 4: Probability p of A playing a1 as a function of r for various levels of meta-gamesmanship The (equilibrium) value v of the game can then be expressed as the expected payoff, E[(p, 1 − p), (q, 1 − q)], when both players play their equilibrium strategies: is significantly asymmetric). We might also be able to make some empirical estimation of general values for reward and risk thresholds. In addition, other work on irrationality in v = E[(p, 1 − p), (q, 1 − q)] game theory has been proposed, though much of it appears = pq(1) + (1 − p)q(−1) + p(1 − q)(−1) to treat the case of extensive games (Krepps et al. 1982; + (1 − p)(1 − q)(r) Neyman 1985; Pettit and Sugden 1989; Aumann 1992). Can = pq − (1 − p)q − p(1 − q) + (1 − p)(1 − q)r any of this be applied to the strategic case? Or, can (and r2 + 2r + 1 2r + 2 2r + 2 should) this work be extended to extensive games so that a = − − comparison with these theories can be attempted? r2 + 6r + 9 r2 + 6r + 9 r2 + 6r + 9 4 + r References r2 + 6r + 9 Aumann, R. 1992. Irrationality in game theory. Economic r2 + 2r − 3 = Analysis of Markets and Games, Essays in Honor of Frank r2 + 6r + 9 Hahn 214–227. r − 1 = Krepps, D.; Milgrom, P.; Roberts, J.; and Wilson, R. r + 3 1982. Rational cooperation in the finitely repeated prisoners dilemma. Journal of Economic Theory 27:245–252. Neyman, A. 1985. Bounded complexity justifies coopera- tion in the finitely repeated prisoner’s dilemma. Economics B Decomposing value of Domination game Letters 19:227–229. into risk and reward Pettit, P., and Sugden, R. 1989. The backward induction paradox. The Journal of Philosophy 86:169–182. For convenience, we note the following equalities, which are easily verified. The first group is expectations based on bi- partisan rationality: Appendices q∗ ∗ ∗ q∗ T E [p |q] = p · xq = [p, (1 − p)][1, −1] A Equilibrium of Domination game r − 1 = 2p − 1 = To find A’s equilibrium strategy, we compute r + 3 q∗ ∗ ∗ q∗ T E[p|q = 1] = p(1) + (1 − p)(−1) = 2p − 1 E [p |1 − q] = p · x1−q = [p, (1 − p)][−1, r] E[p|q = 0] = p(−1) + (1 − p)(r) = −p − pr + r r − 1 = −p + (1 − p)r = Settting these equal and solving for p gives A’s equilibrium r + 3 ∗ p∗ ∗ ∗ p∗ T strategy p = (p, 1 − p): E [q |p] = q · xp = [q, (1 − q)][1, −1] r − 1 2p − 1 = −p − pr + r = 2q − 1 = 3p + pr = r + 1 r + 3 p∗ ∗ ∗ p∗ T p(r + 3) = r + 1 E [q |1 − p] = q · x1−p = [q, (1 − q)][−1, r] r + 1 r − 1 p = = −q + (1 − q)r = r + 3 r + 3

88 The second group is expectations based on unilateral irra- and with the second, for unilateral irrationality: tionality: 2 ∗ ∗  r + 1   1 − r  1 − r πq = qEq [p@|q] = = p@ r + 3 r + 3 r2 + 6r + 9 ∗ ∗ q @ @ q T 2 E [p |q] = p · x = [(1 − p), p][1, −1] ∗ ∗  2   r + r − 2  q ρq = (1 − q)Eq [p@|1 − q] = 1 − r p@ r + 3 r + 3 = 1 − 2p = 2r2 + 2r − 4 r + 3 = ∗ ∗ 2 Eq [p@|1 − q] = p@ · xq = [(1 − p), p][−1, r]T r + 6r + 9 1−q 2 ∗ ∗  r + 1   1 − r  1 − r 2 p p @ r + r − 2 π @ = pE [q |p] = = = p + pr − 1 = q r + 3 r + 3 r2 + 6r + 9 2 r + 3 ∗ ∗     ∗ ∗ p p @ 2 r + r − 2 p @ @ p T ρ @ = (1 − p)E [q |1 − p] = E [q |p] = q · xp = [(1 − q), q][1, −1] q r + 3 r + 3 1 − r 2r2 + 2r − 4 = 1 − 2q = = r + 3 r2 + 6r + 9 p∗ @ @ p∗ T E [q |1 − p] = q · x1−p = [(1 − q), q][−1, r] and with the third, the case for bipartisan irrationality: 2    2  r + r − 2 @ @ r + 1 r + r − 2 = q + qr − 1 = πq = qEq [p@|q] = r + 3 p@ r + 3 r + 3 r3 + 2r2 − r − 2 = And the last group is expectations based on bipartisan irra- r2 + 6r + 9 tionality:     @ @ 2 1 − r ρq = (1 − q)Eq [p@|1 − q] = p@ r + 3 r + 3 q@ @ @ q@ T 2 − 2r E [p |q] = p · xq = [(1 − p), p][−1, r] = 2 r2 + r − 2 r + 6r + 9    2  = p + pr − 1 = @ @ r + 1 r + r − 2 r + 3 p p @ πq@ = pE [q |p] = @ @ r + 3 r + 3 Eq [p@|1 − q] = p@ · xq = [(1 − p), p][1, −1]T 1−q r3 + 2r2 − r − 2 1 − r = = 1 − 2p = r2 + 6r + 9 r + 3 @ @     @ @ p p @ 2 1 − r p @ @ p T ρ @ = (1 − p)E [q |1 − p] = E [q |p] = q · xp = [(1 − q), q][−1, r] q r + 3 r + 3 r2 + r − 2 2 − 2r = q + qr − 1 = = r + 3 r2 + 6r + 9 p@ @ @ p@ T E [q |1 − p] = q · x1−p = [(1 − q), q][1, −1] Finally, we can see that for bipartisan rationality, the value 1 − r v of the game is the sum of the rational reward and the ratio- = 1 − 2q = nal risk (from either player’s perspective): r + 3 2 q∗ q∗ p∗ p∗ r − 1 2r − 2 π ∗ + ρ ∗ = π ∗ + ρ ∗ = + p p q q r2 + 6r + 9 r2 + 6r + 9 Given the first group, we can now compute rewards and r2 + 2r − 3 risks for bipartisan rationality: = r2 + 6r + 9 r − 1     2 = q∗ q∗ ∗ r + 1 r − 1 r − 1 r + 3 π ∗ = qE [p |q] = = p r + 3 r + 3 r2 + 6r + 9 = v     q∗ q∗ ∗ 2 r − 1 and that, in fact, the same thing may be said for either case ρp∗ = (1 − q)E [p |1 − q] = r + 3 r + 3 of unilateral irrationality: 2r − 2 = 2 2 r2 + 6r + 9 q∗ q∗ p∗ p∗ 1 − r 2r + 2r − 4 π @ + ρ @ = π @ + ρ @ = +     2 p p q q r2 + 6r + 9 r2 + 6r + 9 p∗ p∗ ∗ r + 1 r − 1 r − 1 π ∗ = pE [q |p] = = 2 q 2 r + 2r − 3 r + 3 r + 3 r + 6r + 9 =     2 p∗ p∗ ∗ 2 r − 1 r + 6r + 9 ρq∗ = (1 − p)E [q |1 − p] = r − 1 r + 3 r + 3 = 2r − 2 r + 3 = r2 + 6r + 9 = v

89 However, bipartisan irrationality does not conserve the D Revisiting value of Insanity game game’s value, instead scaling it by r: v1 = p1q1(1) + (1 − p1)q1(−1) + p1(1 − q1)(−1)

@ @ @ @ + (1 − p )(1 − q )(r) πq + ρq = πp + ρp 1 1 p@ p@ q@ q@ = [(˜pp + (1 − p˜)(1 − p))(˜qq + (1 − q˜)(1 − q))](1) 3 2 r + 2r − r − 2 2 − 2r + [(1 − (˜pp + (1 − p˜)(1 − p))) = 2 + 2 r + 6r + 9 r + 6r + 9 (˜qq + (1 − q˜)(1 − q))](−1) r3 + 2r2 − 3r = + [(˜pp + (1 − p˜)(1 − p)) r2 + 6r + 9 (1 − (˜qq + (1 − q˜)(1 − q)))](−1) 2 r − r + [(1 − (˜pp + (1 − p˜)(1 − p))) = r + 3 (1 − (˜qq + (1 − q˜)(1 − q)))](r) r(r − 1) = [(˜pp + (1 − p˜)(1 − p))(˜qq + (1 − q˜)(1 − q))](1) = r + 3 +[(˜p(1 − p) + (1 − p˜)p)(˜qq + (1 − q˜)(1 − q))](−1) = rv +[(˜pp + (1 − p˜)(1 − p))(˜q(1 − q) + (1 − q˜)q)](−1) +[(˜p(1 − p) + (1 − p˜)p)(˜q(1 − q) + (1 − q˜)q)](r) C Equilibrium of Insanity game =pp ˜ qq˜ +pp ˜ (1 − q˜)(1 − q) + (1 − p˜)(1 − p)˜qq +(1 − p˜)(1 − p)(1 − q˜)(1 − q) − p˜(1 − p)˜qq −p˜(1 − p)(1 − q˜)(1 − q) − (1 − p˜)pqq˜ To find A’s equilibrium strategy, we compute −(1 − p˜)p(1 − q˜)(1 − q) − pp˜ q˜(1 − q) +pp ˜ (1 − q˜)q −(1 − p˜)(1 − p)˜q(1 − q) − (1 − p˜)(1 − p)(1 − q˜)q E[˜p|q˜ = 1] =pv ˜ − (1 − p˜)v p˜(1 − p)˜q(1 − q)r +p ˜(1 − p)(1 − q˜)qr +(1 − p˜)pq˜(1 − q)r + (1 − p˜)p(1 − q˜)qr E[˜p|q˜ = 0] = −pv˜ + (1 − p˜)rv =p ˜q˜(pq − (1 − p)q − p(1 − q) + (1 − p)(1 − q)r) p˜(1 − q˜)(p(1 − q) − (1 − p)(1 − q) − pq + (1 − p)qr) Setting these equal and solving for p˜ gives A’s equilibrium (1 − p˜)˜q((1 − p)q − pq − (1 − p)(1 − q) + p(1 − q)r) strategy p˜∗ = (˜p, 1 − p˜): (1 − p˜)(1 − q˜)((1 − p)(1 − q) − p(1 − q) − (1 − p)q + pqr) q∗ q∗ p∗ p∗ =p ˜q˜(π ∗ + ρ ∗ ) − p˜(1 − q˜)(π + ρ ) p p q@ q@ ∗ ∗ @ @ pv˜ − (1 − p˜)v = −pv˜ + (1 − p˜)rv −(1 − p˜)˜q(πq + ρq ) + (1 − p˜)(1 − q˜)(πq + ρq ) p@ p@ p@ p@ pv˜ − v +pv ˜ = −pv˜ + rv − prv˜ =p ˜qv˜ − p˜(1 − q˜)v − (1 − p˜)˜qv + (1 − p˜)(1 − q˜)rv 3˜pv +prv ˜ = v + rv = v(˜pq˜ − p˜(1 − q˜) − (1 − p˜)˜q + (1 − p˜)(1 − q˜)r) p˜(rv + 3v) = rv + v = v(pq − p(1 − q) − (1 − p)q + (1 − p)(1 − q)r) rv + v = v2 p˜ = rv + 3v where in the third step we used the identity r + 1 p˜ = r + 3 1−(˜pp+(1−p˜)(1−p)) = 1−pp˜ −1+˜p+p−pp˜ =p ˜(1−p)+(1−p˜)p the substitutions used in the sixth step come from Ap- Since the game is symmetric, we have again that q˜ =p ˜ pendix B and can be verified with some additional algebra ∗ and B’s equilibrium strategy q1 = (˜q, 1 − q˜) is again the and in the ninth step we used the identities r+1 same as A’s: q˜ = r+3 = q. The (equilibrium) value v˜ of the game can then be expressed as the expected payoff, p˜ = p and q˜ = q E[(˜p, 1 − p˜), (˜q, 1 − q˜)], when both players play their equi- librium strategies:

v˜ = E[(˜p, 1 − p˜), (˜q, 1 − q˜)] =p ˜q˜(v) + (1 − p˜)˜q(−v) +p ˜(1 − q˜)(−v) + (1 − p˜)(1 − q˜)(rv) = v(˜pq˜ − (1 − p˜)˜q − p˜(1 − q˜) + (1 − p˜)(1 − q˜)r) = v2 with the last step possible because p˜ = p and q˜ = q and us- ing the third equality from the derivation of v in Appendix A.

90