MT368, Continuous Games Begun
Total Page:16
File Type:pdf, Size:1020Kb
MT369, Continuous games 1
MT369, Part 6 Continuous games on the unit square, begun Notation. Some of this is non-standard. IX denotes the unit interval 0 x 1 and IY the interval 0 y 1 . The unit square written is the square IX IY in the X – Y plane: I will write u to mean that u is a point in the square. The interior of ( written ) is the set of points 0 < x < 1 and 0 < y < 1 . A continuous game on the unit square is played by 2 persons A and B . A secretly chooses a number x IX and B secretly chooses a number y IY . Then they reveal their choices, and A wins an amount z from B where z is some predetermined function of x and y . The payoff function is written P( x , y ) and is usually called the payoff kernel . The traditional example is Example. A duel . A and B each have a pistol with one bullet. At time t = 0 they face each other & start walking towards each other. If nobody gets shot, they meet at time t = 1. If A fires at time t=x , he hits B (payoff 1) with probability x . Likewise if B fires at time t=y he hits B (payoff 1) with probability y . A player who is hit cannot then shoot. If neither is hit, or if both hit, the payoff is 0 . There are 2 versions of this: 1. With silencers. Neither player knows if the other has fired unless he is hit. 2. No silencers. If A shoots first & misses, then B can wait until t = 1. Ex. Find the payoff functions under the 2 rules. Done in class.
Definition. A point ( u , v ) is a saddle point if P( x , v ) P( u , v ) P ( u , y ) (5.1) for all other x IX and y IY . If so, then ( u , v ) are optimum strategies for A and B . Note that this is not the normal definition of a S.P . Normally P has a SP if P has a local max in any direction and a local min in any other direction, but in Game Theory P must have local max in the X direction and min in the Y direction. Hence Theorem. Let P be twice differentiable for all points in with a saddle point at ( u , v ) . Then either u = 0 and P/x < 0 or u = 1 and P/x > 0 or P/x = 0 and 2P/x2 0 and similarly for y with < and > signs switched.
1 Example: P = xy – x/3 – y/2 has a SP at ( ½ , /3 ) Example: P = xy + x + y . SP at ( 1 , 0 ) .
Given the function P we try to solve the game as follows: first look for a SP in . If that fails, we may need to look at the boundaries of : there are 4 sides and 4 corners. If that fails, we must try to find a solution in mixed strategies, once we have decided what a mixed strategy is. MT369, Continuous games 2
5.2 Distributions Definition. A mixed strategy for A is a decision by A to choose numbers x from IX by some random process according to a selected probability distribution ( , say) . (examples) Definition. A probability distribution ( P.D. for short) is an increasing function on IX such that (0) = 0 and (1) = 1 . Given a random process that produces numbers x in IX , the distribution function of that process is given by prob( x ) . For the moment, assume that is differentiable. Then the density function of is d/dx . A value of x is active if d/dx > 0 and non-active if d/dx = 0 . The support or spectrum of is the set of active values of x . A distribution might be discontinuous. For example if A decided to choose the exact value x = ¼ with prob. ½ , then the corresponding jumps up by ½ at that point. is a step point of if jumps in this way. Because it can only have simple jumps. All the step points are active also. So when A is playing a mixed strategy, he first chooses the function . Then each time the game is played, he chooses x according to that distribution. Imagine as mapping IX into IZ . The easiest way for A to choose numbers x according to the distribution is: first choose IZ (all numbers equally likely) and then x = sup{ u IX: (u) } . (examples) Suppose that A has chosen a strategy represented by the dist. ( a map from IX into IZ ) and B chooses to play y . What is A's expected payoff? If IX was a finite set, then the expected payoff would be EP( y )= Prob( x = a ) P ( a , y ) a IX (5.2) Since IX is actually a slice of the real line, we must replace this sum by an integral. Now for any given , the probability that a random x is actually equal to is 0 . So we must replace Prob( x = ) by the probability that x + x . This prob. is equal to ( x ) . If is differentiable, this is approximately x'() . So 1 EP(q , y )= P ( x , y ) dq dx (5.3) x=0 dx If jumps by an amount at x = , that means that A chooses the exact value x = with probability . Then we must modify the RHS of (5.3) by adding a term P( , x ) ; and we need one such term for each step point. Then the integral on the RHS of (5.2) has to be replaced by: 1 EP( y ,q )= P ( x , y )dq dy + P ( a , y ) dq ( a ) x=0 dy (5.4) step points x=a
1 We write this mixture of sum and integral thus: EP( y ,q )= P ( x , y ) d q . (5.5) x=0 Definition. Let be a distribution on IX and u(x) a function defined on IX . If A chooses 1 points in IX with P.D. , then the mean value of u is written u= u( x ) dq , interpreted as x=0 follows: 1 1 u= u( x ) dq = u ( x )dq dx + u ( a ) dq ( a ) 蝌x=0 x = 0 dx (5.6) step points x=a So (5.6) means that we integrate when is smooth and sum where it jumps. MT369, Continuous games 3
Now suppose that A chooses strategy and B chooses . For each y the expected payoff is given by (5.5). Then the overall mean payoff will be got by a 2nd integrate-or-sum operation: 1 1 EP(q , f )= P ( x , y ) d q d f (5.7) 蝌y=0 x = 0
Definition. Let and be a pair of strategies for the two players A , B , respectively. Then * and * are an equilibrium pair of strategies if EP(l , f *)# EP ( q *, f *) EP ( q *, m ) (5.8) for every other strategy for A and for B . Neither player can improve her payoff by deviating. Ville’s Theorem. (not proved, beyond scope) Let P( x , y ) be continuous on . Then the corresponding game has a solution in mixed strategies; that is, there exist equilibrium strategies and as defined above.
The opt. strategies are guaranteed to exist, but there is no known method to find them. There is one special case.
Definition. Let f(x) be a function defined on an interval I R . Then f is a convex function if f(a p+ (1 - a ) q )� a f ( p ) - (1 a ) f ( q ) (5.9) for all p , q I and all in the range 0 < < 1 . f is strictly convex if (5.9) holds with < in place of .
This means that if we take 2 points on the graph of f , namely ( p , f(p) ) and ( q , f(q) ) and join by a line, the graph of u sags below the line. If f is twice differentiable, then f convex implies that f " 0 in I . Roughly speaking: a convex function either goes downhill (getting flatter) , or uphill (getting steeper) or downhill then flat , then uphill . Examples: ex , ex , x2 1/x . Non-example: (x-1)(x-2)(x-3) (draw graph & observe bends). Definition. f(x) is (strictly ) concave if f(x) is (strict) convex . Non-standard notation: from a book on another subject: convex or concave makes it easier to remember which is which. Definition. A game P( x , y ) is convex if P( x , y ) is a strictly convex function of y for each fixed x IX . Note: non –symmetrical definition.
Theorem. (proof will be omitted unless there is time, in any case the proof is not examinable) Let f(x) be convex and bounded in an interval I : a x b . Then there exist numbers c and d such that a c d b and 1 if a < c then f(x) is strictly decreasing in [ a , c ] 2 if c < d then f(x) is constant ( = , say) in ( c , d ) . If c = d, put = f(c) . 3 if d < b then f(x) is strictly increasing in [ d , b ] and is the minimum value of f(x) in I . 4 If f is strict convex, then c = d i.e. the min at c is unique. MT369, Continuous games 4
Proof. (outline only, omitting the difficult parts). Let = min value of f in I and c and d the upper and lower bounds of points where f(x) = . Then f(c) = f(d) = . If c < x < d , then x = c + ( 1 )d for some . Then (5.9) says that f(x) . But = min value f(x) = for all x in ( c , d ) . This proves (2). But if f is strict convex, then (5.9) says that f(x) < , contradiction. Proves (4) . To prove (1). Let a p < q c , to prove that f(p) > f(q) . Since = min (f) and c = least x for which f = , f(p) > , say f(p) = . Now q = p + ( 1 )c for some . Then (5.9) says that f(q) f(p) Proves 1 . (3) is similar.
Definition. A game P( x , y ) is convex if P( x , y ) is a strictly convex function of y for each fixed x IX . Theorem. Assuming that optimum strategies exist, if P( x , y ) is convex then B has a unique opt. strategy which is a pure strategy. i.e. B should play y=u (fixed) every time. Proof. For each fixed x , P( x , y) is a strict-convex function of y . Let p < q < r be any 3 values of y in IY , say q = p + r . By (5.9): P( x , q ) < P( x , p ) + P( x , r ) P( x , p ) + P( x , r ) P( x , q ) > 0 (5.10) Now suppose that is an optimum strategy for A . Multiply (5.10) by d and integrate. Recall 1 that EP( , y ) P ( x , y ) d is the expected payoff to A if A plays and B x0 plays y . Applying this to (5.10) we get: P( , p ) + P( , r ) EP( , q ) > 0 EP( , q ) < P( , p ) + P( , r ) So the expected payoff EP( , y ) is a strictly convex function of y it has a unique minimum, say at y = u . Since this min. is unique, we have EP( , u ) < P( , y ) (5.11) for all y IY different from u . Since is opt. for A we have also EP( , u ) P( , u ) (5.12) where is any deviant strategy for A . But now A might have two opt strategies, say and . By the above argument, EP( , y ) is strict convex, it has a unique min, say at y = v . Required to prove that u = v . Suppose not. Since and y=v are a second pair of opt. strategies for A and B , they must satisfy the equivalent of (5.11) and (5.12) which are: EP( , v ) P( , v ) < P( , y ) (5.13, 5.14) where is any deviant strategy for A and y v. But now EP( , u ) < P( , v ) (by 5.11 with y = v ) P( , v ) ( 5.13 with < P( , u ) (5.14 with y = u ) EP( , u ) (5.12 with
So EP( , u ) < EP( , u ) Contradiction, so u = v .