MT368, Continuous Games Begun

MT368, Continuous Games Begun

<p> MT369, Continuous games 1</p><p>MT369, Part 6 Continuous games on the unit square, begun Notation. Some of this is non-standard. IX denotes the unit interval 0  x  1 and IY the interval 0  y  1 . The unit square written  is the square IX  IY in the X – Y plane: I will write u   to mean that u is a point in the square. The interior of  ( written  ) is the set of points 0 < x < 1 and 0 < y < 1 . A continuous game on the unit square is played by 2 persons A and B . A secretly chooses a number x  IX and B secretly chooses a number y  IY . Then they reveal their choices, and A wins an amount z from B where z is some predetermined function of x and y . The payoff function is written P( x , y ) and is usually called the payoff kernel . The traditional example is Example. A duel . A and B each have a pistol with one bullet. At time t = 0 they face each other & start walking towards each other. If nobody gets shot, they meet at time t = 1. If A fires at time t=x , he hits B (payoff 1) with probability x . Likewise if B fires at time t=y he hits B (payoff 1) with probability y . A player who is hit cannot then shoot. If neither is hit, or if both hit, the payoff is 0 . There are 2 versions of this: 1. With silencers. Neither player knows if the other has fired unless he is hit. 2. No silencers. If A shoots first & misses, then B can wait until t = 1. Ex. Find the payoff functions under the 2 rules. Done in class. </p><p>Definition. A point ( u , v )   is a saddle point if P( x , v )  P( u , v )  P ( u , y ) (5.1) for all other x  IX and y  IY . If so, then ( u , v ) are optimum strategies for A and B . Note that this is not the normal definition of a S.P . Normally P has a SP if P has a local max in any direction and a local min in any other direction, but in Game Theory P must have local max in the X direction and min in the Y direction. Hence Theorem. Let P be twice differentiable for all points in  with a saddle point at ( u , v ) . Then either u = 0 and P/x < 0 or u = 1 and P/x > 0 or P/x = 0 and 2P/x2  0 and similarly for y with < and > signs switched.</p><p>1 Example: P = xy – x/3 – y/2 has a SP at ( ½ , /3 ) Example: P = xy + x + y . SP at ( 1 , 0 ) .</p><p>Given the function P we try to solve the game as follows: first look for a SP in  . If that fails, we may need to look at the boundaries of  : there are 4 sides and 4 corners. If that fails, we must try to find a solution in mixed strategies, once we have decided what a mixed strategy is. MT369, Continuous games 2</p><p>5.2 Distributions Definition. A mixed strategy for A is a decision by A to choose numbers x from IX by some random process according to a selected probability distribution (  , say) . (examples) Definition. A probability distribution ( P.D. for short) is an increasing function  on IX such that (0) = 0 and (1) = 1 . Given a random process that produces numbers x in IX , the distribution function of that process is given by  prob( x   ) . For the moment, assume that  is differentiable. Then the density function of  is d/dx . A value of x is active if d/dx > 0 and non-active if d/dx = 0 . The support or spectrum of  is the set of active values of x . A distribution might be discontinuous. For example if A decided to choose the exact value x = ¼ with prob. ½ , then the corresponding  jumps up by ½ at that point.  is a step point of  if  jumps in this way. Because   it can only have simple jumps. All the step points are active also. So when A is playing a mixed strategy, he first chooses the function  . Then each time the game is played, he chooses x according to that distribution. Imagine  as mapping IX into IZ . The easiest way for A to choose numbers x according to the distribution  is: first choose   IZ (all numbers equally likely) and then x = sup{ u  IX: (u)   } . (examples) Suppose that A has chosen a strategy represented by the dist.  ( a  map from IX into IZ ) and B chooses to play y . What is A's expected payoff? If IX was a finite set, then the expected payoff would be EP( y )= Prob( x = a ) P ( a , y ) a IX (5.2) Since IX is actually a slice of the real line, we must replace this sum by an integral. Now for any given  , the probability that a random x is actually equal to  is 0 . So we must replace Prob( x =  ) by the probability that   x   + x . This prob. is equal to ( x )  . If  is differentiable, this is approximately x'() . So 1 EP(q , y )= P ( x , y ) dq dx (5.3) x=0 dx If  jumps by an amount  at x =  , that means that A chooses the exact value x =  with probability  . Then we must modify the RHS of (5.3) by adding a term P(  , x )  ; and we need one such term for each step point. Then the integral on the RHS of (5.2) has to be replaced by: 1 EP( y ,q )= P ( x , y )dq dy + P ( a , y ) dq ( a ) x=0 dy (5.4) step points x=a</p><p>1 We write this mixture of sum and integral thus: EP( y ,q )= P ( x , y ) d q . (5.5) x=0 Definition. Let  be a distribution on IX and u(x) a function defined on IX . If A chooses 1 points in IX with P.D.  , then the mean value of u is written u= u( x ) dq , interpreted as x=0 follows: 1 1 u= u( x ) dq = u ( x )dq dx + u ( a ) dq ( a ) 蝌x=0 x = 0 dx (5.6) step points x=a So (5.6) means that we integrate when  is smooth and sum where it jumps. MT369, Continuous games 3</p><p>Now suppose that A chooses strategy  and B chooses  . For each y the expected payoff is given by (5.5). Then the overall mean payoff will be got by a 2nd integrate-or-sum operation: 1 1 EP(q , f )= P ( x , y ) d q d f (5.7) 蝌y=0 x = 0</p><p>Definition. Let  and  be a pair of strategies for the two players A , B , respectively. Then * and * are an equilibrium pair of strategies if EP(l , f *)# EP ( q *, f *) EP ( q *, m ) (5.8) for every other strategy for A and  for B . Neither player can improve her payoff by deviating. Ville’s Theorem. (not proved, beyond scope) Let P( x , y ) be continuous on  . Then the corresponding game has a solution in mixed strategies; that is, there exist equilibrium strategies  and  as defined above.</p><p>The opt. strategies are guaranteed to exist, but there is no known method to find them. There is one special case.</p><p>Definition. Let f(x) be a function defined on an interval I  R . Then f is a convex function if f(a p+ (1 - a ) q )� a f ( p ) - (1 a ) f ( q ) (5.9) for all p , q  I and all  in the range 0 <  < 1 . f is strictly convex if (5.9) holds with < in place of  . </p><p>This means that if we take 2 points on the graph of f , namely ( p , f(p) ) and ( q , f(q) ) and join by a line, the graph of u sags below the line. If f is twice differentiable, then f convex implies that f "  0 in I . Roughly speaking: a convex function either goes downhill (getting flatter) , or uphill (getting steeper) or downhill then flat , then uphill . Examples: ex , ex , x2 1/x . Non-example: (x-1)(x-2)(x-3) (draw graph & observe bends). Definition. f(x) is (strictly ) concave if f(x) is (strict) convex . Non-standard notation: from a book on another subject: convex  or concave  makes it easier to remember which is which. Definition. A game P( x , y ) is convex if P( x , y ) is a strictly convex  function of y for each fixed x  IX . Note: non –symmetrical definition.</p><p>Theorem. (proof will be omitted unless there is time, in any case the proof is not examinable) Let f(x) be convex  and bounded in an interval I : a  x  b . Then there exist numbers c and d such that a  c  d  b and 1 if a < c then f(x) is strictly decreasing in [ a , c ] 2 if c < d then f(x) is constant ( = , say) in ( c , d ) . If c = d, put  = f(c) . 3 if d < b then f(x) is strictly increasing in [ d , b ] and  is the minimum value of f(x) in I . 4 If f is strict convex, then c = d i.e. the min at c is unique. MT369, Continuous games 4</p><p>Proof. (outline only, omitting the difficult parts). Let  = min value of f in I and c and d the upper and lower bounds of points where f(x) =  . Then f(c) = f(d) =  . If c < x < d , then x = c + ( 1 )d for some  . Then (5.9) says that f(x)   . But  = min value  f(x) =  for all x in ( c , d ) . This proves (2). But if f is strict convex, then (5.9) says that f(x) <  , contradiction. Proves (4) . To prove (1). Let a  p < q  c , to prove that f(p) > f(q) . Since  = min (f) and c = least x for which f =  , f(p) >  , say f(p) =  . Now q = p + ( 1 )c for some  . Then (5.9) says that f(q)   f(p) Proves 1 . (3) is similar.</p><p>Definition. A game P( x , y ) is convex if P( x , y ) is a strictly convex  function of y for each fixed x  IX . Theorem. Assuming that optimum strategies exist, if P( x , y ) is convex then B has a unique opt. strategy which is a pure strategy. i.e. B should play y=u (fixed) every time. Proof. For each fixed x , P( x , y) is a strict-convex function of y . Let p < q < r be any 3 values of y in IY , say q = p + r . By (5.9): P( x , q ) < P( x , p ) + P( x , r )  P( x , p ) + P( x , r )  P( x , q ) > 0 (5.10) Now suppose that  is an optimum strategy for A . Multiply (5.10) by d and integrate. Recall 1 that EP( , y ) P ( x , y ) d  is the expected payoff to A if A plays  and B x0 plays y . Applying this to (5.10) we get: P(  , p ) + P(  , r )  EP(  , q ) > 0  EP(  , q ) < P(  , p ) + P(  , r ) So the expected payoff EP(  , y ) is a strictly convex function of y  it has a unique minimum, say at y = u . Since this min. is unique, we have EP(  , u ) < P(  , y ) (5.11) for all y  IY different from u . Since  is opt. for A we have also EP(  , u )  P(  , u ) (5.12) where  is any deviant strategy for A . But now A might have two opt strategies, say  and  . By the above argument, EP(  , y ) is strict convex,  it has a unique min, say at y = v . Required to prove that u = v . Suppose not. Since  and y=v are a second pair of opt. strategies for A and B , they must satisfy the equivalent of (5.11) and (5.12) which are: EP(  , v )  P(  , v ) < P(  , y ) (5.13, 5.14) where  is any deviant strategy for A and y  v. But now EP(  , u ) < P(  , v ) (by 5.11 with y = v )  P(  , v ) ( 5.13 with  < P(  , u ) (5.14 with y = u )  EP(  , u ) (5.12 with </p><p>So EP(  , u ) < EP(  , u ) Contradiction, so u = v . </p>

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    4 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us