The Unofficial Solution Manual to A Primer in Game Theory by RA Gibbons Unfinished Draft Navin Kumar Delhi School of Economics 2 This version is an unreleased and unfinished manuscript. The author can be reached at navin.ksrk@gmail.com Last Updated: January 20, 2013 Typeset using LATEX and the Tufte book class. This work is not subject to copyright. Feel free to reproduce, distribute or falsely claim authorship of it in part or whole. 3 This is strictly a beta version. Two thirds of it are missing and there are errors aplenty. You have been warned. On a more positive note, if you do find an error, please email me at navin.ksrk@gmail.com, or tell me in person. - Navin Kumar Static Games of Complete Information Answer 1.1 See text. Answer 1.2 B is strictly dominated by T. C is now strictly dominated by R. The strategies (T,M) and (L,R) survive the iterated elimination of strictly dominated strategies. The Nash Equilibrium are (T,R) and (M,L). Answer 1.3 For whatever value Individual 1 chooses (denoted by S1 ), Individual 2’s best response is S2 = B2 (S1 ) = 1 − S2 . Conversely, S1 = B1 (S2 ) = 1 − S1 . We know this because if S2 < 1 − S1 , then there is money left on the table and Individual 2 could increase his or her payoff by asking for more. If, however, S2 > 1 − S1 , Individual 2 earns nothing and can increase his payoff by reducing his demands sufficiently. Thus the Nash Equilibrium is S1 + S2 = 1. Answer 1.4 The market price of the commodity is determined by the formula P = a − Q in which Q is determined Q = q1 + ... + qn The cost for an individual company is given by Ci = c · qi . The profit made by a single firm is πi = ( p − c) · qi = ( a − Q − c) · qi = ( a − q1∗ − ... − q∗n − c) · qi Where q∗j is the profit-maximizing quantity produced by firm j in equilibrium. This profit is maximized at dπi = ( a − q1∗ − ... − qi∗ − ... − q∗n − c) − qi∗ = 0 dqi ⇒ a − q1∗ − ... − 2 · qi∗ − ... − q∗n − c = 0 ⇒ a − c = q1∗ + ... + 2 · qi∗ + ... + q∗n For all i=1, ... ,n. We could solve this question using matrices and Cramer’s rule, but a simpler method would be to observe that since all firms are symmetric, their equilibrium quantity will be the same i.e. q1∗ = q2∗ = ... = qi∗ = ... = q∗n Which means the preceding equation becomes. a − c = (n + 1)qi∗ ⇒ qi∗ = a−c n+1 Static Games of Complete Information 6 A similar argument applies to all other firms. Answer 1.5 Let qm be the amount produced by a monopolist. Thus, if the two were colluding, they’d each produce q1m = q2m = qm a−c = 2 4 In such a scenario, the profits earned by Firm 1 (and, symmetrically, Firm 2) is a−c 1 1 1 πmm = ( P − c) · qm = ( a − Q − c) · qm = a − − c · q1m 2 = a−c a−c ( a − c )2 · = = 0.13 · ( a − c)2 2 4 8 If both are playing the Cournot equilibrium quantity, than the profits earned by Firm 1 (and Firm 2) are: 1 1 πcc = ( P − c) · q1c = ( a − 1 P = a−Q 2( a − c ) a + 2c − c) · q1c = ( − c) · q1c 3 3 = a − q1cc − q2cc a−c = a−2· 3 a+c = 2 ( a − c )2 9 = 0.11 · ( a − c)2 = What if one of the firms (say Firm 1) plays the Cournot quantity and the other plays the Monopoly quantity? 2 Firm 1’s profits are: 1 πcm = ( P − c) · q1c = ( 5a 7c a−c 5 + − c) · = · ( a − c )2 12 12 3 36 = 0.14 · ( a − c)2 And Firm 2’s profits are: 2 πcm = ( P − c) · q2m = ( 5a 7c a−c 5 + − c) · = · ( a − c )2 12 12 4 48 = 0.10 · ( a − c)2 For notational simplicity, let α ≡ ( a − c )2 Their profits are reversed when their production is. Thus, the payoffs are: Player 2 Player 1 qm qc qm 0.13α, 0.13α 0.10α, 0.14α qc 0.14α, 0.10α 0.11α, 0.11α As you can see, we have a classic Prisoner’s Dilemma: regardless of the other firm’s choice, both firm’s would maximize their payoff by choosing to produce the Cournot quantity. Each firm has a strictly q dominated strategy ( 2m ) and they’re both worse off in equilibrium (where they make 0.11 · ( a − c)2 in profits) than they would have The price is determined by 2 The price is given by P = a−Q = a − q1c − q2m a−c a−c = a− − 3 4 7 = a− · ( a − c) 12 5a 7c = + 12 12 7 been had they cooperated by producing qm together (which would have earned them 0.13 · ( a − c)2 ). Answer 1.6 Price is determined by P = a − Q and Q is determined by Q = q1 + q2 . Thus the profit generated for Firm 1 is given by: π1 = ( P − c1 ) · q1 = ( a − Q − c1 ) · q1 = ( a − q1 − q2 − c1 ) · q1 At the maximum level of profit, dπ1 a − c1 − q2 = ( a − c1 − q1 − q2 ) + q1 · (−1) = 0 ⇒ q1 = dq1 2 And by a similar deduction, q2 = a − c2 − q1 2 Plugging the above equation into it’s predecessor, a − c1 − q1 = 2 a − c2 − q1 2 = a − 2c1 + c2 3 And by a similar deduction, q2 = a − 2c2 + c1 3 Now, 2c2 > a + c1 ⇒ 0 > a − 2c2 + c1 ⇒ 0 > a − 2c2 + c1 ⇒ 0 > q2 3 ⇒ q2 = 0 Since quantities cannot be be negative. Thus, under certain conditions, a sufficient difference in costs can drive one of the firms to shut down. Answer 1.7 We know that, a − pi qi = a−2pi 0 if pi < p j , if pi = p j , if pi > p j . We must now prove pi = p j = c is the Nash Equilibrium of this game. To this end, let’s consider the alternatives exhaustively. If pi > p j = c, qi = 0 and π j = 0. In this scenario, Firm j can increase profits by charging p j + ε where ε > 0 and ε < pi − p j , raising profits. Thus this scenario is not a Nash Equilibrium. If pi > p j > c, qi = 0 and πi = 0 and Firm i can make positive profits by charging p j − ε > c. Thus, this cannot be a Nash Equilibrium. a− p If pi = p j > c, πi = ( pi − c) · 2 i . Firm i can increase profits by charging pi − ε such that p j > pi − ε > 0, grab the entire market and a larger profit, provided ( pi − ε − c ) · ( a − pi − ε ) > ( pi − c ) · ( a − pi ) 2 Static Games of Complete Information 8 Therefore, this is not a Nash Equilibrium If pi = p j = c, then πi = π j = 0. Neither firm has any reason to deviate; if FFirm i were to reduce pi , πi would become negative. If Firm i were to raise pi , qi = πi = 0 and he would be no better off. Thus Firm i (and, symmetrically, Firm j) have no incentive to deviate, making this a Nash Equilibrium. Answer 1.8 The share of votes received by a candidate is given by Si = xi + 1 2 · ( x j − xi ) 1 2 if xi < x j , if xi = x j , 1 2 (1 − xi ) + · ( xi − x j ) if xi > x j . We aim to prove the Nash Equilibrium is ( 12 , 21 ). Let us exhaustively consider the alternatives. Suppose xi = 12 and x j > 12 i.e. One candidate is a centrist while the other (Candidate j) isn’t. In such a case, Candidate j can increase his share of the vote by moving to the left i.e. reducing x j . If x j < 21 , Candidate j can increase his share of the vote by moving to the right. Suppose xi > x j > 12 ; Candidate i can gain a larger share by moving to a point x j > xi > 12 . Thus this is not a Nash Equilibrium. Suppose xi = x j = 21 ; the share of i and j are 12 . If Candidate i were to deviate to a point (say) xi > 12 , his share of the vote would decline. Thus ( 12 , 12 ) is a unique Nash Equilibrium. This is the famous Median Voter Theorm, used extensively in the study of politics. It explains why, for example, presidential candidates in the US veer sharply to the center as election day approaches. Answer 1.9 See text. Answer 1.10 (a) Prisoner’s Dilemma Player 2 Player 1 ( p) Mum (1 − p) Fink (q) Mum −1, −1 −9, 0 (1 − q) Fink 0, −9 −6, −6 Prisoner’s Dilemma In a mixed strategy equilibrium, Player 1 would choose q such that Player 2 would be indifferent between Mum and Fink. The payoff from playing Mum and Fink must be equal. i.e. −1 · q + −9 · (1 − q) = 0 · q + −6 · (1 − q) ⇒ q = −3.5 This is impossible. 9 (b) Player 2 Player 1 Le f t(q0 ) Middle(q1 ) Right(1 − q0 − q1 ) U p( p) 1, 0 1, 2 0, 1 Down(1 − p) 0, 3 0,1 2, 0 Figure 1.1.1. Here, Player 1 must set p so that Player 2 is indifferent between Left, Middle and Right. The payoffs from Left and Middle, for example, have to be equal. i.e. p · 0 + (1 − p ) · 3 = p · 2 + (1 − p ) · 1 ⇒ p = 0.5 Similarly, the payoffs from Middle and Right have to be equal 2 · p + 1 · (1 − p ) = 1 · p + 0 · (1 − p ) ⇒ p = −0.5 Which, besides contradicting the previous result, is quite impossible. (c) Player 2 Player 1 L ( q0 ) C ( q1 ) R (1 − q0 − q1 ) T ( p0 ) 0, 5 4, 0 5, 3 M ( p1 ) 4, 0 0, 4 5, 3 B (1 − p0 − p1 ) 3, 5 3, 5 6, 6 Figure 1.1.4. In a mixed equilibrium, Player 1 sets p0 and p1 so that Player 2 would be indifferent between L, C and R. The payoffs to L and C must, for example, be equal i.e. 4 · p0 + 0 · p1 + 5 · (1 − p0 − p1 ) = 0 · p0 + 4 · p1 + 5 · (1 − p0 − p1 ) ⇒ p0 = p1 Similarly, 0 · p0 + 4 · p1 + 5 · (1 − p0 − p1 ) = 3 · p0 + 3 · p1 + 6 · (1 − p0 − p1 ) ⇒ p1 = 2.5 − p0 Which violates p0 = p1 . Static Games of Complete Information 10 Answer 1.11 This game can be written as Player 2 L ( q0 ) C ( q1 ) R (1 − q0 − q1 ) T ( p0 ) 2, 0 1, 1 4, 2 M ( p1 ) 3, 4 1, 2 2, 3 B (1 − p0 − p1 ) 1, 3 0, 2 3, 0 Player 1 In a mixed Nash Equilibrium, Player 1 sets p0 and p1 so that the expected payoffs from L and C are the same. i.e. E2 ( L) = E2 (C ) ⇒ 0 · p0 + 4 · p1 + 3 · (1 − p0 − p1 ) = 1 · p0 + 2 · p1 + 2 · (1 − p0 − p1 ) ⇒ p1 = 2 · p0 − 1 Similarly, E2 (C ) = E2 ( R) ⇒ 1 · p0 + 2 · p1 + 2 · (1 − p0 − p1 ) = 2 · p0 + 3 · p1 + 0 · (1 − p0 − p1 ) 2 ⇒ p1 = − p0 3 Combining these, 5 2 − p0 ⇒ p0 = 3 9 5 1 ∴ p1 = 2 · p0 − 1 = 2 · − 1 = 9 9 5 1 3 ∴1 − p0 − p1 = 1 − − = 9 9 9 2 · p0 − 1 = Now we must calculate q0 and q1 . Player 2 will set them such that E1 ( T ) = E1 ( M ) ⇒ 2 · q0 + 1 · q1 + 4 · (1 − q0 − q1 ) = 3 · q0 + 1 · q1 + 2 · (1 − q0 − q1 ) ⇒ q1 = 1 − 1.5 · q0 And E1 ( M) = E1 ( B) ⇒ 3 · q0 + 1 · q1 + 2 · (1 − q0 − q1 ) = 1 · q0 + 0 · q1 + 3 · (1 − q0 − q1 ) Qed Answer 1.12 ( q ) L2 (1 − q ) R2 ( p) T1 2, 1 0, 2 (1 − p) B1 1, 2 3, 0 11 Player 1 will set p such that E2 ( L) = E2 ( R) ⇒ 1 · p + 2 · (1 − p ) = 2 · p + 0 · (1 − p ) ⇒p= 2 3 Player 2 will set q such that E1 ( T ) = E1 ( B) ⇒ 2 · q + 0 · (1 − q ) = 1 · q + 3 · (1 − q ) ⇒q= 3 4 Answer 1.13 (q)Apply1 to Firm 1 (1 − q)Apply1 to Firm 2 ( p)Apply2 to Firm 1 1 1 2 w1 , 2 w1 w1 ,w2 (1 − p)Apply2 to Firm 2 w2 ,w1 1 1 2 w2 , 2 w2 There are two pure strategy Nash Equilibrium (Apply to Firm 1, Apply to Firm 2) and (Apply to Firm 2, Apply to Firm 1). In a mixed-strategy equilibrium, Player 1 sets p such that Player 2 is indifferent between Applying to Firm 1 and Applying to Firm 2. E2 (Firm 1) = E2 (Firm 2) 1 1 ⇒ p · w1 + ( 1 − p ) · w1 = p · w2 + ( 1 − p ) · w2 2 2 ⇒p= 2w1 − w2 w1 + w2 Since 2w1 > w2 , 2w1 − w2 is positive and p > 0. For p < 1 to be true, it must be the case that 2w1 − w2 1 < 1 ⇒ w1 < w2 w1 + w2 2 Which is true. And since the payoffs are symmetric, a similar analysis would reveal that q= Answer 1.14 2w1 − w2 w1 + w2 Dynamic Games of Complete Information Answer 2.1 The total family income is given by IC ( A) + IP ( A) This is maximized at d( IC ( A) + IP ( A)) dI ( A) dI ( A) =0⇒ C =− P dA dA dA The utility function of the parents is given by V ( IP − B) + kU ( IC + B) This is maximized at k ⇒k dU ( IC + B) dV ( IP − B) + =0 dB db dU ( IC + B) d( IC + B) dV ( IP − B) IP − B · + · =0 d( IC + B) dB d ( IP − B ) dB ⇒ kU 0 ( IC + B) − V 0 ( IP − B) = 0 ⇒ V 0 ( IP − B∗ ) = kU 0 ( IC + B∗ ) Where B∗ is the maximizing level of the bequest. We know it exists because a) there are no restrictions on B and b) V() and U() are concave and increasing The child’s utility function is given by U ( IC ( A) − B∗ ( A)). This is maximized at dU ( IC ( A) + B∗ ( A)) =0 dA ⇒ U 0 ( IC ( A) + B∗ ( A)) · ⇒ dIC ( A) dB∗ ( A) + dA dA =0 dIC ( A) dB∗ ( A) + = 0 ⇒ IC0 ( A) = − B0∗ ( A) dA dA We now have only to prove B0∗ ( A) = IP0 ( A) dV ( IP ( A) − B∗ ( A)) =0 dA Dynamic Games of Complete Information 14 ⇒V 0 ( IP ( A)∗B ( A)) · dIP ( A) dB∗ ( A) − dA dA =0 ⇒ IP0 ( A) = B0∗ ( A) Answer 2.2 The utility function of the parent is given by V ( IP − B) + k[U1 ( IC − S) + U2 ( B + S)]. This is maximized at d · {V ( IP − B) + k[U1 ( IC − S) + U2 ( B + S)]} =0 dB ⇒ −V 0 + [U10 (−S0B ) + U20 (S0B + 1)] = 0 ⇒ V 0 = −kU10 S0 + U20 S0 + U20 The utility of the child is given by U1 ( IC − S) + U2 (S + B). This is maximized at: d[U1 ( IC − S) + U2 (S + B)] = 0 ⇒ U10 = U20 (1 + B0 ) dS Total utility is given by V ( I p − B) + k(U1 ( IC − S) + U2 ( B + S)) + U1 ( IC − S) + U2 ( B + S) = V ( I p − B) + (1 + k)(U1 ( IC − S) + U2 ( B + S)) This is maximized (w.r.t S) at: V 0 · (− BS0 ) + (1 + K ) · [U10 (−1) + U20 (1 + BS0 )] = 0 ⇒ U10 = U20 (1 + BS0 ) − V 0 BS0 1+k as opposed to U10 = U20 (1 + BS0 ), which is the equilibrium condition. V 0 BS0 1+ k > 0, the equilibrium U10 is ‘too high’ which means that S 0 - the level of savings - must be too low (since dU dS < 0). It should be higher. Since Answer 2.3 To be done Answer 2.4 Let’s suppose c2 = R − c1 , partner 2’s payoff is V − ( R − c1 )2 . If c1 ≥ R, partner 2’s best response is to put in 0 and pocket V. If c1 < R and partner 2 responds with some c2 such that c2 < R − c1 , his payoff is −c22 . If he puts in nothing, he will, receive a payoff of zero. There is, therefore, no reason to put in such a low amount. There is - obviously - also no reason to put in any c2 > R − c1 . He will put in R − c1 if it’s better than putting in nothing i.e. √ V − ( R − c1 )2 ≥ 0 ⇒ c1 ≥ R − V √ √ For player 1, any c1 > R − V is dominated by c1 = R − V. Now, player 1 will do this if the benefit exceeds the cost: √ δV ≥ ( R − V )2 15 ⇒δ≥ R √ −1 V 2 If R2 ≥ 4V, δ would have to be greater than one, which is impossi 2 ble. Therefore, if δ ≥ √R − 1 and R2 ≥ 4V (i.e. the cost is not √ √ V ‘too high’), c1 = R − V and c2 = V. Otherwise, c2 = 0 and c1 = 0. Answer 2.5 Let the ‘wage premium’ be p = w D − wE , where p ∈ (−∞, ∞). In order to get the worker to acquire the skill, the firm has to credibly promise to promote him if he acquires the skill - and not promote him if he doesn’t. Let’s say that he hasn’t acquired the skill. The firm will not promote him iff the returns to the firm are such that: y D0 − w D ≤ y E0 − wE ⇒ y D0 − y E0 ≤ w D − wE = p If he does acquire the skill, the firm will promote if the returns to the firm are such that: y DS − w D ≥ y ES − wE ⇒ y DS − y ES ≥ w D − wE = p Thus the condition which the firm behaves as it ought to in the desired equilibrium is: y D0 − y E0 ≤ p ≤ y DS − y ES Given this condition, the worker will acquire the promotion iff he acquires the skill. He will acquire the skill iff the benefit outweighs the cost i.e. wD − C ≥ wE ⇒ wE + p − C ≥ wE ⇒ p ≥ c That is, the premium paid by the company must cover the cost of training. The company wishes (obviously) to minimize the premium, which occurs at: C if C ≥ y D0 − y E0 , wD − wE = p = y D0 − y E0 if C < y D0 − y E0 A final condition is that the wages must be greater than the alternative i.e. wE ≥ 0 and w D ≥ 0. The firm seeks to maximize yij − wi , which happens at wE = 0 and w D = p. Answer 2.6 The price of the good is determined by P ( Q ) = a − q1 − q2 − q3 The profit earned by a firm is given by πi = ( p − c) · qi . For Firm 2, for example π2 = ( a − q1∗ − q2 − q3 − c) · q2 Dynamic Games of Complete Information 16 which is maximized at dπ2 = ( a − q1∗ − q2 − q3 − c) + q2 (−1) = 0 dq2 ⇒ q2 = a − q1∗ − q3 − c 2 Symmetrically, q3 = a − q1∗ − q2 − c 2 Putting these two together, q2 = a − q1∗ − a−q1∗ −q2 −c 2 −c 2 ⇒ q2 = a − c − q1∗ 3 Which, symmetrically, is equal to q3 . a − c − q1 − c ) · q1 = ∴ π1 = ( a − q1 − q2 − q3 − c ) · q1 = ( a − q1 − 2 · 3 This is maximized at dπ1 a − q1 − c − q1 a−c = + = 0 ⇒ q1∗ = dq1 3 3 2 Plugging this into the previous equations, we get q2 = q3 = a−c 6 Answer 2.7 The profit earned by firm i is πi = ( p − w ) · Li Which is maximized at dπi d ( p − w ) · Li d ( a − Q − w ) Li = = =0 dLi dLi dLi ⇒ d ( a − L1 − . . . − L i − . . . − L n − w ) · L i =0 dLi ⇒ ( a − L1 − . . . − Li − . . . − Ln − w) + Li (−1) = 0 ⇒ L1 + . . . + 2Li + . . . + Ln = a − w ∀ i = 1, . . . , n This generates a system of equation similar to the system in Question (1.4) 2 1 ... 1 L1 a−w 1 2 . . . 1 L2 a − w · = .. .. .. .. .. . . . . . . . . 1 1 ... 2 Ln a−w a − q1 − c 3 · q1 17 Which resolves to Li = a−w n+1 Thus, total labor demand is given by L = L1 + L2 + . . . + L n = a−w a−w n +...+ = · ( a − w) n+1 n+1 n+1 The labor union aims to maximize U = (w − wa ) · L = (w − wa ) · n n · ( a − w) = · ( aw − awa − w2 + wwa ) n+1 n+1 The union maximize U by setting w. This is maximized at dU n a + wa = ( a − 2w + wa ) · =0⇒w= dw n+1 2 The sub-game perfect equilibrium is Li = an−+w1 and w = a+2wa . Although the wage doesn’t change with n, the union’s utility (U) n is an increasing function of n+ 1 , which is an increasing function of n. This is so because the more firms there are in the market, the greater the quantity produced: more workers are hired to produce this larger quantity, increasing employment and the utility of the labor union. Answer 2.8 Answer 2.9 From section 2.2.C, we know that exports from country i can be driven to 0 iff ei∗ = a − c − 2t j a−c = 0 ⇒ tj = 3 2 which, symmetrically, is equal to ti . Note that in this model c = wi and we will only be using wi from now, for simplicity. What happens to domestic sales? ∴ hi = a − wi + a − wi + t i = 3 3 a − wi 2 a − wi 2 = Which is the monopoly amount. Now, in the monopoly-union bargaining model, the quantity produced equals the labor demanded. Thus, Lie=0 = a − wi 2 The profit earned by firm i in this case is given by πie=0 = ( pi − wi )hi = ( a − wi − hi )hi = a − wi − a − wi 2 a − wi 2 The union function’s payoff is defined by U e =0 = ( wi − w a ) L i = ( wi − w a ) a − wi 2 = wi a − wa − wi2 + wi wa 2 = a − wi 2 2 Dynamic Games of Complete Information 18 The union sets wages to maximize utility with the following condition: dU e=0 a + wa a − 2wi + wa = 0 ⇒ wi = = dwi 2 2 Now, tariffs decline to zero. In this situation, t j = 0 and therefore h j = hi = a − wi 3 e j = ei = a − wi 3 and Due to this, prices fall: Pi = Pj = a − Q = a − hi − e j = a − a − wi a − wi a + 2wi − = 3 3 3 So what happens to profits? πit=0 = ( pi − wi )hi + ( p j − wi )e j = ⇒ πit=0 =2 a + 2wi − wi 3 a − wi 3 2 a − wi 3 + a + 2wi − wi 3 = 2 · πie=0 Thus, profits are higher at zero tariffs. What happens to employment? 2 4 a − wi 4 t =0 L i = q i = h i + ei = · ( a − wi ) = · = · Lie=0 3 3 2 3 Employment rises. And what happens to the wage? That depends on the payoff the union now faces: U t =0 = ( wi − w a ) · L i = 2 2 (w − wa )( a − wi ) = · (wi a − wa a − wi2 + wa wi ) 3 i 3 This is maximized at 2 a + wa dU t=0 = ( a − 2wi + wa ) = 0 ⇒ wi = dwi 3 2 Which is the same as before. Answer 2.10 Note that ( P1 , P2 ), ( R1 , R2 ) and (S1 , S2 ) are Nash Equilibrium. P2 Q2 R2 S2 P1 2, 2 x, 0 -1, 0 0,0 Q1 0, x 4, 4 -1, 0 0, 0 R1 0, 0 0, 0 0, 2 0, 0 S2 0, -1 0, -1 -1, -1 2, 0 j So what are player 1’s payoff from playing the strategy? Let POi ( X1 , X2 ) denote player i’s payoff in round j when player 1 plays X1 and a − wi 3 19 player 2 plays X2 . If player 2 doesn’t deviate, he earns the sum of payoffs from two rounds of gaming: PO11 ( Q1 , Q2 ) + PO12 ( P1 , P2 ) = 4 + 2 = 6 And if player 2 deviates from the strategy, player 1 earns: PO11 ( Q1 , P2 ) + PO12 (S1 , S2 ) = 0 + 2 = 2 Now, let’s look at his payoff from deviating, when 2 doesn’t: PO11 ( P1 , Q2 ) + PO12 ( R1 , R2 ) = x + 0 = x And when they both deviate: PO11 ( P1 , P2 ) + PO12 ( P1 , P2 ) = 2 + 2 = 4 Thus, if player 2 deviates, player 1’s best response is to deviate for a payoff of 4 (as opposed to the 2 he’d get when not deviating). If, however, player 2 doesn’t deviate, player 1 gets a payoff of 6 when playing the strategy and x when he doesn’t. Thus, he (player 1) will play the strategy i f f x < 6. A symmetric argument applies to player 2. Thus the condition for which the strategy is a sub-game perfect Nash Equilibrium is 4<x<6 Answer 2.11 The only pure-strategy subgame-perfect Nash Equilibrium in this game are ( T, L) and ( M, C ). L C R T 3,1 0,0 5,0 M 2,1 1,2 3,1 B 1,2 0,1 4,4 Unfortunately, the payoff (4,4) - which comes from the actions (B,R) - cannot be maintained. Player 2 would play R, if player 1 plays B player 1 however would deviate to T to earn a payoff of 5. Consider, however, the following strategy for player 2: • In Round 1, play R. • In Round 2, if Round 1 was (B,R), play L. Else, play M. Player 1’s best response in Round 2 is obviously T or M, depending on what player 2 does. But what should he do in Round 1? If he plays T, his playoff is: 3 PO11 ( T, R) + PO12 ( M, C ) = 5 + 1 = 6 If he plays B: PO11 ( B, R) + PO12 ( T, L) = 4 + 3 = 7 Thus, as long as player 2 is following the strategy given above, we can induce (B,R). If you do not understand the notation, see Answer 2.10 3 Dynamic Games of Complete Information 20 To get a intuitive idea of how we constructed the strategy, note that there are two Nash equilibrium that player 2 can play in the second round: a "reward" equilibrium in which player 1 gets 3 and "punishment" equilibrium in which player 1 gets 1. By (credibly) threatening to punish player 1 in Round 2, player 2 induces "good" behavior in Round 1. Answer 2.12 See text. c Answer 2.13 The monopoly quantity is a− 2 . The monopoly price is, therefore: a−c a+c P = a−Q = q− = 2 2 The players can construct the following strategy. • In Round 1, play pi = a+2 c • In Round t6=1, if the previous Round had p j 6= a+2 c , play pi = c c (the Bertrand equilibrium). Else, play pi = a+ 2 c Now, if player i deviates by charging a− 2 − ε where ε > 0, he secures the entire market and earns monopoly profits for one round and Bertrand profits (which are 0) for all future rounds which totals to a−c a−c ( a − c )2 πdeviate = a − −c · + δ · 0 + δ2 · 0 + . . . = 2 2 4 The payoff from sticking to the strategy (in which both firms proc duce a− 4 ) the payoff is a−c a−c a−c a−c π f ollow = a − −c · +δ· a− −c · +... 2 4 2 4 = 1 1−δ a−c a−c 1 ( a − c )2 · · = · 2 4 1−δ 8 The strategy is stable if πdeviate ≤ π f ollow ( a − c )2 ⇒ ≤ 4 1 1−δ ⇒δ≥ · ( a − c )2 2 1 2 Q.E.D. Answer 2.14 The monopoly quantity when demand is high is a H −c which makes the monopoly price 2 aH − c a +c pH = aH − = H 2 2 21 This is the price that the firms have to maintain when demand is high. Conversely, when demand is low aL + c 2 pL = Let p M be the monopoly price, defined as p if ai = a H H pM = pL if ai = a L Consider the following strategy for firm i: • In Round 1, set pi = p M . • In Round t 6= 1, if p j = p M in the previous round, play p M , else play pi = c The payoff received from deviating is the monopoly profit for one round 4 and then zero profits in all future rounds: πdeviate = See the previous question for the derivation of one round monopoly profits 4 ( a i − c )2 ( a − c )2 + δ · 0 + δ2 · 0 + . . . = i 4 4 If he follows the strategy, he earns: π f ollow ( a H − c )2 ( a L − c )2 ( a i − c )2 +δ· π· + (1 − π ) · +... = 8 8 8 ⇒ π f ollow = ( a i − c )2 δ ( a − c )2 ( a − c )2 + · π· H + (1 − π ) · L 8 1−δ 8 8 The strategy is stable if πdeviate ≤ π f ollow ⇒ ( a i − c )2 ( a − c )2 δ ( a − c )2 ( a − c )2 ≤ i + · π· H + (1 − π ) · L 4 8 1−δ 8 8 c Answer 2.15 If the quantity produced by a monopolist is a− 2 , the quantity produced by a single company in a successful cartel is qm n = a−c 2n Therefore, the profit earned by one of these companies is πm = ( p − c)qm n = (a − Q − c)qm n = a−c a− −c 2 The Cournot oligopoly equilibrium quantity5 is that the profit earned at this equilibrium is πc = ( a − Q − c)qcn = a−n a−c 1+ n a−c 2n 1 = · n which means a−c a−c a−c 2 −c · = 1+n 1+n 1+n A grim trigger strategy for a single company here is • In Round t=1, produce qm n a−c 2 5 2 See Answer 1.4 Dynamic Games of Complete Information 22 • In Round t > 1, if the total quantity produced in t − 1 is n · qm n, m c produce qn , else produce qn . Now, the best response to everyone else producing qm n is deter0 mined by finding q which renders profit a−c n+1 0 0 0 π = ( a − Q − c)q = a − · ( n − 1) − q − c q 0 = · ( a − c ) q 0 − q 02 2n 2n Which is maximized at dπ 0 n+1 a+1 = · ( a − c) − 2q0 = 0 ⇒ q0 = ( a − c) 0 dq 2n 4n The profit at q0 (cheating gain) is π0 = a−c− 2 n+1 n+1 n+1 a−c · ( n − 1) − · ( a − c) · · ( a − c) = · ( a − c) 2n 4n 4n 4n If the firm deviates, it earns the cheating gain for one round and Cournot profits for all future rounds i.e. the gain from deviating from the strategy is πdeviate = n+1 · ( a − c) 4n " ⇒ πdeviate = 2 n+1 4n +δ 2 a−c 1+n δ + 1+δ 2 + δ2 1 1+n a−c 1+n 2 # 2 +... ( a − c )2 If the firm follows the strategy, it’s payoff is πm for all rounds: π f ollow 1 = n a−c 2 2 1 +δ· n a−c 2 2 1 +δ · n 2 a−c 2 2 1 n+1 2 # 1 1 +... = · 1−δ n The strategy is stable if π f ollow ≥ πdeviate 1 1 ⇒ · 1−δ n a−c 2 " 2 ≥ n+1 4n ⇒ δ∗ ≤ 2 δ + 1−δ ( a − c )2 n2 + 2n + 1 n2 + 6n + 1 Thus as n rises, δ∗ falls. If you want to know more, see Rotember and Saloner (1986). Answer 2.16 Answer 2.17 Answer 2.18 See text. Answer 2.19 In a one period game, player 1 would get a payoff of 1 and player 2 would get a payoff of 0 i.e. (1,0). In a two period a−c 2 2 23 game, if the two players can’t agree, the game goes to the second stage, at which point, player 2 gets 1 and player 1 gets 0. This payoff of 1 in the second round is worth δ to player 2 in the first round. If player 1 offers δ to player 2 in the first round, player 2 will accept, getting a payoff of 1 − δ i.e. (1 − δ, δ). In a three period game, if player 2 rejects the offer in the first round, they go on to the second round, at which point it becomes a two period game and player 2 gets a payoff of 1 − δ and player 1 gets δ. This payoff (δ,1 − δ) is worth (δ2 ,δ[1 − δ]) to the players in the first round. Thus, if player 1 makes an offer of (1 − δ[1 − δ],δ[1 − δ])=(1 − δ + δ2 ,δ − δ2 ), player 2 would accept and player 1 would secure a higher payoff. 1 δ Answer 2.20 In round 1, player A offers ( 1− δ , 1+δ ) to player B, δ 1 ∗ which player B accepts since 1+δ ≥ δs = δ 1−δ . What if A deviates from the equilibrium, offers less and B refuses? The game then goes into the next round and B offers A 1+δ δ , 1 1 which will be accepted, leaving 1+ δ for B. This is worth δ · 1+δ to B in the first round (which is why B will refuse anything less than 2 this amount) and 1δ+δ to A in the first round. This is less than the 1 1+δ A would have made had he not deviated. Answer 2.21 Answer 2.22 In the first round, investors can either withdraw (w) or not (d). A strategy can represented as x1 x2 where x1 is what the investor does in the first round and x2 is what the investor does in the second round. The game can be represented by the following table: ww wd dd dw ww r,r r,r D,2r-D D,2r-D wd r,r r,r D,2r-D D,2r-D dd 2r-D,D 2r-D,D R,R D,2R-D dw 2r-D,D 2r-D,D 2R-D,D R,R There are 5 Nash Equilibria: (dw,dw) , (ww,ww) , (ww,wd) , (wd,ww) and (wd,wd). Of these, (ww,wd), (wd,ww) and (wd,wd) are not Subgame Perfect Equilibria since there is no subgame in which both or either player doesn’t withdraw his or her funds in the second round. Answer 2.23 The optimal investment is given by d(v + I − p − I 2 ) 1 = 0 ⇒ I∗ = dI 2 The boost in the value added is If the buyer had played ‘Invest’, buyer will buy if v + I − p − I2 ≥ − I2 ⇒ p ≤ v + I Dynamic Games of Complete Information 24 Thus, the highest possible price that the buyer will pay at this point is p = v + I. If, however, the buyer doesn’t invest, the buyer will buy if v−p ≥ 0 ⇒ v ≥ p Thus the buyer would be willing to pay p = v. Thus investment is I ⊆ {0, 21 }. The price is drawn from p ⊆ {v, v + 12 }. There is no gain from charging anything other than these prices. v+ 1 2 1 2, A 1 2, R − 41 , v + 0, A − 21 , v + v 1 2 1 4, v − 41 , 0 1 2 0, v − 14 , 0 0, R 0, 0 0, 0 As you can see from this (complicated) table, if I = 21 , A weakly dominates R. And if I = 0 R weakly dominates A. Thus we can collapse the above table into a simpler one: 1 2 − 14 , v + 21 (1 − q ) v 0, 0 0, 0 (q)v + ( p) I = 1 2 Accept (1 − p) I = 0Reject 1 4, v The only pure Nash Equilibrium is for the buyer to not invest and the Static Games of Incomplete Information Answer 3.1 See text Answer 3.2 Firm 1 aims to maximize π1 = ( p − c ) q1 = ( a i − c − q1 − q2 ) q1 Which is done by a − c − q2 dπ1 = ai − c − q1 − q2 + q1 (−1) = 0 ⇒ q1 = i dq1 2 Thus, the strategy for firm 1 is a H − c − q2 q1 = a −c2−q 2 L if ai = a H if ai = a L 2 Now, the firm 2 aims to maximize π2 = ( p − c ) q2 = ( a − c − q1 − q2 ) q2 This is maximized at dπ2 a − c − q1 θa + (1 − θ ) a L − c − q1 = a − c − q1 − q2 + q2 (−1) = 0 ⇒ q2 = = H dq2 2 2 Plugging in q1 , we get q2 = θa H + (1 − θ ) a L − c − h θa H +(1−θ ) a L −c−q2 2 i 2 ⇒ q2 = θa H + (1 − θ ) a L − c 3 Now, we need to find out what firm 1’s output would be. If ai = aH , q1H = aH − c − θa H +(1−θ ) a L −c 3 2 = (3 − θ ) a H − (1 − θ ) a L − 2c 6 But what if ai = a L ? q1L = aL − c − θa H +(1−θ ) a L −c 3 2 = (2 + θ ) a L − θa H − 2c 6 Static Games of Incomplete Information 26 Now, based on these results, the constraints for non-negativity are: q2 ≥ 0 ⇒ θa H + (1 − θ ) a L − c ≥ 0 ⇒ θ ≥ c − aL aH − aL Which also requires θ ≤ 1 ⇒ 1 ≥ c − aL ⇒ aL ≥ c − 1 Furthermore, q1L ≥ 0 ⇒ (2 + θ ) a L − θa H − 2c ≥ c − aL ≥ 0 ⇒ θ ≤ 2· 6 aH − aL Which subsumes the last-but-one result. And finally, q1H ≥ 0 ⇒ θ ≤ 2c − 3a H + a L aH − aL Answer 3.3 The profits earned by firm 1 is given by π1 = ( p1 − c)q1 = ( p1 − c)( a − p1 − b1 p2 ) This is maximized at dπ1 a − b1 p2 a − b1 [θ p H + (1 − θ ) p L ] = a − p1 − b1 p2 + p1 (−1) = 0 ⇒ p1 = = dp1 2 2 Now, what if b1 = b H ? To start with, p1 = p H and pH = a − b H [ θ p H + (1 − θ ) p L ] a − (1 − θ ) b H p L ⇒ pH = 2 2 + θb H And if b1 = bL : pL = a − θbL p H a − b L [ θ p H + (1 − θ ) p L ] = 2 2 + b L (1 − θ ) Which means that, a − (1 − θ ) pH = h a−θb L p H 2+(1−θ )b L i 2 + θb H ⇒ pH = a (1 − [1 − θ ] b H ) 4 + 2(1 − θ )bL + 2θb H Similarly, pL = a(1 − θbL ) 4 + 2(1 − θ )bL + 2θb H Answer 3.4 Game 1 is played with 0.5 probability: L R (q)T 1,1 0,0 (1-q)B 0,0 0,0 27 If nature picks game 2, which 0.5 probability: L R (q)T 0,0 0,0 (1-q)B 0,0 2,2 If nature picks game 2, player 1 will always play B, since it weakly dominates T and player 2 will play R, since it weakly dominates L. Now, if nature chooses game 1, player 1 will play T. If nature chooses game 2, player 1 will play B. Furthermore, if player 2 plays L with probability p: 1 1 1 1 1 π2 = p[ · 0 + · 1] + (1 − p)[ · 0 + · 2] = 1 − · p 2 2 2 2 2 This is maximized at p = 0 i.e. player 2 will always play R. Thus the Pure-strategy Bayesian Nash equilibrium is PSNE = {(1, T, R), (2, B, R)} Answer 3.5 Answer 3.6 The payoff is given by vi − bi if bi > b j ∀ j = 1, 2, . . . , i − 1, i + 1, . . . , n ui = vi − b j m if bi = b j 0 if bi < b j for any j = 1, 2, . . . , i − 1, i + 1, . . . , n The beliefs are: v j is uniformly distributed on [0,1]. Actions are given by bi ⊆ [0, 1] and types are given by vi ⊆ [0, 1]. The strategy is given by bi = ai + ci vi . Thus, the aim is to maximize πi = (vi − bi ) · P(bi > b j ∀ j = 1, . . . , i − 1, i + 1, . . . , n) = (vi − bi ) · [P(bi > b j )]n−1 " ⇒ πi = (vi − bi ) · [P(bi > a j + c j v j )] n −1 bi − a j = ( v i − bi ) · P v j < cj This is maximized at dπi = (−1) · dbi bi − a j cj bi − a j cj ! n −1 + ( v i − bi ) ! n −2 · n−1 cj ! a j + (n − 1)vi − nbi cj bi − a j cj ! =0 This requires that either bi − a j = 0 ⇒ bi = a j cj Or that a j + (n − 1)vi − nbi aj n−1 = 0 ⇒ bi = + · vi cj n n ! n −2 =0 !#n−1 = ( v i − bi ) bi − a j cj ! n −1 Static Games of Incomplete Information 28 Now, we know that bi = ai + ci vi . Here, we know that ci = ai = aj a a an ⇒ a1 = 2 = 3 = . . . = n n n n Which is only possible if a1 = a2 = . . . = an = 0. Thus, bi = Answer 3.7 Answer 3.8 n−1 · vi n n −1 n and Dynamic Games of Incomplete Information Answer 4.1 (a) (q)L’ (1-q)R’ L 4,1 0,0 M 3,0 0,1 R 2,2 2,2 The Nash Equilibria are ( L, L0 ), ( R, R0 ) and they are both sub-game perfect equilibria. Now, the payoff to player 2 from playing L0 is π2 ( L 0 ) = 1 · p + 0 · (1 − p ) = p The payoff from playing R0 π2 ( R 0 ) = 0 · p + 1 · (1 − p ) = 1 − p Player 2 will always play L0 if π2 ( L 0 ) > π2 ( R 0 ) ⇒ p > 1 − p ⇒ p > 1 2 The playoff to player 1 from playing L is π1 ( L) = 4 · q + 0 · (1 − q) = 4q And the payoff from playing M is π1 ( M ) = 3 · q + 0 · (1 − q) = 3q Player 1 will always play L if π1 ( L) > π1 ( M ) ⇒ 4q > 3q Which is true. Thus p = 1. In which case, player 2 will always play L0 . Thus the outcome ( R, R0 ) violates Requirements 1 and 2. (b) L0 M0 R0 L 1,3 1,2 4,0 M 4,0 0,2 3,3 R 2,4 2,4 2,4 Dynamic Games of Incomplete Information 30 The expected values of the payoffs to player 2 are: π2 ( L0 ) = 3 · p + 0 · (1 − p) = 3p π2 ( M 0 ) = 2 · p + 2 · (1 − p ) = 2 π2 ( R0 ) = 0 · p + 3 · (1 − p) = 3 − 3p And the payoffs to player 1 are: π1 ( R ) = 2 π1 ( L) = 1 · q1 + 1 · q2 + 4 · (1 − q1 − q2 ) = 4 − 3q1 − 3q2 π1 ( M) = 4 · q1 + 0 · q2 + 3 · (1 − q1 − q2 ) = 3 + q1 − 3q2 The only Nash Equilibrium is ( R, M0 ); it is also sub-game perfect. To be a Perfect Bayesian Equilibrium, player 2 must believe that π2 ( M0 ) > π2 ( L0 ) ⇒ 2 > 3p ⇒ 2 >p 3 and π2 ( M0 ) > π2 ( R0 ) ⇒ 2 > 3 − 3p ⇒ p > 1 3 Furthermore, player 1 must believe π1 ( R) > π1 ( L) ⇒ 2 > 4 − 3q1 − 3q2 ⇒ q1 > 2 − q2 3 Since q1 > 0, this implies that 2 2 − q2 > 0 ⇒ > q2 3 3 and π1 ( R) > π1 ( M ) ⇒ 2 > 3 + q1 − 3q2 ⇒ 3q2 − 1 > q1 Which, in turn, requires 3q2 − 1 > 2 5 − q2 ⇒ q2 > 3 12 The pure Bayesian Nash equilibrium is 1 2 2 2 5 ( R, M), > p > , 3q2 − 1 > q1 > − q2 , > q2 > 3 3 3 3 12 Answer 4.2 (q) L0 (1 − q ) R 0 ( p) L 3,0 0,1 (1 − p ) M 0,1 3,0 R 2,2 2,2 31 As you can see, there is no Pure Nash Equilibrium. But, we need rigorous proof: A pure strategy Nash Equilibrium exists if (a) Player 1 always picks either L or M. For example, player 1 will always play L if π1 ( L ) > π1 ( M ) ⇒ 3 · q + 0 · (1 − q ) > 0 · q + 3 · (1 − q ) ⇒ q > 1 2 Thus if q > 0.5, p = 1. (b) Player 2 always picks either L0 or R0 ; player 2 will always play 0 L if π2 ( L 0 ) > π2 ( R 0 ) ⇒ 0 · p + 1 · (1 − p ) > 1 · p + 0 · (1 − p ) ⇒ 1 >p 2 Thus, if p < 0.5, q = 1. This violates the condition we uncovered in part (a), proving that there is no PSNE. In a mixed strategy BE, player 1 plays L with probability p and player 2 plays L0 with probability q. In equilibrium, player 2 is indifferent between L0 and R0 : 1 2 π2 ( L 0 ) = π2 ( R 0 ) ⇒ 0 · p + 1 · (1 − p ) = 1 · p + 0 · (1 − p ) ⇒ p = And similarly, for player 1: π1 ( L ) = π1 ( M ) ⇒ 3 · q + 0 · (1 − q ) = 0 · q + 3 · (1 − q ) ⇒ q = 1 2 Thus, in a mixed strategy equilibrium, player 1 plays R with p = 0.5 and player 2 plays L0 with probability q = 0.5. Answer 4.3 (a) Let’s start with the pooling equilibrium ( R, R). In this situation, p = 0.5. Now, the payoff to the receiver is π R ( R, u) = 0.5 · (1) + 0.5 · (0) = 0.5 π R ( R, d) = 0.5 · (0) + 0.5 · (2) = 1 Thus, if the sender plays R, the receiver will play d. We have to test two strategies for the receiver: (u, d) and(d, d). Under the strategy (d, d) π1 ( L, d) = 2 and π1 ( R, d) = 3 π2 ( L, d) = 3 and π2 ( R, d) = 2 There is no incentive for type 1 to deviate and play L, but there is an incentive for type 2 to do so. Under the strategy (u, d), π1 ( L, u) = 1 and π1 ( R, d) = 3 π2 ( L, u) = 0 and π2 ( R, d) = 2 Neither type 1 nor type 2 have any reason to L instead of R. Thus, we have the following pooling equilibrium: [( R, R), (u, d), p = 0.5, 1 ≥ q ≥ 0] Dynamic Games of Incomplete Information 32 (b) We must find a pooling equilibrium in which the sender plays ( L, L, L). For the receiver, the payoffs are π R ( L, u) = 1 1 1 ·1+ ·1+ ·1 = 1 3 3 3 π R ( L, d) = 1 1 1 ·0+ ·0+ ·0 = 0 3 3 3 There are two strategies: (u, u) and (u, d). Under (u, u): π1 ( L, u) = 1 and π1 ( R, u) = 0 π2 ( L, u) = 2 and π2 ( R, u) = 1 π3 ( L, u) = 1 and π3 ( R, u) = 0 None of the three types have an incentive to send R instead of L. Thus, we have the following equilibrium: 1 [( L, L, L), (u, u), p = , 1 ≥ q ≥ 0] 3 Answer 4.4 (a) Let’s examine pooling equilibrium ( L, L). p = 0.5. π R ( L, u) = π R ( L, d). Thus, it doesn’t matter for the receiver whether he/she plays u or d. • Under (u, u), π1 ( L, u) = 1 < π1 ( R, u) = 2, making it unsustainable. • Under (u, d), π2 ( L, u) = 0 < π2 ( R, d) = 1, making it unsustainable. • Under (d, d), π2 ( L, d) = 0 < π2 ( R, d) = 1, making it unsustainable. • Under (d, u), π2 ( L, u) = 0 < π2 ( R, d) = 1, making it unsustainable. Thus, ( L, L) is not a sustainable equilibrium. Let’s examine separating equilibrium ( L, R). The best response to this is (u, d)6 . Let’s see if either of the types have an incentive to deviate: • For type 1, π1 ( L, u) = 1 > π1 ( R, d) = 0 i.e. no reason to play R instead of L. • For type 2, π2 ( L, u) = 0 < π2 ( R, d) = 1 i.e. no reason to play L instead of R. Let’s examine pooling equilibrium ( R, R). π R ( R, u) = 1 > π R ( R, d) = 0.5. Therefore, the two strategies that can be followed by the receiver are (u, u) and (d, u). • Under (u, u), π ( L, u) < π ( R, u) for both types. • Under (d, u), π ( L, d) ≤ π ( R, u) for both types. Let’s examine pooling equilibrium ( R, L). The best response to this is (d, u)7 . • For type 1, π1 ( L, d) = 2 ≥ π1 ( R, u) = 2, i.e. will play L. • For type 2, π2 ( L, d) = 0 ≤ π2 ( R, u) = 1, i.e. will play R. Since π R (1, L, u) > π R (1, L, d) and π R (2, R, u) > π R (2, R, d) 6 Since π R (1, R, u) = 2 > π R (1, R, d) = 0 and π R (2, L, u) = 0 < π R (2, L, d) = 1 7 33 Thus the perfect Bayesian equilibrium are: [( L, R), (u, d), p, q] [( R, R), (u, u), p, q = 0.5] [( R, R), (d, u), p, q = 0.5] [( R, L), (d, u), p, q] (b) Let’s examine pooling equilibrium is ( L, L). π R ( L, u) = 1.5 > π R ( L, d) = 1, therefore player 2 will respond to L with u. The two strategies are (u, u) and (u, d). • Under (u, u), π ( L, u) > π ( R, u) for both types. • Under (u, d), π1 ( L, u) = 3 < π1 ( R, d) = 4, making it unsustainable. Let’s examine separating equilibrium ( L, R). The best response to this is (d, u). • For type 1, π1 ( L, d) = 1 > π1 ( R, u) = 0 i.e. type 1 will play L. • For type 2, π2 ( L, d) = 0 < π2 ( R, u) = 1 i.e. type 2 will play R. Let’s examine pooling equilibrium ( R, R). π R ( R, u) = 1 > π R ( L, d) = 0.5, therefore player 2 will respond to R with u. The two strategies are (u, u) and (d, u). • Under (u, u), π1 ( L, u) = 3 > π1 ( R, u) = 0, making ( R, R) unsustainable. • Under (d, u), π1 ( L, d) = 1 > π1 ( R, u) = 0, making ( R, R) unsustainable. Let’s examine separating equilibrium ( R, L). The best response to this is (d, u). • For type 1, π1 ( L, d) = 1 > π1 ( R, u) = 0, making ( R, L) unsustainable. • For type 2, π2 ( L, d) = 0 < π2 ( R, u) = 1, which doesn’t conflict with the equilibrium. The perfect Bayesian Equilibria are [( L, L), (u, u), p = 0.5, q] [( L, R), (d, u), p = 1, q = 0] Answer 4.5 Let’s examine 4.3(a). We’ve already tested equilibrium ( R, R). Let’s try another pooling equilibrium ( L, L). q = 0.5. π R ( L, u) = 1 > π R ( L, d) = 0.5. Thus, the receiver’s response to L will always be u. We have to test two strategies for the receiver: (d, u) and(u, u). • Under the strategy (d, u), π ( L, d) = 2 > π ( R, u) = 0 i.e. there is no incentive for either type to play R instead of L. • Under the strategy (u, u), π2 ( L, u) = 0 < π2 ( R, u) = 1, making ( L, L) unsustainable. 34 Dynamic Games of Incomplete Information Let’s examine ( L, R). The best response to this is (u, d). In response to this, • For type 1, π1 ( L, u) = 1 < π1 ( R, d) = 3 i.e. type 1 will play R which violates the equilibrium • For type 2, π2 ( L, u) = 0 < π2 ( R, d) = 2 i.e. type 2 will play R which doesn’t violate the equilibrium. Let’s examine ( R, L). The best response to this is (u, d). In response to this, • For type 1, π1 ( L, u) = 1 < π1 ( R, d) = 3, i.e. type 1 will play R. • For type 2, π2 ( L, u) = 0 < π2 ( R, d) = 2, i.e. type 2 will play R, violating the equilibrium. The perfect Bayesian Equilibrium is [( L, L), (d, u), p, q = 0.5] Now, let’s examine 4.3(b). There is one pooling equilibrium other than the ( L, L, L): ( R, R, R). There are six pooling equilibria: 1.( L, L, R), 2.( L, R, L), 3.( R, L, L), 4.( L, R, R), 5.( R, L, R) and 6.( R, R, L). Let’s start with pooling equilibrium ( R, R, R). π R ( R, u) = 23 > π R ( R, d) = 13 . Thus, receiver will play the strategy (u, u) or (d, u). • For strategy (u, u), π ( L, u) < π ( R, u) for all types, making it unsustainable. • For strategy (d, u), π1 ( L, d) < π1 ( R, u), making the equilibrium unsustainable. Let’s examine the various separating the equilibrium. 1. ( L, L, R). The best response to this is (u, d), 8 • For type 1, πS (1, L, u) = 1 > πS (1, R, d) = 0 i.e. type 1 will play L. • For type 2, πS (2, L, u) = 2 > πS (2, R, d) = 1 i.e. type 2 will play L. • For type 3, πS (3, L, u) = 1 < πS (3, R, d) = 2 i.e. type 3 will play R. Thus, this is a viable equilibrium. 2. ( L, R, L). The best response to this is (u, u), 9 • For type 1, πS (1, L, u) = 1 > πS (1, R, u) = 0 i.e. type 1 will play L. • For type 2, πS (2, L, u) = 2 > πS (2, R, u) = 1 i.e. type 2 will play L, instead of R. • For type 3, πS (3, L, u) = 1 > πS (3, R, u) = 0 i.e. type 3 will play L. Thus, this is not a viable equilibrium. 3. ( R, L, L). The best response to this is (u, u), 10 • For type 1, πS (1, L, u) = 1 > πS (1, R, u) = 0 i.e. type 1 will play L, instead of R. • For type 2, πS (2, L, u) = 2 > πS (2, R, u) = 1 i.e. type 2 will play L. • For type 3, πS (3, L, u) = 1 > πS (3, R, u) = 0 i.e. type 3 will play L. Thus, this is a not viable equilibrium. 8 Since π R (3, R, u) = 0 < π R (3, R, d) = 1 and 0.5 · π R (1, L, u) + 0.5 · π R (2, L, u) = 1 > 0.5 · π R (1, L, d) + 0.5 · π R (1, L, d) = 0 since π R (2, R, u) = 1 > π R (2, R, d) = 0 and 0.5 · π R (1, L, u) + 0.5 · π R (3, L, u) = 1 > 0.5 · π R (1, L, d) + 0.5 · π R (3, L, d) = 0 9 since π R (1, R, u) = 1 < π R (1, R, d) = 0 and 0.5 · π R (2, L, u) + 0.5 · π R (3, L, u) = 1 > 0.5 · π R (2, L, d) + 0.5 · π R (3, L, d) = 0 10 35 4. ( L, R, R). The best response to this is (u, u) and (u, d) 11 Let’s test (u, u): • For type 1, πS (1, L, u) = 1 > πS (1, R, u) = 0 i.e. type 1 will play since π R (1, L, u) = 1 > π R (1, L, d) = 0 and 0.5 · π R (2, R, u) + 0.5 · π R (3, R, u) = 0.5 = 0.5 · π R (2, R, d) + 0.5 · π R (3, R, d) = 0.5 11 L. • For type 2, πS (2, L, u) = 2 > πS (2, R, u) = 1 i.e. type 2 will play L, instead of R. • For type 3, πS (3, L, u) = 1 > πS (3, R, u) = 0 i.e. type 3 will play L, instead of R. Thus, this is not a viable equilibrium. Let’s test (u, d) • For type 1, πS (1, L, u) = 1 > πS (1, R, d) = 0 i.e. type 1 will play L. • For type 2, πS (2, L, u) = 2 > πS (2, R, d) = 1 i.e. type 2 will play L, instead of R. • For type 3, πS (3, L, u) = 1 < πS (3, R, d) = 2 i.e. type 3 will play R. Thus, this is not a viable equilibrium. 5. ( R, L, R). The best response to this is either (u, u) or (u, d) 12 Let’s test (u, u). • For type 1, πS (1, L, u) = 1 > πS (1, R, u) = 0 i.e. type 1 will play L, instead of R. • For type 2, πS (2, L, u) = 2 > πS (2, R, u) = 1 i.e. type 2 will play L. • For type 3, πS (3, L, u) = 1 > πS (3, R, u) = 0 i.e. type 3 will play L, instead of R. Thus, this is not a viable equilibrium. Let’s test (u, d) • For type 1, πS (1, L, u) = 1 > πS (1, R, d) = 0 i.e. type 1 will play L, instead of R. • For type 2, πS (2, L, u) = 2 > πS (2, R, d) = 0 i.e. type 2 will play L. • For type 3, πS (3, L, u) = 1 = πS (3, R, d) = 1 i.e. type 3 can play R. Thus, this is not a viable equilibrium. 6. ( R, R, L). The best response to this is (u, u), 13 • For type 1, πS (1, L, u) = 1 > πS (1, R, u) = 0 i.e. type 1 will play L, instead of R • For type 2, πS (2, L, u) = 2 > πS (2, R, u) = 1 i.e. type 2 will play L, instead of R • For type 3, πS (3, L, u) = 1 > πS (3, R, u) = 0 i.e. type 3 will play L. Thus, this is not a viable equilibrium. The only other perfect Bayesian Equilibrium is 1 1 ( L, L, R), (u, d), p, q0 = , q1 = 3 3 12 since π R (2, L, u) = 1 > π R (2, L, d) = 0 and 0.5 · π R (2, L, u) + 0.5 · π R (3, L, u) = 0.5 = 0.5 · π R (2, L, d) + 0.5 · π R (3, L, d) = 0.5 13 since π R (3, L, u) = 1 > π R (3, L, d) = 0 and 0.5 · π R (1, R, u) + 0.5 · π R (2, R, u) = 0.5 > 0.5 · π R (1, R, d) + 0.5 · π R (2, R, d) = 0 Answer 4.6 Type 2 will always play R since πS (2, R, a) > πS (2, R, u) and πS (2, R, a) > πS (2, R, d). Thus if the Receiver gets the message L, he knows that 14 it can only be type 1. In such a case, the Receiver plays u14 , creating in fact, π R ( x, L, u) > π R ( x, L, d) for both types, so the Receiver will always play u 36 Dynamic Games of Incomplete Information a payoff of (2, 1). This gives type 1 a higher payoff than if he played R, which would have given him a payoff of 1. Thus, the perfect Bayesian Equilibrium is [( L, R), (u, a), p = 1, q = 0]