Page 1 of 6 ECMC02H Prof. G. Cleveland Problem Set for Game Theory 1. The game of “chicken” is played by two macho teenagers (with more money and spare time than brains) who speed towards each other on a single-lane road. The one who swerves to avoid a collision is labelled a chicken, whereas the one who does not swerve is considered to be fearless and admirable. If neither swerves, both will be hurt (perhaps die) but if both swerve, both participants lose some face. Payoffs to the chicken game are shown below: Person #2 Person #1 Swerve Do not swerve Swerve (1,1) (0,10) Do not swerve (10, 0) (-100, -100) (a) Does this game have a Nash equilibrium? What is it and why? (b) Is a threat by either player not to swerve a credible threat? Why or why not? (c) Would the ability of one player to firmly commit to a no-chicken strategy (for example, by locking the steering wheel into the straight-ahead position and throwing away the key) be desirable for that player? 2. Two firms are considering bringing out competing brands of cigarettes with a special ingredient. Payoffs to the companies are given below: Firm #2 Produce Firm #1 Do not produce Produce (3, 3) (5, 4) Do not produce (4, 5) (2, 2) (a) Does this game have a Nash equilibrium? What is it and why? (b) If one of these players could move first (in a sequential game), would there be any first-mover advantages for either player? Explain (c) Might Firm #2 find it in its interests to bribe Firm #1 to stay out of the market? 3. The Wave Energy Technology company has a monopoly on the production of a particular Page 2 of 6 type of highly-desirable waterbed. Demand for these beds is relatively inelastic…at a price of $1000 per bed, 25,000 will be sold while at a price of $600, 30,000 will be sold. The only costs associated with waterbed production are the initial costs of building a plant. This company has already invested in a plant capable of producing up to 25,000 beds and this is a sunk cost (and therefore irrelevant for current pricing decisions). (a) suppose a would-be entrant to this industry could always be assured of getting half the market, but would have to invest $10 million in a plant to accomplish this. Construct a sequential game showing the payoffs to this firm from the entry/no entry decision. What is the Nash equilibrium and why? (b) Suppose the Wave Energy Technology firm could invest $5 million more to enlarge its existing plant so that it would produce 40,000 beds (another sunk cost). Would this strategy be a profitable way to deter entry by its rival? Explain. 4. Assume now that mixed strategies are possible – not just pure strategies. Take a simple “zero-sum” game. In zero-sum games, the payoffs to both players are assumed to sum to zero, so that one player's gain is inevitably another player's loss. These games correspond to most of the sports which we play and/or watch, since generally one side either wins or loses (or there is a tie). Thus, if we designate the outcome as +1 for a win, 0 for a tie, and -1 for a loss, you can see that two players must always have their results "sum to zero". These games do not allow for collusion, since there is no possible agreement that will make both parties better off. a) Begin with a simple game of "matching" in which two players simultaneously put out either one or two fingers. One of the players - Jon - has chosen "evens" in advance; the other player - Sue - has chosen "odds." If the number of fingers put out by the two players matches, then Jon wins; if they do not match, then Sue wins. Set out the payoff matrix for this game. Is there a Nash equilibrium in pure strategies? b) Now suppose that we allow each player to choose a mixed strategy. In a mixed strategy, the player chooses to play randomly, with a probability assigned to each "pure" strategy. Surprisingly, this is often the best alternative. Suppose that Jon plays the mixed strategy of putting out one finger with probability PJ and two fingers with probability (1-PJ). Show that if Sue can figure out this strategy and respond with the best possible strategy, that Jon's maximin will occur at PJ = 0.5. Then show that if the positions are reversed and Sue chooses PS (defined the same way for her as PJ is for Jon), that her best choice is PS = 0.5. Explain the "common sense" of this result. 5. Suppose that in a football game, the offense can choose one of two strategies - run or pass while the defense can choose one of two strategies - run defense or pass defense. The results of the play are as follows: if the offense runs, it gains 3 yards against a run defense and 9 yards against a pass defense; if the offense passes, it gains 15 yards against a run defense and 0 yards against a pass defense. Show that a mixed strategy is best for both sides. Explain the common sense of this result. Page 3 of 6 Solutions 1(a). There are actually two Nash equilibria; in each of these equilibrium positions, one player swerves and the other does not. Consider player #1’s incentives in the case that player #2 chooses swerve. His incentives are to not swerve and get the payoff of 10. Consider player #1’s incentives in the case that player #2 chooses to not swerve. Now player #1’s incentive is to swerve in order to get the payoff of 0, rather than minus 100. We can use the same logic to analyze player #2’s incentives. In the case where player #1 chooses to swerve, player #2 has the incentive to not swerve, to get the payoff of 10. However, if player #1 chooses to not swerve, then player #2 has the incentive to swerve, to get the payoff of 0, rather than minus 100. Either of these two equilibria is stable. (b) We would expect these players to use any possible threat to try to convince the other player that he (i.e., the one making the threat) is definitely not going to swerve. This might involve a lot of bragging and looking very tough, and not showing the least doubt. However, the threat is not credible. Each player has enormous incentives to avoid ending up in the southeast cell of the payoff matrix (do not swerve and do not swerve). This will affect the behaviour of the “tough guy” no matter what he threatens before the race. (c) Locking the steering wheel into position (as long as it kept the car going straight) would be very effective, if and only if the other player knew about this locking. If player #2 knows that player #1 has a locked steering wheel, then player #2 knows that player #1 cannot swerve. So, player #2 will have to decide to swerve, (unless he is crazy…). Therefore, player #1 will end with a payoff of 10 and player #2 will end up with zero. 2(a). Just like the previous question, there are two Nash equilibria; in both of them one firm will produce and the other firm will not produce. Firm #1 has the incentive not to produce if firm #2 produces, but it has the incentive to produce if firm #2 does not produce. In similar fashion, firm #2 is better off not producing if firm #1 chooses to produce the cigarettes and firm #2 is better off producing the cigarettes if firm #1 decides not to produce. (b) Consider a sequential game with player #1 moving first. Prod 3, 3 Don’t Player #1 5, 4 #2 Prod 4, 5 Player #1 knows that if it decides to produce this brand of cigarettes, then Firm #2 will not Don’t produce. Player #1 can also see that if it decides not to produce cigarettes, then Firm #2 will 2, 2 decide to produce cigarettes. In the former case, Firm #1’s payoff will be 5; in the latter case, Page 4 of 6 Firm #1’s payoff will be 4. Therefore, Firm #1 will choose the former move…it will decide to produce cigarettes. The Nash equilibrium will be that Firm #1 produces cigarettes and Firm #2 does not. In other words, there clearly would be first-mover advantages for Firm #1 in order to determine that the most favourable of the Nash equilibria from the simultaneous game is the one which prevails in a sequential game. If Firm #2 were the first mover, a similar set of advantages would come to Firm #2. 2(c). If Firm #2 were able to bribe Firm #1 to stay out of the cigarette market, it could get a payoff of 5 rather than a payoff of 4. Any bribe less than 1 which was successful would be worth it from Firm #2’s point of view. However, since Firm #1 loses 1 unit by not producing this brand of cigarettes, it is unlikely to accept the bribe. 3. The diagram below shows the sequential decision of Firm #1 (the potential entrant) and Firm #2 (the incumbent monopolist). If Firm #1 enters, then the incumbent firm has the choice of producing at a price of $1000 (and gaining $12.5 million of revenue) or $600 (and gaining $9 million). The incumbent will prefer the $1000 price. If the potential entrant does not enter, it earns nothing and the incumbent firm gains either $25 million or $15 million. The entrant will enter and the incumbent will take the path of least resistance and share the market at the most profitable price of $1000. $1000 Player #1 $600 2.5, 12.5 12.512.5 3, 3 -1, 9 #2 $1000 0, 25 $600 0, 15 (c) If the WET company were to expand its existing plant, this would not affect its incentives in the case of entry. Its incentives would still be to sell at the price of $1000, so any threat to expand its output, and price at the lower price, would not be credible. 4. The matching game is summarized in the zero sum payoff matrix below (where the payoffs to Jon are shown first): Page 5 of 6 Sue One Finger Jon Two Fingers One Finger +1, -1 -1, +1 Two Fingers -1, +1 +1, -1 There is no Nash equilibrium; this accords with our intuition. Each player will refuse to continue with a simple strategy if the other person is winning, so no simple strategy can be a stable result. Suppose instead that Jon plays the mixed strategy of putting out one finger with probability PJ and two fingers with probability (1-PJ). Then Sue's payoff will be R1 = (-1)(PJ) + (+1)(1-PJ) = 1 - 2PJ if she puts out one finger, and R2 = (+1)(PJ) + (-1)(1-PJ) = 2PJ - 1 if she puts out two fingers. If PJ < 0.5, then Sue obtains a positive result by always putting out one finger; if PJ > 0.5, then Sue obtains a positive result by always putting out two fingers. Only if Jon chooses PJ = 0.5 will Sue end up with a zero result (which also means a zero result for Jon, since this is a zero sum game). The argument for Sue is identical, but it is worth going through it. Suppose that Sue plays the mixed strategy of putting out one finger with probability PS and two fingers with probability (1PS). Then Jon's payoff will be R1 = (+1)(PS) + (-1)(1-PS) = 2PS - 1 if he puts out one finger, and R2 = (-1)(PS) + (+1)(1-PS) = 1 - 2PS if he puts out two fingers. If PS < 0.5, then Jon obtains a positive result by always putting out two fingers; if PS > 0.5, then Jon obtains a positive result by always putting out one finger. Only if Sue chooses PS = 0.5 will Jon end up with a zero result (which also means a zero result for Sue). The common sense of this is obvious. In the game of matching, if you can pick up any pattern in the other player's choices, then you can exploit that pattern. If you are playing "evens" and, at any point, you figure out that the other player is "likely" to put out one finger (meaning that the probability of putting out one finger is greater than 0.5), then of course you put out one finger so as to win most of the time. The only way to avoid this is for each player to employ a completely random strategy. 5. Consider first the decisions by the defense. If the defenders defend against the run, you always pass; if the defenders defend against the pass, you always run. Of course, no football team would behave so stupidly on defense (that is, no one would reveal exactly how they are going to play). Suppose now that the defense chooses the run defense with probability PRD and thus chooses the pass defense with probability 1 - PRD (of course, the two probabilities must add to one). Then, if the offense runs, the expected gain is Page 6 of 6 GR = (+3)(PRD) + (+9)(1-PRD) = 9 - 6PRD If the offense passes, the expected gain is GP = (+15)(PRD) + (+0)(1-PRD) = 15PRD The two results are equal (that is, GR=GP) if 9 - 6PRD = 15PRD or PRD = 3/7 When PRD = 3/7, GR = GP = 6.43 The defense will always choose PRD = 3/7 because if they do not, the offense can do better than 6.43 (assuming that the offense is only interested in the expected result - for football fans, you know that there are short yardage situations and end of game situations that dictate different strategies!). If the defense chooses PRD > 3/7, then the offense will always pass because GP > GR, and this will hurt the defense because GP > 6.43 On the other hand, if the defense chooses PRD < 3/7, then the offense will always run because GR > GP, and this will hurt the defense because GR > 6.43. Thus the best that the defense can do is to choose PRD = 3/7. Now consider the decisions by the offense. If the offense runs, you always defend against the run; if the offense passes, you always defend against the pass. Again, no football team would behave so stupidly on offense. Suppose now that the offense the run with probability PR and thus chooses the pass with probability 1 - PR Then, if the defense chooses the run defense, the expected gain is GRD = (+3)(PR) + (+15)(1-PRD) = 15 - 12PRD If the defense chooses the pass defense, the expected gain is GPD = (+9)(PR) + (+0)(1-PR) = 9PR The two results are equal (that is, GRD=GPD) if 15 - 12PR = 9PR or PR = 5/7 When PR = 5/7, GRD = GPD = 6.43 The offense will always choose PR = 5/7 because if they do not, the defense can do better than 6.43 If the offense chooses PR > 5/7, then the defense will always defend against the run because GRD < GPD, and this will hurt the offense because GRD < 6.43 On the other hand, if the defense chooses PR < 5/7, then the defense will always defend against the pass because GRD > GPD, and this will hurt the offense because GPD < 6.43. Thus the best that the offense can do is to choose PR = 5/7. The common sense of the mixed strategy is that it will generally hurt you in any sport to be too predictable!