Digitized by the Internet Archive in 2011 with funding from Boston Library Consortium Member Libraries http://www.archive.org/details/existenceofsubgaOOharr HB31 .M415 no. S'U f economics 1991 THE EXISTENCE OF SUBGAME- PERFECT EQUILIBRIUM IN GAMES WITH SIMULTANEOUS MOVES Christopher Harris Number 570 December 1990 massachusetts institute of technology - 50 memorial drive Cambridge, mass. 02139 THE EXISTENCE OF SUBGAME- PERFECT EQUILIBRIUM IN GAMES WITH SIMULTANEOUS MOVES Christopher Harris Number 570 December 1990 ft-'.. ' ' M.l.T. LiBi apr o 1 nm fl "-' Iffl The Existence in of Subgame—Perfect Equilibrium Games with Simultaneous Moves: a Case for Extensive—Form Correlation Christopher Harris Nuffield College, Oxford December 1990^ Abstract The starting point of this paper subgame-perfect equilibrium existence would be restored result of the if fails is to exist. a simple, regular, dynamic Examination of this game in which example shows that players were allowed to observe public signals. The main paper shows that allowing the observation of public signals yields existence, not only in the example, but also in an extremely large class of dynamic games. should like to thank Drew Fudenberg, Daniel Maldoom, Andreu Mas— Colell, Eric Maskin, Jean—Francois Mertens, Meg Meyer, Jean Tirole, John Vickers, and participants at seminars and workshops at Oxford, Stonybrook and Harvard for helpful comments on the earlier paper from which the present paper is drawn. The present paper is drawn from an earlier paper entitled "The Existence of Subgame—Perfect Equilibrium with and without Markov Strategies: a Case for Extensive Form Correlation". I 1 Introduction Consider the following example of a dynamic game. Firms set out with exogenously specified capacities (which are known to all). In period one they simultaneously choose investment levels (possibly on a random basis), and are then informed of one another's choices. The result is a change in capacities. In period two firms simultaneously choose production levels (within their capacity constraints), and are then informed of one another's choices. The result simultaneously choose prices. their products, but so is a change in inventories. In period three firms The vector of prices chosen affects the vector of demands for do certain exogenous random factors. Firms are informed of one another's chosen prices and of the final realised demands. second change in inventories.) The (The demands bring about a three period cycle of choices the begins afresh. And so on. This There is game a finite given period: all is an example of a game of the following general type. Time number of active players. There players (both active and passive) periods; the set of actions available to is is also a passive player, Nature. know the outcomes any active player is discrete. In any of all previous compact, and depends continuously on the outcomes of the previous periods; the distribution of Nature's action (which is given exogenously) depends continuously on the outcomes of the previous periods; the players (active and passive) choose their actions simultaneously; and the outcome of the period is simply the vector of actions chosen. The outcome of the game as a whole the (possibly infinite) sequence of outcomes of all periods, is and players' payoffs are bounded and depend continuously on the outcome of the game. This is as regular a class of dynamic games as one could ask for. A counter-example shows, however, that games in this class need not have a subgame—perfect equilibrium. is therefore necessary to extend the equilibrium concept in such a way that existence It is restored. l The problem is reminiscent of that of the non-existence of Nash equilibrium in pure What answer to is the most natural extension of the equilibrium concept? this question is One provided by the following considerations. First, if clue as to the we simplify the class of games under consideration by requiring that players' action sets are always finite, then subgame—perfect equilibrium always appears to relate specifically to the fact that we exists. Hence the non-existence problem allow a continuum of actions. Secondly, suppose instead that we simplify the class of games under consideration by assuming that players' action sets are independent of the way of approximating a large but finite game numbers of is to consider subsets of players' action sets that consist of it is Moreover, closely spaced actions. 2 increasingly fine approximations, and a approximations, then outcomes of previous periods. Then a natural if one takes a sequence of subgame—perfect equilibrium of each of the natural to expect that any limit point of the sequence of equilibrium paths so obtained will be an equilibrium path of the original game. 3 One can, however, find examples in which the limit point involves a special form of extensive—form correlation: at the outset of each period agents observe a public signal. 4 strategies. But there is one important difference: then subgame—perfect equilibrium does if agents' action sets are always finite, exist. 2 Hell wig and Leininger (1986) were the first to introduce such an approximation, in the context of finite—horizon games of perfect information. They proved an upper—semicontinuity result: they showed that any limit point of equilibrium paths of the finite approximations is an equilibrium path of the original game. Their result can, however, be understood in terms of the results of Harris (1985a). Indeed, in that paper it is shown that the point—to—set mapping from period—t subgames of a game of perfect information to period—t equilibrium paths is upper semicontinuous. Hence, in order to obtain their result, one need only introduce a dummy player who chooses n e IN U {qd} at the outset of the game. If n < od then the remaining players will be restricted to actions chosen from the n approximation to their action choose actions from their original action sets. sets. If n = od then they will be free to Convergence of equilibrium paths, as used implicitly in Harris (1985a) and explicitly and Leininger (1986) and Borgers (1989) seems more relevant than the convergence of strategies considered by Fudenberg and Levine (1983) and Harris (1985b). 3 in Hellwig 4 Fudenberg and Tirole (1985) consider two such games. They point out that the obvious discretisations of these games have a unique symmetric subgame—perfect equilibrium, and that the limiting equilibria obtained as the discretisation becomes arbitrarily fine involve correlation. Their games do not, however, yield counterexamples to the existence of subgame—perfect equilibrium. For both games possess asymmetric equilibria which do not involve correlation. These considerations suggest that, at the very least, the equilibrium concept should be extended to allow the observation of public signals. This minimal extension sufficient. For, first, it restores existence. Secondly, with correspondence of a continuous family of games limit, is it, is the equilibrium game whose action sets are an equilibrium path of the game. Thirdly, any limit point of subgame-perfect equilibrium paths of a The also upper semicontinuous. In particular, any point of equilibrium paths of finite approximations to a continua is game is e- an equilibrium path in the extended sense. 5 basic structure of the proof of the existence of subgame-perfect equilibrium in the extended sense is the same as the structure of the proof of the existence of subgame-perfect pure—strategy equilibrium in games of perfect information given in Harris (1985a). It breaks down The backwards program into is two main most horizon T. In such a game, it equilibrium paths of the game. period—T subgames. (The parts, the easily explained in the context of a More set of precisely, one solves first for (The set of is the convex for equilibrium paths of a a set of probability distributions over period—(T— 1) histories, And so reached, and a set of probability distributions over period— 1 histories There is period—T outcomes.) Then one solves a joint distribution over period—(T— 1) and period—T outcomes.) is finite the equilibrium paths of equilibrium paths of a period—T subgame the equilibrium paths of period—(T— 1) subgames. is game with what turn out to be the consists in solving recursively for hull of the set of probability distributions over period—(T— 1) subgame backwards and the forwards programs. on is i.e. until period 1 obtained. of course no general guarantee that the paths obtained in the course of carrying out the backwards program really are equilibrium paths until equilibrium strategies generating them have been constructed. To construct such purpose of the forwards program. The essential idea is this. 5 Its is the Pick any of the period— paths obtained as the final product of the backwards program. distribution over period— 1 histories. strategies Such a path is a probability marginal over period— 1 outcomes can be used to Chakrabarti (1988) claims that subgame-perfect e-equilibria exist. claim, but are unable to vouch for the proof. We believe this construct period— 1 behaviour strategies for the players. Its conditional over period— be followed from period 2 onwards in the event that histories yields continuation paths to players do not deviate from their period— 1 strategies. In order to obtain continuation paths in the event that some player does deviate from his period— 1 strategy, to choose from among the period—2 paths obtained in the course of the some path which minimises the continuation payoff of that continuation paths for all way period—2 subgames are obtained. Period—2 strategies are then obtained from the marginals of these paths over period—2 outcomes. finally sufficient backwards program In this player. it is And so on until period—T strategies are obtained. Implementing this proof does, however, involve two significant new complications. First, for technical reasons, it is necessary to mapping from period—t subgames maintain the induction that the point—to—set to period—t equilibrium paths This follows from a generalisation of the work of Simon and because players may are measurable. 6 randomise, That this is Zame upper semicontinuous. (1990). Secondly, essential to ensure that all the strategies constructed it is requirement can be met emerges as a natural corollary of the construction employed: one can always construct a measurable family of conditional distributions from a measurable family of probability distributions. 7 The organisation to the existence of game with two of the paper subgame—perfect is as follows. equilibrium. players in each period. Section 2 sets out the counterexample This example involves a three—period In this game, each player's action set is compact 6 Hellwig and Leininger (1987) were the first to draw attention to the desirability of ensuring that strategies are measurable. Their approach to the existence of subgame—perfect equilibrium is not, however, necessary in order to obtain measurability. Indeed, if the assumptions of Harris (1985a) are specialised appropriately (by assuming that the embedding spaces for players' action sets are compact metric instead of compact Hausdorff), then it is simple to show that exactly the construction given there can be carried out in a measurable way. Indeed, in the notation of Harris (1985a; p. 625), all that is necessary is to ensure that the functions gj this is possible follows Lemma can be chosen to be measurable; and that from the argument given in the first paragraph of the proof of 4.1 in the present paper. 7 Had we employed a dynamic—programming approach, the problem of ensuring measurability of strategies would have been much harder — see below for further discussion. Indeed, we do not know how to solve the measurable selection problem that arises when such an approach is employed. and independent of the outcomes of previous periods, and each player's payoff function continuous. is Section 3 formulates the basic framework used in this paper for the analysis of dynamic games. Section 4 shows that when the dynamics are continuous, players' payoff functions are continuous, and the framework subgame—perfect equilibrium signals, 4.1 provides a detailed for the exists. is extended to allow the observation of public This is the main result of the paper. Section overview of the proof. Section 4.2 develops the analysis necessary backwards program. Section 4.3 develops the analysis necessary for the forwards program. Section 4.4 exploits the analysis of Sections 4.2 and 4.3 to prove the main showing in particular how to deal with the infinite—horizon develop a perspective on the main result. types of game in and zero-sum games. It is also Section 5 attempts to shown there that there are which the introduction of public signals existence of subgame-perfect equilibrium: all It is case. games of is result, at least three not necessary to obtain the perfect information, finite-action games, shown that the introduction of public signals accommodates the equilibria that one can obtain by taking finite-action approximations to a game, or by considering e—equilibria of a game. signals that is it is. We do not know whether the introduction of public the minimal extension with these two properties, but it seems plausible to argue The Counterexample 2 A counterexample to the existence of subgame—perfect equilibrium in the dynamic games described First of for all, it must have in the introduction at least class of must have a certain minimum complexity. two periods. Otherwise the standard existence theorem Nash equilibrium would apply. Secondly, there must be at least two players in the second period. For with only one player in the second period, the correspondence describing the continuation payoffs available to the player or players in period one would be convex valued. between which he (The player is in period two can randomise over any set of alternatives So the existence result of Simon and Zame (1990) would indifferent.) apply to period one. Thirdly, there must be at least two players in period one. For the continuation—payoff correspondence for period one will be upper semicontinuous, and a an optimum. single player will therefore be able to find The counterexample presented periods, with two players in in this section each period. It example would class of not this minimal. has three For, aside from being in each period. dynamic games considered settle the question as to It would be of considerable interest to find an example with only two periods and two players minimal in relation to the is in this paper, such an whether equilibrium exists in two—stage games or not. The cast of the counter—example, period one two punters two, two greyhounds A and B pick a C and D each their attitude to the race. and They pick their choice variables, are as follows. 6 [0,1] and b € [0,1] receive an injection of size a c e [0,1] and d 6 [0,1] , they complete the course. In period three each of two referees winner, picking e 6 {C,D} and f e respectively. {C,D} respectively. + b , In In period which changes which are the times in which E and F must declare a In each period choices are made simultaneously, and players in later periods observe the actions taken by players in earlier periods. Punter A obtains a payoff of both referees, and D declare —1 — to win, —a —1 — b and and both want a They would = C and 1 — (a + — c) b)(l e if whether he or the other greyhound would prefer to run the race =D . That is, e if = D and 1 — (a E in the verdict of referee winner and c It will if — b)(l d) if e = C E Lastly, referee . D he declares : both referees wants C to win, to C greyhound 2c is if E way he but either , Also, he would prefer to be first rather as slowly as possible. + A if also like to keep their declared the winner by referee is —b 1 the form of his payoff depends on than second provided that he does not have to run too 2d declared the winner by obtains The payoff contributions to the injection as small as possible. e B is In other words, otherwise. result. C greyhound if Similarly, punter a otherwise. to be the winner, B wants D 1 The payoff fast. greyhound C like gets payoff d to be the winner. F Referee be seen that, in this game, each player's action he greyhound is only interested C he declares is 's , D to is to be the payoffs are identical. set is compact and independent of the actions chosen by earlier players, and that players' payoffs are continuous. The game c centres around the two greyhounds. < d and D when c > d , the race between the discontinuous payoffs (2c, c > d — (a + 1 b)(l them — d)) is < c if weights in the convex combination can depend on (c,d) therefore yield the following Lemma 2.1 Suppose that a distribution function + (a , (1 + b)(l-x)) for G + b > c (a = + d - b)(l c), 2d) if (Note that the . Standard considerations Then both C and D use mixed . given by G(x) xe .) - of timing with lemma. [$,1] = for x e [0,|] and G(x) = actions with (2x - l)/(2x - 1 . Notice that this mixed action its d C when chooses game essentially a and some convex combination of these two payoffs when , E Because referee is non-atomic, that probability mass concentrates at £ as a + b -» 0+ . its Not support is [j,l] , and that surprisingly, then, we all obtain: Lemma 2.2 one. d + Suppose that a The remainder of the b = argument straightforward. is d with probability zero. Hence both referees + indifferent as to C with b = , then c = d = choose a If + b $ with probability > then c will equal always agree. Also, each greyhound will will A win exactly half the time. The payoffs to other hand, a D Then both C and . —b. and B are therefore —a and with probability one. Hence each referee \ on the If, is which greyhound he declares to be the winner. Suppose that E opts D probability p and D probability q and with probability with probability 1 - q 1 —p A Then . and that F opts , 's payoff 2pq is for — for C with and B 1 's is 2(l-p)(l-q)-l. It is game with easy to see, however, that the no equilibrium. Indeed, any a > is strictly So the only possibility for an equilibrium B b to set = — otherwise A two is with probability one. But would raise a from zero. dominated A for if these payoffs between b to set = then Similarly, 2(1 A for a = it , A any b as is B and > for has B . with probability one and must be that 2pq — p)(l — q) — 1 > . — 1 > And these inequalities are mutually inconsistent. 8 This contradiction establishes the counter- example. The essence of the counterexample use strictly mixed actions. is this. As long as a + b > , both greyhounds This behaviour on their part generates a public signal endogenously within the game. The two referees use this public signal to co-ordinate their actions. in When a + b = , however, this signal suddenly degenerates, and the only which the referees can coordinate their actions one or both choosing D is with probability one. But way by both choosing C with probability if they both choose C then B gets A and B is see this, note that the sum of the payoffs of the two players if e ^ f. Hence the expected value of this sum is non—positive. If the expected value is strictly negative, then the payoff of at least one player is also strictly negative. If the expected value of the sum is zero, then e and f must be perfectly To 8 e = f and —1 correlated. In this case the payoff of the player against whom the decision goes is —1. if —1 . then So he would prefer the truly random outcome. Similarly, A will if both referees choose D wish to restore the truly random outcome. In the light of this explanation, it is players to observe suitable public signals. natural to try to restore existence by allowing In the present example, this would amount to allowing the two referees to toss a coin to determine the winner in the event of a tie (which would certainly restore existence). That such an extension yields existence case will be demonstrated in Section 4. in the general 10 The Basic Framework 3 The Data 3.1 There a is non-empty whose index also a passive player, Nature, is J= therefore J^i U {0} Time . J<A of active players, indexed by finite set is eX i e J must choose an and Jn «J= game is = j The . indexed by The . se J"= t 6 = STjd set of actions available to her or — X by a point—to—set mapping A..: -» Y,. — The vector x . is {0,1,2,...} . each player {1,2,...} him depends on which being played, and on the outcomes of previous periods. This situation is There . overall set of players or t j players play one of a family of In each active period . The action. or i discrete, For notational reasons, we assume that games, parameterised by x is or i e X is modelled — lists the parameter of the game being played, and the outcomes of any preceding periods, while A, .(x take C ) it period Y, A.^x that is t is - — ) = — Y,^ x for all (x ) = A,.(x x A X ,y, ) eX~ such that x payoff of any active player outcomes of vectors t > 1 . x all = periods. ~~ The 9 . (x ,Xp...) e X,0 EX - Y. set of . where Y. , X And outcomes possible y, e A,(x ) modelled by a function x is = simply the It is and Y.. x in set of all pairs — and , i.e. depends on which game x i It is — .: x for all ) — ~ X € nothing more than the set of profiles of actions that players can take. modelled by the point-to-set mapping A,(x For reasons of expositional economy, we the set of actions available. such that x Y. x u-: the graph of A, e Xw — t-1 = X is -» IR . . Finally, the played, and the Here / X 00 is N (x ,x. ,..,x,_.) n the set of V t-1 eX~ . for all (Nature does not have a payoff function.) We make the following standing assumptions about these data: (i) X (ii) for all i € X4 and all (iii) for all i e J<A , u- is , and all of the sets Y.. t > 1 are , , non-empty complete separable metric A..: X — - J6(Y,-) is spaces; measurable; bounded and measurable. This assumption does in principle involve a slight loss of generality. However, since any case fix Nature's strategy below, there seems to be little advantage in constraining her actions as well. 9 we shall in 11 mean (In this paper, 'measurable' will always Borel measurable unless explicitly stated to •%(•) will always denote the space of non-empty compact subsets of the contrary; and endowed with the Hausdorff metric.) These assumptions are designed • to ensure that every player possesses at least one strategy, and that associated with every strategy profile and every subgame there is a well defined payoff for every active player (see below). Stronger assumptions are needed to ensure the existence of equilibrium. Strategies 3.2 and extended strategies In the standard version of our model, the evolution of the beginning of period informed of x 1. players are told which 1, They then simultaneously choose . This completes period period 1. That is, sets for period 2. game they they are told y, . 2, 2. And so be identified with the set of subgames in period For each measurable function x t-1 , 6 f,.: t X > 1 and each ~~ -> is, 9Jl^{ i e J t , At the they are players are told the outcome of it goes on. — Definition 3.1 That are playing. can be summarised by a vector x t as follows. They then choose simultaneously from This completes period information in period is actions from their action sets for period At the beginning of period 1. game It their action follows that players' eX~ — X and that , can . a strategy for player in period i t is a _ 4 -) such that supp[f,.(x )] C A,.(x ) for all v t-l X Here 9Ji{XA denotes the set of probability measures over Y,. 'probability measure' will always mean . (In this paper, Borel probability measure; and spaces of probability measures will be taken to be endowed with the weak topology 10 unless explicitly stated to the contrary.) 10 {A all In the terminology of Parthasarathy (1967), a sequence of probability measures } C &M(Y.) converges in the weak topology to A continuous bounded x- Y -» IR f . iff Jx dX n converges to /xdA for 12 The For each are behaviour strategies. f.- thought of as the randomising device that player him available to in period when x t is t , and x i will use to i f, , .(x among choose the actions As such, the previous history of the game. independent of the devices used by the other players in period t and of , can be ) it is devices used in all preceding and subsequent periods. all Strategies and strategy profiles certainly exist under our standing assumptions. Moreover, given any strategy profile and any subgame, the payoffs of active players can be We may calculated in the natural way. therefore define subgame—perfect equilibrium as follows. Definition 3.2 such that, for A subgame—perfect — all x t > 1, all payoff in subgame x e equilibrium We {£. |t > 1} is £, > 1 . Then the information available to — ~~ =X period t C t > 1, e i Jj&> G X^" , player i cannot improve his in his strategy. subgame—perfect equilibrium Suppose that, = them when they make ,£ ) each active period outcome of the preceding period, but — (x at the beginning of [0,1] , where £ = their choices can be (£-,,..., of such vectors as the set of extended also of t ) We refer to the set subgames of the game in . Definition 3.3 period i — x [0,1] 1 a sequence of signals, drawn independently from summarised by a vector h H all 2 shows, however, players are informed not only of the , 1 therefore introduce the concept of an extended strategy. according to the uniform distribution. t and , by a unilateral change need not exist in general. a strategy profile <f,- — X As the counterexample of Section Suppose that is is A ti (x t_1 ) For each t and each > 1 a measurable function for all (x t_1 ,^) e f,.: H t_1 . H i e J , an extended strategy for player i in — — -> &J((Y..) such that supp[f,.(x ,£ )] 13 — Once again the the randomising device that player in period t , when x signals observed So are behaviour strategies. f,. i will use to choose — the previous history of the is up to and including period t . But f..(x among this <f. strategy profile 6 Jj6 , player i |t > 1, £ is the vector of public time this device t and of all him is independent, devices used in all the public signals. all A subgame—perfect Definition 3.4 can be thought of as ) the actions available to game and not only of the devices used by the other players in period other periods, but also of ,£ i e equilibrium in extended strategies X^> is an extended- — such that, for all t > 1 , all h cannot improve his payoff in extended subgame h eH~ , and all i by a unilateral change in his strategy. A subgame—perfect coordination: after the equilibrium in extended strategies allows a limited amount of outcome of a period has been realised, the players the continuation equilibrium to be played following that public signal. correlation is It is can coordinate on outcome by exploiting the next therefore a limited kind of extensive—form correlated equilibrium. limited in that the players observe a privately observing a single common component of a vector of signal, rather signals.) than each (The 14 The Main Result 4 We need the following assumptions: (Al) for all t > 1 (A2) for all t > 1 is given, and and f,*: J^i for all Taken in conjunction G J^i e i a strategy , (A3) i all , X u- f.* — -* the , mapping A.-: X ~~ -» J£(T..) is continuous; (and not an extended strategy) for Nature in period 9M (Y,,-.) is t continuous; bounded and continuous. is with the standing assumptions made in respect of the basic framework, they guarantee the existence of a subgame—perfect equilibrium in extended strategies. The present section is devoted to a proof of this fact. begins, in Section 4.1, with It an overview of the proof. The proof involves two main steps. The analysis necessary the first of these, namely the backwards program, is The given in Section 4.2. necessary for the other, namely the forwards program, is given in Section 4.3. analysis Section 4.4 then integrates the backwards and the forwards programs to complete the proof. reader may 4.1 Overview wish to read Section 4.1 and then skip to Section Imagine Then the proof from stage T for the finding , two The 5. purposes of the present subsection that the divides into essentially for steps. In the first step, game has horizon T . one programs backwards what turn out to be the equilibrium paths of the game. In the second, one programs forwards, constructing strategies that generate these paths and that constitute an equilibrium. The x T— 1 e basic ideas of the backwards X T— , find the set of Nash program are equilibria (including strategies) that can occur in the final stage of the probability distributions over sets are Y™ as follows. Nash equilibria in game. Let C™(x that result from these Nash compact and payoff functions are continuous, C™(x each history First, for ) equilibria. T— ) is a mixed be the set of Because action non-empty compact 15 set 9JC{Xrr) contained in point—to—set mapping G A T_ (x T— ^(^^(Yrp)) to , the upper semicontinuous. 11 is T— 1 . the set of continuation paths consistent with equilibrium For any v™ - given by is o <p ,y rp_ 1 . )] (This set consists, in effect, of those probability distributions over Nash that can be obtained by randomising over selection c T (x , • ) ^p 2 Crp(x , •) (For a general c,y(x . <p be at least one with Crp(x 2 c,y,(x existence theorem of T— , • that equilibrium ) is , T— 1 when • ) . , •) , Let mapping 2 ,•) which , it is this set This not. (1990).) may Crp_-. (x T— 1 be the is C™, a non-empty compact from X p 2 to )] , And is for For each .) find the set of game Nash specified is by essentially the content of the each Nash equilibrium associated Y™, * Y™ that results when and the continuation paths are given by 2 ) • ) be empty. But there will always find the probability distribution over played in period , Crp(xrj,_, the continuation of the set of such distributions obtained when both and the Nash equilibrium associated with T— ) T Simon and Zame "p , T— Cy,_|(x for ,•) p<_2 Crp(x equilibria in from the correspondence co[Cy(x equilibria that can occur in period Crp(x T— 1 co[C T (x YT , X T— from depend continuously on x be the history of the game prior to period x let ) C™ sets o •p Next, Because actions . subset of is Then are allowed to vary. &J£(Yrr<_i*Yrp) ^(^^(Y^ixYrp)) essentially the content of the generalisation of it, , and the point—to—set upper semicontinuous. (This Simon and Zame's theorem given is in Section 4.2.) The relevance of the assumption that players' action sets in period T depend continuously on the history of the game prior to period T emerges clearly in the context 11 of a two—period game. For each n game when the outcome of period 1 > 1 is , let y, , a? be a Nash equilibrium in period 2 of such a and suppose that (y^a^) -» (y^a^) as n -» <d i 's action set is upper semicontinuous in the outcome of period 1, the actually available to him in period 2 when the outcome of period 1 is y-, . Because player action a„j is . Also, because his action set in period 2 is lower semicontinuous in the outcome of period any deviation open to him in period 2 when the outcome of period 1 is y, can be approximated by deviations open to him when the outcome of period really is an optimal action. 1 is y, . Hence 1, a^. 16 Finally, iterate until period 1 T X...Y. distributions over is obtained for is C.(x reached, and a set eX x all game x out to be the set of equilibrium paths of the (The . of probability ) co[C,(x set will )] turn This completes the backwards .) program. Consider c,(x ) now co[C.(x 6 c..(x ) we can and each £ i definition of (i) C,(x ), there the f,-(x ,£ are given ) by c«(x of e, (x ,£ e [0,1] ) , ) given that y 1 f,-(x ,£ 12 1 < t < ) and cJx T and i e 2 2 ,£ J^£ intended to produce is ) . when is for all (ii) And so ,£ 6 c~(x ,-,£ — ,-) is always c,(x always the conditional distribution of e,(x ) This ensures that no single—period deviation is profitable follows is -» C,(x over Y-. ) ) For . By such that: ,-)] one obtains d„(x ,^ ), set of strategies eJx f,- ,( ), for all do produce the paths they are The essential points are that the — — ) , given ,£ with the conditional distribution and that y, . ) constitute a profitable. c, , .(x ,y,,^ ) is That they constitute an — equilibrium follows from the fact that the f,-(x ) when the continuation paths is ) )] ,f ,£ are given 1. 2 strategies really ,£ 2 in period 1 co[C (x ,£ we , ,•): [0,1] from co[C (x ) on until one has a complete That these X given Lebesgue measure. is y,, c«(x ,y-.,£ ) 6 over C,(x be the marginal of e,(x ) the outcome in period cJx ) variable e,(x [0,1] primarily a technical matter. expectation of e,(x . ) Nash equilibrium Similarly, beginning with f„.(x ,£ each x can be viewed as a randomisation over )] random exists a selection and ; find a d,(x is ) let constitute a ,-,£ we can ) whose distribution over C,(x each for associate a probability measure d,(x Also, for each d,(x . Suppose that, Since any path in co[C,(x . )] paths in C.(x ), the forwards program. Nash equilibrium. That no more complex deviation from a standard unravelling argument. This completes the forwards program. Overall, the essential ingredients of the theorem are the following: players action sets are compact; players' payoffs are continuous; Nature's strategy is action sets in a given period depend continuously on the history of the period; players observe public signals; continuous; players' game prior to that and players' payoffs are bounded. The role played 17 by the The three of these should be obvious. first role played by the fourth ensure that the point—to—set mapping from subgames in period those subgames upper semicontinuous. The role of the is point—to—set mapping from subgames in period convex valued. And the sixth in addition, we do that is t roughly, to is, to equilibrium paths of fifth is to ensure that the subgames to equilibrium paths of those t needed only is, for the relatively technical reason not insist that Nature's behaviour strategies have compact support. The backwards program 4.2 we develop In this subsection the analysis necessary for the backwards program. To this end: X, Y- for (i) let (ii) for all i (iii) let f*: X (iv) let Y= (v) let C (vi) for The : X4 6 -» all i e let , A-: 9>Jt (Y*) "jgjYj and gr(A) each i e -» J and Z be non-empty complete separable metric spaces; , X -> Jfc(Y-) be a continuous point—to—set mapping; be a continuous mapping; A(x) let = Y^ x ^Aj(x) x x for all £ X ; Jf(Z) be an upper semicontinuous point—to—set mapping; J^i , let u-: gr(C) -» R be a continuous function. similarity of this notation to that of Section 3 is deliberate, and is intended to be helpful, not confusing. With reference to the model of Section 3 and the discussion of Section A. of past histories; describes the dependence of player f* describes Nature's behaviour; u- is the payoff of active player what we have is C i . i 's 4: X is action set on the past history; describes the set of possible continuation paths; In terms of a the set and somewhat more abstract perspective, a continuous family of games parameterised by x e X . The three essential ingredients of such a family are: the continuity of the action sets; the upper semicontinuity of the continuation paths; and the continuity of the payoffs in the parameter, the outcome and the continuation. ^C(x) For each x 6 &Jt(Y Z) c * X we are interested in the set of equilibrium paths that are obtained when one is free to choose any randomisation over 18 C(x,y) as the continuation path following the outcome y G A(x) € Jc(x) the marginal \i (ii) the marginal /jg (iii) A possesses an Y of A over of Y* over \i a product measure with supp[/x] C A(x) is is • | C(x,y) for y)] C moreover the marginals of \i- y all e be helpful to state more explicitly what probability measures over A.(x) for fj,- supp[^(- |y)] C C(x,y) for that y all A(x) 6 is /u-(x,y,z)dA(y,z) improve Lemma Proof = his payoff 4.1 We #C , and the by changing is Then . i in game x semicontinuous. Let n- = measurable selection from P. n C(x,y)} . for all closed (By D definition p Z . y all . Suppose we are given (iv). and a transition probability v such Y«Z Then the payoff of an . Nash equilibrium map from X if active player no such player can p-(x,y) . is i , x , into Jf( 3>JC{Y and y A(x) 6 , let * Z)) . p-(x,y) the lowest payoff that can be imposed on Standard considerations show that {z|z e C(x,y) and u-(x,y,z) To . (x,y) semicontinuous. Hence P~ (D) -• when = p-(x,y)} We . p. is shall lower need a Following Deimling (1985; Proposition 24.3 and Theorem . C y) for • | with v 24.3), the existence of such a selection will follow if measurable v( is Let A be the measure over For each following outcome y P:(x,y) ; his probability measure.) begin with some notation. | player . an upper semicontinuous min{u.(x,y,z) z 6 C(x,y)} J constitute a \i- A(x) meant by is \£ all obtained by combining the product of the i v over Z over the Y- constitute a Nash equilibrium (j, the continuation path following outcome y may ; frt(*|x); r.c.p.d. (regular conditional probability distribution) such that supp[i/( (It A formally, iff: (i) (iv) More . Z be the resulting measurable = P7 (D) = this end, define oo if D {(x,y)|p.(x,y) selection. = (x,y) p n C(x,y) = = .) p. (x,y)} {(x,y)|D n P-(x,y) = min{u-(x,y,z)|z Like is p. , p • is j- 0} 6 D is lower measurable. Let q.: gr(A) 19 The proof now paths of game x that are consistent with Nature's the proof shows that this J6{9Jl^i Z)) * mapping B This It is is That is, that A(K) B clear that that is it is we need only show > 1— e that ffl(K) > 1— e A for all The second step. that it is that, for all B(x) e . £*( B(x) be the set of jx) • The . #C e is that X , part of into Since this graph closed. graph its is x this end, let tight in order to > first #C is is non—empty step accounts for the bulk of the proof. compact valued. To we need only show closed, let , to demonstrate that is non-empty valued, and is X mixed action that the graph of that remains all done in the third is need therefore show B(x) , e an upper semicontinuous mapping from is The second shows . clearly contained in that of valued. For each x divides into three parts. closed. e X show that exists a we be given. Since it is compact. K Y there exists a compact But there certainly All that compact c K C * Z such Y* such (V K to and we may therefore , ieX* AjW K let This completes the . be the graph of the restriction of C(x, first step. In order to begin the second step, suppose that -» A (x,A) . For . From all j i # , let j J and be player ^C(x (y )dA n (y,z) j all = continuous Xa ' = jy Y -» K supp(A )cgr(C(x ,.)) n n i 's (x ,A n' payoff from A , n y) 6 and gr^C] let it. , and that (x ,A n' n') be his payoff from one can deduce that: ) [/xi (y i )dA n (y,z)J |/x (y )dA n (y z) j j > continuous A y. i IX0(y0)dA n (y,z) for all r- the definition of /Xi (y i )x for all i • : Y. i -» yR and A (y )df (y ]xn ) J : .(5.1) Y. J ...(5.2) ; .(5.3), 20 where C(x (C(x n is n ,y) • , empty y t A(x n )) iff /Xi (y i )u i (x,y,z)dA n (y,z) for all active r M and i and i Equation a all . ni X\ :Y. -» R ; ...(5.4) and z) A.(x e (5.1) follows (5.2) follows ; continuous ) p ...(5.5) ) . n' from the independence of the marginals of A from the A fact that is its (5.4) holds follows conditional payoff. (• |y) is supported on C(x because the expectation of u-(x from the strategy u (It is ,y) to a set of actions, each of supported on A(x 6 A(x is ir y mass of player which yields him important to avoid the word 'support' here. For player him r. need not be and no matter what the if . , and a.s. 's ) , Equation . ) i This mixed his equilibrium i 's continuation so the set of pure strategies Inequality (5.5) states that player closed.) payoff must be at least what he would get this deviation for all is conditional on y. ,-,-) payoff /u-(x ,y,z)di/ (z|y) need not be continuous in y yielding of A \i fact that, in equilibrium, the probability must be confined over the Y. consistent with Nature's strategy. Relation (5.3) follows from the facts that the marginal and that Z to = ^/^(y^dA^z) /p i (x n ,y\a ni )dA n (y > for all active Equation all Y empty—valued) correspondence from regarded as a (possibly is ) he were to deviate to a • i and 's if, equilibrium following realised choices of the other players, the worst conceivable continuation path from his point of view occurred. idea that no redistribution of probability mass Equation (5.4) captures the among equilibrium actions is worthwhile. Inequality (5.5) captures the idea that no redistribution of probability mass from equilibrium to out—of—equilibrium actions Now Also, since it is ic is worthwhile. easy to see that relations (5.1), (5.2) and (5.3) are preserved in the limit. = /u-(x ,y,z)dA (y,z) , x . -» ir. . Hence (5.4) is preserved too. Lastly, for 21 any A.(x) a- e of A. there exist a , Since also w- . A-(x e . such that a ) lower semicontinuous, (5.5) too is has a closed graph will therefore be complete x and A (5.5) for , -» . which we a- , The proof preserved. is we can show if by the lower semi continuity #C that that the analogues of (5.1)- imply that A shall refer to as (5.1')-(5.5'), 6 ^C(x) . Using standard considerations we can deduce from (5.1'), (5.2') and (5.3') that the marginal Y* over /x /j Y of A over f*(-jx) is , and that A has an supported on C(x,y) for specified by v a product measure supported on A(x) is all y A(x) 6 The marginals . \i- of need not constitute a Nash equilibrium for this game. However, by (5.4'), the set Y- of deviation by active player from i to /x. 6 , is game with continuation paths consider the over the Y. \l that the marginal of v over Z such that £(-|y) r.c.p.d. Now . , a- 6 A.(x) such that the Dirac measure concentrated at a- a- 1 will , 1 i raise his payoff that remains an it = K- K- 17) ^ but i one i. above y. 6 Y- r.c.p.d. of a ly) y. 6 It is . for all i of no importance , and K-|y) how we we therefore constitute a Jc(x) * 7j e if q (X|y) . v{ • | if y) y. e Yj\Yj set for all j Y. for more than deviation no longer is profitable. The //• . non-empty valued. To dense in A.(x) game x . $C has closed graph, this end, fix For each . in the natural existing , Nash equilibrium with these continuations, and therefore A Having proved that N} We ^(-|y) equal to £(-|y) in this case. Inequality (5.5') ensures that, set with the new continuation paths specified by v < define = way v in such a but such deviations are no longer attractive. A, Y\Y therefore alter (Such a y corresponds to a coordinated deviation by two or more players.) For definiteness, € We may has //.—measure zero. x. 1 < way from The family N < x and od , that remains each find, for construct a finite the existing of all game x games obtained as . i show that \?C a sequence {y game with And N , to is for varies N= is is that action sets a> continuous, and non-empty when {y is - 1 1 < n simply take the equilibrium correspondence therefore has compact values and a closed graph. standard results show that the set of equilibria } is N < od . its Moreover It follows that 22 it is non-empty Lemma N= for cd too. 4.1 captures the essence of the backwards program: we begin with an upper C depending on semicontinuous correspondence the sequence of outcomes up to and including the current period, and end up with an upper semicontinuous correspondence depending on the sequence of outcomes preceding the current period. There further complication. When we come is however one theorem in Section to complete the proof of the 'I'C 4.4, the continuation paths in Z will themselves be probability measures over sequences of future outcomes. = ^^(W) , In the present, where W more a complete separable metric space. is itself *C equilibrium paths described by by writing Z abstract, setting this can be captured 9JC{Y lie in $>JC(W)) x With this notation, the This means that . What cannot serve directly as input for the next iteration of the backwards program. required instead &JC{Y of x W) 9>Jl[X those of is More : -> 9JC{Y x W) (The point . is is that elements are probability measures over sequences of future outcomes, whereas x 3>JC{yi)) PJC^&Jt^)) elements of ^C X a correspondence #C Such a correspondence can be obtained by replacing are not.) with their expectations. xGX formally, for each ^C(x) let c ,?1(Y x W) be the set of measures A such that: (i) (ii) the marginal \i the marginal \l* A possesses an (iii) A(x) e over of A of over \i is Y* v over r.c. p. d. a product measure with supp(/z) C A(x) is W fg(x) ; such that it v{ /j. of \i over the Y. may be helpful to expand on (iv). is &JC(k.{xj) and a transition probability v such that € A(x) of the (jl- Then . y) e co[C(x,y)] for all active player The //• i 's constitute a payoff is constitute a y r/(- |y) for all y . u(- |y) G co[C(x,y)] /u-(x,y,^(- |y))d/i(y) Nash equilibrium Nash equilibrium when Suppose that we are given 6 . • | the continuation path following outcome y (Once again ; ; moreover the marginals (iv) Y if , where /z is \i- for all y the product no such player can improve his payoff 23 by changing Lemma his p. .) Suppose that the domain of definition of u-(x,y,z) 4.2 suppose that u(x,y,z) X semi continuous mapping from It on linear in z is *C domain. Then Jtp(?JC(Y into should be noted that, when its we come W)) x convex in is an upper z and , . Lemma to apply is u-(x,y,z) 4.2, defined as the expectation with respect to z of an underlying function of (x,y,w) convexity of the domain of u.(x,y,z) in z and the , natural consequences of the problem structure. #C roughly speaking, that Lemma 4.2 will be in z , . So the will be proved by showing, the image of tfC under the projection that consists of is &J£($ J((W)) by > replacing elements of linearity of u-(x,y,z) be will their expectations, and that this projection is continuous. Proof Let A G r.c.p.d of j/(-|y) fi ^JC{Y Z) x over Z. Let for all y£Y. , fj, let = fi Then the measure with marginal , be the marginal of A over and let u{- |y) X projection and conditional v \i of the r.c.p.d. chosen for \i A The .) . = /z d^(z|y) of A onto (It , and is v be an let be the expectation of M(Y « W) is X should be clear that step of the proof first Y the probability is independent to show that projection 6 9JC{Y is continuous. To that {A~ n} this end, let C ?JC(Y /x(yMw) for all x {A n} W) dA~ (y,w) n C 9Jt{X x converges to -. /x(yMw) continuous functions x- Y"K converge to A Z) Xe^(YxW), it x Z) suffices to . Then, to show show that dl(y,w) and ifr. continuous linear functional defined by 1(k) W -» IR . Let t 3>JC(W) = /^w)d«(w) . Then -» K be the 24 /x(yMw)d!n (y,w) = Mw)di/n (w|y) / x(y)d** (y) n (by Fubini's theorem) = (by definition of = / K\(-\v)) x(y)d/* (y) n €} / /<z)di/ (z|y) x(y)d/x (y) n n (by definition of the expectation of u (• |y)) /x(y)^(z)dA (y )Z ) n (by Fubini again, and because = // // But the ). latter expression converges to /x(y)4 z )dA(y,z) by definition of weak convergence, and this integral is equal in turn to /x(y)V(w)dA"(y,z) by reversing the above argument. This completes the proof of continuity. To complete the proof, it suffices to show that gr(^C) under the projection that takes (x,A) into (x,X) Then the payoff from A /Uj(x,y,z)dA(y,z) Suppose first the image of gr(^C) that is = / /u (x,y z)di<z|y) > d/x(y) i = /u (x,y,/zd^(z|y))d/i(y) i (by the linearity of u-) . is (x,A) G gr(^C) . 25 = /u (x,y,K-|y)) d /i(y) i (by definition of V). But u{- |y) G co[C(x,y)] because supp[*/(- |y)] C C(x,y) /u-(x,y,P(- |y))d/z(y) e gr(*C) is by definition the payoff is To the projection. G gr(#C) We . standard considerations show that C(x,y) gr(co[C]) u{ let | (for y) -» = c(x,y,F( | example, one could y)) for all y set v{- |y) = u- , Lemma 4.2 is Then C has a . non-empty is G (x,A~) gr(#C) of which 9Ji{V) with closed graph, and z G co[C(x,y)] iff The standard G A(x) &-, , and x , let • | such for y). y) it to the work existence theorem for each player's strategy set is a of C Let . to gr(co[C]) , be arbitrary for y £ A(x) Then it can be checked as A obtained by combining the marginal ^C(x) lies in v{ . the basic result justifying the backwards program. present subsection by relating If need to find (x,A) that the measure with the transition probability v follows. easy to see that it is &JC(7j) be a measurable selection from the restriction of above, using the linearity of Jl So . C(x,y,z) be the set of k G this end, let expectation z and such that supp[/t] C C(x,y) c: A~ and . Suppose now that (x,X) (x,A~) from , Simon and Zame We conclude the (1990). normal—form games can be expressed compact metric space, and if there is as a continuous function associating a vector of payoffs with each vector of strategies, then there exists a vector of (possibly mixed) strategies that forms a (1990) extended this result by showing that correspondence associating a convex exists a selection from this if, Nash equilibrium. Simon and Zame instead, there set of payoffs helpful to reinterpret their an upper semi continuous with each vector of strategies, then there correspondence and a vector of (possibly mixed) strategies that forms a Nash equilibrium when the payoff function In order to explain is how Lemma work slightly. is given by this selection. 4.2 extends the result of Much as we do Simon and Zame, in the proof of consider a sequence of finite approximations to their basic game. Lemma Suppose these it is 4.1, they 26 approximations are indexed by Nash equilibrium of game N N for , suppose that the mixed strategies some selection equilibrium payoffs of this equilibrium be strategies 7iVr- , and payoffs /a- 7iYr- . from its Then /a^- payoff correspondence, and their proof shows that, constitute a limit point of the strategies /a^- ir- constitute a the the mixed and payoffs then there exists a selection from the payoff correspondence of the original game such that the constitute a \i- function Nash equilibrium with equilibrium payoffs when the payoff ir- given by this selection. In other words, they demonstrate a limited form of is upper semicontinuity Lemma first is relatively family of games. for a particular continuous family of 4.2 therefore extends the result of minor: it games. Simon and Zame shows that upper semicontinuity holds The second is more substantive. The in two ways. 12 The little Lemma continuous for a general Simon and Zame result of about the behaviour of the equilibrium strategies and equilibrium payoffs, but very if let tells us it tells us about the selections from the payoff correspondence that enforce these strategies. 4.2, by contrast, concerns a comprehensive description of equilibrium. Such a description of equilibrium is not without intrinsic interest. the purposes of the present paper problem that would otherwise is arise that it But its main significance for allows us to avoid a measurable selection when we come to program forwards. 13 As it stands Lemma 4.2 is not a direct generalisation of the result of Simon and Zame. But it can easily be adapted to obtain one. To do so, view Z not as a space of probability measures over a compact metric space, but more generally as a compact metrisable subset of a locally convex topological vector space. (This perspective is more 12 , and includes the general, subset of IR , Z possibility that then one can define u-(x,y,z) is = a compact subset of z. .) And IR ^C(x) define . If in a Z is a compact more limited way, to consist of pairs (/a,7t) G &J((Y) * IR where /a is the distribution of a Nash equilibrium and ir is its equilibrium payoff vector. (This more limited definition of $C is the price one must pay for the more general Z .) Then a proof almost identical to that of , Lemma 4.2 shows that $C is a continuous projection of $C, and therefore upper semi continuous. 13 -» IR It would be tempting to conduct the backwards program as follows. Let II: gr(A) be an upper—hemicontinuous convex—valued correspondence describing possible continuation payoffs. Then $11: X -» IR the correspondence describing possible Nash equilibrium payoffs for the current period, is upper hemicontinuous; and so therefore is co^II] the continuation—payoff correspondence for the next iteration of the backwards program. This simple program is, however, difficult to reverse. For, in order to reverse , , it, 27 4.3 The Forwards Program In this subsection this we develop the analysis necessary for the forwards program. purpose we shall need a measurable selection c from #C . This function can be thought of as part of the output of previous iterations of the forwards program. describes the equilibrium paths that For It must be followed from the current period onwards if the strategies calculated for earlier periods by the previous iterations of the forwards program are really to constitute part of Lemma There x [0,1] -» (i) 4.3 an equilibrium. measurable functions exist $>JC{W) such that, for all (x,£) e X C A(x) themarginal^0(-|x,£)of/z(-|x,£) over (iii) i/(-|x,-,£) for all y 6 an -» ^J({Y * W) Y r.c.p.d. of Y A(«|x,£) over is W ity{-\x); such that u{- |x,y,£) v{- |x,y,£) Moreover /A(- |x,£)d£ = following outcome y c(x) for all x e X c(x) all, conditions (i)—(iv) imply that therefore tells us that, on the basis of the current public signal £ if is Nash equilibrium for all y e A(x). . 4.3 captures the essence of the forwards First of = € co[C(x,y)] A(x); when the continuation path /A(- |x,£)d£ Y a product measure such that is constitute a explanation. * : the marginals /j.(-|x,£) of /x(»|x,£) over the Y- Lemma and w.X ; (ii) is x [0,1] * [0,1] the marginal /x(-|x,£) of A(-|x,£) over supp[/x(-|x,0] (iv) X A: program. A(- |x,£) € the equilibrium path e [0,1] , It requires ^C(x) is some The . condition chosen from ^C(x) then the overall path will be precisely one must be able to find, for any measurable selection ir from #11 a measurable selection $7r from II such that the game with payoff function $7r(x,-) has a Nash equilibrium with payoff 7r(x) This is not a standard measurable—selection problem. For what is actually required is that we select, for each x a function $7t(x,-) and that these choices of functions be coordinated in x to obtain a single measurable function 3>7r (A version of this non—standard problem did arise for Mertens and Parthasarathy (1987; Lemma 1 of Section 6, pp 41—42). In their case, however, the functions $7r(x,-j could be taken to be continuous.) , . , , . 28 c(x) Lebesgue measure (If . conditional distribution over Y x x W ?JC{Y -» [0,1] Secondly, the .) assertion \i-\ period. Y lemma x W , then /A(-|x,^)d^ [0,1] and , A(- |x,-) if the marginal distribution over is goes beyond the simple assertion that there exists such that A(- |x,£) G *C(x) for X , the proposition delivers a v that (x,£) all x [0,1] -• &J((Y.) are precisely the strategies of the players Fourthly, the restriction $c of v to gr(A) X A: For, while this . measurable in (x,y,£) is the is would automatically imply the existence of a v that was measurable in y each (x,£) the W) x the marginal distribution over is for Thirdly, jointly. for the current describes the equilibrium x [0,1] paths that must be followed from the next period onwards. the input for the It serves as next iteration of the forwards program. Most Proof Lemmas of 4.1 The x X G , of the considerations necessary for the proof have already arisen in the proof, and first step supp(d(x)) The 4.2. c is present proof will therefore be brief. X to find a measurable d: \PC(x) and the expectation of d(x) an argument very similar to one used in the proof of measurable mapping 9JL ( $>JC(Y -• A: X x [0,1] &J£(Y -» x W) Lemma x W) is precisely d(x) 'measurable' representation theorem, The third step £(-|x,-,£) an is e.g. to find a measurable is x for all r.c.p.d. of A(- |x,£) |x,«): [0,1] . W)) such -» that, for all This can be done using we need Next, 4.2. such that, when measure, the distribution of the random variable A(- ^Ji(Y c(x) is * given Lebesgue is [0,1] 9Jl{Y a x YV) over Such a mapping can be obtained by applying a . Gihman and Skorohod mapping for all u: XxY (x,£) e X x [0,1] x [0,1] . (1979; -» Lemma 1.2, &Jt(W) such In other words, p 9). that we need 'measurable' version of the standard result which asserts the existence, for each (x,£) an r.c.p.d. v{- |x,-,£) of A(- |x,£) forwards program, and since the literature, Let we /x(- . Since this we have not been is, , a of in effect, the central step of the able to find precisely the result we need in sketch a proof of this. |x,£) be the marginal of A(« |x,£) over sequence of sets that generates the Borel a—algebra on Y Y . . Let {B For each 1 N 1 < , let n < <d} be a 5?™ be the 29 a-algebra generated by {B &** Since 5?-^ Y on N} < And . . give an explicit formula for an r.c.p.d. of A(-|x,£) For each y . ^l#(W) v be a fixed element of let Y G and each measurable W W we C set A(YxW|x,Q -x^L = -L-- - ,,;, n 1 < we can is finite a—algebra given the 1 " (W|x,y,£) l N M(Y|x,0 Y if z, = &N of It is containing y and /i(Y|x,£) immediate from these formulae that and therefore that the convergence , K lim!/v,on and v Certainly u is — v outside K what im M (- 1 is, i/ u{- |x, •,£) There are two u(- |x,-,£) e co[C(x,y)] . . The To obtain A(x) Lemma of u-^ is |x,-,^) exists is an &„ |x,£) //(• r.c.p.d. of difficulties for all y the . To A(x) measurable. Set v , is effectively a version of the the martingale convergence theorem — a.s., and that u(-\x,',£) given the Borel ; stands: it it may and the marginals is a version of a—algebra on not be the case that /z(-|x,£) when the continuation path first property, let of A(-|x,£) need following outcome y to be q(x,y) q be any measurable selection from co[C] wherever obtain the second: for each 4.1; let measurable in jointly A(-|x,£). This completes the third step. with u as 6 is is fourth and final step constructs a v that does have these properties from redefine i>(-|x,y,£) G K set measurable. Also, since i/^(-|x,-,£) not constitute a Nash equilibrium u(- |x,y,£) and u-^(- |x,y,£) effectively the conditional expectation of A(- |x,£) is Y. That v , . conditional expectation of A(- |x,£) given implies that > = KW) N (W|x,y,£) otherwise. (x,y,£) atom the is 7r.(x,£) = i , it let does not q. lie in co[C(x,y)] for , and some y be the function defined in the proof of Ju-(x,y,v(- |x,y,^))d/i(y|x,^) ; and let Y.(x,£) be the set of a- 30 6 such that /u.(x,y\a.,^(- |x,y,£))d/z(y|x,£) A-(x) v{- = |x,y;0 = (x,y) if qi 7 £(• |x,y,£j otherwise. and the Y^YjCx.O 6 for all j * ^(x,^) but yj i Then we may . Y(x,0 and £ , Standard considerations show that the resulting v A(- |x,£) e C(x) fact that A(-|x,£). . > implies that u(- |x,»,£) is still an v{- |x,y,£) is measurable, r.c.p.d. of o Completion of the proof 4.4 Before proceeding, we need to define the equilibrium correspondence for a family of We games. In making this definition we encounter a minor technicality. equilibrium as a strategy profile that is Nash equilibrium in in every strategy profile specifies players' behaviour for the entire family of Hence an equilibrium face. 6 set X We therefore . A such that A of paths equilibrium. by x t— 1 e Theorem from define X t_1 E.. (x ) , , we can the set of equilibrium paths of correspondences E. for Suppose that (Al)—(A3) hold. Then E, games t > is for 2 each game x game x , to be the set by some as a family parameterised . map an upper semi continuous J6{9Jt{**_ Yj). into . The main (see Section 5). games that they might game x the equilibrium path generated in define, subgame. But a whole family of paths, one Similarly, reinterpreting our basic family of X t— 4.1 is will generate a have defined an interest of Theorem 4.1 derives However upper semicontinuity from the is itself fact that of interest: implies existence it it is reassuring to know that our concept of equilibrium possesses this standard regularity property. Proof The proof proceeds period t B. is — in three steps. Let B.(x that are consistent with Nature's strategy. an upper semicontinuous map from X — into I sequence of correspondences {C. 1 1 < t < cd} is ) The be the first set of paths beginning in part of the proof shows that ®Y JGi^JC^ S — )). X s constructed, each of which In the second, a is upper 31 semicontinuous. In the third shown that it is = E. B For notational convenience, we show that for all co[C.] t . To upper semicontinuous. is this 1 B?(x end, let be the set of marginals over ) backwards programming argument similar and 4.1 shows that B, 4.2 Y J6{9Jt{* _, _.Y This completes the converge for ) $, (C for each T . , , ) T 1 < t < C let , T step, let T ^("/iY.) So B, , too, is converges upper semicontinuous. be given. For each > 1 T (x t— > T t C Now T = define C. n t Hence C, too ™,C. for each is , T It is A [^ t (C. 6 paths from C, [^ (C^ t follows. 1 < T < +1 , It is also S ^ the case that (C, od t_1 ) T and A can be enforced by continuation definition, T . T belong a fortiori to C, C = n C t T=i*t( T+l^ Since . ,) , together with C,,, parameters x By . C. ,, in for all C. C *+(C. Finally, ) To ,) C C.. , 1 But paths . )](x )](x , Then ,] .) cC. whenever T obvious that C, t see this, let . upper semicontinuous. t > 1. upper semicontinuous. is B. T= onwards when the continuation paths can be chosen from co[C, t T 6 T = C. let , the set of equilibrium paths that are obtained is ) . the backwards program ensures that each C, A iff its T for period > S. } C into be defined inductively by the formula . C (In other words, . t. all map from X Lemmas first step. Turning to the second And, Then a ). but simpler than, that developed in to, an upper semicontinuous is B,(x of measures in But a sequence of measures {A )). PJCi* marginals over ,Y « To . ' . Hence the rec* uired conclusion see this note that the correspondences T= for ,-, And Lemma C T . , for , form a continuous family of games with cd 4.2 shows that the equilibrium correspondence of — such a family is upper semicontinuous. certainly a limit point of points in Overall, then, {C.| 1 < = co[C,] t < a) } for all such that C, t . We We show first c, from co[CJ we obtain is do that It follows that any [^.(C. A 6 C. (x ), also lies in ,)](x ) , [* t (C, which , is ,)](x ). a sequence of upper semicontinuous correspondences = ^ t (C. this for co[CJ also a selection t C , = E, for all ,) 1 . only. t . 1 So remains only to show that E, The other It suffices from E It . let to C. cases are analogous. show that any measurable selection be any such selection. Then 32 repeated application of the forwards program starting with strategies and a measurable selection c.,, from co[C.,J such f,. , the f.-(x ,£ continuation paths are given by onward when the f.. c, ,(x , ,•>£ are employed in period c. It . programming argument that the strategy follows from (ii) We now show that E, c E,. C co[CJ any subgame-perfect equilibrium -» Y 3>JC(X * ) in f is given. Now ) (i) for all t when follows from To game x is certainly c™,, is the t and a standard dynamic (i) f t And an equilibrium. precisely c, (x this end, let describe the paths followed from period = <f.. |t > 1, > 1 t , This . ) i it X<> € let H c ,: be ~~ onwards when the strategy inherits measurability c, in period and the continuation paths are given by profile f constitutes . , and (") the paths obtained from period extended strategies. For each employed. The mapping profile t that the equilibrium path in completes the proof that co[CJ Nash equilibrium constitute a ) are precisely those required by c.,, that: t > 1 — H~ ,()e (x yields, for all c, from f . T Let be > 1 T a selection from co[Cy,,]. Also, suppose that c,, is a 1 selection Nash at T J from co[C. equilibria for period once that that c, is for t some 1 < t < is a selection from co[C a selection from co[C, a selection from co[CJ . Since f is when the continuation paths T c, T. ] . T , ] . an equilibrium, the are given by c. , f,- . It specify follows Proceeding inductively we therefore conclude Moreover, allowing This completes the proof. T to vary, we obtain that C, is 33 5 Existence with and without Public Signals This section provides a systematic discussion of the question of existence of equilibrium in the dynamic model introduced in Section namely of the main games satisfying the standing assumptions result for this case, when that, 3. It begins with a recapitulation public signals are allowed, any family of (Al)—(A3) possesses an equilibrium. However, the section also addresses two further questions. Under what circumstances can existence And be obtained without the use of public signals? subgame—perfect equilibrium obtained by allowing is the extension of the concept of the observation of public signals sufficiently rich? 5.1 Existence with Public Signals The Theorem following result 5.1 an immediate corollary of Theorem is 4.1: Suppose that (Al)—(A3) hold. Then there exists a subgame-perfect equilibrium in extended strategies. Proof Pick any x eX . Then, by definition of E..(x By Theorem ) , game x . potential drawback of this result equilibrium in what is E,(x ) is non-empty. Pick any A E.(x In particular, there exists an equilibrium! is that, in order to obtain the existence intended to be a purely non-cooperative situation, extraneous element into the model, namely public signals. to discover circumstances under 6 there exists an equilibrium of the family of games that generates equilibrium path A in The 4.1, which equilibrium exists It is it therefore of introduces an some interest even without this element. ). 34 Existence without public signals 5.2 There are existence: games of which public signals are not needed in Theorems Then a subgame—perfect equilibrium Oddly enough, Theorem (Intuitively in and zero—sum games. The and 5.2, 5.3, for 5.4 respectively. Suppose that (Al)—(A3) hold, and that the family of games has perfect 5.2 information. Proof game perfect information, finite—action games, games are summarised results for such Theorem at least three classes of it 5.2 would seem obvious is a exists. not an immediate corollary of when only one player that, Theorem 5.1. acts at a time, public signals do not allow players to achieve anything they could not achieve using mixed strategies. But there a minor complication: players observe not only the current signal, but also is However, an equilibrium can be constructed using the correspondences C, past signals.) used in the proof of Theorem and define f,.(x selection co[C 2] L.(x We • begin with any measurable selection to be the marginal of c,(x from co[C J that enforces = C^ ) ) 4.1. Hence c„ is also a to be the marginal of c (x 2 c, . over ) Since the Y... game Next, . is a game measurable selection from C« ) over Y,. . And so on. . let c« c, from C-. , be a measurable of perfect information, So we may define o Indeed, one can even show that there exists a subgame-perfect equilibrium in which players employ pure actions only. We do not demonstrate this here, since the argument does not appear to be a simple corollary of the construction that we have employed. 14 Note that the proof of Theorem 5.2 shows incidentally that allowing public signals in a game of perfect information does not enlarge the set of equilibrium paths. 14 See Harris (1985) for a treatment of the case of perfect information. The argument used there can easily be adapted to the present context. Alternatively, one can work directly with the apparatus used here, taking care to purify the selections as one programs forward. 35 Let us say that a family of games t and > 1 Theorem all 1 x" 6 * A^x 1 , *) is of finite action a finite of equilibrium paths Theorem Proof if X is finite, 5.3 is compact. is more exists. or less well known, and for all if, set. Suppose that (Al)—(A3) hold, and that the family of games 5.3 Then a subgame—perfect equilibrium action. it. X1 is Moreover, and there is (See Fudenberg and Levine (1983) for a different proof.) for is each x of finite X 6 , the set more than one way of proving So we merely note that can it be proved by a simple adaptation of the methods of this paper. The basic idea can be expressed in terms of the notation of Section 4.2 as follows. Instead of defining #C(x) to be the set of equilibrium paths that can be obtained by choosing continuation paths from co[C] one defines , it to be the set of equilibrium paths that can be obtained continuation paths from x e X this , It that X new C. Because definition still yields X is finite, by choosing and because the A.(x) are finite for all an upper semicontinuous correspondence. should be noted that both finiteness assumptions are needed. Indeed, suppose but that one or more of the A-(x) are infinite for some is finite happen that there is a sequence of equilibria of signal generated endogenously game x, x. Then it can each of which involves a public by players' randomisation over their action sets A.(x), and that this sequence converges to a limiting equilibrium which does not generate any such signal. infinite. there is Suppose, on the other hand, that the A.(x) are finite for Then it at least can happen that there one player coalesce in the limit. for If this whom is all xgX a convergent sequence of games in at least one player for whom at least but that X , X and that two actions player randomises over his two actions, then he generates a public signal which disappears in the limit. It is precisely in order to compensate for the potential disappearance of such endogenously generated public signals that the exogenous public signals are used in our basic existence theorem. o is 36 To complete our discussion of the existence of subgame—perfect consider the case of a family of zero—sum games. (A family of games contains precisely two elements, and = Theorem if E. j/^\ Then a subgame—perfect equilibrium period zero—sum games. It is known well convex in such a game. But the if it is J<A .) helpful to recall is of zero sum. some facts about oneis regarded as the set of probability set of equilibria, Hence, even in a zero—sum game, the set of The set of players are allowed to observe a public signal. equilibrium payoffs, on the other hand, will not. to co-ordinate if that each players' set of optimal strategies distributions over outcomes, need not be. equilibria will be enlarged sum we exists. In order to understand this result, equilibria of zero Suppose that (Al)—(A3) hold, and that the family of games 5.4 two players is equilibrium, For the public signal merely allows the on a choice of Nash equilibrium have the same equilibrium payoff (namely (v, for the —v) , game, and where v is all Nash the value of the game). Proof Note first that our proof of the existence of equilibrium shows, in particular, that the extended family of games has a Nash equilibrium. It follows immediately from standard considerations that this family also has a value. That measurable mapping v, (Indeed, since v.,(x ) is X : such that v, (x simply player upper semicontinuous, v, Similarly, there exist -> [R is mappings v X is there exists a the value of extended payoff from any path in E,(x actually continuous. extended game beginning in period With l's ) is, ~~ -» IR We do not need this such that v.(x ) is ), game x . and since E, fact, is however.) the value of the t. these considerations in mind, we can construct a equilibrium of the original family of games as follows. subgame—perfect The construction follows the construction of an equilibrium of the extended game, but takes care to avoid the need for a 37 public signal. The step first to find a measurable selection is measurable selection from C, may product measures. So we be the marginal of c,(x rather than co[C,] ) define strategies over from co[CJ that enforces the Y... f,. ) 2 C9 But . continuation payoff, namely v (x Similarly, . X -» ^^(Y.-) by strategies £.: X t_1 ?JC{Y.^ -» Since c, . A e co[C (x are already setting f,.(x to ) c« if Cr, generate the same )] 2 a is generates a measurable selection c, a measurable selection from is then (L generates the same continuation payoffs as c« Hence the • f, • can be (The continuation paths specified by c« do not • necessarily constitute an r.c.p.d. of c, Iterating this of course.) , for all > t 2 argument we obtain o as well, Approximation by finite—action games Suppose that we are given a single game singleton) satisfying (Al)—(A3). Then a roughly speaking, a new family corresponding to n = is oo of X a family of games for which (i.e. game finite—action approximation to this games indexed by n e IN the original game; for each n 6 of finite action; and taken as a whole the is marginals over Y, In particular, . enforced by <L just as well as by c« 5.3 all its , f,.: from C, c, new family of U{od} IN , is, such that: the game the games a is game corresponding satisfies (Al)-( A3). to n (Here (Al) ensures that players' action sets in the original game are continuously approximated; (A2) ensures that Nature's behaviour is continuously approximated; and (A3) ensures that active players' payoffs are continuously approximated.) Finite-action approximations to a single Indeed, we can even To do (Al)-(A3). i 6 J, {yjj|n let game satisfying (Al)-(A3) certainly exist. construct a finite-action approximation to a family of games satisfying this, we need the following ingredients. Yt be dense in 6 IN} Secondly, for each n € IN , Y|| let = - ; and m {y | 1 < {yj|n let m< n} . each First, for 6 IN} be dense in Thirdly, for each 6 IN, o£j:X let -» Y.. be a continuous function such that, for — — a..(x ) A..(x ). YQ = t > 1, — — n belongs to A,.(x ) , and ay(x all x and each > 1 t e X X°. i J and G — : — ) is as close to y?. as any other point — (Such a function exists because A,.:X -» J£(Y..) is continuous. In in 38 Nature's case the function -. MY ti be defined by ) Fourthly, for each such that, for close to f rt(x simply constant with value y.* ^(x 1-1 = and n t > 1 x all is e 6 X : as another ) IN, ) {aj^x* let ^\*: supp [</> measure X -1 — j(x C )] {y™0 ^.(x 1 1 < 1-1 ) m< n} . Yj = X Finally: let = A (x t_1 ti ) and ; for all t > In terms of these ingredients Q 1 {y A\ . &J((Y. *) for all t > 1 f ; and x we can n} for all x v X t-l n a.-: let t_1 X t_1 e . &JC(Y.q) be a continuous function -» t( in m< )! 1 < And .) < m < n}; whose support t_1 e , 6 i X t_1 and is ^ (x rt(x is ) as contained in Jj6 and x let <P t t_1 t_1 6 = ) X t_1 (x f t0 let , t_1 ) . define a larger family of games, which can be interpreted as a sequence of families of finite-action games approximating the original ssr\ family, as follows. First, let Secondly, suppose that X ^t— X be the union over n t the restriction of aj-(-) to > 1 , f.0 is the restriction of in the obvious Now X — s\ 1 ; fl^-) A. to t > 1 = A x- X . • and X. {n} of the sets Then: for all * Yn ie 7, A,. = gr(A.) . Thirdly, for all — . Finally, payoff functions u- are defined suppose that we are given a finite-action approximation to a single game. is simply a family of games parameterised by n e is it IN U {qd} upper semicontinuous in any subgame-perfect equilibrium path of any game n equilibrium path of that game, 6 IN is, n. in particular, and Hence, an follows that any limit point of subgame-perfect equilibrium paths of the finite-action games is an equilibrium path of the original game. our solution concept (namely subgame-perfect equilibrium in extended strategies) includes everything that can be obtained by taking finite-action approximations to a (or, . ^ /\ /\ s\ satisfying (Al)-(A3), its equilibrium correspondence And U {qd} way. Since this approximation since IN has been defined for some S\i is 6 more precisely, everything that can be obtained game by taking continuous finite-action approximations to a game satisfying (Al)-(A3)). 5.4 Approximation by e-equilibria One can a given single also game is show that any limit point of subgame-perfect e-equilibrium paths of an equilibrium path of the game, by combining the techniques of this 39 paper with those of Borgers (1989). Indeed, suppose that we are given a family of games satisfying (A1)-(A3). For each t > 1, x t_1 e X t_1 and e > , let E (e,x t_1 ) be the set t of equilibrium paths of except that: (i) subgame x each player i G Jji of a family of is it is identical to the original replaced by a sequence {t-|t > 1} of agents; each agent chooses actions that are at or within Then games e game (ii) of being optimal in each subgame. easy to adapt the methods of Section 4 to show that E. is upper semicontinuous. But this implies the required result. For any subgame-perfect e—equilibrium path of the original game path is is also an e-equilibrium path; any e—equilibrium an e—equilibrium path of the game in which each active player sequence of agents; and the 0—equilibrium paths of the game in is replaced by a which each active player is replaced by a sequence of agents coincide with the equilibrium paths of the original game. 40 Conclusion 7 The general results of this paper concerning the existence of subgame-perfect equilibrium in dynamic games create a small dilemma. If one adopts the point of view that the observation of public signals violates the spirit of non-cooperative interpretation of the counterexample of Section 2 continuum of actions are somehow wholly non—pathological in all is If, theory, then the presumably that games with a intrinsically pathological. other respects. game After all, that example seems on the other hand, one thinks that games with a continuum of actions are an indispensable part of non-cooperative game theory, then presumably one must accept that games must be extended to allow the observation of public signals. Faced with dilemma this There actions as pathological. too simplistic. First, are, tempting to dismiss games with a continuum of however, at least two reasons why may this reaction be can be argued that a solution concept should be robust to small it errors in the data of the equilibria that occur in it is games games to which it applies, in the sense that arbitrarily close to the game under one should not rule out consideration. 15 (In other words, the equilibrium correspondence should be upper semicontinous.) Provided that such errors are confined to payoffs, the solution concept of subgame-perfect equilibrium satisfies this requirement when applied to finite—action games. present in the extensive form semicontinuity do violate the is itself, to be satisfied. spirit of If, however, such errors then public signals must be introduced Secondly, it is if may be upper not entirely clear that public signals really non-cooperative game theory. After all, players must in any case coordinate on a specific Nash equilibrium to play; and public signals merely increase the extent to which they can coordinate, by allowing them to coordinate on a equilibrium. 15 See Fudenberg, Kreps and Levine (1988) for example. random Nash 41 References Borgers, T (1989): "Perfect Equilibrium Histories of Finite— 47, 218-227. and Infinite—Horizon Games", Journal of Economic Theory Chakrabarti, S K (1988): "On the Existence of Equilibria in a Class of Discrete-Time of Economics, Dynamic Games with Imperfect Information", mimeo, Department Indiana University. A Mas—Colell Duffie, D, J Geanakoplos, Equilibria", Research &; A McLennan (1989): "Stationary Markov Paper 1079, Graduate School of Business, Stanford University. Forges, "An Approach F (1986): 1385. to Communication Equilibria", Econometrica 54, 1375- Fudenberg, D, D Kreps & D Levine (1988): "On the Robustness of Equilibrium Refinements", Journal of Economic Theory 44, 354—380. Fudenberg, D, & D Levine (1983): "Subgame-Perfect Equilibria of Finite— and InfiniteHorizon Games", Journal of Economic Theory 31, 251—268. Fudenberg, D, and J Tirole (1985): "Pre— Emption and Rent Equalization in the Adoption of a New Technology", Review of Economic Studies 52, 383-401. Gihman, Harris, Harris, 1 1, and AV Skorohod (1979): Controlled Stochastic Processes Springer—Verlag. , C (1985a): "Existence and Characterisation of Perfect Equilibrium in Perfect Information", Econometrica 53, 613-628. C (1985b): "A Characterisation of of the Perfect Equilibria of Infinite-Horizon Games", Journal of Economic Theory Harris, Games 37, 99—125. C (1990): "The Existence of Subgame-Perfect Equilibrium with and without Strategies: a Case for Extensive— Form Correlation", Discussion Paper 53, Nuffield College, Oxford. Markov W M, &; Leininger (1986):"A Note on the Relation between Discrete and Continuous Games of Perfect Information", Discussion Paper A—92, University of Bonn. Hellwig, W Hellwig, M, & Leininger (1987): "On the Existence of Subgame-Perfect Equilibrium in Infinite—Action Games of Perfect Information", Journal of Economic Theory 43, 55-75. Mertens, J—F, & T Parthasarathy (1987): "Equilibria for Discounted Stochastic Games", Research Paper 8750, CORE, Universite Catholique de Louvain. Parthasarathy, KR W (1967): Probability Measures & R Zame (1990): "Discontinuous Rules", Econometrica 58, 861-872. Simon, L K, on Metric Spaces Academic Press. , Games and Endogenous Sharing R QQ§ Ui Date Due MIT LIBRARIES DUPL 3 IDflD 00b72155 t