MIT LIBRARIES 3 9080 02618 1369 iP*"^''- RjffitlUiti: II •! .M415 no.06--28 2006 Digitized by the Internet Archive in 2011 with funding from Boston Library Consortium IVIember Libraries http://www.archive.org/details/learningdisagreeOOacem DEWEY HB31 .M415 Massachusetts Institute of Technology Department of Economics Working Paper Series LEARNING AND DISAGREEMENT UNCERTAIN WORLD IN AN Daron Acemoglu Victor Chernozhukov Muhammad Yildiz Working Paper 06-28 October 20, 2006 Room E52-251 50 Memorial Drive Cambridge, MA 02142 This paper can be downloaded without charge from the Social Science Research Network Paper Collection http://ssrn.com/abstract=9391 69 at '""iEii?.Li&-^ J=!§RmiEs Learning and Disagreement in an Uncertain World* Daron Acemoglu, Victor Chernozhukov, and Muhamet Yildiz^ October, 2006. Abstract Most economic analyses presume that there are limited differences in the prior beliefs of most often justified by the argument that sufficient common experiences and observations will eliminate disagreements. We investigate this claim using a simple model of Bayesian learning. Two individuals with different priors observe the same infinite sequence of signals about some underlying parameter. Existing results in the literature individuals, an assumption establish that when individuals are certain about the interpretation of signals, under very mild conditions there will be asymptotic agreement — their assessments will eventually agree. In con- we look at an environment in which individuals are uncertain about the interpretation of meaning that they have non-degenerate probability distributions over the conditional distribution of signals given the underlying parameter. When priors on the parameter and the trast, signals, conditional distribution of signals have full we prove the following results: (1) Indisame infinite sequence of signals. (2) Before support, viduals will never agree, even after observing the observing the signals, they believe with probability 1 that their posteriors about the underlying (3) Observing the same sequence of signals may lead to a divergence of opinion rather than the typically-presumed convergence. We then characterize the parameter will fail to converge. conditions for asymptotic agreement under "approximate certainty" limit — i.e., where uncertainty about the interpretation of the signals disappears. as we look at the When the family of probability distributions of signals given the parameter has "rapidly-varying tails" (such as the normal or the exponential distributions), approximate certainty restores asymptotic agreement. However, when the family of probability distributions has "regularly-varying (such as the Pareto, the log-normal, and the t-distributions) , tails" asymptotic agreement does not obtain even in the limit as the amount of uncertainty disappears. Lack of common circumstances. We economic behavior priors has important implications for illustrate economic behavior in a range of the type of learning outlined in this paper interacts with in various different situations, including nation, asset trading games of common interest, coordi- and bargaining. Keywords: asymptotic JEL how disagreement, Bayesian learning, merging of opinions. Classification: Oil, C72, D83. *We thank Eduardo Faingold, Greg Fisher, Drew Fudenberg, Alessandro Lizzeri, Giuseppe Moscarini, Mar- ciano Sinscalchi, Robert Wilson and seminar participants at the University of British Columbia, University of Urbana-Champaign and Yale for useful comments and suggestions. Department of Economics, Massachusetts Institute of Technology. Illinois at * Introduction 1 The common prior assumption is one of the cornerstones of modern economic analysis. Most in a models postulate that the players precisely, that they for example, they known have a all common agree that game have the "same model prior about the some 6. The may own common prior view is among d, p. A some this process will lead to A 6. strong version of this 48) as the statement that a Bayesian individual, does not assign zero probability to "the truth," are informative about the truth. drawn from a assumption comes from and individuals about the distribution of the vector expressed in Savage (1954, is experiences and the communication of others, will have access to a history of events informative about the vector "agreement" more distributions also have additional information about typical justification for the learning; individuals, through their game form and payoff payoff-relevant parameter vector 6 distribution G, even though each components of of the world," or will learn who eventually as long as the signals it more sophisticated version of this conclusion also follows from Blackwell and Dubins' (1962) theorem about the "merging of opinions".^ Despite these powerful intuitions and theorems, disagreement exception in practice. Just to mention a few instances, there ment even among economists working on a certain topic. is the rule rather than the typically considerable disagree- is For example, economists routinely disagree about the role of monetary policy, the impact of subsidies on investment or the mag- nitude of the returns to schooling. Similarly, there are deep divides within populations with shared experiences, and finally, there about religious beliefs was recently considerable dis- agreement among experts with access to the same data about whether Iraq had weapons of mass destruction. In none of these cases, can the disagreements be traced to individuals having access to different histories of observations. Rather particular, it it is their interpretations that differ. seems that an estimate showing that subsidies increase investment is In interpreted very differently by two economists starting with different priors; for example, an economist believing that subsidies have no effect on investment appears more likely to judge the data or the methods leading to this estimate to be unreliable and thus to attach less importance to this evidence. Similarly, those who believed in the existence of weapons of mass destruction in Iraq Blackwell and Dubins' (1962) theorem shows that two probability measures are absolutely continuous with same events), then as the number of observations goes to infinity, their predictions about future frequencies will agree. This is also related to Doob's (1948) consistency theorem for Bayesian posteriors, which we discuss and use below. if respect to each other (meaning that they assign positive probability to the presumably interpreted the evidence from inspectors and journalists indicating the opposite as biased rather than informative. In this paper, we show that type of behavior will be the outcome of learning by this Bayesian individuals with different priors when they are uncertain about the informativeness of signals. In particular, we consider the following simple environment: one or two individuals with given priors observe a sequence of signals, {st}"^Q, and form their posteriors about some underlying state variable (or parameter) is may be that these individuals 9. The only non-standard uncertain about the distribution of signals conditional on In the simplest case where the state 9 e {A,B}, and {a, &}, this implies that Pr(st and the the underlying state. st G = 9 individuals' subjective probability distribution This distribution, which can differ and to among uncertainty about the informativeness of signals. are non- degenerate, individuals will have some its 9) \ We but individuals also have a prior over p0, say given by Fg. density. feature of the environment — pg signal are binary, e.g., is not a known number, refer to this distribution Fg as density fg as subjective (probability) individuals, When is a natural measure of their subjective probability distributions latitude in interpreting the sequence of signals they observe. We identify conditions under which Bayesian updating leads to asymptotic learning (indi- viduals learning, or believing that they will be learning, the true value of 9 with probability after observing infinitely many assessments of the value of 9). signals) We asymptotic learning and agreement. and asymptotic agreement (convergence between first are guaranteed. probability 1 p' > Second, First, we show that 1/2 (with possibly p^ we learning under certainty leads to as- ^ p^), > i is sure that pg = p' for then asymptotic learning and agreement establish the stronger results that to the event that pg their provide some generalizations of existing results on ymptotic learning and agreement. In particular, when each individual some known number 1 when both individuals attach 1/2 for 9 G {A, B}, then there will again be asymptotic learning and agreement. These positive be less 1. results do not hold, however, when there than 1/2. In particular, when Fg has a There will not full is support positive probability that pg might for each 9, we show that: be asymptotic learning. Instead each individual's posterior of 9 continues to be a function of his prior. 2. There will not be asymptotic agreement; two individuals with different priors observing of signals will reach different posterior beliefs even after observing same sequence the infinitely many signals. Moreover, individuals attach ex ante probability 1 that they will disagree after observing the sequence of signals. 3. Two individuals may disagree they did so previously. In the full more fact, for it common sequence of signals than any model of learning under uncertainty that satisfies support assumption, there exists an open set of pairs of priors such that the disagreement between the two individuals While a after observing may appear the event that pe < will necessarily grow starting from these priors. plausible that the individuals should not attach zero probability to 1/2, it is also reasonable to expect that the probability of such events should be relatively small. This raises the question of whether the results regarding the lack of asymptotic learning of uncertainty. Put and agreement results We and agreement under uncertainty survive when there differently, we would like to is a small understand whether the asymptotic learning under certainty are robust to a small amount of uncertainty. investigate this issue by studying learning under "approximate certainty," sidering a family of subjective density functions {fm} that around a single point amount —thus converging to i.e., by con- become more and more concentrated full certainty. It is straightforward to see that as each individual becomes more and more certain about the interpretation of the signals, as- ymptotic learning obtains. Interestingly, however, the conditions much more demanding than those for cisymptotic learning. for asymptotic agreement are Consequently, even though each individual expects to learn the payoff-relevant parameters, asymptotic agreement obtain. This implies that asymptotic agreement under certainty point of a general model of learning under uncertainty. is the case depends on the tail this family has regularly-varying tails (such as the tions), even under approximate certainty there to a discontinuous limit show that whether or not this will Pareto or the log-normal distribu- be asymptotic disagreement. When {fm} (such as the normal distribution), there will be asymptotic agreement has rapidly- varying tails under approximate certainty. approximate certainty will learn the payoff- relevant will fail to learn. fail properties of the family of subjective density functions {fm}- When Intuitively, We may be may Whether is sufficient to parameter, but they or not they believe this make each may still individual believe that they believe that the other individual depends on how an individual reacts when a frequency of signals different from the one he expects with "almost certainty" occurs. If this event prevents the individual under approximate certainty. trusts his own model with his model. This of the world When from learning, then there will be asymptotic disagreement because under approximate certainty, each individual is and thus expects the limiting frequencies to be consistent the other individual's model of the world differs, he expects the other Then whether individual to be surprised by the limiting frequencies of the signals. asymptotic agreement will obtain depends on whether this surprise other individual from learning, which in turn depends on the tail is or not sufficient to prevent the properties of the family of subjective density functions {/m}- Lack of asymptotic agreement has important implications tions. We illustrate some individuals observe the we of these for a by considering a number of simple environments where two same sequence of signals before or while playing a game. In particular, games discuss the implications of learning in uncertain environments for games of common interest, bargaining, games of than they would in environments with common priors will play these establish that contrary to standard results, individuals more information about observing the same sequence of signals may games — and also differently than ments without common priors but where learning takes place under interests before receiving of coordination, payoffs. certainty. in environ- For example, may wish to play games of We show how the also common possibility of lead individuals to trade only after they observe Stokey, 1982) and existing results on asset trading without common provide a simple example illustrating a potential reason why individuals (e.g., priors, assume learning under certainty (Harrison and Kreps, 1978, and Morris, 1996). about informativeness of signals show differently the public information. This result contrasts with both standard no-trade theorems grom and We communication and asset trading. how, when they are learning under uncertainty, individuals we range of economic situa- Mil- which Finally, may be we uncertain —the strategic behavior of other agents trying to manipulate their beliefs. Our results cast learning. In many doubt on the idea that the common prior assumption may be environments, even when there is little justified by uncertainty so that each individual believes that he will learn the true state, learning does not necessarily imply agreement about the relevant parameters. Consequently, the strategic outcome from that of the common-prior environment.^ Whether a this may be assumption significantly different is warranted therefore For previous arguments on whether game-theoretic models should be formulated with all individuals having prior, see, for example, Aumann (1986, 1998) and Gul (1998). Gul (1998), for example, questions common depends on the specific setting and what type of information individuals are trying to glean from the data. Relating our results to the famous Blackwell-Dubins (1962) theorem essence. As briefly mentioned on zero-probability events in (i.e., other), asymptotically, they will Our results do not contradict Footnote 1, may theorem shows that when two agents agree this their priors are absolutely continuous make the same this help clarify their with respect to each predictions about future frequencies of signals. theorem, since we impose absolute continuity throughout. Instead, our results rely on the fact that agreeing about future frequencies not the same is Put as agreeing about the underlying state (or the underlying payoff relevant parameters).'^ differently, under uncertainty, there is an "identification problem" making it impossible for individuals to infer the underlying state from limiting frequencies and this leads to different interpretations of the nomic situations, same what is signal sequence important is by individuals with different priors. most In eco- not future frequencies of signals but some payoff-relevant parameter. For example, what was essential for the debate on the weapons of mass destruction was not the frequency is of news about such weapons but whether or not they relevant for economists trying to evaluate a policy effect of similar policies (and if) is may be What not the frequency of estimates on the from other researchers, but the impact of implemented. Similarly, what existed. this specific policy relevant in trading assets is when not the frequency of information about the dividend process, but the actual dividend that the asset will pay. Thus, many situations will which individuals need to learn about a parameter or state that in determine their ultimate payoff as a function of their action falls within the realm of the analysis here. In this respect, our work differs from papers, such as Freedman (1963, 1965) and Miller and Sanchirico (1999), that question the applicability of the absolute continuity assumption in the Blackwell-Dubins theorem in statistical and economic settings (see also Diaconis and Freedman, 1986, Stinchcombe, 2005). for Similarly, a number of important theorems in statistics, example. Berk (1966), show that under certain conditions, limiting posteriors their support on the limiting distribution). whether the common set of all identifiable values Our results are different prior assumption maizes sense 'in this respect, our paper is also related to agree about long-run frequencies, but their beliefs fail may fail there is no ex ante (1994, 1996), to who merge because have to converge to a from those of Berk both because when Kurz (though they will in our model stage. considers a situation in which agents of the non-stationarity of the world. individuals always place positive probability on the truth and also because characterization of the conditions for lack of asymptotic learning Our paper also closely related to recent independent is "common knowledge" when signals. They show how the signal space is infinite. how will Ely, Mailath and be "common learning" by individual learning may not lead to Cripps, Ely, Mailath and Samuelson's analysis focuses on the case in which the agents start with certainty (though they note tight and agreement. work by Cripps, Samuelson (2006), who study the conditions under which there two agents observing correlated we provide a common priors and learn under can be extended to the case of non-common their results Consequently, our emphasis on learning under uncertainty and the results on learning priors). under conditions of approximate certainty are not shared by this paper. Nevertheless, there is a close connection between our result that under approximate certainty each agent expects that he will learn the payoff-relevant parameters but that he will disagree with the other agent and Cripps, Ely, common Mailath and Samuelson's finding of lack of learning with infinite- dimensional signal spaces. The rest of the paper is organized as follows. Section 2 provides all our main results in the context of a two-state two-signal setup. Section 3 provides generalizations of these results to an environment with of our results, K states and L> K signals. and Section Section 4 considers a variety of applications 5 concludes. The Two-State Model 2 Environment 2.1 We model with binary start with a two-state signals. This model is sufficient to establish all our main results in the simplest possible setting. These results are generalized to arbitrary number of states and signal values in Section There are two individuals, denoted by {st}"_Q where ability i.e., tt' 6 £ St (0, \) {a, b]. to 9 = The underlying A. The i 3. = state 1 is and ^ i — 2, who observe a sequence of signals e {A, B], and agent individuals believe that, given 9, i assigns ex ante prob- the signals are exchangeable, they are independently and identically distributed with an unknown distribution.^ That "See, for example, Billingsley (1995). If there were only one state, then our model would be identical to De example, Savage, 1954). In the context of this model, De Finetti's theorem provides a Bayesian foundation for classical probability theory by showing that exchangeability (i.e., invariance under permutations of the order of signals) is equivalent to having an independent identical unknown distrib- Finetti's canonical model (see, for ution and implies that posteriors converge to long-run frequencies. De Finetti's decomposition of probability the probability of is, of = Si 6 given =5 6* = sj is = A a given 9 is B.n unknown number p^; an unknown number pB — as b PA 1 1 -PA -PB PB Our main departure from the standard models uncertain about pA and to individual is i — We pB is we allow the that individuals to be denote the cumulative distribution function of po according his subjective probability distribution i.e., in the following table: B A a shown likewise, the probability —by Fg. In the standard models, F^ degenerate and puts probability 1 at some pg. In contrast, for most of the analysis, we will impose the following assumption: Assumption For each 1 i and 9, Fg has a continuous, non-zero and finite density fg over [0,1]. The assumption implies that Fg has full support over assumption allows Fg this whether or not this is so the distribution o( pg). (p) (i.e., and Fg (p) to differ, for [0, 1]. many worth noting that while It is of our results Remark 2, not important common whether or not the two individuals have a Moreover, as discussed in it is Assumption 1 is prior about stronger than necessary for our results, but simplifies the exposition. In addition, throughout We we assume that consider infinite sequences s The sequences. = 7r\ tt^, Fg and Fg are known to both individuals.^ {sf}^^ of signals and write posterior belief of individual i S for the set of all about 9 after observing the first n such signals {st]^^^ is where Pr^ {9 =A \ {sf}JLj) denotes the posterior probability that 9 signals {sf}"^i under prior tt* — A and subjective probability distribution Fg given a sequence of (see footnote 7 for a formal definition). Throughout, without loss of generality, we suppose that in reality 9 = A. The two questions of interest for us are: extended by Jackson, Kalai and Smorodinsky (1999) to cover cases without exchangeability. player 1 knows the prior and probability assessment of player 2 regarding the distribution of signals given the state is used in the "asymptotic agreement" results and in applications. Since our purpose is to understand whether learning justifies tlie common prior assumption, we assume that agents do not change their views because the beliefs of others differ from theirs. distributions ° is The assumption that 1. Asymptotic 2. Asymptotic agreement: whether lecirning: whether Pr' (Hm„^oo = (Ki (s) Pr' (Hm„_,oo l\9 \'Pn i^) ~ = A) 'Pn{^)\ = — 1 for O) z = = 1, 2. 1 for z = 1, 2. Notice that both asymptotic learning and agreement are defined in terms of the ex ante probabihty assessments of the two individuals. Therefore, asymptotic learning implies that an individual, believes that he or she will ultimately learn the truth, while asymptotic agreement implies that both individuals believe that their assessments will eventually converge.^ Asymptotic Lecirning and Disagreement 2.2 The following theorem gives the well-known result, which applies A hold. version of this result is when Assumption 1 does not stated in Savage (1954) and also follows from Blackwell and Dubins' (1962) more general theorem applied to this case. uses different arguments than those presented below and Since the proof of this theorem tangential to our focus here, is it is relegated to the Appendix. Theorem Fg (p') = 1 1 Assume and Fg that for (p) = some p^,fP' for each p = = A) < = p^. £ Pr^(lim„_oo<^Us) 2. Fv' (lim„^oo 1^^ (s) -<t>lis)\=0)==l. Theorem 1 is will learn the the i — 1 on p% i.e., 1,2, l. a slightly generalized version of the standard theorem where the individual truth with experience (almost surely as same sequence will necessarily agree. The that p^ = follow. The theorem The the probability oi intuition of this st theorem states that even = if is ^ oo) and two individuals observing is differ from p^ (while Savage, 1954, assumes useful for understanding the results that will the two individuals have different expectations about a conditional on 9 these beliefs with certainty n generalization arises from the fact that learning and agreement take place even though p^ may p~). each Fg puts probability Then, for each 1. l|0 (1/2,1], sufficient for = A, the fact that p* > 1/2 and that they hold asymptotic learning and agreement. For example, We formulate asymptotic learning in terms of each individual's initial probability measure so as not to take a position on what the "objective" for "true" probability measure is. In terms of asymptotic agreement, we will see that Pr' (lim„_oo {(p^ (*) "~ "^n (*)| — O) = 1 also implies =0 limn—oo |0j, (s) — 4>'^{s)\ for almost all sample paths, thus individual totic agreement coincide with asymptotic agreement (and vice versa). beliefs that there will be asymp- consider an individual = 6 is A. 1 p\ in learning applies which case he will — he will conclude that 6 that this individual is = conclude that the true state with probability 9 is = Given that p > and sufficient to observe that this Next to see why asymptotic agreement obtains, suppose 1. confronted with a frequency p > p^ of when a signals. Since he believes with the state 9 is =A and 1 — p'' when B, he will interpret the frequency p as resulting from samphng variation. p^, this sampling variation 1 much much more is frequency of a signals p' when even though p^ the state ^ fP is ^ =A — A. the state is 9 ^A Asymptotic agreement believes that individual j will observe a and expects that he and (as long as p' i when likely to the event that 9 then follows from the observation that individual =A the underlying state B. Therefore, this individual believes that he will learn therefore, he will attach probability that 9 is when 1/2 A, or he will observe a limiting frequency of certainty that the frequency of signals should be p* the state it > sure that he will be confronted either with a limiting frequency of a signals equal is — p', and expects a frequency of a signals p* why asymptotic First, to see individual to who fp are will conclude from this both greater than 1/2 as assumed in the theorem). We next generalize Theorem 1 about the signal distribution but to the case where the individuals are not necessarily certain their subjective distributions do not satisfy the full support feature of Assumption 1 Theorem i = 2 Assume that each Fq has a density fg and satisfies Fg (1/2) — 0. Then, for each l,2, 1. Pr'(lim„^oo<^Us) 2. Pr' (lim„^oo This theorem Fq (1/2) The = \4>l is) will = l|e -<pl{s)\^Q)^l. be proved together with Theorem implies that pg intuition for this result probability 1 = A)-l. > is 3. It is evident that the assumption 1/2, contradicting the full support feature in similar to that of to the event that pg the limiting distribution will be > Sj Theorem 1: when both Assumption 1. individuals attach 1/2, they will believe that the majority of the signals in = a when 6 = A. Thus, each believes that both he and the other individual will learn the underlying state with probability 1 (even though they both be uncertain about the exact distribution of signals conditional on the underlying may state). This theorem shows that results on asymptotic learning and agreement are substantially more 9 Nevertheless, the result relies on the feature that general than Savage's original theorem. = F^ (1/2) each for i = 1,2 and each This implies that both individuals attach zero 6. probabihty to a range of possible models of the world be less than 1/2. each individual may It may instead be — more reasonable i.e., to they are certain that pe cannot presume under uncertainty, that, attach positive (though perhaps small) probabihty to encapsulated by Assumption 1. We next general circumstances where Fg has full impose this all values of pg as assumption and show that under the more support, there will be neither asymptotic learning nor asymptotic agreement. Theorem 3 Suppose Assumption 1. Pr' (lim„^oo 2. Pr' (lim„-,oo 4^n \<l>li 1 holds {s)^l\e^A)=l (s) - (t>n {s)\¥'^) for for = 1 i i = = 1,2. Then, 1,2; whenever tt^ ^ tt^ and F^ = F^ e for each {A,B}. This theorem contrasts with Theorems will fail to learn the true state if 1 and with probability 2 1. and implies that the individual The second in question part of the theorem states that the individuals' prior beliefs about the state differ (but they interpret the signals in the same way), then probability that there 1 is their posteriors will eventually disagree, and moreover, they to the event that their beliefs will eventually diverge. Put both attach differently, this implies "agreement to eventually disagree" between the two individuals, in the sense that they both believe ex ante that after observing the signals they will will will to agree. This feature fail play an important role in the applications in Section 4 below. Remark 1 The assumption that Fg — Fg in this theorem is adopted for simplicity. Even in the absence of this condition, there will typically be no asymptotic agreement. Theorem 6 in the next section generalizes this theorem to a situation with multiple states and multiple signals and also dispenses with the assumption that Fg — Fg. It establishes that the set of priors subjective probability distributions that leads to asymptotic agreement Remark 2 Assumption only for simplicity. It 1 is considerably stronger than necessary for is of "measure zero". Theorem can be verified that for lack of asymptotic learning it is 3 and is adopted sufficient (but not necessary) that the measures generated by the distribution functions F\'{p) and 10 and Fg (1 — p) be absolutely continuous with respect to each other. Similarly, for lack of asymptotic agreement, it is not necessary) that the measures generated by sufficient (but and Fq (1 - that pb is pA is (p), F^ (1 For example, p) be absolutely continuous with respect each other. viduals beheve that F\ — if p), F^ both either 0.3 or 0.7 (with the latter receiving greater probability) (p) indi- and also either 0.3 or 0.7 (with the former receiving greater probability), then there will be neither asymptotic learning nor asymptotic agreement. Throughout we use Assumption both because it and because simplifies the notation a natural assumption it is 1 when we turn to the analysis of asymptotic agreement under approximate certainty below. Towards proving the above theorems, we now introduce some notation, which throughout the paper. Recall that the sequence of signals, s, is will be used generated by an exchangeable process, so that the order of the signals does not matter for the posterior. Let = H^{t< rn{s) be the number of times r-n (s) St = /n converges to some p a out of (s) e first n n\st = a) By signals.^ the strong law of large numbers, almost surely according to both individuals. Defining [0, 1] the set S= {s e this observation implies that Pr' (s sample paths G S lim„-,oo Tn (s) : = S") 1 for i = /n 1, 2. exists} We Now, a straightforward application or with probability 1 where Pr' the probability of observing the signal is . ^ (s) = lim„^oo i'li (5) for all Given the definition of r^ = sample paths - (b^'^-) = {(9,s') \s't St our results for all = st for ""' (1 each [ - P""*'* (^ Tr') (1 " = a exactly r„ times out of - < p)'-"^^' n}, where s 11 B} x S = p-^-'^'/b \s {st}^i and s' (p) = n useful formula for P)""""*"' Ja (p) dp, and I' t Bayes rule gives s in S. the probabihty distribution Pr' on {A, Pr' (e^'"''^'^ Pr' at each event e"'"'" {s\, of the The next lemma provides a very signals with respect to the distribution Fq. (j!)'^ will often state which equivalently implies that these statements are true almost surely s in S, (r„|6') (1) , dp {sjjj^j. Lemma 1 Suppose Assumption = <Pl [p (s)) where p (s) = lim„_»oo ^n (s) /n, Then for holds. 1 4 lim and Wp G all s E S, = [s) } (3) [0, 1], Proof. Write W{rn\e^B) Pr' irn\e = A) J^p--{l-pr-'-fB{l-p)dp ^ JoP'H^-Pr-'-fA{p)dp /q p''"(l-p)"-'""dp ~ /n'p'-"(l-p)"-'-"/A(p)dp E^lfBJl Here, the first same term. Jb (1 — {sf}"^Q. equaUty The obtained by dividing the numerator and the denominator by the is resulting expression on the numerator p) given r„ Denoting - p)\rn] under the this by the conditional expectation of Ja the conditional expectation of (Lebesgue) prior on p and the Bernoulli distribution on flat — p)|r„], and E'^[/j3(l is (p), by the denominator, which we obtain the E'**[/4(p)|r„], is similarly defined as last equality. consistency theorem for Bayesian posterior expectation of the parameter, as rn -+ that E^[/b(1 -p)|r„] ^ /s(l - and E^[/A(p)|r„] p) and Ramamoorthi, 2003, Theorem 1.3.2). ^ /^(p) (see, e.g., By Doob's p, we have Doob, 1949, Ghosh This establishes PTHrn\e = B) , as defined in (4). Equation (3) then follows from (2). In equation (4), ff (p) the true state is B versus this likelihood ratio An immediate proofs of the asymptotic likelihood ratio of observing frequency p of a when and Bayes it is = 4>lo Theorems 2 A. rule to implication of 4>l iP is)) The is (P is)) Lemma compute Lemma if 1 1 is and only states that, asymptotically, individual his posterior beliefs about when i uses 6. that given any s & S, if ^^i?^ [p (s)) = ^—^R" {p (s)) and 3 now follow from 12 Lemma 1 and equation (5). (5) Proof of Theorem ment Lemma in when p (s) < 1 I for = = z 1 still applies, Given 9 1/2. cording to both Under the assumption that Fg 2. i == 1 and — 2. and we have A, then r„ Hence, Pr' Similarly, Pi' 1,2. = = {p (s)) {(j)]^ = = 0\e theorem, the argu- when p (s) > 1/2 and K' = l\e B) in the = A) = (s) > = {p (s)) oo 1/2 almost surely ac- = Pr' (<^^ (p (s)) = Pr' (<^^ (p(s)) O|0 1\0 = A) = B) = = for 1 1,2, establishing the second part. Proof of Theorem Hence, by Lemma 1, Since fg (1 3. (f>^^ (p(s)) ¥" follows from equation (5), since n^ and thus Pr' rf>l is), Intuitively, is {p (s)) = /n converges to some p (s) {p {s)) {(f)]^ i?' (1/2) {\cf>l [s) - ^^ when Assumption - > p{s)) and /a (p(s)) is finite, for each s, establishing the first part. 1 ^ tv' (s)| 1 (in = and Fq ^ O) = 1 Fq implies that for i = for each > R' (p(s)) The second s E S, part (f)^ (s) his views particular, the full support feature) holds, about pe (the informativeness of the signals) as well as underlying state. For example, even when signal a B, a very high frequency of a because he may is more an individual a. about the state and (3), i?' (p) us tells how way informativeness of the signals in the same always be reflected In contrast, if were sure that PA he initially in state is beheved, and they be determined by A, may his prior beliefs (i.e., R^ When — two individuals interpret the R^), the differences in their priors an individual were sure about the informativeness of the signals — PB = to a value different from ^ than in their posteriors. P^ for some p* question the informativeness of the signals p{s) A the individual updates his beliefs about the informativeness of the signals as he observes the signals. will views about the defined in (4) never takes the value zero or infinity. his posterior beliefs will by R^, which also likely in state will Therefore, the individual never becomes certain about the state, captured by the fact that Consequently, as shown in his not necessarily convince him that the true state infer that the signals are not as reliable as instead be biased towards which will is ^ 1, 2. never sure about the exact interpretation of the sequence of signals he observes and update 0. p'or 1 — p'. > 1/2) as in Theorem 1, (i.e., if i then he would never —even when the limiting frequency of a converges Consequently, in this case, for each sample path with 1/2 both individuals would learn the true state and their posterior beliefs would agree asymptotically. As noted above, an important implication of Theorem 3 is that there will tj^aically be "agreement to eventually disagree" between the individuals. In other words, given their 13 priors, both individuals the same infinite sequence of signals they will will agree that after seeing disagree (with probability This implication 1). is interesting in part because the common still prior assumption, typically justified by learning, leads to the celebrated "no agreement to disagree" result (Aumann, 1976, 1998), which states that if the individuals' posterior beliefs are common knowledge, then they must be equal.^ In contrast, in the limit of the learning process here, individuals' beliefs are common knowledge different with probability as in Assumption beliefs 1. This is (as there is because in the presence of uncertainty and both individuals understand that 1, no private information), but they are support full on their priors will have an effect even asymptotically; thus they expect to disagree. Many of the applications we their discuss in Section 4 exploit this feature. Divergence of Opinions 2.3 Theorem 3 established that the differences in priors are reflected in the posteriors even in the limit as n —+ oo. two individuals. The first does not, however, quantify the possible disagreement between the rest of this section investigates different aspects of this question. show that two individuals that observe the same sequence of posteriors, so that Theorem satisfy p G It common 4 Suppose Assumption 1 may have that subjective probability distributions are given by Fg and diverging information can increase disagreement. that there exists e > such that \B} Then, there exists an open set of priors n^ and [0, 1]. signals We lim n—>(X) n'^, {p) — R'^ {p)\ such that for all s and Fq > e that for each & S, III \4>i{s)-4>l{s)\>y-7r^\; ' in particular, Pr'(lim \cPi{s)-4>l[s)\>y-n^\)=l. Proof. that \R^ (p) while \n^ ^ Fix Fg and Fq and take - i?2 (p)| '^'^\ — ^- > e for each p e tt^ = [0, 1], tt^ = 1/2. lim„^c)o \(t>n. By Lemma (s) - (p^ (s)| > Since both expressions are continuous in n^ and neighborhood of 1/2 such that the above inequality uniformly holds and the hypothesis I for e' tt^, for some there is e' > 0, an open each p whenever n^ Note, however, that the "no agreement to disagree" result derives from individuals' updating their beliefs because those of others differ from their own (Geanakoplos and Polemarchakis, 1982), whereas here individuals only update their beliefs by learning. 14 and n~ are in this and signals, the if the original disagreement is (s G 5) = 1. will interpret signals relatively small, after almost all sequences of disagreement between the two individuals grows. Consequently, the observation common of a statement follows from the fact that Pr' last even a small difference in priors ensures that individuals Intuitively, differently, neighborhood. The sequence of signals causes an initial difference of opinion between individuals to widen (instead of the standard merging of opinions under certainty). Theorem 4 also shows that observing the same both individuals are certain ex ante that their posteriors will diverge after sequence of signals, because they understand that they will interpret the signals differently. This strengthens our results further and shows that for some priors individuals will "agree to eventually disagree even more" An orems interesting implication of 1 and when 2, there Theorem 4 is worth noting. As demonstrated by The- also learning under certainty individuals initially disagree, but each is individual also believes that they will eventually agree (and in fact, that they will converge to his beliefs). ically, let This implies that each individual expects the other to "learn more". More specif- 1b=a be the indicator function be a measure of learning is - A^] — A for individual i,and let E' Under the probability measure Pr*). so that E' [A' for 9 = - (tt' - tt^)^ certainty. < = and A' (tt* — — Iq^a) ((f)^^ be the expectation of individual Theorem and thus E' 1 imphes that W < [A'] [A^] . 4>lo = 'Poo — 10=a) (under i ~ Ie=A, Under uncertainty, this not necessarily true. In particular. Theorem 4 implies that, under the assumptions of the theorem, there exists an open subset of the interval this subset, individual we have j. W The reason also expects to learn [A'] is > E' [A-^] , [0, 1] so that individual that individual i is more from the sequence will tt^ are in would expect to learn more than initial guess tt', but of signals than individual j, because he believes The fact that an individual may expect play an important role in some of the applications in Section 4. Non-monotonicity of the Likelihood Ratio 2.4 We more than others i not only confident about his that individual j has the "wrong model of the world." to learn such that whenever n^ and next illustrate that the asymptotic likelihood ratio, R' that when an (p), may be non-monotone, meaning individual observes a high frequency of signals taking the value that the signals are biased towards a and may put have done with a lower frequency of a among the 15 a, lower probability on state signals. he A may conclude than he would This feature not only illustrates the types of behavior that are possible important for the when individuals are learning under uncertainty but apphcations we discuss in Section also is 4. Inspection of expression (3) establishes the following: Lemma 2 For any s € 5, </)^ (s) is decreasing at p (s) Proof. This follows immediately from equation When FC and only if B} increasing at p is (s). above. (3) non-monotone, even a small amount of uncertainty about the informativeness is may of the signals if The next example lead to significant differences in limit posteriors. trates this point, while the second example shows that there can be illus- "reversals" in individuals' assessments, meaning that after observing a sequence "favorable" to state A, the individual may have a lower posterior about this state than his prior. on asymptotic agreement Example and pb 1 be more systematically studied will (Non-monotonicity) Each More individual p' > (1 precisely, for p' > (1 4- are in a (^-neighborhood of are not informative. The impact some -I- i in the next subsection. thinks that with probability 5) /2, (5) of small uncertainty but with probability /2, e > < and 5 |j3^ > e 0, I — e, pA the signals — p^|, we have P(„^i^+{^-^)/S ifpe{f-5/2,f + 6/2) ^^^^' for each 9 and Now, by i. R'{p{s)) = \ tfp{s)^{f-5/2,p^ + i^^ ^fp{s)e{l-p^-5/2^-p^ + 1 otherwise. is 1 when p{s) is (s) will also 0J,Q (s) is the informative suggesting that the state (/>to (s) is small, is (s) B, thus ^^^ informative, but this time favoring state A, so uninformative and 4>^^[s) falls back to tt*. </>^ (s) Intuitively, 16 = (s) goes back to 5/2) and 5 > to 1 1. — — > 0. p', The goes likelihood down By Lemmas 1 to and 1 2, the signals are not informative, thus In contrast, around as the prior, 7r\ become uninformative again and — p\ and then jumps back be non-monotone: when p same 8/2) small, takes a very high value at afterwards, becomes nearly zero around </'oo for e 1 ' is TTiffz^y This and other relevant functions are plotted in Figure ratio FC {p{s)) ^ the asymptotic likelihood ratio (4), { otherwise e 7r\ ^ 1 0. — p', become very After this point, the signals Around 1. the signals p', the signals are again Finally, signals again when p{s) is around 1 — become p' or p', the R' 4 Yl-'PI <I>L 1 l-Tt' 71' Tt^-7l' 1-7:2 7r' 4-^ i-p' \-p' Figure 1: The as •kL-' -1—+ P P three panels show, respectively, the approximate values of R} (p), (f>^^, and 0. e individual assigns very high probability to the true state, but outside of this region, he sticks to his prior, concluding that the signals are not informative. The important observation first range of limiting frequencies, as event that he will learn as e ^ and either 1 — j5* 5 or — > 0, ^ e This 9. that even though is and (5 — equal to the prior for a large is (?!)^ each individual attaches probability > 1 to the because as illustrated by the discussion after Theorem is 1, each individual becomes convinced that the limiting frequencies will be p^'. However, asymptotic learning individual also understands that since 5 region where he learns that 9 = Each considerably weaker than asymptotic agreement. is < |p^ — p^|, when the long-run frequency is in a A, the other individual will conclude that the signals are uninformative and adhere to his prior Consequently, he expects the posterior beliefs belief. of the other individual to be always far from his. Put differently, as e — > and 5 — > 0, each individual believes that he will learn the value of 9 himself but that the other individual will fail to learn, thus attaches probability the third panel of Figure to learn, 1;. at each 1 to the event that they disagree. This can be seen sample path and the difference between their limiting posteriors following "objective" TF^ = 1/3 and one of the individuals will will fail be uniformly higher than the bound mm {n\7T^,l-7r\l When in S, at least from tt^ — 2/3, this bound is ttMttI -1}- equal to 1/3. In fact, the belief of each indi- vidual regarding potential disagreement can be greater than this; each individual believes 17 that he will learn but the other individual will Pr' (lim„^oo \(l>i (s) - > Z) > <Pl (s)| I - where e, fail as e to do so. -> 0, Z Consequently, for each ^ min {n^n^, 1-7t\1~- i, tt^}. This "subjective" bound can be as high as 1/2. The next example shows an even more extreme phenomenon, whereby a high frequency s = a among the signals Example may (Reversal) 2 reduce the individual's posterior that 9 Now —A below of his prior. suppose that individuals' subjective probability densities are given by for each 9 and = i 1,2, r (l-e-e2)/(5 ifp^-6/2<p<f + S/2 [ e^ otherwise where > e 0, p' > < and 1/2, 6 < - p^ p'^. Clearly, as e — 0, (4) > gives: tfp{s)<l-f-5/2, orl-f + 5/2<pis)<l/2, or f -5/2<p (s) <f + (5/2 ff(p(5))oo otherwise. Hence, the asymptotic posterior probability that 9 = A'lS ifp{s)<l-f-6/2, or 1 - p' + S/2 < p{s) < 1/2, orf -S/2 < p{s) <f- + 5/2 1 <t^oo{p{s))^{ otherwise. Consequently, in this case observing a sufficiently high frequency of posterior that 9 =A below the prior. monotone, at agree. when To least it is the state is crucial that the likelihood ratio R} > fP Given the form of R^ A, p{s) will be close to high probability to the event that 9 j A would also converge to 1 as e will learn the true state and that He p'. — A when would assign even higher probability to on \4>]^ is one of the individuals would expect that their see this, take p* — > = a may Moreover, the individuals assign probability there will be extreme asymptotic disagreement in the sense that In both examples, s 0. yl p (p (s)) 1 — e (?i^ (p (s)) not monotone. If i?' | that ^ 1. were beliefs will asymptotically i is almost certain that, also understands that j would assign a very (s) (p), = p' > at p (s) = p' individual p'. If R^ were monotone, individual and thus Therefore, in this case i his probability will assessment be almost certain that j their beliefs will agree asymptotically. 18 — reduce the Theorem 1 established that there will be asymptotic agreement under certainty. have thought that as e —+ and uncertainty disappears, the same conclusion would apply. In show that even as each F^ converges to a Dirac distribution (that contrast, the above examples mass to a assigns a unit point), there two individuals. Notably and p^ (5 — — + p^ One might this is may be significant asymptotic disagreement when true not only there is between the negligible uncertainty, but also when the individuals' subjective distributions are nearly identical, 0, —+ This suggests that the result of asymptotic agreement in Theorem . — e i.e., i.e., may 1 > as not be a continuous limit point of a more general model of learning under uncertainty.^ Instead, we next subsection that whether or not there will see in the is asymptotic agreement under becomes more and more concentrated around a point) approximate certainty (i.e., determined by the properties of the family of distributions Fg tail Agreement 2.5 In this subsection, is Disagreement with Approximate Certainty cind we as Fg characterize the conditions under which "approximate certainty" ensures asymptotic agreement. More specifically, we will study the behavior of asymptotic beliefs as the subjective probability distribution Fg converges to a Dirac distribution and the uncertainty As already about the interpretation of the signals disappears. Fq converges to a Dirac distribution, each individual he will learn agreement particular, in the limit on their tail as is considerably more learning, this does not guarantee that the individuals will believe that they will also agree on totic 1, become increasingly convinced that However, because asymptotic agreement the true state. demanding than asymptotic will by Example illustrated We 6. will demonstrate that whether or not there is asymp- depends on the family of distributions converging to certainty properties. For many natural distributions, a small tainty about the informativeness of the signals is amount —in of uncer- sufficient to lead to significant differences in posteriors. To state and prove our main density functions that Fg.^ —^Fg^ there are many fg^ for where different i = Fg^ ways result in this case, consider 1,2, 9 £ {^,-B} and assigns probability in 1 m to p a family of subjective probability € Z+, such that as = p' e for some p' m^ cxd, we have (1/2, 1). Naturally, which a family of subjective probability distributions may 'Nevertheless, it is also not the case that asymptotic agreement under approximate certainty requires the support of the distribution of each Fg to converge to a set as in Theorem 2 (that does not assign positive probability to p\ < 1/2). See Theorem 5 below. 19 converge to such a limiting distribution. Both for tractabihty and to make the analysis more concrete, we focus on families of subjective probability distributions We a determining density function /. (i) (ii) / is symmetric around < there exists x < /^ ^^ impose the following conditions on parameterized by > /: zero; oo such that / [x] is decreasing for x>x\ all (iii) R{x,y)^ exists in Conditions oo] at all (x, y) [0, and (i) introduces the function (ii) — (x /m, which y) scale e K^.^° which down will arise naturally in the the real line around y by the factor 1/m. densities for individuals' beliefs about transformation x i-^ [x — p') /m}^ each 9 and i fff^isa proper hm„->oo (Pn^m (^) where c' (m) individual as m— > is i ^^^' 1/ Jq f [m [p scaled 2 below). The of the form dp is for [0, 1] fg^- In x i— family of subjective we consider the following family of — p^)) (iii) ^^ determined by / and the densities (8) a correction factor to ensure that We each m. ^ the limiting posterior distribution of individual is and = d{m)f{m{p-f)) probability density function on probability density of signals about Pa = pA and pB, {fom}' In particular, fl^ip) for 1 we consider mappings of uncertainty, Condition study of asymptotic agreement asymptotic statistics (see Definitions in amount In order to vary the (7) are natural and serve to simplify the notation. R (x, y), and has a natural meaning (j^ lim m-»oo j [my) i also define when he (f)]^ ^ = believes that the this family of subjective densities, the uncertainty down by 1/m, and /^^ converges to unit becomes sure about the informativeness of the mass at p' as 1 oo, so that signals in the limit. In other words, oo, this family of subjective probability distributions leads to (and ensures asymptotic learning; see the proof of Part m^ of Theorem approximate certainty 5). '° Convergence will be uniform in most cases in view of the results discussed following Definition 1 below (and of Egorov's Theorem, which links pointwise convergence of a family of functions to a limiting function to uniform convergence, see, for example, Billingsley, 1995, Section 13). '^This formulation assumes that p\ and pg are equal. We can easily assume these to be different, but do not for such differences in the context introduce this generality here to simplify the exposition. Theorem 8 allows of the more general model with multiple states and multiple 20 signals. The next theorem characterizes the class of determining functions / for which the resulting famil}' of the subjective densities s /^ leads to asymptotic agreement under approximate ^jj [ certainty. Theorem 5 Suppose that Assumption subjective densities (iii) + p^ — 1,1?^ 1. lim„^^oo 2. Suppose that rh G "^e/zned in (8) for fgmf Suppose that f [mx) above. of {p^ I — P^\)- (</'j>o,m R (p^ > R (^p^ 0, there exists = — — 1, \p^ ^/ ^"^^ °^^y 1,2, 1/2, with to consider the family of f satisfying conditions (i)- R{x, y) over a neighborhood ^ {P^ + P^ - ^f p^\) — 4>l,m {s)\>e)<S 0. Then for every l^;,,^, (5) + p^ — ttt. - l,\p^ e Z+ such — p'^\) ^0. Then e > 1, - P^l) = \f and > S 0. 0, there exists (Proof of Part Hence, by (5), definition, we lim^^oo there exists e = > 1,2). such that for each >l-5 = (Vm > m,2 1,2). Let R!^ (p) be the asymptotic likelihood ratio as defined in 1) One can with subjective density fg ^. (4) associated (Vm > m,i that: Pr'flim \cl>l^is)-cf>l^{s)\>e) Proof. — such that Z-i- Suppose that 5 > p* {my) uniformly converges <PL,m, (P')) +p'^ Pr' ( lim 3. some i Then, - (P') /f For each holds. 1 {(Plc,m (p') - ^,m (p')) "= ^ easily check that limm->oo Rm RL (p') if and only if lim^^oo (pO = 0. ~ '^• By have: m-^oo m-^oo fym-ypp)) = R{l-p'-p\f-p^) = i?(pl+p2_l_|pl_p2Q^ where the last equality follows that lim,„-^oo Rln (p') = by condition the symmetry of the function /. This establishes (i), (and thus lim^^oo ('/''oo.m {f) - </4>,m {?')) = 0) and only if if .R(pi+p2-l,|pi-p2|) =0. (Proof of Part 2) Take any By Lemma 1, there exists e' > e > and 6 > such that 0, and assume that </)Jx3.m 21 (P i^)) > 1 R [p^ + p^ — — e whenever 1, \f?- ff [p (s)) — < p^|) e'. = 0. There also exists xo such that /Xo f{x)dx>l- 5. (9) Xo Let K, = min^-gr ^u^j;;,] > (x) / 0. Since / monotonically decreases to zero in the < above), there exists xi such tl>at / {x) 0. Then, for any m > mi and p{s) G c'k (p' whenever \x\ > Let \xi\. mi = — xo/m,p' + xo/m), we have + xi) / (xq — \p (s) tails (see (ii) 1 ~^) > (2p^ > xi/m, +p*| and hence Therefore, for all m> mi and p{s) £ {p^ — xo/m,p' + xo/m), we have that 0L,m(p(s))>l-e- Lemma 1, Again, by e". Now, for there exists e" > (10) such that 4^^^^ iP (*)) > ^~^ whenever Bin {p (s)) < each p{s), \im B?^{p{s)) m—>oo = R{p{s)+f-l,\p{s)-f\). " (11) ' Moreover, by the uniform convergence assumption, there exists uniformly converges to B[p{s) + fP - — fp\) on \,\p{s) (p' — > 77 T],p^ such that Bin {p (s)) + r)) and B{p{s)+p'-l,\p{s)-f\)<e"/2 for each p uous at 0). — (s) in (p' (p^ + p^ - +t]). Moreover, uniform convergence also implies that ri,p^ 1, |p^ ~ P^l) < Hence, there exists m,2 Bi, [p Therefore, for all m > m,2 (s)) (ai^d in this part of the proof, by hypothesis, 00 such that for m > m.^ all <B{p [s) +fP-l, and p [s] £ (p' — \p + T],p^ and p {s)-fP\)+ rj), (s) 6 e"/2 (p' < — it B is contin- takes the value r],p'^ + r]), e". we have (t^oo,m{p{s))>l-e. Set (p' m s Then, by max{7Tii,m2,77/xo}. - xo/m,p' +xo/m), we have Pr^ (|(^'^^„ ip{s)) - establishes that Pr* <jP^^^^ {p{s))\ {\<Plo,m. (p (•s)) (10) iP (s)) - <t\e^A) > I \<t>lc^m - <Pio,m (pi^)) 22 | and (12) (12), for < £ (pi^^m iP is))\ < By S. e) > 1 - the (^ any m Then, symmetry for m > > m. fh and p{s) e (9) implies that of A and B, this (Proof of Part 3) Since linim—oo-Rm strictly positive, Vimm-^oo 4>io,m < (f) = R[p^ ^p^ - (p*) We 1- set = e (l - 1, \p^ ~P^|) limm-.oo is assumed to be /2 and use <?!4),m (P*)) similar arguments to those in the proof of Part 2 to obtain the desired conclusion. Theorem complete characterization of the conditions under which approximate 5 provides a certainty will lead to asymptotic agreement. In particular, certainty ensures asymptotic learning, may it agreement under tainty may be The full certainty. Theorem 5, shows that, while that there will always be asymptotic 1 shows that even a small amount of uncer- instead, cause disagreement between the individuals. sufficient to part of the theorem provides a simple condition on the first of the distribution tail / that determines whether the asymptotic difference between the posteriors approximate uncertainty. This condition can be expressed R{p'+f-l,p'-p')^ The theorem shows that approximate not be sufficient to guarantee asymptotic agree- This contrasts with the result in Theorems ment. it if this condition lim \, small under as: ,./ -.2^^ -"• 13 then as uncertainty about the informa- satisfied, is is tiveness of the signals disappears the difference between the posteriors of the two individuals will become negligible. Notice that condition (13) Intuitively, condition (13) individual will learn. is is symmetric and does not depend on related to the beliefs of one individual on whether the other Under approximate certainty, we always have that limm->oo R\n (p') so that each agent believes that he will learn the value of 9 with probability agreement (or lack thereof) the value of When R 0. +fP'~l, |p^ ~P^|) frequency of p^ in the asymptotic distribution frequency be p^ when , p^ is still R(j)^ . Therefore, individual 1, who = 0; is l,\p^ — fP'\) ^ 0, and thus expects to learn R (p^ + J? — — \, \p^ learn the true state p^l) when ^0, an individual limiting frequency the limiting frequency will be p^ (or 1 — p^), 0; Asymptotic will learn when the hmiting almost certain that the limiting frequency will individual signals will be p^ — an individual who expects a limiting will still learn the true state believes that individual 2 will reach the + f? — 1. depends on whether he also believes the other individual {p^ i. 1 same still is himself. who inference as himself. In contrast, certain that limiting frequency of However, he understands that, when expects a limiting frequency of p^ will happens to be p^. Since he is fail to almost certain that he expects the other agent not to learn the truth and thus he expects the disagreement between them to persist asymptotically. Parts 2 and 3 of the theorem then exploit this result and the continuity of 23 R to show that the individuals will attach probability when their beliefs will disappear disagreement when (13) mate fails to the eveift that the asymptotic difference between 1 (13) holds, and they will attach probability 1 to Thus the behavior to hold. asymptotic of asymptotic beliefs under approxi- certainty are completely determined by condition (13). Theorem on whether means 5 establishes t^at R (p^ +p^ — whether or not there ~ 1, \p^ equal to i^ P^]) determining distributions /. for will We 0. be asymptotic agreement depends next investigate what this condition Clearly, this will depend on the tail behavior of /, which, in turn, determines the behavior of the family of subjective densities ifg^}- Suppose X = p^ + p'^ — 1 > p^ — p^ = y > Then, condition (13) can be expressed as 0. lim m—>oo = 41^ / [my) 0. This condition holds for distributions with exponential On normal distributions. the other hand, it fails for distributions example, consider the Pareto distribution, where / (x) Then, for such as the exponential or the tails, is with polynomial proportional to |xp" for some will a negligible amount 1. be independent of m to converge even fail when of uncertainty. In fact, for this distribution, the asymptotic beliefs (since R\^ does not =A the asymptotic posterior probability oi d ,,: ( ( depend on m). according to If we take tt^ = = tt^ 1/2, then is i (p(s)-p')"" \\- +(p(s)+p'-l) [p{s)-p') for a > >o This implies that for the Pareto distribution, individuals' beliefs will is For each m, fjmx) ^ (^\~°' there tails. any m. As illustrated in Figure 2, in this case (p^^ the previous subsection explained to breakdown). To case, (/>^„,(p(s)) see the is why it ^ is not monotone (in fact, the discussion in had to be non-monotone asymptotic agreement for magnitude of asymptotic disagreement, consider p approximately 1, and <Pio,m (P i^)) is (s) = p'. approximately y~"/ (x~" In that -f Hence, both individuals believe that the difference between their asymptotic posteriors j/~°). will be ™— I'rcxj.m This asymptotic difference is r'oo.Tnl C-" -I- J/~" increasing with the difference y = p^ ~ P'^, which corresponds to the difference in the individuals' views on which frequencies of signals are most 24 likely. It is Figure limn^oo 2: 4>n i^) fo^ Pareto distribution as a function of p (s) \a — = 2, p^ 3/4.] ^ also clear from this expression that this asymptotic difference will converge to zero as y (i.e., — as p^ p'^). > Proposition D= A {{x,y)\ > and This 1 In — me 1 last statement Theorem < X < < l,\y\ any such p^ and assume y} for some y \p^ > 0. — p^\ >e this proposition, notation in that proof. Since A in addition, lim Pr' X, there exists indeed generally true such that whenever (0, cx)) Proof. To prove 5, is R is R that is when continuous: > e and > 5 (Vm > m,i = <<5 of Part 2 of continuous on the compact set D 1,2). Theorem and R {x, 0) = p^. Then, by the uniform convergence assumption, there exists ~P^|) < ^"/^ whenever R (p (s) + p* — 1, \p{s) - p' |) on (p* \p^ — rj r],p'^ each for ~ P^\ < such that that R^i {p (s)) uniformly converges to and use the 5 > 1, \p^ there exist 0, A, we modify the proof R (p^ +p^ — is continuous on the set Then for every < R ^- > Fix such + rj) and Ri^p[s)+p' -l,\p{s)-fP\) <e"l2 for each p Theorem (s) in (p' — ri,p^ + tj). The rest of the proof is identical to the proof of Part 2 in 5. This proposition impHes that of signals, then if the individuals are almost certain about the informativeness any significant difference in their asymptotic beliefs must be due to a significant difference in their subjective densities regarding the signal distribution that when \p^ p^ — p^l = p^, is not small). In particular, the continuity of we must have R (p^ +p^ — 1, \p^ 25 — p'^j) — 0, R in (i.e., it must be the case Proposition 1 implies that and thus, from Theorem 5, there pi significant differences in asymptotic beliefs. Notably, however, the requirement that be no will = p2 be asymptotic agreement for all p^jP^ > 1 established that mider certainty there will 1/2. R worth noting that the assumption that It is also relevant range is important illustrated a situation in bounded away from We Theorem rather strong. For example. jg for the results in Proposition which this assumption next focus on the case where p^ ^ and the asymptotic failed Example differences 1 remained p'^. agreement under approximate certainty. first define: A Definition 1 density function f has regularly- varying tails if satisfies l^ = Mm > for any x The condition in Definition from the fact that H{x)H{y) 1 that = it H (x) is only possible any compact subset of (0,oo). tails, is implies that in the limit, the function H{xy), which M e H (•) has unbounded support and : if H{x) H{x) 1 = x""' for This implies that we have assumption) are satisfied. same expression as for the Pareto distribution, = x"" for but nevertheless has a £ (0, oo). a G This follows Moreover, Seneta (0,00).-^^ holds locally uniformly, then the assumptions imposed in Theorem 5 In fact, relatively weak, must be a solution to the functional equation (1976) shows that the convergence in Definition in H{x) e it 0. important implications. In particular, X continuous in the p^ and provide a further characterization of which classes of determining functions lead to asymptotic We (p) ^^ In particular, recall that 1. gap between p^ and zero, irrespective of the Rm or lirn^^o if i.e., uniformly for a density / has regularly-varying (in particular, that, in this case, R the uniform convergence defined in (7) is given by the X R{x.y) \y and is everywhere continuous. As this expression suggests, densities with regularly- varying behave approximately varying . tails '"To see this, like can be written as /(x) Haan = in the tails; indeed a density £(x)x~" note that since limm^oo {f{mx)/f{m)) //(.y)= See de power functions (lp^)= lim ^-^00 \ f(m) J for = H (x) lim 26 (x) with regularly- some slowly-varying function C (with £ R, we have ffe)^) =//(.) Tn-.^\f(my) f {m) J (1970) or Feller (1971). / tails ff(,) = limm->oo 'C(mx)/£ (m) and 1). Many common distributions, including the Pareto, log-normal, t-distributions, have regularly-varying densities. Definition 2 A We also define: density function f has rapidly- varying tails x>l if 4t^ lim " " / (m) ^ ^ - x-°° = < As > if oo if ' x = ^ X < , . •' for any x 1 if it satisfies l , I 0. above convergence holds locally uniformly (uniformly in Definition 1, the compact subset that excludes Examples 1). x over any in of densities with rapidly- varying tails include the exponential and the normal densities. Prom totic these definitions, the following corollary to agreement under approximate certainty to the Theorem tail 5 is immediate and links asymp- behavior of the determining density function. Corollary 1. 1 Suppose that 1 holds and p^ y^P^- Suppose that in Theorem 5 f has regularly-varying that for each 5 > 0, there exists rh Pr'fhm 2. Assumption S W (s)|>e) |(^^,„(s)-</.2 varying tails there exists (Vm > m,i = >1_<5 e > such tails. Then for every e 1, 2). > and 5 > 0, 1^+ such that ( lim |</.^,„ (s) - <„ {s)\>e)<5 This corollary therefore implies that whether there on whether the family Then ^J^j^ such that Suppose that in Theorem 5 f has rapidly-varying there exists fh tails. will (Vm > m,i = 1,2). be asymptotic agreement depends of subjective densities converging to "certainty" has regularly or rapidly- (provided thatp-^ ^ p^). Returning to the intuition above. Corollary that the failure of asymptotic agreement about limiting frequencies, i.e., p^ ^ fP, is 1 and the previous definitions make it clear related to disagreement between the two individuals together with sufficiently thick probability distribution so that an individual who 27 tails of the subjective expects p^ should have sufficient uncertainty when confronted with a Umiting frequency of p^ this sufficient for is both individuals to believe that they but that the other individual become Along the . do will fail to lines of the intuition given there, will learn the true value of Q themselves, Rapidly-varying so. imply that individuals tails model of the world and thus when individual relatively certain of their by sampling limiting frequency p close to jDut different from p', he will interpret this as driven ~ variation and attach a high probability io 9 A. This will guarantee asymptotic agreement between the two individuals. In contrast, with regularly-varying from certainty, limiting frequencies different but as potential evidence for 6 = The previous be interpreted not as a sampling variation, B, preventing asymptotic agreement. section provided our signals. In this section, and L > states K To = signals. The main an environment with two states and two an environment with results generalize to e and results parallel those of Section 2, all K>2 the proofs for Appendix. Q= {A^, ...,A^^ is a set K >2 distinct elements. We refer to a generic element of the set by A'^. Similarly, ...,a^}, with L > K signal values. As before, define s = {st}^^, and for each {a-^, 1, ...,L, let r^(s) be the number of times the signal large numbers imphes surely converges to pis) = the set p' (s) G = [0, 1] (pi(s),...,p^(s)), where a' out of n first a'| both individuals, with YliLi A(L) = {p - ~ P^ i^) (pi, . . . Once signals. for ^- each I again, the strong law of = 1, Define p{s) e ,p^) e [0, 1]^ L, r^ (s) /n almost ..., A (L) as the vector Eti p' = : 1 } . and let S be With analogy individual a} St = #{i<n|si = that, according to some 5= — results in generalize our results to this environment, let 9 E Q, where containing let St main we show that our main this section are contained in the s p' will even under approximate tails, Generalizations 3 / observes a i i < s : lim„_>oo ?"i to the two-state-two-signal assigns to ^ when 6 5 = A'', tt' the true state is 9. = (s) /n model exists for each immediate generalizations of Theorems 1 = in Section 2, let 7r| (ttj, ...,7r^), When I 1, ..., > L (14) \ be the prior probability and pg be the frequency of observing signal players are certain about pg's as in usual models, and 2 apply. With analogy to the joint subjective probability distribution of conditional frequencies p 28 before, = we define Fg as (pg, ...,p^) according to individual to Since our focus i. Assumption Assumption and is 2 For each and i the distribution Fq over 0, = also define (pl^ni^) probability that 6 — P^* [9 — lines discussed in {st}"=o) ^°^ ^^^^ ^ A'^ \ 4,oo(p(s))= lim n—>oo ' it is Remark ~ 2 above. l,...,K as the posterior observing the sequence of signals {st}^^Q, and A'' after this structure, A(L) has a continuous, non-zero A(L). This assumption can be weakened along the Given similar 1. finite density pg over We we impose an assumption learning under uncertainty, 4>l^nis)' straightforward to generalize the results in Section define the transformation T^ : 2. Let us now M.^ -^ M.^~^, such that Tk{x)= (^,k'e{l,...,K}\k Here T^ (x) is taken as a column vector. theorems and the proofs. In particular, This transformation this transformation will will play a useful role in the be apphed to the vector priors to determine the ratio of priors assigned the different states by individual define the norm ||x|| The next lemma Lemma — max; |2;|' generalizes for x — Lemma (x^, . . . i. tt^ of Let us also ,x^) G M^. 1: 3 Suppose Assumption 2 holds. Then for all s G 5, Ek'^k^lfLApi.^)) 1 Our first theorem Theorem in this section parallels 2 there will be lack of asymptotic learning, and under a 3 and shows that under Assumption relatively weak additional condition, there will also asymptotic disagreement. Theorem 1. 6 Suppose Assumption 2 holds fori VT'{<t^,^{p{s))^l\e = A^) = = l,and 29 1,2, then for each k — 1, .-.^K, and for each 2. Pr^ {\<i>l^ (pis)) 0) The 0) = 0, = and F^ - ^ <Pl^ ipis))\ = F^ for each 6* = 0) 1 wheneverPi^in {n')-n {ir''))'n{f{p{s)) = W{{Tk {TT'^))'Tk{f{p{s)) = e 0. Theorem additional condition in part 2 of plays the role of differences in priors in of the vector in question). In particular, relative asymptotic hkelihood of some 6, that Theorem [iv^) 3 (here " -Tk ' " denotes the transpose then at some p(s), the this condition did not hold, if states could be the same according to two individuals with different priors and they would interpret at least some sequences of signals in a similar manner and achieve asymptotic agreement. Pr^((Tfc (tt^) it - Tfc (7r2))'Tfc(/'(/9(s)) = = 0) did not hold, a small perturbation of is or tt^ important to note that the condition that It is tt^ weak and holds generically— i.e., relatively would restore The Part it.^*^ 2 of Theorem if 6 therefore implies that asymptotic disagreement occmrs generically. The next theorem shows that small differences in priors can again widen after observing the same sequence of signals. Theorem each p G 7 Under Assumption — each k [0, 1], vectors n^ and tt-^, \<Pk,oo 1, 2, ,K, assume 1' = where 1 [Tk (1, ..., - Ufg (p))^g0] 1)'. Then, there exists an open set of prior Tk f (/| (p))^gej j ¥= for such that iP (s)) -9^fc,oo(p(s))| > kfc-7r|| for eachk = l,...,K and s e S and Pr' The (kioo condition 1' (P (s)) ( condition in part 2 of Tfc ( - '/'loo (P(s))| (/g ip))g^Q) Theorem 6, generically. Finally, the following — > KJ - Tk ( 7r||) = 1 (/| {p))g^Q) ) 7^ and as with that condition, is it is theorem generalizes Theorem tion of the families of probability densities is = for each k 5. 1, ..., K. similar to the additional relatively weak and holds The appropriate construc- also provided in the theorem. '^More formally, the set of solutions S = {(7^^7r^p) € l\{Lf {Tk (tt^) - Tk (7r^))'Tfc(/'(p)) = 0} has Lebesgue measure 0. This is a consequence of the Preimage Theorem and Sard's Theorem in differential topology (see, for example, Guillemin and Pollack, 1974, pp. 21 and 39). The Preimage Theorem implies that — y, then f~^{y) is a submanifold of if y is a regular value of a map / with dimension equal to dimX — dim Y. In our context, this implies that if is a regular value of the map {Tk (tt') —Tk (7i""))'Tfc(/'(p)), then the set S is a two dimensional submanifold of A(L)^ and thus has Lebesgue measure 0. Sard's theorem : : implies that is X X generically a regular value. 30 Theorem 8 Suppose that Assumption 2 holds. subjective density fg ^ ever 9 ^ 9' , = '^/ and f Jp^^in f M^ — : Q and m 6 Z_|_, define the by = cii,e,m)f{mip-p{i,e))) fi^{p) where c{i,9,m) For each 6 E ('7^ ]R is > (p - p{i,6))) G dp, p{i,6) A (L) (16) withp{i,9) ^p when- {i,0') a positive, continuous probability density function that satisfies the following conditions: (i) limft^oo niax|3..||^||>/j} (x) / = 0, R{x,y)^ (17) and exists at all x,y, (Hi) ^^^ lim m^oo / (my) convergence in (1 7) holds uniformly over a neighborhood of each {p{i,9)-p{j,9'),p{i,9)-p{j,9)). Also let 4>k^oo,m (P i^)) subjective density fg 1. ^ ^^ ^^^ asymptotic posterior of individual i with Then, . lim^^oo (fk,oo,m — l™n^oo 4>knm i^) {P ih A^)) - ^fc,oo,m {p (^'^'')) j = ^ «/ ""^ Only if R(p{i,A^)-p(^j,A'''yp{i,A'')-p{j,A''))^Oforeachk'y^k. 2. Suppose that for every e R [p > {i, 9) and 5 > —p (j, 9') — p (j, 9)) = ,p{i,9) m 0, there exists G Z+ —p Suppose that R [p (i, there exists e > such that for each 5 9) Pr^ (l|0L,m(5) (j, 0') - ,p{i,9) > <^^,^(5)1| (Vm>7n,2 = These theorems therefore show that the — p {j, 9)) ^ 0, there exists > e) > results 9' . Then such that W[\\4>l^,m{s)-4L,m{s)\\>e)<5 3. for each distinct 9 and 1 - <5 l,2). for each distinct 9 and m ^1.+ 9' . Then such that iym>m,i = 1,2). about lack of asymptotic learning and asymptotic agreement derived in the previous section do not depend on the assumption that 31 there are only two states and binary signals. 1 and Corollary The 1 It is also to the case with multiple states straightforward to generalize Proposition and signals; results in this section are stated for the case in and states are finite. and states, They can we omit this to avoid repetition. which both the number of signal values also be generalized to the case of a continuum of signal values but this introduces a range of technical issues that are not central to our focus here. Applications 4 In this section we are chosen to show various discuss a number different under uncertainty. Throughout, we illustrates how application and shows how fourth example illustrates in bargaining. Damme how arise (1993). some simple The third example is the most substantial The example shows how a special case of our model of learning when there is information transmission by a potentially biased 4.1 Value of Information in Common-Interest Consider a common-interest game interest from basic game learning under uncertainty can affect the timing of agreement outlet.^"* common example learning can act as an equilibrium selection media cally in insights first learning under uncertainty affects speculative asset trading. Finally, the last under uncertainty can The strive to choose the simplest examples. learning under uncertainty can overturn device as in Carlsson and van applications economic consequences from learning and disagreement The second example shows how such theory. The of applications of the results derived so far. in Games which the players have identical payoff functions. Typi- games information is valuable in the sense that with about underlying parameters, the value of the game in the best equihbrium would therefore expect players to more information will We be higher. collect or at least wait for the arrival of additional informa- '^In this section, except for the example on equihbrium selection manipulation, we study complete-information games with possibly and belief structure in these n < oo, consider the information partition /" games can be described = {/" as follows. (s) = and the last non-common example of the game of priors. Fix the state space {{d,s') \s[ = st^t < n} |s belief Formally, information Q =Q x S, and € 5} that is for each common for both players. For n = oo, we introduce the common information partition /°° = {/°° (s) = 9 x {s} \s e S}. At each /" (s), player i = 1,2 assigns probability (f>l^ (s) to the state 9 = A and probability 1 — 05, (s) to the sate 6 = B. Since the players have a common partition at each s and n, their beliefs are common knowledge. Notice that, under certainty, ip^^ (s) = 0^ (s) e {0, 1}, so that after observing s, both players assign probabihty 1 to the same 6. In that case, there will be common certainty of 9, or loosely speaking, 9 becomes "common knowledge." This is not necessarily the case under uncertainty. 32 We now tion before playing such games. show that when there is learning under uncertainty, additional information can be harmful in common-interest games, and thus the agents game prefer to play the To before additional information arrives. matrix illustrate these issues, consider the payoff a a where = ^ G 6' 1 is TT (3 29,29 1/2,1/2 1/2,1/2 1-9,1-9 and the agents have a common prior on 9 according to which probability of {0, 1}, e (1/2, 1). When there value of the payoff from (a, a) from (/3,/3) is strictly less is no information, a dominates than 1/2). In the dominant-strategy equilibrium, First, consider the implications of learning = 9\9) = p' > the agents will learn will learn that a, and hence P and (a, a) Theorem 1/2. 9. 9—1; If 1 = the frequency p (s) = 0. the dominant strategy equilibrium. = if 1. greater than 1/2, they 1 is Up (s) < 1/2, P strictly dominates > 1/2, a strictly dominates If p{s) Consequently, when they certainty before playing the game, the expected payoff to each player This implies that, each player {st}^j, where each agent believes that of signal with sj the dominant strategy equilibrium. is > {a, a), then implies that after observing the sequence of signals, otherwise they will learn that 9 (/3,/3) is (since the expected under certainty. Suppose that the agents are allowed to observe an infinite sequence of signals s [st /? than 1/2 and the expected value of the payoff strictly greater is strictly receives 29 with probability n, thus achieve an expected payoff of 27r Pr' may is 27r -|- (1 learn under — tt) > 27r. they have the option, the players would prefer to wait for the arrival of public information before playing the game. Let us next turn to learning under uncertainty. In particular, suppose that the agents do not know the signal distribution and their subjective densities are similar to those in Example 2: (l - e - e^) ^ for each 9, where the limit where e < -^ 6 (5 — > if f -5/2<p<f + 5/2 (18) otherwise < p^—p^ and and /5 ifp<l/2 e 0, e and 6 are taken to be arbitrarily small or loosely speaking, where 33 e = and 6 = (i.e., we consider 0). Recall from Example 2 that when e = and 5 = the asymptotic posterior probabihty of ^ 0, a 1 'PLiPis)) < l-p'-5/2, + 6/2 < p{s) < p{s) -p' or 1 or f- 5/2 <p{s) <f + 6/2, = 1 is 1/2, otherwise. As discussed above, when value of 9, e ^ and 6 = each agent beheves that he will learn the true 0, while the other agent will reach the opposite conclusion. agents expect that one of them will have (/>^ Consequently, the unique equilibrium will be (a, payoff of 1/2, which is strictly less arrival of information (which agents may is 27r). prefer to play the game and van Damme may be due (1993), we now 4>]^ ip{s)) = 0. giving both agents an ex ante expected there is learning under uncertainty, both Games initial difference in players' beliefs it while the other has before the arrival of public information. The prior; when Therefore, Selection in Coordination common /3), 1 than the expected payoff to playing the game before the 4.2 of — ip{s)) This implies that both about the signal distribution need not be due to lack on an example by Carlsson to private information. Building illustrate that distribution, small differences in beliefs, on the outcome of the game and may when the players are uncertain about the signal combined with learning, may have a select significant effect one of the multiple equilibria of the game. Consider a game with the payoff matrix _I where 9 ~ A/" (0,1). where G {0, 1} St The 77 9,9 9-1,0 N 0,9-1 0,0 players observe an infinite sequence of public signals s ~ A/'(0, 1). = l\9) = l/{l + exp{-i9 + Ui is {st}^o> rj))), + Ui uniformly distributed on [— e/2, e/2] for some small Let us define k (19) In addition, each player observes a private signal Xi=r] where = and PTist with N I = \og{p{s)) the players infer that 9 + t] = k. — log(l — e > 0. p{s)). Equation (19) implies that after observing For small e, conditional on Xi, 34 77 is s, distributed approximately uniformly on [xi and ditional on Xi Now — e/2,Xi 9 s, is + e/2] approximately uniformly distributed on note that with the reverse order on extremal rationalizable strategy Nash tion / Xj if < X* and x, A'^ if between the two actions, O (e) where is This x*. = such that lime_<o when < K- 1/2 and — Xi — e/2, k— + Xi e/2]. supermodular. Therefore, there exist is monotone, symmetric Bayesian also constitute a cutoff value, x*, such that the equilibrium ac- cutoff, x*, small, the N Xi > k + > Pr(xj O (e) = e is if is game [k defined such that player is i is indifferent i.e., X* Therefore, the which profiles, > K.-X* Xi Xi, Equilibria. In each equilibrium, there is and van Damme, 1993). This implies that con- (see Carlsson game = X*) = 1/2 + (e) This establishes that 0. = x*\xi - AC is 1/2 - O (e) dominance . solvable, and each player i plays / if 1/2. In this game, learning under certainty has very different implications from those above. Suppose instead that the players knew the conditional signal distribution so that we common are in a world of learning under certainty. Then after s how they knew observed, 9 would is knowledge, and there would be multiple equilibria whenever 6 G therefore illustrates (i.e., (0, 1). rj), become This example learning under uncertainty can lead to the selection of one of the equilibria in a coordination game. A 4.3 One Simple Model of Asset Trade of the trading. most interesting applications of the ideas developed here Models of assets trading with different priors have Harrison and Kreps (1978) and Morris (1996). "speculative asset trading" for the . We now They models of asset by, among different priors others, about establish the possibility of investigate the implications of learning under uncertainty pattern of speculative asset trading. Consider an asset that pays 2 been studied These works assume the dividend process and allow for learning under certainty. to is owns the asset, but Agent 1 1 if the state may wish to is buy A it. and We if the state is B. Assume that Agent have two dates, r = the agents observe a sequence of signals between these dates. For simplicity, to be an infinite sequence s = {st}'^-^- We also simplify this 35 and t — I, and we again take this example by assuming that Agent 1 has the bargaining power: at either date, all if he wants to buy the asset, Agent a take-it-or-leave-it price offer P^, and trade occurs at price Pr Assume also that Agent would 1 > tt^ like to tt^, so that Agent purchase the game. this We asset. more 1 is 1, Pa both agents recognize same both of them: for at each history p{s). 1 — Pb = at r = it is that at r > p* In that case, from 1/2. = 1, for each p ,0 (s) > 1/2 and 1 if Therefore, and the continuation value of Agent is 0, Agent 0, 2 accepts offer the price Pq = '""^ not offer if p (s) — know p.4 and pB for and only = Agent Agent approximately 1 is 1 To 1, = n'^, 1. 0, his with ir^ = 4>to = 1 (P i^))) the continuation value of 0, they trade at price Pq, then the If Since . tt^ > tt'^, Agent 1 is happy to Therefore, with learning under place. 0. e ^ 0). trading, which — 0. > is Now, continuation value 1 = <^^ {p (s)) 1 — 1 is tt'^, p^ — 5/2 and the value < p (s) < of the asset 2 at Since the equilibrium payoff shows that when they do not at least (1 - TT^) value of Agent 2 must be at least sum 1, if buys the asset from Agent tt^. Therefore, they can trade at date 7r\ exceeds the asset at date at date in all other contingencies, this only tt^, if of these continuation values, n^ (l impossible, there will be no trade at r buy the e now have consider a simple example where first Hence, at such p{s), Agent The continuation is we enjoying gains from trade equal to of never selling his asset. to Pq > Agent 2 Tfl is = [p (s)) (j)]^ r Pq and Pq, respectively. This imphes that at illustrate this, Example must be non- negative trade at date (when if — and trade takes 5/2, then the value of the asset for price Pi (p(s)) of tt^ is tt^. 1/2. Hence, at In contrast to the case of learning under certainty, the agents subjective densities are as in + = < next turn to the case of learning under uncertainty and suppose that the agents do an incentive to delay trading. p^ if at date r be 2 will be immediate trade at r certainty, there will We an and 1 2 Theorem the value of the asset will the (s), trade does not occur at r if Suppose that each learning under certainty. some number P^ for be worth will continuation value of agents date assumption ensures that optimistic. This the agents will be indifferent between trading the asset (at price Pi Agent offer. ^ certain that is Agent 2 accepts the if are interested in subgame-perfect equilibrium of Let us start with the case in which there agent makes 1 = (provided that p 0. (s) Instead, Agent 1 will as he has the option the total payoff from — tt'^) + tt^. Since this wait for the information turns out to be in a range where he concludes 36 that the asset pays 1). This example exploits the general intuition discussed after Theorem uncertain about the informativeness of the signals, each agent may 4: if the agents are expect to learn more from the signals than the other agent. In fact, this example has the extreme feature whereby each agent believes that he will definitely learn the true state, but the other agent will This induces the agents to wait so. towards common and make trades beliefs learning under certainty and .trade at the less likely. the reason is common why do to information before trading. for the arrival of additional This contrasts with the intuition that observation of fail information should take agents This intuition is correct in models of previous models have generated speculative beginning (Harrison and Kreps, 1978, and Morris, 1996). Instead, here there is delayed speculative trading. The next result characterizes the conditions for delayed asset trading Proposition 2 In any subgame-perfect E2[,/,y That is, when 'PIo (p('S)) > > n'^ E^ [min (plc (p(s))>' tt^ < E^ Agent his valuation of the asset take place at the price Pq At date 1, Agent max{(/)^ {p{s)) only 1 — = n'^, [min {(^^, f/)^}], Agent (j?^ is Pr^ {6 =A while trade at r buys the asset if — delayed to t does not buy at t 1 Proof. In any subgame-perfect equilibrium. Agent 2 and hence is generally: 1 if and only if =7r2>Ei[min{^^,0^}]. {<?^^,(/'oo}]' when equilibrium, trade more and only = | is = and buys buys at t 1 indiiTerent = at if (p]^ will between trading and not, be at the price Pi (p {p (s)) {p{s)) ,0}. This implies that Agent > 0^ 1 is 1 if 0. Information). Therefore, trade at r 1 = t (s)) = — 0^ can {p (s)). {p (s)), yielding the payoff of willing to buy at r = if and if TT^-TT^ > Ei[max{(/.i,(p(s))-,^^(p(s)),0}] = B^ [</)L = tt' -E' (P (s)) - min {4>l, {p is)) , 4>l {p is))}] [min {(Plip is)), 4>lip{s))}], as claimed. Since n^ ~ E^ [<?^io] difference in beliefs, tt^ — ^^ [min — tt^, in {(/>^, (fto]\ > ^^^^ result provides terms of the differences 37 a cutoff value for the initial in the agents' interpretation of the signals. The cutoff value [max is E-^ - 0^ [cf)^ {p (s)) (p (s)) = lower than this value, then agents will wait until r , O}] the If . initial difference is to trade; otherwise, they will trade 1 immediately. Consistent with the above example, delay in trading becomes more likely the agents interpret the signals more differently, which This reasoning' also suggests that cutoff value. if evident from the expression for the is = Fg Fg each 6 (so that the agents for The next interpret the signals in a similar fashion), ^^ then trade should occur immediately. lemma shows when that each agent believes that additional information will bring the other agent's expectations closer to his own and will = be used to prove that Fg Fg indeed implies immediate trading. Lemma 4 > If n^ tt^ = and Fg Fg for each El then 9, > n\ [4>l] Proof. Recall that ex ante expectation of individual = E^C,] + f\-K^f\{p)<i>>^{p) i regarding (j>'^ can be written as {l-i:')fh{l-p)4>^oo{p)]dp (20) Jo ~ where the first line f^ n^f^{p) ^^fAip) Jo + {l-n^)fs{l-p) = fe (P) = Now fe (p) ^°^ ^11 p. r(^\^^^'= f (20), E^ [(p^,] increasing in n. = I {n^) ^/-4 (P) and tt^ and + (1 = - ^) /b fact that since = Fg Fg, = E^ [0^] (1 - P) . f.,dp. - p)\JA [P) 2\ t l^ TT^) /s (1 / (tt^). ^^^^^ 'Recall from and the (4) Hence, it suffices to show that / is Now, JO Moreover, /^ ^^ define ^2t (r^^l^ ^ JA [P) + (1 I Jo From ^ uses the definition of ex ante expectation under the probability measure Pr', while the second line exploits equations (3) fe (P) ^ + il-n^)fBil-py^^'^'^'' r2/.4(p) TT-^ (p) / [tt^Ja (p) Theorem + (l 3 that even {fA{p)-fB{l-p))dp. + (l-7r2)/B(l-p) - tt^) when Fq /b = (1 - p)] > 1 if and only if /^ (p) > /b Fg, agents interpret signals differently because 38 tt' (1 ^ - tt'. p). Hence, I'i^) fA{p) 2f ( \^(C'^\f n P)-AfA{p)-fB{l-p))dp {^-T^-) fBi.1JjA>fB'^ fA[p) + = [ -/ 2, (r.^^!^^^''\^f n JfA<fB ^ /^ yP) + (1 - ^ ) JB (1 > . Ub{1- p)- fA[p))dp P) UB{l-p)-fA{p))dp UA{p)-fB{l-p))dp- f [ JfA<fB JfA>fB = - / ifA{p)-fBil-p))dp^O. Jo Together with the previous proposition, this lemma yields the following result establishing that delay in asset trading can only occur when subjective probability distributions differ across individuals. Proposition 3 atT = // Fq 4. Fq for each 9, then m any subgame-perfect equilibrium, trade occurs 0. Proof. Since n^ each p = (s). > iP' Then, E^ [min and {<?!>^, Therefore, by Proposition Fl} — (pto}] 2, Agent i?^, ~ 1 E^ Lemma [^to] - buys at r 1 imphes that (l)^{p{s)) > 4>1oip{s)) ""^i = where the last inequality is by for Lemma 0. This proposition establishes that when the two agents have the same subjective probability distributions, there will be when Fg ^ Fg, delayed when no delay in trading. speculative trading is However, as the example above possible. The intuition is given by illustrates, Lemma 4: agents have the same subjective probability distribution but different priors, each will believe that additional information will bring the other agent's beliefs closer to his own. This leads to early trading. However, when the distributions, they expect to learn Theorem agents differ in terms of their subjective probability more from new information (because, as discussed after 4 above, they believe that they have the "correct model of the world"). Consequently, they delay trading. Learning under uncertainty does not necessarily lead to additional delay actions, however. Whether it in economic trans- does so or not depends on the effect of the extent of disagreement on the timing of economic transactions. We the presence of learning under uncertainty will next see that, in the context of bargaining, may be a force towards immediate agreement rather than delay. 39 Bargaining With Outside Options 4.4 Consider two agents bargaining over the division of a and Agent ^ and the value of 9 ai can be thought to be more = 1 Agent 2 accepts the receives the remaining 1 — the is costly, so that if game Agent r. 1 offers a share Wr to Agent ends. Agent 2 receives the proposal, Wr, and Agent Wt- If Agent 2 rejects the offer, = negotiations continue until date r Finally, as in Yildiz (2003), the agents are y she decides whether to take her = E^ Agent 1 > [61] assume that incurs a cost c assumed to be "optimistic," - E^ [0] 1, We Agent 2 believing that her outside option is 0. 0. higher than Agent — and y parameterizes the extent of optimism > in the sense that In other words, they differ in their expectations of 6 on the outside option of Agent 2 option where 1, under 6l- likely offer, {0, 1}, dates, the agents outside option, terminating the game, or wait for the next stage of the game. delay re £ {a^, an}, where the signal {stjt^i with sj Bargaining follows a simple protocol: at each date If Between the two unknown. initially is observe an infinite sequence of public signals s 2. There are two dates, an outside option 6 6 {^Li^h} that expires at the end of date 2 has < 9h < Oi dollar. in this 1 's — -with assessment of this outside game. We assume that the game form and beliefs are common knowledge and look for the subgameperfect equilibrium of this simple bargaining game. By backward 1 is E^ [9\p (s)] < induction, at date r 1, for 1, and hence she accepts an = therefore offers wi = E^ [9\p (s)]. If there is any p{s), the value of outside option offer wi and only if no agreement if wi > for E'^ [9\p (s)]. Agent Agent 2 at date 0, the continuation values of the two agents are: yi which uses the = 1 _ c - E^ fact that there is no cost of delay the agents will delay the agreement to date r e2 Here, E^ [E^ [0|/9(s)]] is Agent after observing the signals s. V^ and [E^ {9\p (s)]] If I's [9] - for = E^ [E^ 1 if = Agent 1 2. and only [e\p (s)]] expectation about Agent E^ [E^ > [9\p (s)]] = E^ Since they have [9] 1 dollar in total, if c. how Agent 2 will update her beliefs expects that the information will reduce Agent 2's 40 expectation of her outside option more than the cost of waiting, then Agent This description makes on Agent I's When it clear that assessment of each agent is whether there how Agent will wilhng to wait. 1 is be agreement at date r = depends 2 will interpret the (public) signals. certain about the informativeness of the signals, they agree ex ante Lemma 4 in that they will interpret the information correctly. Consequently, as in the previous subsection. Agent I's Bayesian updating wall indicate that the public information will reveal him Yildiz (2004) has to be right. shown that "wait to persuade" Agent 2 that her outside option that each agent may i.e., differ. i is certain that Pr* Then, from Theorem Pr' (E2 {9\p{s)] ^ 9) = 1 1, for delays the agreement to date r = — (sj 9\9) Agent this reasoning gives is — p^ > relatively low. More 1/2 for some p^ and an incentive to 1 specifically, p-^, assume where p^ and p^ the agents agree that Agent 2 will learn her outside option, each 1 if i. Hence, E^ [E^ [9\p{s)]] and only = E^ [6]. Therefore, Agent 1 if y>c, i.e., and only if the level of optimism if is higher than the cost of waiting. This discussion therefore indicates that the arrival of public information can create a reason for delay in bargaining games. We now motive show that when agents for delay is are uncertain about the informativeness of the signals, this reduced and there can be immediate agreement. understands that the same signals will Intuitively, each agent be interpreted differently by the other agent and thus expects that they are less likely to persuade the other agent. This decreases the incentives to delay agreement. This result is illustrated starkly here, with an example where a small amount of uncer- tainty about the informativeness of signals removes that the agents' beliefs are again as in bility [l more than —p— (5/2, 1 1 — p^ — -I- e Example 1 all with e small. Now Agent to the event that that p (s) will be either in [p (5/2] , inducing Agent 2 to stick to her prior. that Agent 2 will not update her prior by much. In particular, E^ [e2[0|p(s)]] =E^[9] + 0{e). [9] - E^ [e2 [^Ip (s)]] 41 = -O (e) < c. — 1 assigns proba- 6/2,p^ + Hence, Agent we have Thus E^ Suppose incentives to delay agreement. 5/2] or in 1 expects This implies that agents will agree at the beginning of the game. Therefore, the same forces that led to delayed asset trading in the previous subsection in bargaining when agents can also induce immediate agreement are "optimistic" Manipulation and Uncertainty 4.5 Our final example how intended to show is the pattern of uncertainty used in the body of the paper can result from game theoretic interactions between an agent and an informed party, example as in cheap talk this possibility, games (Crawford and we choose the simplest environment discussion to the single agent setting is Sobel, 1982). Since our purpose to is for to illustrate communicate these ideas and limit the —the generalization to the case with two or more agents straightforward. The environment with a prior belief decision x e [0, 1], tt is as follows. G (0, 1) and The = that his payoff — is state of the world at 1 (x — = t 0) 0. G is At time i = {0, 1}, 1, this agent has to Thus the agent would . and the agent like to starts make a form as accurate an expectation about Q as possible. The other = s' player {sj}^j with with G St € s\ The {0, 1}. a media outlet, is {0, 1}, M, which observes a large (infinite) number of signals and makes a sequence of reports to the agent an exchangeable sequence, we can represent are as denoted by I's, it p' G [0, 1], and similarly s it, (tm: [0,1]^ A ([0,1]) is puts probability Otherwise, i.e., if [0, 1]. t^ i, there is Since s' This is convenient A ([0,1]), [0,1]. Let on the identity mapping, thus corresponding to ctm articles. s' media as a mapping the set of probability distributions on 1 {s(}^j as before, with the fraction of signals that represented by p G is allows us to model the mixed strategy of the where = reports s can be thought of as contents of newspaper articles, while correspond to the information that the newspaper collects before writing the is s i be the strategy that M reporting truthfully. manipulation (or misreporting) on the part of the media outlet M.1'5 We 6 — 1 go (p) ' also assume for simplicity that p' has a and go when 9 = > p and go (p) for all p < 0, such that = g\ {p) = for all p continuous distribution with density gi for ed\ > p- p < p and gi (p) > for all p This assumption implies that See Baron (2004) and Gentzkow and Shapiro (2006) for related models of media bias. 42 if > p, when while M reports truthfully, om = i.e., the superscript of type of p' , a and Let us biased towards is and receives now (0, 1), H — honest). A/3 With is M (unobservable to the honest and can only play a^j probability Aq € (0, 1 — A//), the = = and — Xa — \h, I — (where is outlet receives utility equal to x irrespective x. \ i media outlet manipulate the agent to choose high values of equal to utility the media Type a media 1. like to complementary probability 0, 6 A// type for is and hence would towards be asymptotic learning (and suppose instead that there are three different types of player With probabihty agent). will agreement when there are more than one agent). also asymptotic Now ^ then Theorem 2 apphes and there the media outlet is of type j3 With is the biased x. look for the perfect Bayesian equilibrium of the game between the media outlet and the agent. The perfect Bayesian equilibrium can be represented by two reporting functions a%; : [0, 1] function ^A (p : and ([0, 1]) [0,1] -^ sequence of reports cr^^ [Oil]i is : [0, 1] ^A ([0, 1]) for which determines the belief of the agent that and an action function x p, choice of the agent as a function of p (there is no the two biased types of : [0,1] — > [0,1], loss of generality M, and updating 9=1 when the which determines the here in restricting to pure strategies). In equilibrium, x rule given o""^, a^,^ must be optimal and the prior tt; for the agent given and a"^ and ct^^ (p; <p must be derived from Bayes must be optimal for the two biased media outlets given x. Note p', first without that since the payoff to the biased media outlet does not depend on the true loss of generality, shght abuse of notation, two types report let we can restrict <t^ (p) and a^ a^ (p) and a j^^ not to depend on p' . Then, with a be the respective densities with which these p. Second, the optimal choice of the agent after observing a sequence of signals with fraction p being equal to 1 is xip)^(l> for all p e [0, 1], i.e., {p) the agent will choose an action equal to his behef (f>{p). Third, an application of Bayes' rule implies the following belief for the agent: if 4>{p) = p<p (21) { TAHSl(p)+AaCr^(p)+A;30-M{P) 43 The following Lemma p lemma shows that any (perfect Bayesian) equilibrium has a very simple form: 5 In any equilibrium, there exist < p and (p = (p) Proof. Prom (pA /'^^ (j) exists Pi,p2 media type (3 P < P = But {p). (f) It such that (p (p) = (f)^ for all iov p > > (f>{p) < 't'iPi)- p. tt when p > Now Since the media type suppose that the lemma Then we in that case, p. also have a^ (p^) equation (21) implies that constant over p e [0,p). (l){p) is — -- (j) The proof is false [pi) since — 0, for (j){p) analogous. (p, 1] is lemma from this that p ^p, follows immediately and = (p) contradicting the hypothesis. Therefore, being constant over p e p, such that (piPi) {p) < n (f>g P- we have cr^ {p), minimizes x > n when p < (21), 4>{p) <: a maximizes x{p) = and there O'^l' > n and (j)^ that equilibrium beliefs will take the form given in the next proposition: Proposition 4 Suppose then the unique equilibrium actions and beliefs are: ^M x{p) = cp (p) (22) a^(p)^ffo(p) (23) = (24) <^ ^^fiSfe^ I Proof. Consider the case p < constant over p G p) (by [0, this range. Since this range ^M ~ 9o- Similarly, a^^ = p. Lemma is the As 5), = 91 (P) (P) proof of in the ^fp>-p- Lemma 5, equation (21) implies that common a^^ (p) a^ support of the densities gi. Substituting these equalities in (21), a^ is = 0. Since (j) (p) is proportional to go on and 50, it we obtain must be that (24). This proposition implies that the unique equilibrium of the game between the media outlet and the agent leads to a special case of our model of learning under uncertainty. In particulaj, the beliefs in (24) can be obtained by the appropriate choice of the functions /^ from equation analyzed is (3) in in this Section paper is 2. likely to (•) and fs (•) This illustrates that the type of learning under uncertainty emerge in game-theoretic situations where one of the players trying to manipulate the beliefs of others. 44 Concluding Remarks 5 A key assumption of most theoretical analyses is that individuals have a "common prior," game ing that they have beliefs consistent with each other regarding the • and possible distributions of payoff-relevant parameters. This presumption common the argument that sufficient mean- forms, institutions, is often justified by experiences and observations, either through individual observations or transmission of information from others, will eliminate disagreements, taking agents towards common known theorems Dubins This presumption receives support from a number of well- priors. in statistics and economics, for example. Savage (1954) and Blackwell and (1962). Nevertheless, existing theorems apply to environments in which learning occurs under certainty, that is, individuals are certain about the tions, individuals are not only learning meaning of different signals. many In situa- about a payoff-relevant parameter but also about the interpretation of different signals. This takes us to the realm of environments where learning takes place under uncertainty. For example, many signals favoring a particular interpretation might make individuals suspicious that the signals come from a biased source. learning in environments with uncertainty full identification (in the may We show that lead to a situation in which there standard sense of the term in econometrics and situations, information will be useful to individuals but may not lead to in common which two individuals with different priors observe the informative about some underlying parameter. certainty, since individuals same Our environment full learning. probability distributions of both individuals have also consider an environment infinite is will take sequence of signals one of learning under un- have non-degenerate subjective probability distribution over the likelihood of different signals given the values of the parameter. tions), We priors (or asymptotic agreement). lack of In such statistics). This paper investigates the conditions under which learning under uncertainty individuals towards is full We show that when subjective support (or in fact under weaker assump- they will never agree, even after observing the same infinite sequence of signals. show that this corresponds to a result of will agree, before observing the parameter will "agreement to eventually disagree" common understanding that more information to similar beliefs for agents has important implications for a variety of when there is individuals sequence of signals, that their posteriors about the underlying not converge. This models. Instead, ; We no full may not lead games and economic support in subjective probably distributions, asymptotic 45 learning and agreement An may obtain. important implication of this analysis is that after observing the same sequence two Bayesian individuals may end up disagreeing more than they originally signals, result contrasts with the common presumption of did. This that shared information and experiences will take individuals' assessments-^Ioser to each other. We also systematically investigate whether asymptotic agreement obtain as the amount of uncertainty in the environment diminishes (i.e., as we look at families of subjective probability distributions converging to degenerate limit distributions with We all their mass at one point). provide a complete characterization of the conditions under which this will be the case. Asymptotic disagreement may prevail even under "approximate certainty," as long as the family of subjective probability distributions converging to a degenerate distribution (and thus to an environment with certainty) has regularly-varying or the t-distributions) tails (such as for the Pareto, the log-normal In contrast, with rapidly- varying tails (such as the normal and the . exponential distributions), convergence to certainty leads to asymptotic agreement. Lack of common ior in and common priors has important implications a range of circumstances. We economic behavior interacts with tion, beliefs games of common illustrate games interest, bargaining, asset trading show how the possibility of observing the lay" in asset trading among example why illustrating may of signals may individuals that start with similar beliefs. individuals may be of coordina- For wish to play common- more information about same sequence games and games of communication. results, individuals before rather than after receiving economic behav- the type of learning outhned in this paper in various different situations, including example, we show that contrary to standard interest how for payoffs. Similarly, we lead to "speculative de- We also provide a simple uncertain about informativeness of signals —the strategic behavior of other agents trying to manipulate their beliefs. The issues raised here have important implications for statistics as learning in game-theoretic situations. corresponds to one in which there are well-behaved these posteriors and converge is is well As noted above, the environment considered here lack of full identification. Nevertheless, Bayesian posteriors to a limiting distribution. Studying the limiting properties of more generally and how they may be used econometric models and econometrics as an interesting area for research. 46 for inference in under-identified Appendix: Omitted Proofs 6 Proof of Theorem Under the hypothesis 1. of the theorem and with the notation in (2), we have l-2r„/n' which converges to or oo depending on Mmn^aofn/n is greater than 1/2 or less than 1/2. If lim„_oorn (s) /n > 1/2, then by (2), limn^co'Pl (s) = hm„_.c>o (pl (s) = 1, and if lim„_,oor„ (s) /n < 1/2, then lim„_,oo'An (•s) = liin„_,oo "An (*) ~ 0. Since lim^_oo^n (s) /n = 1/2 occurs with probability zero, this shows the second part. The first part follows from the fact that, according to each i, conditional on 6 = A, lim„_,oor„ (s) /n = p^ > 1/2. Lemma Proof of The proof 3. Proof of Theorem identical to that of is 1. 6. Lemma (Parti) This part immediately follows from as each TT\>f^k' {p{s)) 3, is and positive, is finite. ^ifA^ipis)) Assume Fg = Fg (Part 2) only Lemma — if (Tfc (tt^) T^ (tt^)) Tfc ( Then, by each 9 e @. for = (/e(p))gg@) The 0. Lemma 3, (pl^ao (p) ~ "Afc.oolp) has probabiUty latter inequality = if and under both probability measures Pr^ and Pr^ by hypothesis. Proof of Theorem s (1, kI,oo (P is)) is . 1) , . . - <t>ioc continuous in p Now, of 7f tt = (1/A', . . , . = .lf\Ap{s)) i^lf\.[p{s)) where 1 Define 7. l/K). ' V' = First, take n^ {^^' (P(^))).eej tt- = - T. Then, 7f. [{fe (p(^))),,ej J ^ «, and the inequality follows by the hypothesis of the theorem. Hence, by Lemma 3, for each p {s) e [0, 1]. Since [0, 1] is compact and [0^,^ (p (s)) - 0^_^ (p {s))\ (P (s))| > , (s), there exists since I^J-oq (p(s)) - e > such that 0fc^oo (p(s))| \(i>l^oo continuous in is (P (s)) ~ tt^ and for each 0fc,oo tt^, (P (5))| > each p ^ for (s) E there exists a neighborhood [0, 1]. N {n) such that I^I.c^oIpIs)) -<Afc,oo(p(s))| for all n^ , tt'^ € N {ft). Since Pr' (5) Proof of Theorem Lemma 8. = 1, Our proof > kfc-TT^I A; = 1, ..., /-C and s G 5 the last statement in the theorem follows. utilizes the following two lemmas. A. lim m^oo 0loo,m(P) + Zk'^k^RiP-Pih^'')^P-p(^,A>n) ^ Proof. By condition (i), liinm~>oo c (i, A'' ,m) = 1 for each i and k. Hence, for every distinct k and k\ "^^°° Then, /V vP) Lemma A "1-.00 c(i,A'-,m) follows from m^oo Lemma j [m{p — 3. 47 p{i,A''))) V V / ^ V Lemma p with (s) > B. For any i (s) \\p -p (i, > Q and h (^)) - I Proof. R by hypothesis, Since, < K, and k each II |4,oo,m (P Lemma m> m, rh such that for each 0, there exists < h/m, A'') > A, there exists h' lim m —*oo 0U,- (P (^ ^') < ) continuous at each {p{i,6) is ^- (25) I I — 'p{j,9') ,p{i,9) — p{j,9)), by such that 0, hm - Hm <f>l^^^ {p is)) < A'))\ 4>l,oo,m {P {h e/2 (26) I and by condition > there exists rh (iii), h/h' such that UU,-(P(^))- 1™ 4,^,„(p(s))|<£/2. holds uniformly in (Proof of Part A —p (s) ||p 1) Since A'^) (i, R ip < || h' -p A'') (i, = implies that lim„^oo<?ifc,oo,m(p(^^'')) if and only latter holds Part if if limm^oo only m oo <Afc (p (*' ^'^)) R ip (i, A'^) - p if The . (i, ij, A''' A^' = j , j and each for k' (27) =/^ then imply Since each ratio '• j ,p -p A'') (i. A'')] (j. is 7r;^//7r;^ = (25). k (by condition Hence, lim^^oo (^I-.oo.m (p(^.^'')) 1- ~ inequalities in (26) (27) " by each ^ k' Lemma "^i.oo.m (P (^' ^'''))) positive, for (i)), Lemma A, the establishing k, L (Proof of Part Lemma 2) Fix there exists 3, holds for every k' ^ k. > e > e' Now, by -p{i, Pr^ (||p {s) and (i), > 5 such that 0, Fix also any 0. there exists II > /io,fc < ho,k/m\e = A'') > 0fc_oo,m (p(*)) 1 and i — e Since each 7rj.,/7ri- is finite, by e' such that 0, = f A'') k. whenever /'^, {p{s)) / f\k (p(s)) < f {x)dx > - {1 5) Let Qk,m and K = t, There any / k' f min||3,||</(j, e'K/2. {p e By 0. (i), A (L) ; ||p -p (i. A'') II > there exists hi^k \\p{s) —p (i,A'' J > hi^k/Tn, /(m(p(s)-p(i,A'=))) Moreover, since limm^oo , m ) such that, whenever /c (i, ^'^ to) , c (i, ^, < m) = 1 for each 2 for every k' -^ k i and and any m > ||x|| mi,fc, > /ii_fc, / (a;) < p{s) £ Qk,m, and and /(m(p(.)-p(z,^^-'))) c (i, a'' < ho,k/m} exists a sufficiently large constant mi^k such that for we have k, > [x] = ^, m> ^ ^1^^22 K there exists vfii^k Tn2,fc > mi^fc such that This implies />(p(s))/A4p(^))<«', establishing that 4,oo,^(p(s))>l-eNow, Then, by such that (28) assume that R{p(i,9) —p{j,6') ,p{i,d) —p(j,0)) = for each distinct 9 and 9'. ~ Lemma A, limm-.oo 'Pk oo m{P (*' ^'^)) '' ^^'^ hence by Lemma B, there exists ma^fc > TO2,fc for each > m^^k, p{s) 6 Qk,m, for j 7^ i, m 4>ioo,miPis))>l-e. 48 (29) when Notice that we obtain (28) and (29) hold, we have H^iJcm the desired inequaht)' for each Pr^ {\\4>l,m is) - 4>l,m < is)\\ E = e) (') m > fh: ^'' ~ 0L,m - (li-^-,- (^) (s)|| < <m ^- Then, setting (s)\\ < ^1^ = m = naax^t rm^k, A'^) Pt' [9 = A' k<K k<K k<K = 1-6. (Proof of Part 9'. fc. Then, since each 3) 7r^.,/7r^ is - p (j, 6') ,p{i,e)-p {j, 9)) j^ for each distinct 9 and Lemma A imphes that limm_oo (Pi cxmiP (^' ^'^)) ^ ^ ^°^ ^ach R (p Assume that positive, {i, 9) Let K Then, by part •^fc.oo (pC^s)) p{s) 6 > 2, for 1 each /Tt —'OO J m there exists m2,fc such that for every > m2,k and p [s) G Qk.m, we have B, there also exists m^^k > Tn2,k such that for every > m^^k and k, m — £ By Lemma Qfc,„j, €.oo,m (P («)) < ^l™, 4,oc,m This imphes that ||<p^,m (p(s)) IkL.m I. (s) - <Plo,-m {s)\\ < e — ^^.m at the (P (pt''))!! (i, ^')) > ^- +£< Setting 1 tti end of the proof of Part 2 to the desired inequality. 49 - 2e = maxf^ms^fc and changing < 4,^ ||(/>^_„ (s) (p (s)) - 0^_„ - 6. (s)|| > e, we obtain References 7 Aumann, Robert ity," (1986): "Correlated Equilibrium as an Expression of Bayesian Rational- Econometrica, 55, 1-18. Aumann, Robert (1998): Baron, David (2004): "Common "Persistent priors; Media A Reply to Qui," Econometrica, 66-4, 929-938. Bias," Stanford Graduate School of Business, Mimeo. Berk, Robert (1966): "Limiting Behavior of Posterior Distributions When the Model Is Incorrect," Annals of Mathematical Statistics, 37, 51-58. Billingsley, Patrick (1995): Probability and Measure, Third Edition, Wiley and Sons, New York. Blackwell, David and Lester Dubins (1962): "Merging of Opinions with Increasing Infor- mation," Annals of Mathematical Statistics, 33, 882-887. Carlsson, Hans and Eric van Damme (1993): "Global Games and Equilibrium Selection," Econometrica, 61, 981-1018. Crawford, Vincent and Joel Sobel (1982): "Strategic Information Transmission," Econo- metrica, 50:1431-52. Cripps, Martin, Jeffrey Ely, George Mailath and Larry Samuelson (2006): ing" Cowles Foundation Discussion Doob, J. ities et ses "Common Learn- Paper No. 1575, August 2006. L. (1949): "Application of the Theory of Martingales," in Le Calcul des Probabil- Applications, 23-27, Paris. Diaconis Persi and David Freedman (1986): "On Inconsistent Bayesian Estimates of Location" Annals of Feller, II, 2nd Statistics, 14, 68-87. William (1971): ed. Wiley, New An Introduction to Probability Theory and Its Applications. York. Freedman, David (1963): "On the Asymptotic Behavior of Bayes Estimates Case I" Volume Annals of Mathematical in the Discrete Statistics, 34, 1386-1403. Freedman, David (1965): "On the Asymptotic Behavior of Bayes Estimates in the Discrete Case 11" Annals of Mathematical Statistics, 36, 454-456. Geanakoplos, John and Heraklis Polemarchakis (1982): nal of Economic Theory, "We Can't Disagree Forever," Jour- 28, 192-200. Gentzkow, Matthew and Jesse M. Schapiro (2006): "Media Bias and Reputation," Journal 50 of Political Ghosh New Economy, J. 114, 280-317. K. and R. V. Ramamoorthi (2003): Bayesian Nonparametrics, Springer- Verlang, York. "A Comment on Aumann's Bayesian view", Econometrica, Gul, Faruk (1998): 66-4, 923- 927. Guillemin Victor and Alan Pollack (1974): Differential Topology, Prentice Hall, Haan, L. de (1970): On Regular Variation and Sample Extremes, Mathematical Centre Tract 32, Harrison, Michael and David Kreps (1978): Its Application to the Matthew O., Ehud Kalai and Kurz, Mordecai (1994): Theory, 4, De of Mathematics Centre, Amsterdam, Holland. "Speculative Investor Behavior in a Stock Rann Smorodinsky tion of Stochastic Processes under Learning: Jersey. Weak Convergence Market with Heterogenous Expectations," Quarterly Journal of Economics, Jackson, New (1999): 92, 323-36. "Bayesian Representar- Finetti Revisited," Econometrica, 67, 875-893. "On the Structure and Diversity of Rational Beliefs," Economic 877-900. Kurz, Mordecai (1996): "Rational Beliefs and Endogenous Uncertainty: Economic Theory, 8, An Introduction," 383-397. Milgrom, Paul and Nancy Stokey (1982): "Information, Trade and Common Knowledge," Journal of Economic Theory, 26, 177-227. Miller, Ronald I. and Chris W. Sanchirico (1999): "The Role of Option with Continuity Merging of Opinions and Rational Learning," Games and Economic Behavior, in 29, 170-190. Morris, Stephen (1996): "Speculative Investor Behavior and Learning," Quarterly Journal of Economics, Vol. Ill, No. 4, 1111-33. Savage, Leonard J. (1954): The Foundations of Statistics, Dover reprint, New York, 1972. Seneta, Eugene (1976): Regularly Varying Functions, Lecture Notes in Mathematics, 508, Springer, Berhn. Stinchcombe, Maxwell (2005): "The Unbearable Flightness of Bayesians: Generically Erratic Updating," University of Texas, Austin, mimeo. Yildiz, Muhamet (2003): "Bargaining without a Theorem," Econometrica, 71 Yildiz, Muhamet (3), Common Prior - An Immediate Agreement 793-811. (2004): "Waiting to Persuade," Quarterly Journal of Economics, 119 (1), 223-249. 51 2! -77