frontiers of economics in the post-neoclassical era

advertisement
FRONTIERS OF ECONOMICS IN THE POST-NEOCLASSICAL ERA
J. Barkley Rosser, Jr.
James Madison University
rosserjb@jmu.edu
June, 2008
1
I. Introduction
The most important fact about 21st century economics is that it is the postneoclassical era in terms of the frontiers of economic research. One can still find
orthodox, neoclassical theory in most textbooks, especially those at the upper
undergraduate level. However, this no longer reflects the reality of how economists at
the cutting edge of economics are thinking, including those who are in the mainstream of
the profession. The intellectual orthodoxy of neoclassicism has died (Colander, 2000)
and the current thrust of research at the cutting edge of the frontier is the search for the
appropriate alternative to replace it.
It is useful at this point to distinguish between certain categories, following
Colander et al. (2004), notably “orthodoxy,” “mainstream,” and “heterodoxy.” In their
view, now gaining acceptance among many observers (Bateman, 2008), orthodoxy is an
intellectual category, an established structure of ideas or a paradigm, with the
neoclassical orthodoxy being characterized by the trinity of rationality, greed, and
equilibrium. Mainstream is a sociological category, the set of leading individuals in the
profession who control the leading universities, research institutions, journals, and other
centers of power and control. However, heterodoxy is both, with the heterodox both
opposing the dominant intellectual orthodoxy and also operating sociologically on the
fringes or margins of the profession, alienated from the mainstream leaders, sometimes
even to the point of persecution or suppression.1 The crucial point here is that during a
1
A fairly notorious recent example of such suppression has been the case of the well-known heterodox
economics department at Notre Dame University in the United States, where in 2003, the administration
took away the graduate program from the department and gave it to a newly created department openly
described as being “neoclassical.” A number of prominent mainstream economists protested this move
2
period of upheaval when an established orthodoxy has died and is being replaced, the
mainstream can and will step aside from it and pursue this effort, not being tied to any
particular paradigm necessarily. During such periods, they may be open to drawing on
ideas from heterodox schools or approaches, if not always fully acknowledging doing so,
and the cutting edge or frontier of the profession may well be near the interface between
the mainstream and the heterodox.2
Now it must be recognized that many economists, quite possibly most
economists, do not recognize that neoclassical economics has died. However, I argue
that the case is easily made that it has for those at the cutting edge of the frontier, even if
no single new paradigm has clearly arisen to replace it. It is this latter fact that probably
serves most to continue the illusion for many that neoclassical economics has not died
(along with its persistence in lower level textbooks, where change is slow). Another
issue involves both when it died and how.
Colander (2000) pushes the date too far into the past for my taste, arguing in
effect that neoclassical economics was simply the marginalist economics that became
prominent in the German, French, and English language traditions starting in the 1870s
(with the Germans having been ahead of the others), and which culminated with the
publication of Samuelson’s (1947) Foundations of Economic Analysis. In this view, such
subsequent developments as Nash’s (1951) non-cooperative game theory and the closely
related Arrow-Debreu (1954) general equilibrium theory represented a move beyond
(along with many more heterodox ones, unsurprisingly), with Robert Solow being a prime example
(McCloskey, 2003).
2
It must be noted at this point that the heterodox have formed their own schools, of which there are many
quite well known and established ones, such as Marxism, Institutionalism, Post Keynesianism, and
Austrianism, among others. These schools can in turn create their own sub-orthodoxies, with their own
sub-mainstreams that may well enforce their orthodoxies through specialized journals and selected
universities or research entities where followers of these schools may predominate and enforce their views
on potential “heretics.”
3
neoclassicism proper to something else. Certainly the case can be made for this view,
and it is arguably more consistent from a history of economics perspective.
However, for most economists the efforts carried out at such institutions as MIT
to marry this earlier marginalist Samuelsonianism with the later developments was
successful as epitomized by such widely used graduate textbooks as Varian (1992). In
this latter view, the highwater mark of neoclassical orthodoxy came in the 1970s and
early 1980s with the spread in macroeconomics of New Classical micro-founded models
assuming homogeneous (or at least representative) agents possessing rational
expectations. These models assumed Walrasian general equilibrium to hold,3 with the
macro-model simply reflected at large the micro-solution for this representative agent,
thereby overcoming the aggregation problems long posed by such observers as Keynes
(1936) with the well-known fallacy of composition.4 In this view it would be a series of
external shocks and events that would break this orthodoxy down, ranging from the
unexpected stock market crash of 1987 through the collapse of the Soviet Union and its
command socialist empire, through to a further series of financial crises in the late 1990s
and at the turn of the century, accompanied by a series of intellectual discoveries and
breakthroughs that would undermine this orthodoxy, including perhaps most importantly,
the repetition of results and findings in experimental economics that served to ultimately
3
Because of the emphasis on specifically Walrasian general equilibrium in these models, Colander (2006)
has argued for successor approaches in macroeconomics to be labeled “Post Walrasian.” Hoover (2006)
defends a Marshallian approach to the relationship between theory and empirics, with the Walrasian
approach being an “engineering” approach whereas the Marshallian is “archeological” and exemplified by
the econometric methods of Hendry and Krolzig (2001), with Colander (1995) also holding up Marshall as
exempt from many of the alleged sins of Walras. Hoover sees Milton Friedman as defending a Marshallian
methodology against the Walrasianism of Robert Lucas. It should also be noted that whereas “Post
Keynesian” refers to people who admire Keynes, “Post Walrasian” refers to critics of Walras.
4
Arguably the most serious criticisms based on aggregation problems came out of general equilibrium
theory itself, with Debreu (1974) contributing substantially to the analysis of the famous SonnenscheinMantel-Debreu results that showed serious limits to the theory in the aggregate. This would later culminate
in studies showing how even when there exists a representative agent, aggregate behavior may not
correspond to that of this representative agent (Kirman, 1992).
4
undermine this all-knowing, rational individual agent, even if the difficult aggregation
problems could be overcome.
While many experiments would be done that would contribute to the undermining
of the orthodoxy of rationality, greed, and equilibrium (or at least the first two), arguably
the most stunning and influential was the ultimatum game experiment by Güth et al.
(1982). In this experiment, an agent has a sum of money to offer to divide with another
agent. The first agent suggests a division. If the second agent accepts it, they each get
the agreed-upon amounts. However, if the second agent rejects the (ultimatum) offer,
then neither gets anything. The Nash equilibrium is for the first agent to offer the
minimum possible amount, and for the second agent to accept the offer.5 In fact, it is
now well established that generally smaller offers are rejected, and that in the US the
median offers are around 40% of the total, with the modal offer being an even split, and
with most offers less than 20% being rejected, supposedly “irrationally.” Explaining this
result has been the foundation for an enormous amount of ongoing research that has
brought in other disciplines such as psychology into economics.
Another strand of economics that had been developing through the latter part of
the 20th century was complexity economics (Rosser, 1999), which also implied serious
limits to the ability of individuals to rationally assess the economic environment.6 The
combination of these results has led to the revival of an older idea of economics, bounded
More precisely, if the experiment is repeated it is Selten’s (1975) extension of the Nash equilibrium to the
repeated games case, sub-game perfect equilibrium, that is violated, with an irony being that it has been
reported by Ken Binmore (Colander, Holt, and Rosser, 2004, p. 53) that it was Selten himself who
suggested the ultimatum game experiment to Güth. There is evidence that repeating the game or putting it
into a two-stage form, offers and acceptances may become closer to the Nash equilibrium (Binmore et al.,
1985).
6
Davidson (1996) and Rosser (2006) have debated the source of such limits, with Davidson arguing that it
is ontological and axiomatic, with complexity-based theories providing merely epistemological elements as
in the inability of people to compute. Rosser responds that while such factors enter in, complexity in fact
amounts more profoundly to an ontological explanation of its own for fundamental uncertainty.
5
5
rationality, first proposed by Herbert Simon (1957). This combination has also led to a
broader trend that has emerged strongly since the turn of the century, an ever-increasing
influence of other disciplines in economics, ranging from other social sciences such as
psychology and sociology to natural sciences such as physics and biology, as well as the
emerging computer sciences. All of this has led to a push for higher-order
transdisciplinary approaches.
The rest of this paper will focus on these three main themes with some subthemes in each section. First, we shall consider more seriously developments in
experimental and behavioral economics, with neuroeconomics being an especially recent
and promising area arising from these two. Second we shall look at various ideas
associated with the complexity movement, such as agent-based modeling, computational
economics, and econophysics. Implications for this for a successor to the New Classical
approach to macroeconomics will be examined. Then, we shall consider more
biologically based approaches drawing on evolution and ecology, suggesting a revival of
sorts of institutional economics, but in new forms as major areas of focus at the current
cutting edge of the frontiers of economics. Finally, there will be a brief consideration of
the role of the older schools of heterodox schools of thought in the newer developments.
II. Experimental and Behavioral Economic Research
A. Market Mechanisms and a Partial Defense of Orthodoxy
While we have already noted that there are many results coming out of economic
experiments that question the traditional vision of the neoclassical homo oeconomicus,
some strands of experimental economic research have produced results that can be argued
6
to support the ability of properly structured free markets to reach reasonably efficient
equilibria despite these apparent findings about individuals. The key person arguing this
point has been the “father of experimental economics,” Vernon Smith ever since his
finding in 1962 that double auction markets converge very rapidly to equilibria, with this
result being reinforced and found to be efficient and robust by many later studies
(Friedman, 1974). Smith has since become a strongly pro-free market advocate and has
been involved in setting up many actual auction markets, and the double auction has
spread widely into many financial markets (Smith, 2008). Walrasian general equilibrium
may have serious problems, but individual markets, even ones linked with another market
or two, can be set up to function quite well in terms of equilibrating.
Even as Smith has argued this point for decades, he has not argued against the
findings of all sorts of irrationalities or peculiarities in individual behavior, nor has he
argued that all markets everywhere behave well or efficiently. Thus, he was one of the
first to show that speculative bubbles and crashes rather easily happen in experimental
markets, even if learning tends to reduce them (Smith et al. 1988). This result was found
to still hold even when experimental subjects had to provide their own money for the
market activity (Porter and Smith, 1995). Indeed, even though double auctions work
well, many real-world auctions are not easy to set up to follow them, and while Smith and
his allies have argued for simpler rules for large-scale auctions (Banks et al., 2003), many
have not followed this advice, such as in setting up the US power spectrum auctions.
This raises what has been an increasing concern for Smith and his coauthors over
time: how to reconcile this apparent irrationality of individual agents with their ability in
at least certain kinds of situations to act together in an apparently rational and efficient
7
way. In this he has increasingly fallen back on arguments of Adam Smith and Friedrich
Hayek regarding markets as emergent spontaneous orders, and the idea that even though
people have bounded rationality, they are able to intuitively operate in market settings to
achieve efficient results, even as their conscious minds are not carrying out the necessary
calculations. People’s brains and minds have evolved to do these sorts of things, even if
unconsciously (Smith, 2008) as part of more generalized reciprocity, which is
evolutionarily advantageous.
In this regard, while arguing that boundedly rational people can achieve efficient
outcomes in the right institutional setups through intuition or rules of thumb, Smith and
his coauthors have directly confronted traditional game theory that posits high levels of
rationality to individual decisionmakers. Thus, Aumann (1995) has long argued for
assuming common knowledge, with its implication that people regularly use backward
induction to decide how to behave strategically when interacting with others. Such
backward induction depends on believing that others also are thinking ahead and
calculating the best possible responses of other agents into the distant future in response
to possible actions by oneself. As it is, the experimental evidence suggests that most
people do not think very far ahead in such situations (Stahl and Wilson, 1995; Camerer,
2003), and that backward induction is thus largely a purely theoretical construct.
The accumulation of this evidence has put game theory on the defensive. Major
developments coming out of this has been the emergence of both evolutionary game
theory (Binmore and Samuelson, 1999), with its greater emphasis on random forces
pushing agents into different basins of attraction over time and behavioral game theory
that recognizes that agents are not fully rational or perfectly informed (Camerer, 2003).
8
B. Social Preferences and Other Complications
While there had been a long series of experiments documenting the ubiquity of
“anomalies” that undercut the standard model of the rational economic agent, particularly
the expected utility model of decisionmaking under risk (Strotz, 1955; Kahneman and
Tversky, 1979; Thaler, 1981), it was probably ultimately the ultimatum game
experiments mentioned above and the repeated verification of their findings that most
seriously drove the nail into the coffin of the greedily individualistic, rational economic
agent. People were regularly willing to bear costs to punish each other for violating
norms of fairness (Rabin, 1993). This argued for the serious introduction of
psychological and sociological ideas into economics, something which had been
regularly done before the middle of the 20th century, but which had fallen out of fashion
as it was thought that the rational agent model of economics offered the superior
paradigm for the social sciences more broadly. In particular Rabin and others argued that
people have a “taste for fairness” that must be accounted for by economists and which
goes against the standard model of simple greedy maximization.
Ironically this has set up a somewhat curious debate over how best to deal with this
matter. On the one hand is a group led by Rabin, more or less, that argues against the
conventional approach by asserting the reality of the taste for fairness, which may
somehow be viewed as a kind of selfless altruism.7 However, the approach taken by
Rabin is in another sense highly conventional; he assumes a conventional sort of utility
function that an individual maximizes subject to a budget constraint, with that utility
function simply having an additional argument: fairness, which then must be traded off
against other more standard such arguments in the utility maximization calculus. In this
7
Andreoni (1990) has labeled this purer altruism as the “warm-glow effect.”
9
regard, Rabin in his interview in Colander et al. (2004, p. 151) expressed the side of his
views that believes in a more conventional approach.
“In fact, I have various fears about teaching psychological economics. One is of
attracting graduate students who are just hostile to economics. Another is all the people
who want to use evolution and other approaches to explain departures from classical
models rather than using evidence of departures to do better economics.”
This last sentence describes those taking the alternative view, which would include
Vernon Smith (2008) as well as some from the evolutionary game theoretic school such
as Binmore, as expressed in Colander et al., 2004. pp. 65-66, where he complains of
people assuming “exotic preferences” not defensible by evolutionary processes. While
Binmore tends to defend somewhat more conventional approaches, as with his
questioning of the robustness of the ultimatum game findings, Smith pushes a
methodologically more unconventional position, rejecting standard utility maximization
modeling, indeed of assuming a utility function at all. The evidence of people’s lack of
standard calculating and conscious optimization is too strong to be ignored. In the end,
this group argues for assuming an ultimate more selfish approach, that of some form of
reciprocity, even if it is distant and indirect. They argue that this is defensible as having
arisen from evolutionary processes, even if the mind makes the necessary calculations in
the sort of unconscious way described earlier above. It is an ultimately “selfish”
reciprocity motive that underlies ultimatum game results, not some vaguely selfish
altruism or inequity aversion.8
8
This view is also identified with the term reciprocal altruism introduced by the evolutionary biologist,
Robert Trivers (1971).
10
A response by those more willing to allow for such motives has spawned an
enormous amount of research, with Fehr and Schmidt (1999) and Bolton and Ockenfels
(2000) posing ways to distinguish reciprocity from inequity aversion in data. These
proposals have proven controversial and considerable debate has ensued regarding
whether these methods are accurate or not (Charness and Rabin, 2002; Engelmann and
Tyran, 2005). In any case, some of those supporting this approach see more generalized
forms of reciprocity, or strong reciprocity, as indeed being evolutionarily founded (Gintis
et al., 2004) and drawing on work of anthropologists as well as psychologists.9 Such
forms have come to be labeled social preferences.
C. Neuroeconomics and the Struggle of the Laboratory versus the Field
In laboratory experiments these debates have moved more deeply into the brain
itself as the use of magnetic resonance imaging (MRI) has come to be used in economic
experiments to see which parts of the brain are stimulated and how when agents are
engaging in various kinds of decisionmaking or actions (Zak, 2004; Camerer et al. 2005).
One of the first in economics was the study of McCabe et al. (2001) of cooperative
versus non-cooperative behavior in the prisoner’s dilemma. They found that very
different parts of the brain were stimulated when agents were engaging in one or the
other. Interestingly, whereas Nash originally posited that the selfishness of not
cooperating is “rational,” it is when agents are cooperating that they appear to be using
At a certain level this debate can be seen as another round of the old “Adam Smith problem,” how to
reconcile the Adam Smith of The Theory of Moral Sentiments (1759) with the apparently harder nosed
Adam Smith of The Wealth of Nations (1776) Vernon Smith (1998) draws on Hayek to say that it is a
matter of personal dealings within more closely knit groups where moral sympathy predominates versus
more arms-length dealings in larger society as in markets where reciprocity and ultimate selfishness
predominates.
9
11
the pre-frontal cortex parts of their brains that are more usually identified with “higher”
or more “rational” thought by psychologists.
Given that such hard-wiring of the brain is almost certainly the result of deep
evolutionary processes, the debates over reciprocity and altruism have more recently
moved into the laboratories of the neuroeconomists. Thus, Fehr and Gächter (2002) and
Fehr et al. (2002) argue strongly that strong reciprocity involves a pleasure reward in the
brain associated with punishing those who violate social norms. Zak (2005) has
identified the development of trust as being associated with the chemical oxytocin, which
is released in the brain and provides a sensation of pleasure. Zak (2008) and his
associates have pushed this argument to say that oxytocin release and reinforcement of
socially virtuous behavior serves as the foundation of broader morality and social
cohesion within market economies..
Left unresolved here is the degree to which we can associate stimulation of
specific areas of the brain with particular psychological states or perceptions.
Furthermore, there also remains the issue of whether or not what is found in the
laboratory really reflects what goes on in reality, or “in the field.” This has led to deep
debates over the use of laboratory experiments versus so-called field experiments (Levitt
and List, 2007). Advocates of the former stress the ability to more scientifically pin
down circumstances and eliminate extra “noise,: even as advocates of the latter argue that
in laboratories important framing effects can arise from the influence of the
experimenters themselves that can influence and bias the results. A compromise would
seem to be that the two approaches are complementary, but this is a matter of ongoing
dispute, just as is the matter of the degree to which the results of neuroeconomics studies
12
of which parts of the brain are associated with what kinds of decisions really carry over
into reality. In any case, all of those involved in experimental economics hope to achieve
a more truly scientific approach to the studying and understanding human behavior.
III. The Complex Implications of Economic Complexity
A. The Complexity Problem
While both the behavioralist and complexity approaches profoundly undermine
the fully rational and informed and selfish homo oeconomicus of neoclassical economics,
especially in its stricter Walrasian form, they have developed and operated largely
independently. However, they had a common link at the foundation of behavioral
economics with Herbert Simon, already mentioned as the inventor of the bounded
rationality hypothesis. He was also an early developer of complexity theory, especially
its hierarchical version (Simon, 1962). Simon saw the deep link between complexity and
behaviorialism through the role of complexity as a foundation for bounded rationality. It
should be kept in mind that Simon’s later innovative research efforts in pursuit of the
ideal of artificial intelligence was motivated deeply by his awareness of bounded
rationality and its foundation in the inevitable and unavoidable reality of complexity.
Rosser (2008a) poses three levels of thinking about the nature of complexity. At
the highest level is meta-complexity, the broadest conception, which allows for the 45
definitions of Seth Lloyd (Horgan, 1997, p. 303)10 as well as others. Next down is the
dynamic definition provided in Rosser (1999), taken from Day (1994). This “big tent”
definition is that a system is (dynamically) complex if it deterministically and
endogenously exhibits irregular dynamics that do not converge on a point, a limit cycle,
10
Horgan is a critic of the notion in most of its forms and coined the term chaoplexity partly to mock it.
13
or a smooth growth or decline path (thus ruling out irregularities due to the exogenous
shock noise that drives real business cycle models). This “big tent” definition includes
“the four C’s” of cybernetics, catastrophe theory, chaos theory, and the “small tent”
complexity of heterogeneous interacting agent models. It is this last type that constitutes
our lowest level and is what most people think of when they think of “complexity
economics.” This last approach has often been associated with the Santa Fe Institute and
such computer simulation methods as cellular automata (Wolfram, 1984) and artificial
life programs (Langton, 1990; Tesfatsion, 2006).
Of the other definitions of complexity that one can find at the meta-complexity
level, probably the most important one in terms of current research are those based on
computability (Markose, 2005; Velupillai, 2005). There are a variety of these, most of
them ultimately drawing on the information theory of Shannon (1948) and its further
development by Kolmogorov (1965), Chaitin (1987), and others. Although there are
competing versions of these definitions and measures, such well-studied divisions as
between programs that are solvable in polynomial time versus in exponential time versus
those that cannot be solved at all in finite time due to halting problems (Blum et al.,
1998). All of these in some way or another end up being related to the minimum length
of the program required to compute a solution to a problem or system, and while there are
variations, these provide more specifically measurable degrees of complexity. It is this
latter fact (along with the increasing spread of computational economics in various
forms) that has added to the popularity more recently of this sort of definition. Although
this approach now has many followers, it was probably first applied in economics by
Albin (1982), with full discussion in Albin with Foley (1998).
14
B. The Agent-Based Approach
This is one of the most productive and promising areas of current research at the
frontiers of economic research, although Tesfatsion and Judd (2006) provides broad and
in-depth coverage across many areas of economics and Delli Gatti et al. (2008) provide
excellent coverage for macroeconomics more particularly. Among the more useful of the
surveys in Tesfatsion and Judd are those on agent-based models in economics and
computational finance by Hommes (2006) and by LeBaron (2006) and also the one by
Duffy (2006), which examines links between the heterogeneous agent-based approaches
and the behavioral and experimental approaches we have discussed above. An early
approach that has been widely imitated for modeling broader societal development from
agent-based modeling is Epstein and Axtell (1996).
General theoretical modeling of interacting agent-based systems has been done by
Brock and Hommes (1997) and Brock and Durlauf (2001).11 These papers provided a
foundational approach based on mean-field dynamics drawn from statistical physics.
Central to the analysis of these nonlinear systems is consideration of the structure of
bifurcation sets and various forms of complex dynamics that arise as the system crosses
bifurcation. The complex dynamics extend well beyond just complex dynamics to
involve multiple basins of attraction and fractal basin boundaries of these basins.12 The
crucial control parameters determining the bifurcation structures in this set of models turn
out to be degree of interaction between the agents (willingness to herd) and their
willingness to change strategies, with more complex dynamics more likely as the values
11
Föllmer (1974) was the first to present a model of equilibrium of heterogeneous agents based on a
statistical mechanics approach, long before the establishment of the Santa Fe Institute or the development
of econophysics. Aoki and Yoshikawa (2006) show how the approach can be used in macroeconomics.
12
For a fuller analysis of systems with multiple basins of attraction with fractal boundaries and multidimensional chaotic dynamics using global analysis, see Agliari et al. (2002).
15
of these parameters increase. Brock et al. (2008) show how a model based on this
framework can be used to analyze the destabilization of financial markets as hedging
instruments increase, a result counter to that of standard financial economics theory.
Gallegati et al. (2008) also use a variation of this to study financial market dynamics with
the imposition of a Minsky (1972) type of financial constraint that allows for showing the
phenomenon of a “period of financial distress” between the peak and the crash of a
speculative bubble that is the most common pattern seen in historical bubbles
(Kindleberger, 2000, Appendix B), even though again such a pattern has not been
modeled by conventional economic theory. Föllmer et al. (2005) are able to get
somewhat similar results with a related agent-based model.
Agent-based models have also been used to model more specific market dynamics
not involving finance. So the evolving structures of loyalty in a market have been studied
by Kirman and Vriend (2001). How market structures relate to technological change has
also been a major area of study. Fagioli and Dosi (2003) have studied innovation
dynamics on a two-dimensional lattice with a variety of effects and processes. Silverberg
and Verspagen (2005) have looked at the diffusion process in more detailed in a variety
of technology spaces. Dawid and Reimann (2004) have used genetic algorithms to
examine innovation in relation to product life cycles.
A major thrust of many agent-based models has been to study learning dynamics.
This is an area where the agent-based complexity approach has intersected with the
experimental behavioral approach, with experiments increasingly involving subjects
interacting with computerized systems in repeated games or market interactions in which
learning takes place. Probably the first such in this line was Roth and Murnighan (1978)
16
who had subjects playing repeated Prisoners’ Dilemma games against computerized
opponents. Arifovic (1996) had experimental subjects using genetic algorithms to learn
about foreign exchange rate dynamics. She (2001) would later find that boundedly
rational agents could do better than fully rational agents in a complex context. Bullard
and Duffy (2001) studied how learning dynamics can lead to excess volatility, and
Brenner and Witt (2003) further studied game dynamics with learning. Nyarko and
Schotter (2002) saw how eliciting beliefs could enhance learning in laboratory subjects.
Perhaps the ultimate result involves zero-intelligence computerized agents that
nevertheless “learn” in a certain sense to move toward a double auction equilibrium
(Gode and Sunder, 1993). This result can be seen as underpinning the spread of
completely computerized exchanges without human actors, which fully make manifest
the prediction of Mirowski (2007) that markets simply are algorithms that evolve to
become what he labels markomata.
Finally, another area of current interest for studying agent-based models has been
in the emerging area of network analysis. This work, with close links to sociology
derives largely from the study of Watts and Strogatz (1998), but is now burgeoning
widely. Kirman and Vriend (2001), mentioned above, is an example in terms of
networks of market relations. Kranton and Minehart (2001) also study networks in
market relations. A general finding of this literature is that clusterings occur in patterns
that are neither completely random nor fully concentrated. The famous “small worlds”
phenomenon, sometimes described by the famous claim that no two people in the world
are separated by more than six degrees of separation.
B. Econophysics
17
A term coined by H. Eugene Stanley in the mid-1990s, econophysics has been
described by Mantegna and Stanley (2000, vii-ix) as the “multidisciplinary field…that
denotes the activities of physicists who are working on economic problems to test a
variety of new conceptual approaches deriving from the physical sciences.” This
sociologically oriented definition emphasizes who is doing it more than what it is. In
practice much emphasis has been on looking at data first and then trying to find models
that fit the data, in contrast with the usual economics approach of assuming that the
standard theory is generally correct and then applying it to data to see if it or the data can
be sufficiently tweaked to find a matchup. A major focus of much of the econophysics
research has been upon power law distributions that are linear in the mapping of logged
relationships between variables. In terms of returns in financial markets, such power
laws are associated with the presence of more extremal events than would be predicted if
the distribution were Gaussian normal, so that one observes kurtosis or “fat tails.” Many
of the models that the econophysicists bring to bear can be seen as variations on
heterogeneous agent models as many are drawn from statistical mechanics models of
interacting particles. They are thus linked to the broader “small tent” complexity
approach that emphasizes such interacting heterogeneous agent models. The most
fervent defenders of the approach (McCauley, 2004) argue for its superiority over
economic models because of their lack of invariance laws.
Prominent applications in financial models included Bouchaud and Cont (2002),
Farmer and Joshi (2002), and Sornette (2003). Wealth distributions have also been found
to have power law distributions (Levy and Solomon, 1997), although income
distributions appear to have such only at their upper ends (Drăgolescu and Yakovenko,
18
2001; Chatterjee et al. 2005).13 City size distributions have also been estimated by such
means (Gabaix, 1999) as have firm sizes (Axtell, 2001).
A curious aspect of this is that many of these ideas were first developed by
economists, whereas some of the more conventional ideas used by standard economic
theory originated from physicists, an old argument of Mirowski (1989). Thus the first to
suggest the use of power laws for studying income distribution was Pareto (1897).
Whereas the idea of normal distributions of financial returns was initiated by the
mathematician Bachelier (1900), this would become a standard idea in physics as
Brownian motion, and the physicist Osborne (1959) would later suggest its use in
financial economics, which then became standard. It was the mathematician Mandelbrot
(1963) who first suggested applying power laws to asset returns. It was a geographer
who would first suggest that city sizes might be distributed according to a power law
(Zipf, 1941), and Ijiri and Simon (1977) did so for firm sizes a quarter of a century before
Simon’s student Axtell (2001) would confirm this finding. Furthermore, the
mathematical economist Föllmer (1974) suggested applying statistical mechanics models
to studying markets well before the physicists did, and the economist, Duncan Foley
(1993) applied entropy theory to define statistical distributions of prices as a new concept
of equilibrium to replace the neoclassical one.
Given this tangled history it is not surprising that econophysics has engendered
controversy. Gallegati et al. (2006) criticized much work as not properly citing previous
work by economists, as using shoddy statistical methods, asserting universal laws where
13
It has long been argued that labor income tends to be distributed lognormally. To the extent that wealth
represents mostly accumulated capital income, and financial markets show power law distributions, then it
may be perfectly reasonable for the lower part of the income distribution to be lognormal, where labor
income is most important, while the upper end is power law where wealth and capital income predominate.
19
none apply, and lacking in theoretical models or explanations for findings. Rosser
(2008b) argues that most of these criticisms had some merit, but that they are being
overcome or can be by having economists and physicists work together, with Foley
(1993) providing an example of a theoretical model drawn from physics that is of
relevance for explaining economic phenomena. However, Lux (2009) warns that
physicists may get led astray if they work with economists whose ideas are too
conventional, with some econophysicists embarrassingly making incorrect public
forecasts based on inappropriate models drawn from economics.
C. Econobiology
In contrast to econophysics, econobiology is not a self-identified entity created by
its practitioners but rather a term of disapprobation assigned by critics such as McCauley
(2004, chap. 9), who sees biology lacking in invariance laws in the same way that
economics does and that therefore it cannot be a real science in the way that physics
is.(and econophysics might be). However, we can identify at least three sub-areas of
what might be called econobiology that have their own distinct and identities:
evolutionary economics, bioeconomics, and ecological economics.
Of these the first is certainly the oldest in terms of identity, with old institutional
economists taking up the term and the identity (Veblen, 1898) and economics and the
theory of evolution having a complicated interrelationship throughout much of the
nineteenth century as Malthus influenced Darwin, who in turn influenced both Marx and
Marshall (Rosser, 1991, chap. 12). As old institutionalism lost its struggle with
neoclassical economics (Veblen having coined the term, “neoclassical economics”), and
new institutional economics became more important, evolutionary economics fell from
20
favor and attention, although arguments relating to evolution continued to be made in
various parts of economics such as the theory of firms and how they change over time
and in the theory of technological change, although Schumpeter (1936) favored a more
discontinuous view of evolution that contrasted with Darwin and foreshadowed the
punctuated equilibrium theory of Stephen Jay Gould (2002).
However, evolutionary economics is experiencing a revival along two lines. One
is the already discussed area of evolutionary game theory, which has now become the
leading edge of game theory. The other is in work that seeks to overcome the divide
between the old and new institutionalist schools. This approach has been pushed by
Hodgson (2006) who sees in modern evolutionary theory as it has come to be after the
innovations of Gould and others a fresh paradigm for use in economics.
Arguably the latter two are simply variations on each other, although
bioeconomics is the older term, dating to its introduction by Colin W. Clark in 1976,
whereas ecological economics came into usage in the 1980s (Martinez-Alier, 1987).
Both can be viewed as transdisciplinary disciplines (indeed the ecological economists
specifically label themselves as such). Both seek to integrate fully models of biology or
ecology with those of economics. The former are more likely to more specifically model
population dynamics of biological populations in conjunction with the impact of
interactions with economic agents. Clark’s work on fisheries is the model here, and the
fact that supply curves for fish tend to bend backwards opens up the door to a variety of
possible nonlinear and complex dynamics in fisheries, including catastrophic collapses
(Rosser, 2001; Gunderson and Holling, 2002),which we have observed all too many
times in real life unfortunately.
21
Forestry has long been an area of study as well by bioeconomists and ecological
economists, using a variety of methods (Kant and Berry, 2005). The study of both
fisheries and forestry are seen as increasingly important given the role these sectors play
in developing countries and in the pressure that they are both perceived to be facing at the
global level. Similarly lake systems have been studied by joint teams of ecologists and
economists (Carpenter et al. 1999). For many of these systems problems of common
property resource management are the issue (Sethi and Somanathan, 1996), which bring
into relevance the issues studied by experimental and behavioral economists regarding
the ability of people to cooperate or not cooperate, along with the old debate about multilevel evolution (Henrich, 2004). These problems arguably come to ahead in the great
question of managing the global commons in the face of man-induced global warming
(Stern, 2008), which are from easy to resolve.
IV. Conclusion
It must be noted that we have barely scratched the surface of areas of current
research at the cutting edge of economics. However, most of what has been cited here in
terms of research since the turn of the century has moved beyond the orthodox
neoclassical approach, even if much of it has been done by mainstream economists,
although the econophysicists are clearly an exception to that. We have seen that although
they are mostly separate and parallel, our two main streams of experimental/behavioral
economics and complexity economics do have some common roots and also have
recently seemed to intersect and interpenetrate with each other more frequently.
Complexity and bounded rationality are deeply connected with each other.
22
A final question raises itself that we are not going to attempt to answer in any depth.
This involves the nature of the relationship between this newer research and the older
heterodox schools of thought. I shall only note here in conclusion that this period of
post-neoclassical upheaval opens many doors, and there are arguably chances for some
ideas from some of these schools to enter into the current discourse, just as there is also
the chance for some of these ideas to enter into and stimulate work in these schools.
Thus, I would note Chiarella et al. (2005) as an example of Post Keynesian work that has
been strongly influenced by the currents coming from research in nonlinear complex
dynamics, just as older Post Keynesian models were sources of inspiration for some of
the earlier complex dynamics models (Rosser, 2006). Also, certainly the ideas of Keynes
are clearly more in tune with the newer behavioral economics than with the idea of fully
rational and well-informed agents that have underpinned neoclassical economics.
Finally, just as one can find roots of modern approaches in Keynes, likewise
Hayek (1978) foresaw the complexity approach, and the Austrians have long disavowed
standard equilibrium analysis, emphasizing the distinctive nature of individual economic
agents, even as their interactions may bring about a spontaneous order. In this new
environment there may even be possibilities of cross-fertilization of ideas between these
older schools of heterodox thought as they open to providing and receiving influences
from the newer post-neoclassical approaches (Koppl and Rosser, 2002).
References
Agliari, Anna, Gian-Italo Bischi, and Laura Gardini. 2002. “Some Methods for the
Global Analysis of Dynamic Games Represented by Iterated Noninvertible Maps.” In
Tönu Puu and Irina Sushko, eds., Oligopoly Dynamics: Models and Tools. Heidelberg:
Springer, pp. 31-83.
23
Albin, Peter S. 1982. “The Metalogic of Economic Predictions, Calculations and
Propositions.” Mathematical Social Sciences 3, 329-358.
Albin, Peter S. with Duncan K. Foley. 1998. Barriers and Bounds to Rationality: Essays
on Economic Complexity and Dynamics in Interactive Systems. Princeton: Princeton
University Press.
Andreoni, James.1990. “Impure Altruism and Donations to Public Goods: A Theory of
Warm-Glow Giving.” Economic Journal 100, 464-477.
Aoki, Masanao and Hiroshi Yoshikawa. 2006. Restructuring Macroeconomics: A
Perspective from Statistical Physics and Combinatorial Stochastic Processes.
Cambridge: Cambridge University Press.
Arifovic, Jasmina. 1996. “The Behavior of the Exchange Rate in the Genetic Algorithm
and Experimental Economies.” Journal of Political Economy 104, 510-541.
Arifovic, Jasmina. 2001. “Performance of Rational and Boundedly Rational Agents in a
Model with Persistent Exchange Rate Volatility.” Macroeconomic Dynamics 5, 204-234.
Arrow, Kenneth J. and Gérard Debreu. 1954. “Existence of an Equilibrium for a
Competitive Economy.” Econometrica 22, 265-290.
Aumann, Robert. 1995. “Backward Induction and Common Knowledge of Rationality.”
Games and Economic Behavior 8, 6-19.
Axtell, Robert. 2001. “Zipf Distribution of U.S. Firm Sizes.” Science 293, 1818-1820.
Bachelier, Louis. 1900. “Théorie de la Spéculation.” Annales Scientifique de l’École
Supérieure III-I, 21-86.
Banks, Jeffrey M., Mark Olson, David Porter, Stephen Rassenti, and Vernon Smith.
2003. “Theory, Experiment and the Federal Communications Commission Spectrum
Auctions.” Journal of Economic Behavior and Organization 51, 303-350.
Bateman, Bradley W. 2008. “2007 HES Presidential Address: Reflections on the
Secularization of American Economics,” Journal of the History of Economic Thought 30,
1-20.
Binmore, Ken and Larry Samuelson. 1999. “Equilibrium Selection and Evolutionary
Drift.” Review of Economic Studies 66, 393-394.
Binmore, Ken, Avner Shaked, and John Sutton. 1985. “Testing Non-Cooperative Game
Theory: A Preliminary Study.” American Economic Review. 75, 1178-1180.
24
Blum, Lenore, Felipe Cucker, Michael Shub, and Steve Smale. 1998. Complexity and
Real Computation. New York: Springer.
Bolton, Gary E. and Axel Ockenfels. 2000. “ERC: A Theory of Equity, Reciprocity, and
Competition.” American Economic Review 90, 166-193.
Bouchaud, Jean-Philippe and R. Cont. 2002. “ A Langevin Approach to Stock Market
Fluctuations and Crashes.” European Physical Journal B 6, 543-550.
Brenner, Thomas and Ulrich Witt. 2003. “Melioration Learning in Games with Constant
and Frequency-Dependent Payoffs.” Journal of Economic Behavior and Organization 50,
429-448.
Brock, William A. and Steven N. Durlauf. 2001. “Discrete Choice with Social
Interactions.” Review of Economic Studies 68, 235-260.
Brock, William A. and Cars H. Hommes. 1997. “A Rational Route to Randomness.”
Econometrica 65, 1059-1095.
Brock, William A., Cars H. Hommes, and Florian Wagener. 2008. “More Hedging
Instruments may Destabilize Markets.” Mimeo, SSRI University of Wisconsin-Madison,
CeNDEF, University of Amsterdam.
Bullard, James and John Duffy. 2001. “Learning and Excess Volatility.” Macroeconomic
Dynamics 5, 272-302.
Camerer, Colin F. 2003. Behavioral Game Theory: Experiments on Strategic Interaction.
Princeton: Princeton University Press.
Camerer, Colin F., George Lowenstein, and Drazen Prelec. 2005. “Neuroeconomics:
How Neuroeconomics Can Inform Economics.” Journal of Economic Literature 48, 964.
Carpenter, Stephen R., Donald Ludwig, and William A. Brock. 1999. “Management of
Eutrophication for Lakes Subject to Potentially Reversible Change.” Ecological
Applications 9, 751-771.
Chaitin, Gregory J. 1987. Algorithmic Information Theory. Cambridge: Cambridge
University Press.
Charness, Gary and Matthew Rabin. 2002. “Understanding Social Preferences with
Simple Tests.” Quarterly Journal of Economics 117, 817-869.
Chatterjee, Arnab, Sudhakar Yarlagadda, and Bikas K. Chakrabarti, eds. 2005.
Econophysics of Wealth Distributions. Milan: Springer.
25
Chiarella, Carl, Peter Flaschel, and Reiner Franke. 2005. Foundations for a
Disequilibrium Theory of the Business Cycle. Cambridge: Cambridge University Press.
Clark, Colin W. 1976. Mathematical Bioeconomics. New York: Wiley-Interscience (2nd
edition, 1990).
Colander, David. 1995. “Marshallian General Equilibrium Analysis.” Eastern Economic
Journal 21, 281-293.
Colander, David. 2000. “The Death of Neoclassical Economics.” Journal of the History
of Economic Thought 22, 127-143.
Colander, David, ed. 2006. Post Walrasian Macroeconomics: Beyond the Dynamic
Stochastic General Equilibrium Model. New York: Cambridge University Press.
Colander, David, Richard P.F. Holt, and J. Barkley Rosser, Jr. 2004. The Changing Face
of Economics: Conversations with Cutting Edge Economists. Ann Arbor: University of
Michigan Press.
Davidson, Paul. 1996. “Reality and Economic Theory.” Journal of Post Keynesian
Economics 18, 479-508.
Dawid, Herbert and M. Reimann. 2004. “Evaluating Market Attractiveness: Individual
Incentives vs. Industrial Profitability.” Computational Economics 24, 321-355.
Day, Richard H. 1994. Complex Economic Dynamics, Volume I: An Introduction to
Dynamical Systems and Market Mechanisms. Cambridge: MIT Press.
Debreu, Gérard. 1974. “Excess Demand Functions.” Journal of Mathematical Economics
1, 15-23.
Delli Gatti, Domenico, Edoardo Gaffeo, Mauro Gallegati, Gianfranco Giulioni, and
Antonio Palestrini. 2008. Emergent Macroeconomics: An Agent-Based Approach to
Business Fluctuations. Milan: Springer.
Drăgulescu, A.A. and Victor Yakovenko. 2001. “Exponential and Power-Law Probability
Distributions of Wealth and Income in the United Kingdom and the United States.
Physica A 299, 213-231.
Duffy, John. 2006. “Agent-Based Models and Human Subject Experiments.” In
Tesfatsion and Judd, eds., pp. 950-1011.
Engelmann, Dirk and Jean-Robert Tyran. 2005. “To Buy or Not to Buy? An
Experimental Study of Consumer Boycotts in Retail Markets.” Economica 72, 1-16.
26
Epstein, Joshua and Robert Axtell. 1996. Growing Artificial Societies: Social Science
from the Bottom Up. Cambridge: MIT Press.
Fagioli, Giorgio and Giovanni Dosi. 2003. “Exploitation, Exploration and Innovation in a
Model of Endogenous Growth with Locally Interacting Agents.” Structural Change and
Economic Dynamics 14, 237-273.
Farmer, J. Doyne and S. Joshi. 2002. “The Price Dynamics of Common Trading
Strategies.” Journal of Economic Behavior and Organization 49, 149-171.
Fehr, Ernst, Urs Fischbacher, and Simon Gächter. 2002. “Strong Reciprocity, Human
Cooperation and the Enforcement of Social Norms.” Human Nature 13, 1-25.
Fehr, Ernst and Simon Gächter. 2002. “Altruistic Punishment in Humans.” Nature 415,
137-140.
Fehr, Ernst and Klaus M. Schmidt. 1999. “A Theory of Fairness, Competition, and
Cooperation.” Quarterly Journal of Economics 114, 817-868.
Föllmer, Hans. 1974. Random Economies with Many Interacting Agents.” Journal of
Mathematical Economics 1, 51-62.
Föllmer, Hans, Ulrich Horst, and Alan P. Kirman. 2005. “Equilibria in Financial Markets
with Heterogeneous Agents: A Probability Perspective.” Journal of Mathematical
Economics 41, 123-155.
Friedman, Daniel. 1974. “On the Efficiency of Double Auction Markets.” American
Economic Review 74, 60-72.
Gabaix, Xavier. 1999. “Zipf’s Law for Cities: An Explanation.” Quarterly Journal of
Economics 114, 739-767.
Gallegati, Mauro, Steve Keen, Thomas Lux, and Paul Ormerod. 2006. “Worrying Trends
in Econophysics.” Physica A 370, 1-6.
Gallegati, Mauro, Antonio Palestrina, and J. Barkley Rosser, Jr. 2008. “The Period of
Financial Distress in Speculative Markets: Interacting Heterogeneous Agents and
Financial Constraints.” Mimeo, Polytechnic University of the Marches, University of
Teramo, and James Madison University, available at http://cob.jmu.edu/rosserjb.
Gintis, Herbert, Samuel Bowles, Robert Boyd, and Ernst Fehr. 2004. Moral Sentiments
and Material Interests: On the Foundations of Cooperation in Economic Life.
Cambridge: MIT Press.
27
Gode, D.K. and Shyam Sunder. 1993. “Allocative Efficiency of Markets with ZeroIntelligence Traders: Markets as a Partial Substitute for Individual Rationality.” Journal
of Political Economy 101, 119-137.
Gould, Stephen Jay. 2002. The Structure of Evolutionary Theory. Cambridge: Belknap
Press of Harvard University.
Gunderson, Lance H. and C.S. Holling, eds. 2002.Panarchy: Understanding
Transformations in Human and Natural Systems. Washington: Island Press.
Güth, Werner, Rolf Schmittberger, and Bemd Schwarze. 1982. “An Experimental
Analysis of Ultimatum Bargaining.” Journal of Economic Behavior and Organization 3,
367-388.
Hayek, Friedrich A. 1967. “The Theory of Complex Phenomena.” In Friedrich A. Hayek,
Studies in Philosophy, Politics, and Economics.London: Routledge & Kegan Paul, pp.
22-42.
Hendry, D.F. and H.M. Krolzig. 2001. Automatic Econometric Model Selection Using
PcGets 1.0. London: Timberlake Consultants.
Henrich, Joseph. 2004. “Cultural Group Selection, Coevolutionary Processes, and LargeScale Cooperation.” Journal of Economic Behavior and Organization 53, 3-35.
Hodgson, Geoffrey M. 2006. Economics in the Shadow of Darwin and Marx: Essays on
Institutional and Evolutionary Themes. Cheltenham: Edward Elgar.
Hommes, Cars H. “Heterogeneous Agent Models in Economics and Finance.” In
Tesfatsion and Judd, eds., pp. 1109-1186.
Hoover, Kevin D. 2006. “The Past as the Future: The Marshallian Approach to Post
Walrasian Econometrics.” In Colander, ed., (2006), pp. 239-257.
Horgan, John. 1997. The End of Science: Facing the Limits of Knowledge in the Twilight
of the Scientific Age. New York: Broadway Books (paperback).
Ijiri, Y. and Herbert A. Simon. 1977. Skew Distributions and the Sizes of Business Firms.
Amsterdam: North-Holland.
Kahneman, Daniel and Amos Tversky. 1979. “Prospect Theory: An Analysis of Decision
under Risk.” Econometrica 47, 263-291.
Kant, Shashi and R. Albert Berry, eds. 2005. Sustainability, Economics, and Natural
Resources: Economics of Sustainable Forest Management. Dordrecht: Springer.
28
Keynes, John Maynard. 1936. The General Theory of Employment, Interest and Money.
London: Macmillan.
Kindleberger, Charles P. 2000. Manias, Panics, and Crashes: A History of Financial
Crises, 4th edn. New York: Wiley and Sons.
Kirman, Alan P. 1992. “Whom or What Does the Representative Individual Represent?”
Journal of Economic Perspectives 6(2), 117-136.
Kirman, Alan P. and Nicolaas J. Vriend. 2001. “Evolving Market Structure: An ACE
Model of Price Dispersion and Loyalty.” Journal of Economic Dynamics and Control 25,
459-502.
Kolmogorov, Andrei N. 1965. “Three Approaches to the Quantification of the Definition
of Information.” Problems of Information Transmission 1, 4-7.
Koppl, Roger and J. Barkley Rosser, Jr. 2002. “All that I have to Say has Already
Crossed Your Mind.” Metroeconomica 53, 339-360.
Rachel Kranton and D. Minehart. 2001. “A Theory of Buyer-Seller Networks.” American
Economic Review 91, 485-508.
Langton, Christoper G. 1990. “Computation at the Edge of Chaos: Phase Transitions and
Emergent Computation.” Physiica D 42, 12-37.
LeBaron, Blake. 2006. “Agent-Based Computational Finance.” In Tesfatsion and Judd,
eds., pp. 1187-1233.
Levitt, Steven D. and John A. List. 2007. “What Do Laboratory Experiments Measuring
Social Preferences Reveal about the Real World?” Journal of Economic Perspectives 21,
153-174.
Levy, M. and Sorin Solomon. 1997. “New Evidence for the Power-Law Distribution of
Wealth.” Physica A 242, 90-94.
Lux, Thomas. 2009. “Applications of Statistical Physics in Finance and Economics.” In J.
Barkley Rosser, Jr., ed., Handbook of Complexity Research. Cheltenham: Edward Elgar,
in press.
Mandelbrot, Benoit B. 1963. “The Variation of Certain Speculative Prices.” Journal of
Business 36, 394-419.
Mantegna, Rosario N. and H. Eugene Stanley. 2000. An Introduction to Econophysics:
Correlations and Complexity in Finance. Cambridge: Cambridge University Press.
29
Markose, Sheri M. 2005. “Computability and Evolutionary Complexity: Markets as
Complex Adaptive Systems.” Economic Journal 115: F159-F192.
Martinez-Alier, J. 1987. Ecological Economics: Energy, Environment and Society.
Oxford: Blackwell.
McCabe, Kevin, Daniel Houser, Lee Ryan, Vernon. Smith, and Theodore Trouard. 2001.
“A Functional Imaging Study of Cooperation in Two-Person Reciprocal Exchange.”
Proceedings of the [US] National Academy of the Sciences 98, 11832-11835.
McCauley, Joseph L. 2004. Dynamics of Markets: Econophysics and Finance.
Cambridge: Cambridge University Press.
McCloskey, Deirdre N. 2003. “Notre Dame Loses.” Eastern Economic Journal 29, 309315.
Minsky, Hyman P. 1972. “Financial Instability Revisited: The Economics of Disaster.”
Reappraisal of the Federal Reserve Discount Mechanism 3, 97-136.
Mirowski, Philip. 1989. More Heat than Light: Economics as Social Physics, Physics as
Nature’s Economics. Cambridge: Cambridge University Press.
Mirowski, Philip. 2007. “Markets Come to Bits: Evolution, Computation and Markomata
in Economic Science.” Journal of Economic Behavior and Organization 63, 209-242.
Nash, John F., Jr. 1951. “Non-Cooperative Games.” Annals of Mathematics 54, 286-295.
Nyarko, Yaw and Andrew Schotter. 2002. “An Experimental Study of Belief Learning
using Elicited Beliefs.” Econometrica 70, 971-1005.
Osborne, M.F.M. 1959. “Brownian Motion in Stock Markets.” Operations Research 7,
145-173.
Porter, David P. and Vernon L. Smith. 1995. “Futures Contracting and Dividend
Uncertainty.” Journal of Business 68, 509-541.
Rabin, Matthew. 1993. “Incorporating Fairness into Game Theory and Economics.”
American Economic Review 83, 1281-1302.
Rosser, J. Barkley, Jr. 1991. From Catastrophe to Chaos: A General Theory of Economic
Discontinuities. Boston: Kluwer.
Rosser, J. Barkley, Jr. 1999. “On the Complexities of Complex Economic Dynamics.”
Journal of Economic Perspectives 13(4), 169-192.
30
Rosser, J. Barkley, Jr. 2001. “Complex Ecologic-Economic Dynamics and Environmental
Policy.” Ecological Economics 37, 23-37.
Rosser, J. Barkley, Jr. 2006. “Complex Dynamics and Post Keynesian Economics.” In
Mark Setterfield, ed., Complexity, Endogenous Money and Macroeconomics: Essays in
Honour of Basil J. Moore. Cheltenham: Edward Elgar, pp. 74-98.
Rosser, J. Barkley, Jr. 2008a. “Is a Transdisciplinary Perspective on Economic
Complexity Possible?” Mimeo, Department of Economics, James Madison University,
available at http://cob.jmu.edu/rosserjb.
Rosser, J. Barkley, Jr. 2008b. “Debating the Role of Econophysics.” Nonlinear
Dynamics, Psychology, and Life Sciences, in press.
Roth, Alvin E. and J.K. Murnighan. 1978. “Equilibrium Behavior and Repeated Play of
the Prisoner’s Dilemma.” Journal of Mathematical Psychology 17, 189-198.
Samuelson, Paul A. 1947. Foundations of Economic Analysis. Cambridge: Harvard
University Press, enlarged edition, 1983.
Schumpeter, Joseph A. 1936. The Theory of Economic Development. Cambridge:
Harvard University Press.
Selten, Reinhard. 1975. “Reexamination of the Perfectness Concept for Points in
Extensive Games.” International Journal of Game Theory 4, 25-55.
Sethi, Rajiv and Eswaran Somanathan. 1996. “The Evolution of Social Norms in
Common Property Resource Use.” American Economic Review 86, 766-788.
Shannon, Claude E. 1948. “A Mathematical Theory of Communication.” Bell Systems
Technical Journal 27, 379-423, 623-656.
Silverberg, Gerald and Bart Verspagen. 2005. “A Percolation Model of Innovation in
Complex Technology Space.” Journal of Economic Dynamics and Control 29, 225-244.
Simon, Herbert A. 1957. Models of Man. New York: Wiley.
Simon, Herbert A. 1962. “The Architecture of Complexity.” Proceedings of the
American Philosophical Society 106, 467-482.
Smith, Adam. 1759. The Theory of Moral Sentiments. Reprinted 1976, Indianapolis:
Liberty Press.
Smith, Adam. 1776. An Inquiry into the Nature and the Causes of the Wealth of Nations.
London: Strahan and Cadell.
31
Smith, Vernon L. 1962. “An Experimental Study of Competitive Market Behavior.”
Journal of Political Economy 70, 111-137.
Smith, Vernon L. 1998. “The Two Faces of Adam Smith.” Southern Economic Journal
65, 1-19.
Smith, Vernon L. 2008. Rationality in Economics: Constructivist and Ecological Forms.
New York: Cambridge University Press.
Smith, Vernon L., Gerry L. Suchanek, and Arlington W. Williams. 1988. “Bubbles,
Crashes, and Endogenous Expectations.” Econometrica 56, 1119-1131.
Sornette, Didier. 2003. Why Stock Markets Crash: Critical Points in Complex Financial
Systems. Princeton: Princeton University Press.
Stahl, Dale and Paul Wilson. 1995. “On Players’ Models of Other Players: Theory and
Experimental Evidence.” Games and Economic Behavior 10, 218-254.
Stern, Nicholas. 2008. “The Economics of Climate Change.” American Economic
Review, Papers and Proceedings 98, 1-37.
Strotz, Robert H. 1955. “Myopia and Inconsistency in Dynamic Utility Maximization.”
Review of Economic Studies 23, 165-180.
Tesfatsion, Leigh. 2006. “Agent-Based Computational Economics: A Constructive
Approach to Economic Theory.” In Tesfatsion and Judd, pp. 831-880.
Tesfatsion, Leigh. and Kenneth L. Judd, eds. 2006. Handbook of Computational
Economics, Volume 2: Agent-Based Computational Economics. Amsterdam: NorthHolland.
Thaler, Richard. 1981. “Some Empirical Evidence on Dynamic Inconsistency.”
Economics Letters 8, 201-207.
Trivers, Robert L. 1971. “The Evolution of Reciprocal Altruism.” Quarterly Review of
Biology 46, 35-57.
Varian, Hal R. 1992. Microeconomic Analysis, 3rd edition. New York: W.W. Norton.
Veblen, Thorstein. 1898. “Why is Economics not an Evolutionary Science?” Quarterly
Journal of Economics 12, 373-397.
Velupillai, K. Vela, ed. 2005. Computability, Complexity and Constructivity in Economic
Analysis. Victoria: Blackwell.
32
Watts, Duncan and Steven H. Strogatz. 1998. “Collective Dynamics of ‘Small-World’
Networks.” Nature 393, 440-442.
Wolfram, Stephen. 1984. “Uncertainty and Complexity in Cellular Automata.” Physica D
10, 1-35.
Zak, Paul J. 2004. “Neuroeconomics.” Philosophical Transactions of the Royal Society
359, 1737-1748.
Zak, Paul J. 2005. “Trust: A Temporary Human Attachment Facilitated by Oxytocin.”
Behavioral and Brain Sciences 228, 336-368.
Zak, Paul J., ed. 2008. Moral Markets: The Critical Role of Values in the Economy.
Princeton: Princeton University Press.
Zipf, George K. 1941. National Unity and Disunity. Bloomington: Principia Press.
33
Download