Scope and Limits of Predictions by Social Dynamic Models: Crisis, Innovation, Decision Making Péter Érdi1,2 1 : Center for Complex Systems Studies, Kalamazoo College, 1200 Academy Street, Kalamazoo, MI 49006, USA and 2 : Department of Biophysics, KFKI Res. Inst. Nuclear and Particle Physics, Hungarian Academy of Sciences, Budapest, Hungary. Email: perdi@kzoo.edu Abstract Models of social dynamics grew up from methods used on the physical and life sciences. Different types of growth processes, among others those which lead to finite time singularities are reviewed. The scope and limits of the analogy between neurological disorders (such as epilepsy) and financial crises are analyzed by exploiting the concept of dynamical diseases. Governing rules of social dynamics phenomena can be extracted from collected empirical data, and they have some predictive power. Data for evolving patent citation networks, and annual budget changes are analyzed for generating dynamical models. Classification codes Keywords: social dynamics, feedback, causality, predictability, financial crisis, patent citations, innovation, budget changes, power law distribution 1. Introduction Sociodynamics, as a discipline grew up from the application of the theory of dynamical systems. (For illustration, Montroll, 1978, Helbing 1995, Weidlich, 2002, 2006). The theory became a predominant paradigm, and materialized the philosophical concept of causality by mathematical tools. This concept tells that causes imply effects, consequently the present state determines the future. The general idea behind sociodynamics is that we are able to characterize the state of a social system at a fixed point, and there are rules, which govern the change of state of a system. If a rule acting on the actual state is deterministic, there is only “one” future. In case of probabilistic rules, different futures could be predicted with certain probabilities. Dynamical models are able to describe both temporal phenomena (which assume perfect mixing of “players”, i.e. spatial homogenity) and spatiotemporal processes (which lead to homogenization in closed systems, and pattern formation – including waves – in open systems). Models of boundless and bounded growths, of business and other economic cycles and of the emergence of chaotic patterns are now extensively used. Complex systems theory offers a new perspective to understand the mechanism of the emerging patterns. Due to the consequence of natural and social crises, such as earthquakes, tsunamis, crises of financial markets, the occurrence of rare large (“extreme”) events are now in the focus of extensive mathematical analysis (Sornette 2003, Albeverio et al. 2006). In this paper, with the intention to follow the spirit of Mainzer (2009), who discussed the challenges of complexity we further analyze the scope and limits of predictability in social sciences. As Mainzer says, the research program to develop social dynamics is not physicalism. The assumption is that by using the concepts of dynamical systems, spatiotemporal patterns observed in natural and socioeconomical systems are supposed to be subject of causal explanations. One difference, what should be mentioned here is, that while the rules behind the changes are supposed to be fixed in nature, rules governing social dynamics can be subjects of change. Some ingredients of complex systems (Érdi, 2007) are reviewed first (Section 2) by emphasizing the role of feedback, circular causality and super-exponential growth related to uncompensated positive feedback. We will argue that large stock market crashes are social analogues of epileptogenesis. Social networks can be derived as the discrete, somewhat more microscopic approaches to continuous models; the nodes might be people, organizations, communities, documents, etc. Dynamic network models have two extreme forms: (i) propagation on networks, (ii) evolution of networks. An application for this latter case is given in Section 3, where the dynamics of innovation is analyzed. How did innovative technology develop? Are we in the position to predict the emergence of new technological fields? Are there any generative rules for the evolution of technology and innovation and can we extract them from data? Patent systems should promote innovation. Advances in computer technology and the availability of large patent databases have recently made possible to study many aspects of the patent system quantitatively. Since patents and the citations between them can be seen as a growing network, techniques from network theory can usefully be applied to analyze the patent citation networks. The “nodes” of the network are patents and the “links” are citations of one patent. For each patent it is required to cite all relevant prior art related to the patented invention. We were particularly interested in the inverse problem: what is the best rule for describing the evolution of a given network, or a set of networks, in a well-defined mathematical framework? (Csárdi et al, 2009) By extracting the rules which govern the dynamics, there are some possibilities to make predictions. Section 4 is about the dynamics of annual government budgeting. Current empirical data show that frequency distributions characterizing budget change in the United States and elsewhere strongly deviate from the Gaussian distribution. Year to year budget changes (adjusted for inflation) from many nations are all distributed as power functions. We are building a model framework which takes into account such assumptions as: (i) limited attention of political institutions; (ii) friction by collective decision rules, resistance to change, (iii) error accumulation or strong signals, (iv) prioritized preferences. Such kind of model frameworks have predictive power, at least in statistical sense. 2. Some Ingredients of Complex Systems What are simple systems? Before telling anything about complex systems, we should answer this question. Here are some characteristic properties of simple systems: Single cause and single effect A small change in the cause implies a small change in the effects Predictability Common sense thinking and problem solving often adopts the concept of`”single cause and a single effect”. It also suggests that small changes in the cause imply small changes in the effect. It does not literally mean (as sometimes it is mentioned) that there is a linear relationship between the cause and the effect, but it means that the system's behavior will not be surprising, its is predictable. With a somewhat more technical terminology, small changes in the parameters (or in the structure of the system) do not qualitatively alter the system's behavior, i.e. the system is “structurally stable”. Circular causality, feedback, growth processes Circular causalit y in Figure 1. Illustration of linear and circular causality. essence is a sequence of causes and effects whereby the explanation of a pattern leads back to the first cause and either confirms or changes that first cause. The concept itself had a bad reputation in legitimate scientific circles, since it was somehow related to use “vicious circles” in reasoning. It was reintroduced to science by cybernetics (Wiener, 1948) emphasizing feedback. The concept of circular causality is reflected also in the theory of reflexivity, an approach promoted in economics by George Soros (1988). In a feedback system there is no clear discrimination between “causes” and “effects”, since the output influences the input. In particular, feedback between the rate of change (x') and the state (x) controls the dynamics. Growth processes describe the increase of different physical, chemical, biological, economical, etc. state variables, such as size, chemical concentration, population density, price, etc. Assuming that other variables do not influence the dynamics, we might set a polynomial differential equation: x'=x m (1) Different types of growth processes can be described by models relying on different assumptions for the velocity of the specific process (Fig. 2.). Figure 2. Unbounded growth processes: infinite and finite-time singularities The solution of the equation with m>1 is x t =x 0 t c− t tc − 1/ m− 1 If m>1 The equation itself implements large, “higher-than-linear” (HTL) positive feedback, which seems to be a general mechanism behind finite time singularities. It ensures that the “instantaneous doubling time” tends to zero after a finite period. Finite time singularity roughly speaking means that a dynamical variable gets an infinite value during finite time. This phenomenon is qualitatively different from the exponential growth, when infinite value can be attained during infinite time only. In terms of application, there are two types of approaches. If we build assumption-driven models (as we did by setting Equation 1), and ask what can we say about the properties of the model (“about the solution of the equation”), we speak about “direct problems”. Another set of problems is data-driven, and asks for finding out the generating rules. The magic bullet of social dynamic modeling might be related to the sophisticated combination of inverse and direct models. From finite-time singularity to crash The lack of the stabilizing effects of negative feedback mechanisms may lead to catastrophic consequences. If there are no mechanisms to compensate for the effects of HTL positive feedback, the processes lead to finite time singularities. Positive feedback seems to be a general mechanism (Sornette, 2002) behind eruption of earthquakes, stock market crashes, hyperinflation, and epileptic seizures. Conventional economic theory, which is based on the equilibrium between demand and supply, says that decreasing demand is compensated by increasing supply. Complex systems approach, however, suggests that partially due to our susceptibility to imitate each others behavior, there is a period when both demand and supply increase, which finally lead to singularities. Equilibrium theory works well when the negative feedback effects expresses its stabilizing effect to the increase due to positive feedback. While in “normal situations” the activities of “buyers” and “sellers” neutralize each other, in “critical situations” there is a cooperative effect due to the imitative behavior (“everybody wants to buy it since everybody else has already bought it”). So, the positive feedback is HTL. Super-exponential increase (due to the irrational expectations) cannot be continued for “ever” due to the unstable nature of this process and the increase is unsustainable. Consequently, it should be followed by a compensatory process (i.e. stock market crash). Uncompensated Positive Feedback: Epilepsy and Financial Crisis Onset of two types of financial crises, namely stock market crash and hyperinflation (Sornette et al, 2003) are generated by the positive feedback between the actual and the expected growth rate. Large stock market crashes are social analogues of big earthquakes and epileptogenesis (as all analogies, this also has scope and limits). Sornette's view is that the stock market crash is not induced by single local events (such as a raise in the interest values, or other governmental regulations), but due to the unsustainable velocity of price increase. This speculative increase will take the system more and more unstable situation. Finally, the market collapses due to any small disturbance. Unsustainable velocity of price increase resembles to over-excitation of the epileptic brain. Indeed, the existence of vicious circle-type mechanism characterizes both epilepsy and financial crisis. The expression, “Seizures Beget Seizures”, for explaining the progression of epilepsy goes back to Sir William Gowers, a British neurologist in the nineteenth century. Complex systems research puts epilepsy into a broader perspective: As a complex systems disorder, epilepsy poses challenges for identifying single mechanisms or processes that are required for epileptogenesis or progression. An attempt to identify a list of specific molecular or cellular mechanisms responsible for progression of epilepsy is likely to be difficult because emergent phenomena such as network synchronization are not likely to be universally or linearly related to the specific defect (Sutula, 2004, see also Ben-Ari et al. 2008). The stable dynamic operation of the brain is based on the balance of excitatory and inhibitory interactions. The impairment of the inhibitory synaptic transmission implies the onset of epileptic seizures. Epileptic activity occurs in a population of neurons when the membrane potentials of the neurons are “abnormally” synchronized. Both experiments and theoretical studies suggest the existence of a general synchronization mechanism in the hippocampal CA3 region. Synaptic inhibition regulates the spread of firing of pyramidal neurons. In experimental situations inhibition may be reduced by applying drugs to block (mostly) GABA-A receptors. (GABA is the most important inhibitory transmitter). If inhibition falls below a critical level, the degree of synchrony exceeds the threshold of normal patterns, and the the system's dynamics switches to epileptic pattern. Collective phenomena occurring in neural networks, such as the case of disinhibiton-induced epilepsy have been studied successfully by combined physiological and computational methods. (An excellent book for summarizing the basic physiological mechanisms and computational models of epilepsy, see Soltesz and Staley. 2008). How economics can exploit the analogy? There are at least two fields, where the developed methods can be transferred; detecting of precursors and new therapeutic strategies. Methods for epileptic seizure prediction are subject of extensive research. While it is true that “...none of the current epileptic seizure prediction methods can widely be accepted...” (Liu et al, 2009), still the methods developed can be transferred from computaional neurology to economics. The analysis of EEG curves based on dynamical approach offered methods to separate the different qualitative regions before and during a seizure. The challenge is to define good measures. Recurrence plot analysis, see http://www.recurrence-plot.tk/ (11 February, 2010), is a method to analyze complex dynamics. A measure (called recurrence rate) proved to be useful (Li et al 2004) to separate the different qualitative regions before and during a seizure. While by visual inspection only two stages (inter-ictal and ictal periods) can be discriminated, the analysis helped to show the existence of a third stage, the pre-ictal period, which separated the two stages previously mentioned. For details and figures see the original paper. Another method for detecting precursors is Sornette's method based on the phenomenon called logperiodicity. He found that on a refined scale for different systems (from acoustic energy to stock market price) the dynamics leading to 'finite-time singularity'' is modulated with log-periodicity. This log-periodic behavior is characteristic for such kinds of critical events, when it is driven by slowly accumulating, endogenous processes. (Of course, there are events which are responses to exogenous shocks) From the technical point of view, for the stock market crash, the simple price equation showing β finite time singularity, log [p t ]=A+B t c− t , should be replaced by log [p t ]=A+B t c− t β [ 1+C⋅ cos ω⋅ log T − c− t T ] . Very formally, log-periodicity reflects the fact that critical exponents, more generally dimensions can be complex, (of course, here “complex” means such kind of numbers which when squared give negative values). The method was used to see that endogenous price increase due to positive feedback mechanisms of speculative herding should be followed by crashes. This way the log-periodicity serves as a signature of crash, since it precedes the actual drawdown. Therapy Epilepsy, and some other neurological and psychiatric disorders can be understood as dynamical diseases. Such diseases are characterized by sudden changes in qualitative dynamic behavior (Belair et al 1995). We offered (Aradi and Érdi. 2006) a new way of drug discovery by using dynamical models. Efficient drugs (“therapeutic interventions”) should shift the system from pathological dynamical states to normal ones. What we learned from neuroscience is that there are some pathological mechanisms, which by blocking the inhibitory effects, induce a vicious circle of disinhibition. “...This may result in aberrant activity triggering epileptiform discharges...” (Rigo et al, 2002). A large class of antiepileptic drugs acts to break the vicious circle, i.e. to restore the inhibition by acting on GABAergic synaptic transmission. Subprime mortgage crisis is being seen now as a typical example of vicious circle: http://www.viswiki.com/en/Virtuous_circle_and_vicious_circle (February 26, 2010). The self-regulated market collapsed. What to do with it? “...Now we're in the therapy phase. And what therapy are we using? Well, it's very interesting because we're using two quite contradictory courses of therapy. One is the prescription of Dr. Friedman Milton Friedman, that is - which is being administered by the Federal Reserve: massive injections of liquidity to avert the kind of banking crisis that caused the Great Depression of the early 1930s. I'm fine with that. That's the right thing to do. But there is another course of therapy that is simultaneously being administered, which is the therapy prescribed by Dr. Keynes - John Maynard Keynes - and that therapy involves the running of massive fiscal deficits in excess of 12 percent of gross domestic product this year, and the issuance therefore of vast quantities of freshly minted bonds. There is a clear contradiction between these two policies, and we're trying to have it both ways. You can't be a monetarist and a Keynesian simultaneously - at least I can't see how you can, because if the aim of the monetarist policy is to keep interest rates down, to keep liquidity high, the effect of the Keynesian policy must be to drive interest rates up.” (Niall Ferguson, 2009). Each analogy has its limits of course. What we certainly learned from systems theoretical perspective that both the healthy brain and a healthy macroeconomics have cycles. In both cases the cycles are mostly endogenous (of course, self-sustained oscillation occurs in open systems only). Furthermore, in both cases we see pathological rhythms (including arhythmia, i.e chaotic patterns). Neo-Keynesian pioneers (such as Michael Kalecki and Nicholas Kaldor) in the 1930s suggested to use nonlinear dynamical systems to explain the endogenous generation of business cycles. These cycles are natural part of a working economy. The soul of the Kaldor model is that investment and saving are nonlinear functions of income. The mathematical version of the original model leads to limit cycle behavior, i.e. self-regulated oscillations, other variants of the Kaldor and other post-Keynesian models may lead to chaotic patterns (for a review, see Rosser 2006). While chaos has benevolent functional consequences in the neural information processing, it does not seem to be desirable in economic systems. Therapeutic strategies should maintain the “natural” oscillations (superimposed to the growth processes) built-in our actual world economical system. (Of course, economics is a moral science, normative models of economics describe how a system should behave. See also Aruka, 2007). There is some more analogy between economical crises and neurological and psychiatric diseases. The world-wide economic crisis can be interpreted as collective panic disorder, as a group of German psychiatrists suggested (Sperling et al. 2008). Based on this assumption the same authors in another paper (Sperling et al, 2009), suggested therapeutic strategies. Steps of cognitive behavioral therapy can be applied for both the banking system and the general public (including both media and individual persons). Social dynamics: inductive versus deductive Francis Bacon (1561-1626) suggested that the only way of acquiring scientific knowledge is “true and perfect Induction”. Whatever successful this program has been (nobody denies the role of thorough observation and data collection), Bacon underestimated the role of the creative imagination, and the formation of testable (falsifiable, using Karl Popper's terminology) hypotheses. As it is well known, Kepler's theory relied on Tycho Brahe's observations, as searched for simple algebraic relations for the motion of celestial bodies in defense of Copernicus' heliocentric concept, and derived his laws empirically. In the first law Kepler corrected Copernicus' hypothesis about the circular motion of the planets, stating that their orbits are ellipses, and the Sun's center of mass is at one of the two foci of the ellipse. The second law gives some quantitative characterization: The planets move so that the line between the Sun and the planet (i.e. the radius vector) sweeps over equal areas in equal intervals of time in any place of the orbit. Both the first and the second laws refer to the individual motion of the planets. The third law, however, makes quantitative connection between the motions of two planets, by finding relationship 3/2 between the periods and the mean distances of the orbits from the Sun. Specifically, T ~ R is a scaling law1, where T is the period of a planet's orbit, and R is the length of its semi-major axis. It is interesting to note that Sornette's log-periodicity is also based on scaling laws. Discrete scale-invariance governs the dynamics of the stock market. 1 Einstein explained the logical status of the Kepler's laws (Einstein 1927): “But the principal thing is that these laws have reference to motion as a whole, and not to the question of how there is developed from one condition of motion of a system that which immediately follows it in time. They are, in our phraseology of today, integral laws and not differential laws.” In Einstein's remark the term “motion as whole” means that Kepler's laws describe the orbit globally. Kepler did not know the cause of the motions, i.e. he was not able yet to predict the motion of a planet from its actual position and velocity. To put it another way, there were no known local rules to calculate the motion. By defining the law of universal gravitation Newton was able to derive (and slightly modify) Kepler's data-driven global, integral laws by deductive, local, differential rules. Based on his theory it was possible to make predictions which led to the discovery of Neptune and Pluto. How to apply this dynamical approach to social sciences? A lot of data has been accumulated in social sciences, but the overwhelming majority of them have been evaluated and interpreted in a static or equilibrium perspective. The a priori knowledge of the rules, which basically govern the dynamic evolution of the systems is difficult: social dynamics is in the post-Keplerian and pre-Newtonian age (and it may remain there). Still, at least in certain cases, we try to predict our social Neptunes and Plutos2. In the next two sections, sketches of examples are given for developing dynamical models by combining inductive and deductive approaches. 3. Dynamic Modeling of Innovation: Development of the Patent Citation Network Technological development is reflected in patents and in the citations between them. The network of patents connected by citations is an imperfect but faithful representation of technological development. A patent citing another implies that the cited patent represents a piece of previously existing knowledge that the citing patent builds upon. Consequently, a directed link between patents can be interpreted as an idea-flow. Nassim Nicholas Taleb (Taleb 2005) analyses Benoit Mandelbrot's vision on applying the concept of fractals to financial fluctuations: “This would make BM the true Kepler of the social sciences. The analogy to Kepler is at two levels, first in the building of insights rather than mere circuitry, second because you can step on his shoulders – the title of Kepler or “Newton of the social sciences” is one so many thinkers with grand ideas have tried to grab (Marx for one aimed at being the Newton of the sciences of man). I am not in the business of defining genius, but it seems to me that the mark of a genius is the ability to pick up pieces that are fragmented in people’s mind and binding them together in one, a meta-connection of the dots.” 2 Patent citation networks are special cases of evolving networks. The evolution of a network involves the addition (and in the more general case even deletion) of edges and vertices. Gábor Csárdi developed a mathematical framework to solve the ”inverse problem”, which asks: what is the rule which gives the best description of the evolution of a given network, or a set of networks. (Csárdi et al 2007, 2009). Modeling framework: states, and kinetics How to define a mathematical model framework for describing the temporal evolution of the patent network? 1. State variables. The first, obviously somewhat arbitrary step is to decide about the state variables. In our model, each patent is described by two variables: k, the number of citations it has received up to the current time step, and l, the age of the patent, which is simply the difference between the current time step (as measured in patent numbers) and the patent number. Because a given patent may cite more than one other patents, several citations may be made in one time step. These two variables define what we call the “attractiveness” of a patent, A(k,l), which determines the likelihood that the patent will be cited when the next citation is made. 2. Temporal change. The evolution of the network is modeled by a discrete time, discrete space stochastic dynamic system. Time is measured in patent number units, so that each “time step” represents the citations made by a single patent. In every time step the probability that an older patent will be cited is proportional to the older patent's attractiveness multiplied by the number of citations made in that time step. We found that this simple model gives a very good approximation of the observed kinetics of the growth of the patent citation network. The A(k,l) function determines the evolution of the network. It describes the average citation preferences of the citing patents (the inventors and patent examiners in reality). We measured and analysed A(k,l) for the United States patent system during the time period covered by our data and found first that the parametrization by k and l consistently describes the average kinetics of the patent citation network. Of course, there are many hidden relationships behind patent citations like social relationships between patentees, the evaluations of the significance of the cited patent by the patent examiner and the technological relationship between the citing and cited patents that our probabilistic approach does not capture. The way in which these “microscopic dynamics” are translated into the average behavior that we observe remains an open question. Direct and inverse problems can be defined, as Figure 3 shows. Figure 3.The direct and the inverse problems. In the direct problem we have the A(·,·) kernel function (and maybe other kernel functions for more general networks) and we use a kernel-based network generator algorithm to create an artificial network. The question is what kind of structures can we obtain from a given class of kernel functions. In the inverse problem, we examine a real network assembled from observations. The kernel-based measurement algorithm takes this as the input and gives a possible kernel function as the output. If the measurement algorithm is good then the output kernel function is good, i.e. it creates the observed network with high probability when used in a kernel-based generator. The inverse problem can be solved by the intuitive (frequentist) and the better founded maximum likelihood methods. Figure 4. Sections from the in-degree and age based maximum likelihood fitted kernel function for the US patent citation network. Both plots have logarithmic axes. The obtained free-form kernel can be well fitted with a function where the effects of degree and age are separated, A d,l =Ad d ⋅ Al l . The degree-dependent part is the usual preferential attachment form: A d d =d α +a , of course with a different exponent and different a parameter. We fitted the agedependent part with a double Pareto function, this is a unimodal function with an initial power-law increase and a power-law decrease in the tail: Al l = { 1 tp 1 tp β p− 1 if l≤ t p α p− 1 if l>t p } About 96% of the goodness of the free-form model is preserved with the fitted form. Consequently, the preferential attachment times double Pareto form is a very good model if one wants to model the network based on in-degree and age. Change in the Dynamics Rules in social dynamics might be subjects of change. So far we assumed that a single kernel function is valid during the whole development of the network. We showed that the dynamics of the patent network changed around 1990 (supposedly) because of the introduced legal changes. We handled the time dependence of the kernel function by using a sliding time-window of 500,000 patents, i.e. separate kernel functions were estimated for each time window. A model based on in-degree and age was studied. We used the maximum likelihood fitting method to α fit the parameters of a predefined shape: A d,l = d +a ⋅ Al l , where Al(l) is a double Pareto form. Stratificatio n meant that the more citable patents are increasingl y more citable than the less citable patents. More than 400 classes and six big categories were Figure 5.Time evolution of the stratification parameter using in-degree and defined. age. The six categories are Chemical, Computers and Communications, Drugs and Medical, Electrical and Electronic, Mechanical and Others: Stratification proved to be higher in some categories than in others, i.e. . Computers & Communication and Drugs and Medical Patent citation network is highly assortative (i.e. say Chemical cites Chemical). In these categories a “few” highly important patents receive the majority of citations. Based on the assumption that emergence of new technological fields can be detected from nonassortative citatitons, recently, with Kinga Makovi and Péter Volf (in preparation) we started a new analysis. The dynamics of the community structure of the non-assortative patents – those who have citations outside their own industry – allows us to make predictions about the near future by describing the evolution of clusters in terms of growth, shrinking, splitting and merging, which might be the elementary processes found by Palla et al (2007). While the patent citation network is an open system with continuous interaction with its environment, the scopes and limits of predictions are different from that of the natural sciences. Patent laws, habits of patent examinees, the pace of economic growth and many other factors influence the development of the patent network, and vice versa, the creation of patents changes the environment. This limits the scope of predictions, which can be made to only a statistical description: we can predict only patterns of behavior, but not specific events. 4. Dynamic Modeling of Public Policy Decision Making Processes General remarks Institutional decision making mechanisms can be analysed in terms of their information processing, and/or as an input/output system. Annual budgeting has been studied well. Input data are supposed to come from independent sources, and assumed to be characterized by Gaussian distributions. The measured output shows a deviation from the Gaussian distribution, so the general problem is to uncover the dynamic mechanism of the transformation during decision making. By knowing this mechanism we might have some method to make predictions for the future of the national budgets, at least in a statistical sense. Lessons from political science Bryan Jones and Frank Baumgartner have published a set of papers and books focusing on annual budget changes (Jones and Baumgartner 2005). The budget is a scalar number annually (the sum of many elements, i.e. categories). Each year a new point is generated , and the difference is the budget change (adjusted to the inflation). So technically, data on annual budget changes define a set of positive and negative numbers, as values of some random process. Leptokurtic distribution of percentual budget changes were observed in a broad range of settings: small increases and small decreases of budgets and budget components are the most frequent, but time to time large increases and cut-offs are observed as well. Most frequency distributions appeared linear on diagrams with double log axes, which strongly suggests a special case of kurtosis, the power-law distribution. Dynamic mechanisms leading to power laws are one of the hot topics in the theory and practice of complex systems (for a short review, see Érdi 2007, 6.3.2.). In our case a plausible hypothesis is that the skew distribution emerges due to the existence of institutional friction. Figure 6. Category Midpoints of Percentage Budget Real Percent Change Model framework Traditional pure incremental models of budget changes (and their improved versions, Padgett 1980) lead to Gaussian distributions. In a more advanced model (Jones at al 2001) the key was the recognition that humans are disproportionate processors of information - that is, they are incapable of adjusting actions directly to the incoming information that has implications for action. There are a variety of reasons for this inability, but the most important is the inability of people to pay attention to multiple streams of information at the same time. Instead, they juggle among income streams, with emotional arousal being the central adjustment mechanism. This leads to disjoint and episodic shifts in attention to the information streams they face. Punctuations in processing capacity, and hence in any action taken, is a direct result of coping with numerous streams of information in a complex world. In political systems, there are good reasons to expect that this tendency may be magnified. Democracies have built-in error correction mechanisms that if policymakers are not appropriately responding, they may be replaced. Predictive models were built based on the assumptions listed: limited attention of political institutions friction by collective decision rules resistance to change error accumulation prioritized preferences Jones and Baumgartner (2005) developed a single threshold model to account for the strongly leptokurtic budget output distributions they observed. Errors, in the sense of a mismatch between information inputs and policy outputs, would accumulate in a policy area until the errors exceeded a threshold, after which policymakers would attend to the problem and initiate action. Simulations based on the model suggested considerable agreement, but problematic mismatches in some cases. In a cooperation between political scientists and complex systems modelers, László Zalányi made further improvements on the model (Jones et al 2009). A dynamic model with an error-accumulating threshold-based nonlinearity and an infinite memory of earlier signals was defined. A single threshold was set: if the accumulated input is below the threshold, the response is small, if it is above than large. The variables and parameters of the model can be summarized as follows: Rt: response variable St: input signal, (Gaussian random variable) C: threshold λ: efficiency parameter or friction β: amplification parameter. Error accumulation (incorporated as sum of the signals) is a key feature of the model. 0 < λ < 1, the efficiency below the threshold C is lower than one, but definitely should be lower than β. This is one of the characteristic point of the model: below the threshold there are smaller reaction than it is above the threshold. The model has two versions. Error accumulation can or cannot be reset to zero, Rt = β⋅ S t if ∑ S t' >C λ⋅ β⋅ S t otherwise where k is the time of the resetting (k=1 means no reset). { } Simulations results for two situations (without and with resetting) are shown in Fig. 7. Appropriate choice of the λ, β, C parameters makes this model reach the actual kurtosis values observed on real data. C = 1, λ = 0.4 and β = 1 were chosen. Figure 7. A leptokurtic distribution as a simulated response, comparing to a Gaussian distribution with and without resetting the counter What is the difference between the two models? The counter spends a lot of time above the threshold and also below in the first case. The higher the threshold the narrower the output distribution (increasing the kurtosis) but the rarer the “overshooting”, i.e. the big changes. The resulted output distribution is a sum of two Gaussians. In the second case the output is much narrower as we restrict the “time” what the counter can spend above the threshold, and hence the weight of the wider Gaussian (β) in the sum is smaller. However, there is one more difference. It is hardly seen that the output distributions are asymmetric (and hence they are not exactly sums of Gaussians, but can be approximated so very well). This asymmetry comes from the requirements of threshold crossing direction: it can only be crossed in the positive direction to switch into the frictionless state. In summary, the single threshold model cannot explain the higher frequency of the occurrence of some extreme events. A three-threshold model The institutional decision making is obviously a multi-stage process. The simplest way to implement the series of processes into a model is to increase the number of thresholds. Different frictions for negative and positive counter (i.e. demand) were defined. Also it was necessary to smooth the positive side of the output distribution which can be remedied by introducing a second, upper threshold for the real overshooting reactions and a specific friction for the intermediate regime. This model gives different responses in the negative counter depending on the sign of the actual input: the response is different to an increase than to a decrease demand. By incorporating these modifications a much finer fit to the budget change distributions was obtained. C1=-4 C2=1.2 lower threshold middle threshold C3=3 λ1n=0.02 λ1p=0.03 λ2p=0.3 β=1 T upper threshold efficiency for the region of negative accumulated sum efficiency for the region of positive accumulated sum efficiency between the two positive thresholds and denotes the time of last resetting to 0 of the accumulated signal. The response: Rt = { λ1p⋅ β⋅ S t ∑ s t' >C 3 ,T=t λ2p⋅ β⋅ S t if C 3 ∑ st' >C 2 λ 1n⋅ β⋅ S t if ∑ st' 0 ∧ S t 0 λ1n⋅ β⋅ S t if ∑ st' <C 1 ,T=t< 0 if C 2 ∑ st' 0 ∨ 0 ∑ s t' >C 1 ∧ S t β⋅ S t if 0 } Fig. 8 shows the US budget data with the fitted model distribution. The fittings were performed as follows: the six-dimensional parameter space was numerically explored and the set of parameters which gives the maximum likelihood of the dataset was selected. The three-threshold model fits are better than that of the former models. It reproduces generally the shape, the asymmetric feature of the distribution of the data, and gives significant probability to extreme events. However, the model is not perfect: both the mean and standard deviations of the simulated distributions are bigger then the momenta of the datasets. The reason of this deviation partially comes from the features of the model: it is a continuous one which has to fit to a rather sparse dataset. There are several ways to extend or modify the threshold models. To make the next step one has to investigate the decision making procedure of the different countries in details due to the natural limitation of the annual budget data sizes. Figure 8. Transformation of the Gaussian input signals to skew distribution fitted to the US budget data. The upper figure shows the tenfold enlarged input Gaussian sample distribution and the budget distribution histogram. The lower figure shows the measured data histogram and the fitted model output histogram. Further questions Dynamical modeling strategy offers a methodological tool to set and test hypotheses about the budgeting process. Many open questions left to be answered. Can we design institutions that are more efficient? Do we want to? Can we get direct measurements of decision making costs? Why do budgets produce power laws when virtually all other distributions (such as GDP) are less extreme? 5. Conclusions Dynamic systems theory proved to be very efficient in modeling phenomena of the physical and life sciences. Local rules have the power to make casual explanations and predictions. Rules behind different types of growth processes were reviewed. As Sornette argued, finite time singularities are rather general models of different types of crises. Epileptic seizures and certain types of financial crises are the results of uncompensated positive feedback. A little bit more is suggested: the concept of “dynamical diseases” can cautiously be used for understanding undesired macroeconomic irregularities, and offering therapeutic interventions. As Kepler and Newton combined data-driven inductive and concept-driven deductive approaches, governing rules of social dynamics phenomena might be extracted from collected empirical data. Such kinds of rules might have some predictive power. We showed two examples. First, a project based on data for evolving US patent citation networks were mentioned. The analysis led to the construction of a stochastic kinetic model of the development of the citation network. We believe that it is possible to identify technological branches at various times and predict the emergence of new technologies based on the clustering of the citations. Second, comparative data for the annual budget changes are analysed for generating dynamical models for the institutional decision making (i.e. policy making) mechanisms. The basic phenomenon, valid over different countries, is that budget change distributions deviate from the Gaussian. Since it seems to be very plausible (but data are missing) that the input to the policy making system is Gaussian, the systems transforms the Gaussian distribution to either a leptokurtic, or most probably a power law distribution. Institutional frictions and error-accumulation phenomena are taken into account by setting a set of thresholds. This model framework also has some predictive power. Acknowledgments Thanks for motivation to Prof Yuji Aruka: this paper grew out from a lecture presented at Chuo University in October 2008. The “patent citation project” described is a cooperative effort with Gábor Csárdi, Katherine Strandburg and Jan Tobochnik, recently with Kinga Makovi and Péter Volf. The “budget change” project is collaborative work with Bryan D. Jones, Frank Baumgartner, and László Zalányi. Thanks to Mihály Bányai for technical help. General support from the Henry Luce Foundation is acknowledged. References Albeverio S., V. Jentsch and H. Kantz (eds) (2006); Extreme Events in Nature and Society, Springer, Berlin/Heidelberg/New York. Aradi I and Érdi P (2006): Computational neuropharmacology: dynamical approaches in drug discovery. Trends in Pharmacological Sciences 27(240-243) Aruka Y (2007): The Evolution of Moral Science: Economic Rationality in the Complex Social System Evolutionary and Institutional Economics Review 4(217-237) Belair,J, Glass L, An der Heiden, U and Milton J (1995):. Dynamical disease: Identification, temporal aspects and treatment strategies of human illness. Chaos 1995, 5(1-7). Ben-Ari Y., Crepel V and Represa A (2008): “Seizures beget seizures in temporal lobe epilepsies: the boomerang effects of newly formed aberrant kainatergic synapses”, Epilepsy Curr. 3(68-72) Buckmaster PS, Thind K (2005): Quantifying routes of positive-feedback among granule cells in a model of temporal lobe epilepsy. Epilepsia 46:91–131 Csárdi G, Strandburg KJ, Zalányi L, Tobochnik J, Érdi P (2007): Modeling innovation by a kinetic description of the patent citation system Physica A: Statistical and Theoretical Physics 374(1-2) 783793 Csárdi G, KJ Strandburg, J Tobochnik, Érdi P. (2009): The inverse problem of evolving networks with application to social nets. Chapter 10 in Handbook of Large-Scale Random Networks, eds.: Béla Bollobás, Róbert Kozma, Dezső Miklós. Springer-Verlag Einstein A (1927): Isasc Newton. Smithonian Annual Report Érdi P (2007): Complexity Explained, Springer Verlag, Berlin-Heidelberg Fergusson, N (2009): http://www.nybooks.com/articles/22756) Gowers WR. (1881): Epilepsy and other chronic convulsive disorders: their causes, symptoms and treatment. London: J&A Churchill; Ionov D (2009): Self-reinforcing loop mechanism in epilepsy. Med. Hypotheses 73(608-9) Jones B. D. (2001). Politics and the Architecture of Choice . University of Chicago Press. Jones B. D., and Baumgartner F. R. (2005): The Politics of Attention: How Government Prioritizes Problems University of Chicago Press. Jones BD, Zalányi L, Baumgartner F and Érdi P, (2009): Dynamic Threshold Modeling of Budget Changes. Proc. AAAI Fall Symp.; Complex Adaptive Systems and the Threshold Effect: Views from the Natural and Social Sciences, AAA Press, pp. 61-66. Helbing D (1995): Quantitative Sociodynamics. Stochastic Methods and Models of Social Interaction Processes (Kluwer Academic, Dordrecht). Li X., Ouyang, G., Yao X., and Guan, X. (2004): Dynamical characteristics of pre-epileptic seizures in rats with recurrence quantification analysis. Physics Letters A 333 (164-171) Liu D, Pang Z and Wang Z (2009): Epileptic seizure prediction by a system of particle filter associated with a neural network. EURASIP Journal on Advances in Signal Processing Mainzer K (2009): Challenges of Complexity in the 21st Century. Evolutionary and Institutional Economics Review 6(1-22) Montroll EW (1978)L “Social Dynamics and the Quantifying Forces”,. Proc. Natl. Acad. Sci. USA 75:4633-36) Padgett J. F. (1980): Bounded rationality in budgetary research. American Political Science Review 74:354–72. Palla G, Barabási A.-L, Vicsek T (2007): Quantifying social group evolution. Nature 446:7136, 664667 (2007) Rigo J.-M , Hans G et al (2002): The anti-epileptic drug levetiracetam reverses the inhibition by negative allosteric modulators of neuronal GABA- and glycine-gated currents. Br. J.Pharmacol. 36(659-672) Rosser J. B., Jr. (2006): ‘Complex dynamics and post Keynesian economics’, in Setterfield, M. (ed.): Complexity, endogenous money and macroeconomic theory: essays in honour of Basil J. Moore, Edward Elgar, Cheltenham. Sutala T (2004): Mechanisms of epilepsy progression: current theories and perspectives from neuroplasticity in adulthood and development. Epilepsy Research, Volume 60(161-171) Soltesz I and Staley, K (2008): Computational Neuroscience in Epilepsy, Elsevier. Sornette D: Predictability of catastrophic events: Material rupture, earthquakes, turbulence, financial crashes, and human birth, Proc. Natl .Acad. US 2002 Sornette D (2003): Why Stock Markets Crash: Critical Events in Complex Financial Systems. Princetion Univ. Press. Sperling, W, Bleich, S and Reulbach, U (2008) Black Monday on stock markets throughout the world – A new phenomenon of collective panic disorder? A psychiatric approach, Medical Hypotheses, Volume 71, Issue 6, Pages 972-974 Sperling W, Biermann,T and Maler J (2009), Global panic reaction – A therapeutic approach to a world-wide economic crisis. Medical Hypotheses, Volume 73, Issue 2, Pages 167-168 Taleb N (2005): Mandelbrot Makes Sense: A Book Review Essay http://www.fooledbyrandomness.com/mandelbrotandhudson.pdf (March 15th, 2010) Soros G (1988) The Alchemy of Finance: Reading the mind of the market. Simon and Schuster Weidlich W. (2002): Sociodynamics: A Systematic Approach to Mathematical Modelling in the Social Sciences Weidlich W (2006): Intentions and Principles of Economics Review. 2(161-165) Sociodynamics, Evolutionary and Institutional Wiener N (1948) Cybernetics: Or Control and Communication in the Animal and the Machine. MIT Press.Cambridge,