A short note on concepts in rational choice theory Microeconomic theory is probably the most elaborate and well-structured theory in the social sciences. Microeconomic theory is basically a theory of rational choice by individual actors. Microeconomic theory in its original form includes consumer-theory, producer-theory and theory on how consumers and producers interact in markets. However, the core concepts underlying this theory can and has also been used to study a broader range of phenomena. This is due generality of the core assumptions underlying the theory, and I will here speak generally about the theory of rational choice. Not only consumers and producers, but also voters, politicians, lobbyists social classes and states in international affairs have been studied under the same core assumptions and by using similar concepts as those used in microeconomic theory in a narrower sense. That being said, the core rational choice assumptions are too general to be used fruitfully in social science research without putting additional information about social structure and actors into the particular model of interest. I will here dwell mostly on the general core concepts and assumptions, but from time to time illustrate how rational choice analysts put more flesh on the skeleton of the general framework in concrete research.1 Thin and instrumental rationality A “deceptively simple sentence” that summarizes the theory of rational choice is according to Jon Elster: “When faced with several courses of action, people usually do what they believe is likely to have the best overall outcome” (Elster, 1989:22). It is perhaps best to think of rational choice not as one theory, but rather a more general theoretical framework or a family of related theories. However, the notion of instrumental rationality, that individual actors choose those actions that maximize the satisfaction of their goals, is common for this family. Rationality does not say anything particular about the value of the goals actors pursue; rationality is therefore thin in general. Rationality, as used in rational choice theory, relates to how actors use the means and strategies at their disposal to follow these goals. Rationality is instrumental. However, in concrete research settings analysts feed certain preference structures into their models, and ideally, these preferencestructures should resemble central elements of the goals and aims that the real world actors (which the model is supposed to capture) actually have. Politicians might for example be motivated by holding office, grabbing economic rents, increasing their power over society or by promoting specific ideologies. However, the rational choice framework as such does not come with a predisposed list of what the goals should be. Rational choice theory does not require that the actors are self-interested either. It is possible for the actors to have altruistic or sociopathic preferences, even though selfinterested actors are most common in the literature. A basic feature of rational choice theory is that actions are chosen because of intended consequences. Actions are not valuable in their own right, which can be contrasted for example with Kant’s Categorical Imperative. 1 I will not in this paper go into the debate on the advantages, problems and usefulness of rational choice in the social sciences, even though the literature on the appropriateness of rational choice is large and fascinating. This paper only aims at presenting the core theoretical framework. Let me however add that I have trouble with general and bombastic claims on the usefulness or uselessness of rational choice theory in the social sciences. The appropriateness of any methodology or theory hinges among others upon the nature of the particular research question asked. 1 Requirements on preferences There are however some requirements made on the preferences of rational actors. In order for the theory to work, preferences must satisfy at least two criteria; completeness and transitivity. Completeness requires that the actor must be able to rank all possible outcomes. That is, if we have outcomes x and y, we have either that x>y, x=y (indifference) or x<y. > and < indicate a strict preference (better than or worse than), whereas ≥ and ≤ indicate a weak preference (better than or equal to and worse than or equal to). Transitivity requires that if x>y and y>z x>z. If an apple is preferred to a banana and a banana is preferred to a kiwi, an apple must be preferred to a kiwi. These assumptions (together with reflexitivity; an outcome cannot be strictly preferred over itself) are sufficient to utilize a rational choice framework, but as mentioned, in practical research settings one often puts more structure on the preferences of actors. Utility functions If we in addition to the assumptions above assume that preferences are continuous, Gerard Debreu (1959) has shown that there exists at least one utility function that can be constructed to represent these preferences. We can for example have a utility function that depends on income, or on probability of remaining in office, or over location on a policy-dimension (left-right, where extreme left =0 and extreme right=1.) Figure 1: Hypothetical utility function, U(x), where x is income U(x) x Figure 2: Hypothetical utility function, U(x), for centrist party, where x is value on left-right dimension U(x) 2 x The utility functions, as long as there is no uncertainty involved, can be interpreted as ordinal. An ordinal measurement level implies that we can only draw information about ranking, and not relative size, from numbers. That is, if U(1) =2 (the utility of having 1 unit is 2) and U(2)= 6, we cannot with ordinal utility functions say that the utility of receiving two units (for example Euros) is three times the utility of receiving one. The only thing we can say is that the utility of receiving two is higher than the utility of receiving one Euro. If we had a cardinal utility function however, the numbers would have contained information about relative size in addition to rank. Utility functions can take on different forms and researchers often use simple mathematical functions to represent the preferences of actors. The simplest alternative is the linear utility function where U(x) = a + bx. If a is equal to 0 and b is equal to 1, U(x)=x. Another alternative is using a version of U(x)= a+bxc. We can for example have that a is equal to 0, b are equal to 1, and c equal to 1/ 2 U(x)=x1/2=√x. Another common version is U(x)=ln(x); the natural logarithm of x. However, actors often care about more than one thing, and it is therefore common to model a utility function with more than one objective. Let me illustrate a utility function where an actor has two objectives x and y. Then U = U(x,y), and we can for example choose the simple, linear version U(x,y) = x+y to represent preferences, or the common Cobb-Douglass form U(x,y)= xαy1-α. Let us take an example. We have a political party that competes in two district, and the utility of the party depends on votes gathered in district one (x) and district two (y). The party receives 500 votes in the first district and 700 votes in the second district. The linear utility function would then yield a utility: U(500, 700) = 500 + 700 = 1200 When it comes to the Cobb-Douglass version, let us now make a twist and say that the party values votes in district 1 more than votes in district two, and the α is therefore larger than ½. Let us set α=0,7, which implies that 1- α=0,3. This gives us: U(500, 700) = 5000,7 * 7000,3 ≈553 The choice of a specific utility function is not completely arbitrary however, and the mathematical structure depends to a certain extent on what more assumptions we make about actor preferences other than the very general ones presented above. I will come back to this issue, but first we need to introduce some mathematics. Elementary calculus Most Norwegian students learn how to take the derivative of a function (differentiation) in high school, but unfortunately, not much time is spent on this important topic. MA-Political science students and many researchers have often not encountered differentiation in many years. Since differentiation is so central to rational choice-based theory, and utilized in many political-economic studies, we will have to indulge in a crash-course. The derivative of a function can be thought of as the slope of the function. It is most straightforward to think about the derivative of a linear function, for example Y= 3 + 2X. The slope of the function in this case is always 2, as the function Y increases with 2 units as X increases with 1 unit. When it comes to more complex functions, the derivative of a function is not necessarily constant. Take the 3 utility function in Figure 1. The derivative of the function is always positive, as the function increases over the whole interval. However, the derivative decreases as the x-values become larger, because the slope becomes less and less steep. In formal terms: U’(x)>0 (The derivative of U with respect to x is positive). And: U’(xb) <U’(xa) if xb>xa. (The derivative of U at point Xb is smaller than at point xa if xb is larger than Xa). The derivative of the function is decreasing as x gets bigger. The ‘ denotes that we are referring to the derivative of the U function with respect to x. Alternatively, we can write this expression as dU/dx. Now take the utility function in Figure 2. The derivative of the function, U’(x) is positive before the maximum point and negative after the maximum point, since the function is first sloping upwards and then downwards. Here we come to an important point. What is the value of the derivative at the maximum point? Consider Figure 3 below: Figure 3: The derivative of a function in maximum The straight line on top of the function is the slope at the maximum point. Since the line is flat, the slope is equal to zero. This means that at the maximum point of a function, the derivative of a function is equal to zero (the same is true for minimum points). This is extremely important in rational choice considerations. What will it say to make the best choice for an individual? It can be interpreted as maximizing utility (or optimizing). Therefore, when we search for the optimal action, we take the derivative of a utility function and set it equal to zero. That is, we find the x-value that satisfies the equation U’(x)=0. We now introduce some general derivation rules for different functions: Derivative of powers f(x) = xr f’(x) = rxr-1 Example: f(x) = x3 f’(x) = 3x3-1=3x2 4 Derivative of powers, extended f(x) = axr f’(x) = r*axr-1 Example: f(x) = 2x4 f’(x) = 4*2x4-1=8x3 Remember that x1=x and x0=1, such that f(x) =2x f’(x) =1*2X0=2. Similarly, the derivative of f(x)=25x is equal to 25. The derivative of a constant is equal to zero f(x)=5 f’(x)=0 The derivative of sums and differences f(x) + g(x) differentiated is equal to f’(x) + g’(x) and f(x) - g(x) differentiated is equal to f’(x) - g’(x) This means that the derivative of 2x2+5 is equal to the derivative of 2x2 plus the derivative of 5. One differentiates the sum piecewise. The derivative of the function can be expressed as: (2x2+5)’ or d(2x2+5)/dx, and it is equal to 4x+0=4x, following the rules above. At this point in time, you might want to go back and read this section one more time. Alternatively, you might want to look up a text book in mathematics, such as Sydsæter (2000) for Norwegian students or Gill (2006). I will stop the discussion on derivatives so far, and go back to the interpretational aspects, but here are some of the most used rules for differentiation. Even if you don’t remember or understand these, you can look them up if needed when reading articles or books that use differentiation: f(x) = ex f’(x) = ex f(x) = ln(x) f’(x) = 1/x Product rule: f(x)*g(x) derivated is equal to f’(x)*g(x)+f(x)*g’(x) Quotient rule: f(x)/g(x) derivated is equal to (f’(x)*g(x) - f(x)*g’(x))/g(x)2 Chain rule: f(g(x)) derivated is equal to f’(g(x))*g’(x) Example of chain rule f(g(x))= (2x+3)3. g(x) is here 2x+3. Following the chain rule, differentiation gives us: d(f(g(x))/dx= 3(2x+3)2 *g’(x)= 3(2x+3)2 *2 =6(2x+3)2 How to interpret U’(x): marginal utility Remember that the derivative of a function in a point is equal to the slope of the function in that point. The derivative of the utility function is often called marginal utility, and can be interpreted as the increase in utility when you increase x marginally (sometimes interpreted as one unit): How much extra utility does the actor get from one extra unit of consumption, or one extra vote. Consider U(x) = 8x-x2, which could be the utility function of beers on a Monday night student party U’(x)=8 -2x, and this is the marginal utility function. 5 What is the marginal utility of an extra beer if you already have had one beer? U’(1)=8-2=6. When should the rational actor stop drinking beers? The answer is: When she has maximized her utility function. Remember that this requires that U’(x)=0, that is the marginal utility of beer is equal to zero. (There are some other conditions that need to be satisfied as well called second-order conditions, but never mind these here.) How do we find the optimum? Simply by setting U’(x)=8-2x=0 This implies that the optimal amount of beers is 4. What if we have a function that we would like to minimize, for example a cost function, C(x)? The same logic applies. The bottom of the cost function is found where the derivative, C’(x)=0, that is where the marginal cost is equal to zero. If for example C(x)= x2-4x+16 C’(x)= 2x-4 C’(x)=2x-4=0 x=2 What if we have an action, x, say intensity of guerrilla warfare, which bears with it both gains, U(x), and costs, C(x)? Then the rational guerilla would like to optimize the expression U(x) – C(x). Let U(x)=x2 and C(x)=2x We use the differentiation of a difference presented above and set the total expression equal to zero: U’(x) – C’(x) = 0 d(x2)/dx –d(2x)/dx=0 2x-2=0 x=1 X=1 is the optimal intensity of guerilla warfare in this example. Uncertainty and expected utility In most interesting social science questions, there are uncertain outcomes related to taking a specific action. How does rational choice theory deal with such uncertainty? First, under uncertainty, the analyst will have to move from an ordinal interpretation of the utility function to a cardinal, but I will not go deeper into this issue. If actors are not under thick uncertainty, a lá the Rawlsian veil of ignorance, but are able to make predictions on probabilities of different outcomes, the rational choice framework can easily be applied to situations of uncertainty. Actors are assumed to maximize expected utility (EU(x)), given their beliefs about probabilities of outcome. Moreover, given certain specific assumptions (look up a microeconomic textbook), Von Neumann and Morgenstern showed that maximizing expected utility is equivalent to maximizing the following expression: EU(x) = p1U(x1) + p2U(x2) + ……+pnU(xn) In this case, there are n possible outcomes, which are assigned n different probabilities, given that a certain action is taken. Let us take an example by considering a decision on whether to go to war or not. If the actor does not go to war, he receives utility =2 for certain, that is with a probability =1. However, if the actor goes to war, there are three different outcomes, victory, stalemate and loss. The utility of victory is 10, the utility of stalemate is 0 and the utility of loss is -20. Is it rational for the actor to go to war? This depends on the probability beliefs the actor holds. If the estimated 6 probability of victory is 0,7, of stalemate is 0,1 and of loss is 0,2, we can easily find the expected utility of going to war for the actor: EU = 0,7*10 + 0,1*0 + 0,2*-20 = 7+0-4=3 The expected utility of going to war is 3 and the expected utility (actually a certain utility) of not going to war is 2. The rational actor chooses the action with the highest expected utility, namely war. The main, general assumption underlying rational choice analysis under uncertainty is that actors maximize expected utility, given their beliefs on probabilities of outcomes for specific actions. A qualifying statement is that actors must, when they are allowed to update their beliefs after gathering information from observations (for example in dynamic models with more than one time period), form beliefs after the so-called “Bayes’ rule”. Look up any book on game theory to read more on rational belief formation. Here is one simple example: You know that you are facing either a weak or a strong opponent in a war game, and the weak opponent never goes to war. Initially, you believe that you face a weak opponent with a 0,4 probability and a strong with 0,6 probability. If you observe, in the first round that the opponent goes to war, you should update your probability beliefs. Since weak opponent never go to war, the probability of facing a weak opponent has to be 0! Then you know that the probability of facing a strong opponent is equal to 1. You have successfully used Bayes’ rule to update your beliefs! The expected utility framework allows the incorporation of risk aversion among the actors. The risk aversion can be incorporated into the utility function, and one does therefore not have to assume risk neutrality. Interested readers can look up a textbook in microeconomics. Optimization over time In many rational choice models, there is a time-structure, and thereby the number of time periods is larger than one. How do rational actors respond to such time structures? The common way to model such dynamic settings is by presenting a utility function that is additively separable over periods; that is, total utility can be presented as a sum of utilities gained in each period. However, most actors are impatient and would rather take an immediate prize now than the same prize in the future, and we incorporate this trait by discounting the utility in future periods. We then need a discount factor, β, which tells us how much the actor values future gains (one period ahead) relative to immediate gains. β is in the interval between 0 and 1, and the closer it is to 1, the more patient the actor. Let subscript t be time period, and we start in period 0. We now get that total utility, U, is given by: U(xt) = u(x0) + βu(x1) + β2u(x2) + …..+ βnu(xn) =∑ βtu(xt) The∑ , sigma, means nothing other than the sum of the components. It is used regularly as a summation sign, so you should get acquainted with it. Usually, the first time period is written below the sigma (t=0 here), and the last period in the sum is written on top of the sigma (n here). Notice that as the t becomes very large, βt goes towards zero, and this means that pay-offs far into the future is almost neglected by the rational actors. One special and very important case is when we have an infinite stream of payoffs, that is t ∞. If we have short time periods or long lived actors (states?), an infinite horizon might function well as an approximation to the situation, even though no one or nothing lives forever. We can actually calculate a finite discounted pay-off for the actor, even though we do it over an infinite time horizon. 7 The extra assumption needed for the formula below to be correct is that the size of the x is equal in each period. The result draws on the mathematical theory of geometric series (look up a textbook in mathematics for the proof). We have that: U(xt) = u(x0) + βu(x1) + β2u(x2) + …..+ βnu(xn)+……= u(x)/(1-β) If we now go back to the rational actor and his optimal choice, the rational actor would choose the course of action that yields the highest discounted utility. Take one individual considering whether to invest a 1000 of his 1050 Euros, or consume it all right away. If he invests it, he gets an interest of 5% a year (50 Euros), which he consumes in each period. The actor is relatively impatient with a β of 0,6. The utility function for each period is the simple u(x)=x. Let us calculate the utility from the two courses of action: Consume right away: U(xt) = u(1050) + βu(0) + β2u(0)+….=1050 + 0,6*0 +0,62*0 +….=1050 Save 1000 and live of interest: U(xt) = u(50) + βu(50) + β2u(50)+….= 50 + 0,6*50 +0,62*50 +….=50/(1-0,6) =125 Since 1050>125, the rational actor chooses to consume everything at once. If the actor had been very patient, with a β of for example 0,98, he would have chosen the other course of action. Try the calculation yourself! Extensions on calculus: Concavity and convexity, partial differentiation and optimization under constraints In order to read some of the literature in economics and political economy, we need to make some expansions on the calculus-apparatus that was treated above. First, we need to define two concepts that regularly come up in the literature, namely concavity and convexity. Figure 4 shows two examples of concave functions, one increasing concave and one decreasing concave, and figure 5 shows an increasing convex and a decreasing convex function. Figure 4: Concave functions 8 1 2 Figure 5: Convex functions 4 3 A concave function is a function where the first order derivative decreases in absolute value as the value of x increases. Take for example the increasing and concave function, function 1 in Figure 4. This type of function is particularly interesting for us, since most utility functions are assumed to be increasing and concave. The marginal utility of getting one more unit is always positive, since U’(x)>0. More is better. However, the size of the marginal utility is decreasing as we increase x. More is better, but one unit extra is worth more in utility terms when we have little of the good to begin with. Take money for example. An individual would always value an extra hundred Euros positively. However, the extra hundred Euros is worth more to the person in utility terms when he is poor than when he is rich (buying food not too starve versus buying a second bottle of Chardonnay at a fancy club). Decreasing marginal utility means that the U’(x) function is decreasing. Mathematically, we can express this by taking the second-order derivative: We differentiate U’(x) and get U’’(x). When a function, f(x) is concave, f’’(x)<0. When it comes to concave utility functions, U’’(x)<0. A short way to 9 express an increasing concave utility function is then to say that U’(x)>0 and U’’(x)<0. Another equivalent way to write this is: dU/dx>0 and d2U/dx2<0. Convexity of a function is marked by f’’(x)>0 (or alternatively written d2f/dx2>0. We therefore have the following descriptions of figure 1-4: 1) f’(x)>0 and f’’(x)<0 (increasing and concave) 2) f’(x)<0 and f’’(x)<0 (decreasing and concave) 3) f’(x)>0 and f’’(x)>0 (increasing and convex) 4) f’(x)<0 and f’’(x)>0 (decreasing and convex) * We have now established many of the basic tools we need to read and understand much of the mathematically oriented social science literature, but there are some important expansions left. What if an actor does not care only about one type of outcome, but several? A dictator might for example care about both personal wealth and personal power. A union might for example care about both the unemployment rate and wages. A political party might for example care about both ideology and votes. In these cases, the utility functions must be specified so that they capture these features. We therefore need to consider functions of multiple variables. Take the dictator. His utility depends both on wealth, w, and power, p. We can then write his utility function as U(w,p). If we want to express that the marginal utility of wealth is positive, it is common to write first-order derivative expressions in the following way when we have a function of many variables: ∂U(w,p)/ ∂w>0, or just ∂U/∂w>0. Similarly a positive marginal utility of power can be expressed ∂U(w,p)/ ∂p>0, or ∂U/∂p>0. The expressions ∂U(w,p)/ ∂w and ∂U(w,p)/ ∂p are called partial derivatives (or partial first-order derivatives) of the utility function. There are also other ways of writing partial derivatives. For example U’w and U’p, or just Uw and Up. This latter way to express partial derivatives is for example used in Przeworski (2000). Let us take one example of partial differentiation you are actually already familiar with! The example is related to the multiple regression equation. What we are interested in, is often to interpret the different regression coefficients, the βs. We say for example that β1 is the average increase in Y when X increases with one unit, holding all other variables constant. The operation of finding the increase in Y when X increases, while holding other variables constant, is actually similar to finding the partial derivative of the Y-function with respect to X. If we had a bivariate regression function Y(X) given by Y = α +βX,, then dy/dx = β. The regression coefficient in a bivariate regression is equal to the derivative of Y with respect to X. The logic is completely similar in a multiple regression. The only difference is that we now want to find partial derivatives. It is important to know that when you find the partial derivative of a function with respect to a variable, you treat all the other variables as constant when performing the calculation. Let us define such a regression equation with three variables: Y = α +β1X1+β2X2+β3X3+ε 10 If we want to find ∂Y/∂X1, we use the rules of differentiation and remember to hold X2 and X3 constant: We get ∂Y/∂X1= 0 + β1+0+0+0 = β1. The regression coefficients in the multiple regression are nothing other than the partial derivatives of Y with respect to the different Xs. We look for the increase in Y when we increase X marginally (or with one unit since we have a linear function), holding all the other variables constant. This is how we interpret regression coefficients, and this is exactly what we do when we take partial derivatives. * The dictator in our earlier example had a positive marginal utility of wealth and a positive marginal utility of power. What if we assume that the marginal utility of power and wealth are both decreasing (but always positive)? Then we write: ∂2U(w,p)/ ∂w2<0 and ∂2U(w,p)/ ∂p2<0. This is completely similar to the discussion on concavity and convexity of functions of one variable. The only difference is that we now have to explicitly state which variable we are treating when investigating the utility function. It would have made no difference if we expanded the number of variables beyond two; the practice is the same with regards to specifying partial first-order and second orderderivatives. We do however have one extra expansion, and that is the so-called cross-derivative. What if the marginal utility of wealth depends on the level of power a dictator has? We then need to investigate the expression ∂2U(w,p)/ ∂w∂p, the cross-derivative. If ∂2U(w,p)/∂w∂p>0, the marginal utility of wealth increases in the level of power (a powerful dictator values an extra dollar more than a less powerful dictator). If ∂2U(w,p)/∂w∂p<0, the marginal utility of wealth decreases in the level of power a dictator has. Don’t bother thinking too much about this, if you don’t comprehend it (we first differentiate U with respect to w, and then we differentiate this new expression with respect to p). However, it is nice to know the so-called Young’s theorem, which states that ∂2U(w,p)/ ∂w∂p = ∂2U(w,p)/ ∂p∂w, if we have continuous functions. This for example implies that if marginal utility of wealth increases in level of power, then we automatically know that the marginal utility of power increases in the level of wealth. * An expression that often appears in rational choice literature is the marginal rate of substitution (MRS). This concept is expressed, if we use the dictatorial utility function above, as: (∂U/∂w)/(∂U/∂p) Alternatively: Uw/Up The marginal rate of substitution is therefore the ratio between the first-order partial derivatives of the utility function. Why is this expression so important? It can be shown that the marginal rate of substitution is equal to the amount of power a dictator is willing to give up if he achieves one unit of wealth and utility levels are held constant. Expressed in another way: How many units of power is the dictator willing to give up in order to achieve one extra unit of wealth. If Uw/Up =3, then the dictator will have the same utility level if he trades three units of power for one unit of wealth. We can describe such trade-offs between the two goods graphically by so-called indifference curves. Remember that the utility level is held constant when “trading” utility for power, when discussing the MRS. Figure 6 shows such trade-offs with hypothetical indifference curves, and the utility level is 11 constant along each indifference curve. An indifference curve to the northeast in the diagram implies a higher level of utility for the dictator than one to the southwest, since the dictator on this indifference curve has more power and wealth. Do not confuse movement along an indifference curve (trade-offs with same utility-level) with comparison of different indifference curves (different utility-levels). Figure 6: Indifference curves for the rational dictator w U=50 U=30 U=10 p The convex shape of the indifference curve illustrates a common assumption that is often made about an actor’s preferences, when it comes to trade-offs. We see that when the amount of power is relatively low, the indifference curves are steep. This means that an actor needs many units of wealth to compensate for one unit of power when the amount of power is already low. If the amount of power is high however, the dictator needs relatively less wealth to compensate for a loss of one unit of power. This changing trade-off, implies that the MRS is not constant, but changing according to how much of the two goods an actor has. The general rule is that an actor needs more of good 1 to compensate for a loss of good 2, if the actor already has relatively little of good 2, and vice versa. * We need one last bit of mathematical knowledge important for rational choice analysts, and this bit is related to optimization under constraints. Most political and economic decisions are not made in a vacuum. Actors who try to make optimal choices generally face specific constraints, and they need to take these constraints into account when they choose courses of action. Constraints can for example be budget constraints, technological constraints, institutional constraints, or, as we will see below in the section on game theory, constraints that are caused by other actors’ (responding) actions. Take for example the consumer. She wants to maximize her utility from consumption, and generally more consumption is better, that is U’(c)>0. This would alone have implied a rational choice of 12 infinite consumption. As we all know however, consumers face a particular constraint in the market, namely the budget constraint. The consumer wants to optimize consumption, given that she keeps her budget. Similarly a technological constraint binds when firms want to minimize costs. The firm cannot go any lower in costs, given a certain target level of production, than the best available technology allows it to. If we therefore go back to the optimization calculus related to finding the maximum of a utilityfunction, we need to incorporate this feature. Let us take the consumption example, since it is particularly illustrating. Assume that our consumer wants to maximize a utility function that depends on the consumption of two goods x1 and x2: U(x1, x2), with Ux1 and Ux2, the first order derivatives, both positive. However, the consumer has an exogenously given income, m, and faces exogenously given prices, p1 and p2, in the market. The problem then boils down to: Max U( x1, x2) Subject to the constraint: p1 x1 + p2 x2 = m The optimal choice of x1 and x2 can be found graphically in a indifference curve diagram, by rewriting the budget constraint so that x2 = m/p2 - p1/p2*x1. This rewritten constraint represents a straight line, with the slope -p1/p2. Remember now that the utility in such an indifference curve diagram increases to the northeast, and the consumer wants to land on the highest possible utility level. The implication is that the consumer will choose the x1 and x2 combination that coincides with the point where the indifference curve just exactly touches the constraint. In this point the slope of the indifference curve (which is given by the MRS) is equal to the slope of the constraint. That is where U’1/U’2 = p1/p2. See figure 7 below: Figure 7: Utility maximization under a budget constraint x2 x1 13 A way to find the optimum mathematically is by putting in the rewritten budget constraint, where x2 is expressed as a function of x1 into the utility function so that we get: U( x1, x2) = U( x1, m/p2 - p1/p2*x1) We now have a function of one variable, and we can differentiate and set the expression equal to zero. We will then find the optimal allocation. The final answer will be equal to the solution expressed above, namely that the MRS is equal to the price ratio. The interpretation is that in optimum, the subjective utility tradeoff the consumer has when trading units of x1 and x2 is equal to the relative price (p1/p2) of the units in the market. If this was not the case, and the consumer valued x1 relatively more than the market, she would have just buy more of those units, and less of x2. See Przeworski’s (2000) chapter 2 or a microeconomics book, if you want to understand more of the logic. For now, it is sufficient to remember that rational actors face constraints, and in optimum they tend to adjust so that the subjective MRS, their relative evaluation of two goods at the margin, is equal to the objective trade-offs they face between the two goods (those given by budget-market constraints, technological constraints etc). For the interested reader, let us take one example of a particular utility function to show the procedure more concretely. Assume that the utility function in the example above is given by: U( x1, x2) = x1*x2 We can further set p1 =1 and p2=2, and let m = 10. We remember that the constraint p1 x1 + p2 x2 = m can be rewritten as x2 = m/p2 - p1/p2*x1 This gives us: x2 = 10/2 – ½* x1 = 5 - ½* x1 The procedure is now first to substitute for x2 into the utility function. We get: U( x1, x2) =x1*(5 - ½* x1) = 5x1 -1/2 x12 We remember that utility maximization was equivalent to differentiating the utility function, and setting the resulting expression (marginal utility function) equal to zero. First we differentiate: dU/dx1 = d(5x1 -1/2 x12)/d x1 = 5- x1 Then we set the expression equal to zero: 5- x1 = 0 x1=5 The optimal amount of x1 is 5. What about x2? Well, we know that x2 = 5 - ½* x1 from the rewritten budget constraint, so that the optimal amount of x2 is equal to 2,5. Let me sum up the procedure of calculating optimal choice under constraints very quickly. 1) Rewrite constraint 2) Put constraint into the utility function 3) Differentiate utility function with respect to remaining variable 14 4) Set the new expression equal to zero 5) Calculate optimal choice on the variable 6) Calculate the optimal choice on other variable by inserting into the constraint There is another way of calculating optimization under constraints, which is performed by constructing a so-called Lagrange-function. When following this procedure, one arrives at the same result as if one rewrites the constraint and put it into the utility function. You don’t have to worry about learning this technique. Look up a textbook in mathematics (Sydsæter, 2000; Gill, 2006) if you are interested. Przeworski (2003) uses this technique, but don’t bother with the calculations. Just make sure that you remember that it is optimization under constraints that is going on, and be sure to understand the interpretation of the final result! Game theory We will just go through some of the concepts of game theory very quickly here, and interested students should look up a game theory book to get a more complete picture. Very good introductions are Watson (2002) in English and Hovi (2008) in Norwegian, and a more complex presentation related to game theory in political science with several superb examples is McCarty and Meirowitz (2006). Game theory is a theoretical framework on strategic interaction between (rational) actors, also known as players. Optimization in a game theoretical framework requires that the players think through how other players’ actions will affect the pay-off from their own action. That is, the pay-offs you get depend not only on your own actions, but also on what others do. A phrase that characterizes such a setting is strategic interaction. Not only are players assumed to be rational, but the players assume that everybody else is rational as well. You can now look up a qualitative description of a prisoner’s dilemma or a chicken game if you don’t remember these games. Descriptions of these games are everywhere, even on the internet, so I don’t bother giving a description here. This will give you a feeling of what’s going on in a game setting. Remember however that these games are only examples of games (some students tend to think that game theory is equivalent with the prisoner’s dilemma game). Game theory is very general, and is widely used in political economic research. We have descriptions of games between politicians and electorates, games between countries, games between politicians and businesses, games between dictators and democratic movements etc. What is common is (as mentioned) that all players are rational, and that they try to make projections of how other players will act, and bring these beliefs into their calculations. There are different ways of describing a game. In a static game, where we interpret the game as if the players were moving simultaneously (or at least that they cannot observe the other player’s action before moving), we usually model the game by using the normal form representation. Figure 8 shows a normal form game: Figure 8: Game presented in the normal form 15 Player 1\Player 2 Left Right Up (2,0) (3,1) Down (5,2) (2,4) The normal form presents the players of the game, their potential strategies (Up and Down for 1 and Left and Right for 2), and the payoffs (utilities) for the actors given the different outcomes. The payoffs for player 1 is the number to the left of the comma. We see that the pay-offs are not only dependent on the player’s own strategy, but also on the other player’s choice of strategy. A Strategy is a complete plan of action for a player in a game: A strategy expresses a plan of action for every contingency. This is not important in the static game, since there are no contingencies (one cannot observe what the other player does), but in a sequential game, which we will come to later, the importance of having a strategy that specifies what to do given different situations becomes important. A strategy is dominant if the strategy yields higher pay-offs than any other strategy, independent of which action the other player takes (we here assume only two player’s because of simplicity, but there can of course be more than two players in a game). A rational actor always chooses a dominant strategy. In games with more than two strategies, we might have strategies that are dominated by other strategies, even if we do not have any universal dominant strategy. A rational player never chooses dominated strategies, so these strategies can be skipped from further consideration in an analysis. In the prisoner’s dilemma (PD), not telling on the other prisoner is a dominated strategy, and telling is a dominant strategy. Telling always gives a higher pay-off independent of what the other player does, and both prisoners therefore choose to tell. Note that the outcome related to the strategy combination {tell,tell} in the PD, is worse for both prisoner’s than the outcome {not tell, not tell}. It is not given that rational players interacting in a game are able to arrive at the best possible solution. {tell,tell} is not Pareto optimal, since it is possible to make at least one actor better of by moving to another situation, namely {not tell, not tell}. However, {tell, tell} is the equilibrium of the game, since no player has an individual incentive to deviate. Make sure that you are able to separate these concepts. Here are the two definitions of Pareto improvement and Pareto optimality: Given a set of alternative allocations for a set of actors, a movement from one allocation to another that can make at least one individual better off without making any other individual worse off is called a Pareto improvement. An allocation is Pareto optimal when no further Pareto improvements can be made, that is when no one can be made better off without anyone else being made worse off. These two concepts are very important in economics, and Pareto Optimality (PO) is often used as a normative criterion to judge whether a situation is “good”, or more precisely efficient. Notice that 16 the concept does not say much about distribution. PO does however indicate that we do not waste resources, since there are no alternatives that can make at least one actor better off without making anyone else worse off. The PD equilibrium is inefficient in this sense (it is not PO). The concept of equilibrium is not a normative criterion, but a description of a plausible solution to a game (or the outcome in a market, or in a political process). The equilibrium concept entails the notion that there are no individual forces that can push or pull the outcome away from the situation we are in. Equilibria are assumed to have a certain degree of stability to them because of this property. Situations that are not equilibria are assumed to be more unstable, and will ultimately move towards an equilibrium (there can be more than one equilibrium). There are several different specifications of the equilibrium concept, and the most commonly used in simple game theory is the Nash equilibrium. The definition of a Nash equilibrium is that all players have chosen their best strategy, given the other players’ strategies. That is, given the other players’ choices, no one can make themselves individually better off by deviating from their chosen strategy in the Nash equilibrium. No one regrets their choices in a Nash equilibrium, once they observe what the others have done. Another way to express the Nash equilibrium is by saying that all players play their best responses to the other players’ chosen strategies. In the game in Figure 8, {Up, Right} is the only Nash equilibrium. The reason is that Player 1 cannot gain by switching from Up to Down given that 2 plays right (She would go from 3 to 2), and Player 2 cannot gain by going from Right to Left, given that 1 is playing Up (she would go from 1 to 0). Notice that {Down, Left} would have been a Pareto-improvement from {Up, Right}. Both players would have preferred this outcome, but it is not a feasible equilibrium. Why? Well, Player 2 would now have had an incentive to go Right, and the situation is therefore not an equilibrium. When game theorists look for descriptive solutions to a game, they look for equilibria. This does not mean that the equilibria are normatively good, or preferred by the actors themselves. However, in many games, equilibria can be Pareto optimal, but not in the game in Figure 8. * The last concept we are going to deal with is the concept of a sequential game. In a sequential game, there is a sequence to the actions of the players, in the sense that one player can observe the other player’s moves before making his own move. In these games, the Nash equilibrium is not a sufficiently strong equilibrium concept, since it yields some implausible solutions. We need another, stronger equilibrium concept, which is known as Subgame Perfect Nash Equilibrium (SPNE). I will not go into the definition of the concept here, since it requires the understanding of what a subgame is, but just know that this is the solution concept used in sequential games. The reason why Nash equilibrium is too weak as a solution concept in sequential games is that it cannot deal with non-credible threats and non-credible promises. SPNE can. Fortunately, there is a simple way of finding solutions to sequential games, and it is known as backwards induction. The best way to understand this procedure is by looking at a concrete example. Sequential games are modeled by a game tree, which has a very intuitive structure, as seen in Figure 9. The circles are called nodes and the lines are called branches. A specified player acts at a node, and the different alternative actions are illustrated by the branches. 17 Figure 9: A game tree Iran Peace Bomb Israel (1,1) Bomb back Passive (2, -1) (0,0) The game tree illustrates a situation where Iran has a nuclear bomb, and must decide whether to bomb Israel or to keep the peace. If the Iranians decide to bomb, Israel must decide between remaining passive or bombing back. What will Iran do? The best situation for Iran is to nuke Israel if Israel thereafter remains passive. This yields a utility (or pay-off) of 2 for Iran. However, this is not a plausible solution to the sequential game. We can show this by backward induction. We start at the last node, and ask what Israel will do if they are bombed. Remaining passive yields a utility of -1 and bombing back yields a utility of 0. Israel is rational and will opt for bombing back, since this yields the highest utility. Iran knows that Israel is a rational player, and it can therefore eliminate the option of bombing Israel and Israel remaining passive. The two only real options are choosing peace or choosing to engage in a mutual nuclear war. Peace yields a utility of 1 and reciprocal bombing yields a utility of 0. Iran, knowing that Israel will bomb back if they bomb, therefore chooses the peaceoption. The outcome of the game is peace. We have now solved the simple sequential game by backwards induction. We start at the last node in the game, and eliminate non-plausible branches. We then move further up the game tree and eliminate more non-plausible actions, given that the player knows that the next player to move in the game will choose a certain action should a particular node be reached. The players evaluate realistic alternatives, and do not engage in wishful thinking. They are rational, and they know that other players are rational as well. We also have games where the actors are uncertain about the other player’s type; that is, his preference structure. In these games, the solution concept is called Bayesian Nash Equilibrium for static games and Perfect Bayesian Equilibrium for sequential games, but I will not treat these games and solution concepts. 18 Some encouraging words If you have been able to capture a third of the knowledge in this text after the first read-through, you should be happy about yourself! Economists use at least a year, or maybe more, to fully grasp these issues, so you should not expect to understand them completely. However, I advise you to take a break, and go through the text again. Later, you can have this text on the side of your desk when you read substantial articles using rational choice approaches. It is not important to know how to do the calculations yourself at first, but it is important to understand the general concepts. You do not have to perform calculations yourself to check whether scholarly articles are correct mathematically (they probably are). However, you should have some vague sense of what is going, and most importantly: You should be able to understand the authors’ qualitative description of their results. 19 Literature Debreu, Gerard (1959) The Theory of Value. New Haven: Yale University Press Elster, Jon (1989) Nuts and Bolts for the Social Sciences, Cambridge: Cambridge University Press. Gill, Jeff (2006) Essential Mathematics for Political and Social Research Hovi, Jon (2008) Spillteori: En innføring. Oslo: Universitetsforlaget. McCarty, Nolan and Adam Meirowitz (2007) Political Game Theory: An Introduction. Cambridge: Cambridge University Press. Przeworski, Adam (2003) States and Markets – A Primer in Political Economy. Cambridge: Cambridge University Press. Sydsæter, Knut (2000) Matematisk analyse. Bind 1. Oslo: Gyldendal Akademisk. Watson, Joel (2002) Strategy: An Introduction to Game Theory. New York: W.W. Norton. 20