DECISION THEORY We must often make decisions about what to do even when we’re uncertain of what will happen. Suppose a friend offers you the following choice: 1. Toss a coin: heads you win $2; tails you lose $2. OR: 2. Toss a coin twice in a row: two heads, you win $10; otherwise you lose $1. Which one should you choose? 1 Game 1: 1. 50% chance of winning $2; 2. 50% chance of losing $2. 3. Each consequence “cancels the other out”. 4. This game offers you no advantage. Game 2: 1. 25% percent chance of winning: Pr(H1 & H2) = Pr(H1) x Pr(H2) = ¼. 2. 75% chance of losing. 3. But, if you win, you get $10, while if you lose you only lose $1. 4. This game seems more appealing. Let’s get more precise than this. 2 When considering an act, we must take into account both the probability and the utility (i.e. the benefit or value) of each possible consequence. We do this by multiplying the probability of each consequence by its utility and then adding them all up. Game 1: Pr(Win) = 0.5 Pr(Lose) = 0.5 Utility(Win) = +2 Utility(Lose) = -2 Value = 0.5 x 2 + 0.5 x -2 = 0 Game 2: Pr(W) = 0.25 Pr(L) = 0.75 Utility(W) = +10 Utility(L) = -1 Value = 0.25 x 10 + 0.75 x -1 = 1.75 As we suspected, game 2 has a higher “expected value”. On average you will win $1.75 per game. 3 Notation Acts: bold capital letters: A, B, C … Consequences: capital letters: A, B, C … Utility (of a consequence): U(C) Probability of C given A: Pr(C/A) Expected value of an Act: Exp(A) The Expected value of an act is the sum of the products (utilities x probabilities) In other words: Exp(A) = [Pr(Ci)U(Ci)] Where C1, C2, … Cn are the consequences of A and i = 1, 2, … n. Notice that we leave out the conditional probability because it’s clear we’re talking about act A. Technically, it should be there. 4 Example: Suppose your friend charges you $2 to enter game 2. What is the expected value of the game now? A: play game 2 for $2. C1: win U(C1): 8 C2: lose U(C2): -3 Exp(A) = (0.25 x 8) + (0.75 x -3) = -0.25 Now the game is advantageous to your friend, i.e. in the long you will lose 25 cents per game you play. 5 Fair Price A fair price for a risk is the price at which neither buyer nor seller has an advantage, i.e. the price at which: Exp(A) = 0 What is the fair price for game 2? B: play game 2 without entry fee Exp(B) = 1.75. To set Exp(B) = 0 we must set the price at $1.75. In that case: Exp(B) = (0.25 x 8.25)+(0.75 x –2.75) =0 The fair price for game 2 is $1.75. 6 Should you refuse to play for $2? It depends on your values: 1. Perhaps you enjoy the game more than the money you win. 2. Perhaps you enjoy spending time with your friend. 3. Perhaps the thrill of gambling is rewarding to you. 4. Etc… If you value the time with your friend at, say, $5, then even at a charge of $2, playing the game once has an expected value of $4.75. It is not necessarily irrational to act even when the expected value of the act is negative. 7 St. Petersburg Paradox “Paradox”: An argument that apparently derives self-contradictory or absurd conclusions by reasoning from apparently acceptable premises. The game: Toss a coin. - If it falls heads, you win $2. Game stops. - If not, you toss again. - If it falls heads, win $4, game stops. - If not, toss again. - If heads, win $8, game stops. - And so on (until heads is thrown). (Notice you can’t lose money) S: play the game N: game stops on the nth toss. U(N): 2n Pr(N): (½)n What is a fair price to enter this game? 8 We need to calculate Exp(S): Exp(S) = Pr(N1)U(N1) + Pr(N2)U(N2) + … = ½ x 2 + ¼ x 4 + 1/8 x 8 + 1/16 x 16 + … =1+1+1+1+1+… = Infinite! The fair price for this game is that for which Exp(S) = 0, i.e. infinity. You should be willing to pay all the money you, your friends and your family have or will have just to play this game. Are you willing to pay this? What’s gone wrong here? 9 Possible answers: 1. The game is fair at any price. Pay whatever is asked! 2. In reality the game would end in a finite time so you don’t really need to pay an infinite amount. Pay a small amount. 3. Exp(S) is not really defined because there is no upper bound. No real problem has been raised. 4. You shouldn’t be willing to pay an infinite amount because after a finite number of tosses, the value of further winning goes down (diminished marginal utility). Pay a lot, but not too much. 5. The utility of a large win is great, but the chances are extremely small. Don’t pay very much to play. 10 Choosing among possible acts Suppose I go to the store to buy some chips. Why do I do this? 1. I believe that the store sells chips. 2. I want some chips. Decisions depend on: What we believe. What we want. So: We can represent beliefs by probabilities. We can represent desires by utilities. In this way we can calculate the expected value of an act easily enough. How do we use the result? 11 Expected value rule: Act so as to maximize expected value. Lottery A: Lottery B: 100 tickets sold 1,000,000 sold Prize: $10,000 Prize: $100,000 Tickets: $150.00 Tickets: $1 A: buy a ticket for Lottery A B: buy a ticket for Lottery B Exp(A) = 0.01 x 9850 + 0.99 x -150 = -$50 Exp(B) = 0.000001 x 99,999 + 0.999999 x -1 = -0.9 The expected value rule says to buy a ticket for Lottery B over one for Lottery A. 12 Evaluating the rule Consider a game where you toss a coin and win $1000 if it comes up heads and lose $1000 if it comes up tails. The expected value of this game is 0. The expected value of declining to play is 0. The expected value rule says that each choice is of equal value and you should not prefer one to the other. Is this right? 13 What if you’re risk averse, i.e. more afraid of risking money than gaining money? Then you’d choose to decline the game. If you aren’t very wealthy, then losing $1000 is far worse than gaining $1000. So, the risk of losing might outweigh the benefit of winning even though each is equally probable. Some say it is irrational to violate the expected value rule. Is this just arrogance? If not, what’s wrong with the rule? 14 Utiles There are other values than money, for example, the pleasure of playing a game and the desire to remain financially solvent. We shouldn’t only measure utility in dollars. We shall use the term utiles to indicate a general measure of value. For example, you might value $1000 at +1000 utiles, but the risk of gambling might be -100 utiles. In that case, the expected value of the coin toss game is negative (i.e. -200). Perhaps, then, we should follow the expected value rule but broaden our notion of value to include risk aversion. 15 Problem: Utility (valuing something) and risk aversion are difficult if not impossible to compare. It seems artificial to try to combine them into one measure called “utility”. How could you do it? You could give numbers to each, but they appear to be on different dimensions. Do you think we can save the expected value rule in this way? Insurance companies assume that we dislike risk enough to pay more money that we will (on average) get back. 16 We must be careful to consider all possible consequences of an act. If we don’t, and we make the wrong decision, it does not follow that that expected value rule is wrong. We might also disagree about the probability of an event or its utility. Just because we do, it does not mean that the expected value rule is wrong. Accident No Accident Explorers Probability 1/2500 2499/2500 Consequence Not so bad Tremendous Accident No Accident Safety-first Probability 1/430 429/430 Consequence Many deaths Not so great Even a good rule can lead to disagreement. What do you think of the expected value rule? 17 Pascal’s Wager Q: Should you force yourself to believe in God? Pascal: Yes. Here’s why: Either God exists or doesn’t exist (G v ~G) 1. If you believe and God exists, you are rewarded with infinite bliss. 2. If you believe and God doesn’t exist, you don’t lose very much (a few Sunday mornings). 3. If you refuse to believe and God exists, you are punished with infinite torture. 4. If you refuse to believe and God doesn’t exist, you don’t gain very much (a few Sunday mornings). Therefore, you gain more and lose less by believing, whether or not God exists. So, don’t try to convince yourself, just believe! 18 DECISION PROBLEMS: A partition of possible states of affairs. The possible acts one can undertake. The utilities of the consequences of each act in each state of affair. A partition is a set of exhaustive (one must occur) and mutually exclusive (only one may occur) possibilities. Pascal’s Partition: ~B B G - + ~G 0 0 19 If in at least one state of affairs, one act has more utility than every other act, and if in no state of affairs does it have less utility, then this act dominates the others. Clearly, B dominates ~B. Dominance Rule: If one act dominates the others, do it. 20 Warning: The Dominance Rule only works if your acts don’t causally influence the states of affairs on the partition. Why? Well, imagine if refusing to believe in God caused God to respect you and so reward you a little bit. Assume also that believing in God caused God to think you were spineless and punish you a bit. Then: ~B B G + - ~G 0 0 Now ~B dominates. Clearly Pascal assumes that your choice whether to believe doesn’t causally influence the possible states of affairs in the partition. 21 Improved Dominance Rule: If no act in the set of acts has any causal influence on the states of affairs in the partition, then if one act dominates, do it. 22 But what if you aren’t indifferent about going to church, praying, etc.? That is, what if you value the fun you get from free Sunday mornings and disvalue the boredom of attending church? In that case: ~B B G - + ~G + - B doesn’t dominate: it has a better payoff in one state of affairs but a lower payoff in the other. 23 Pascal’s response: use probability assignments and the expected value rule: No matter how small you think it is, there is some finite probability that God exists: Pr(G)>0 Exp(B) = Pr(G)U(B,G) + Pr(~G)U(B,~G) But U(B,G) is infinite: Exp(B) = + Exp(~B) = Pr(G)U(~B,G) + Pr(~G)U(~B,~G) No matter how high, U(~B,~G) is still finite and U(~B,G) infinitely bad: Exp(~B) = - 24 So according to Pascal a decision problem consists of: A partition of possible states of affairs. The possible acts one can undertake. The utilities of the consequences of each act in each state of affair. And: A class of admissible probability assignments to the states of affairs in the partition. 25 Pascal claims: If in every admissible probability distribution, one act has greater expected value than every other act, then this act dominates the others in expected value: Dominant Expected Value Rule If one act dominates all the others in expected value, DO IT! What do you think? Is Pascal’s argument a good one? How would you criticize it? Is belief in God a “live possibility” for you? 26 Homework Do the exercises at the end of chapters 8, 9 and 10. Go over the examples in those chapters to get clear on the ideas from today’s lecture. 27