6.825 Exercise Solutions, Decision Theory 1 Decision Theory I Dr. No has a patient who is very sick. Without further treatment, this patient will die in about 3 months. The only treatment alternative is a risky operation. The patient is expected to live about 1 year if he survives the operation; however, the probability that the patient will not survive the operation is 0.3. 1. Draw a decision tree for this simple decision problem. Show all the probabilities and outcome values. Solution: live (0.7) op U(12) die (0.3) U(0) no op U(3) 2. Let U (x) denote the patient’s utility function, where x is the number of months to live. Assuming that U (12) = 1.0 and U (0) = 0, how low can the patient’s utility for living 3 months be and still have the operation be preferred? For the rest of the problem, assume that U (3) = 0.8. Solution: The operation would be preferred as long as U (3) < 0.7. live (0.7) 0.7 op max(U(3),0.7) 1 die (0.3) 0 no op U(3) 3. Dr. No finds out that there is a less risky test procedure that will provide uncertain information that predicts whether or not the patient will survive the operation. When this test is positive, the probability that the patient will survive the operation is increased. The test has the following characteristics: • True-positive rate: The probability that the results of this test will be positive if the patient will survive the operation is 0.90. • False-positive rate: The probability that the results of this test will be positive if the patient will not survive the operation is 0.10. What is the patient’s probability of surviving the operation if the test is positive? Solution: 1 Pr(survive|pos) = 0.9545. Pr(pos|survive) Pr(survive) Pr(pos) = Pr(pos|survive) Pr(survive) Pr(pos|survive) Pr(survive)+Pr(pos|¬survive) Pr(¬survive) = 0.9·0.7 0.66 = 4. Assuming the patient has the test done, at no cost, and the result is positive, should Dr. No perform the operation? Solution: Yes. EU (op) > 0.8. 5. It turns out that the test may have some fatal complications, i.e., the patient may die during the test. Draw a decision tree showing all the options and consequences of Dr. No’s problem. Solution: For this problem, we need to calculate Pr(survive|neg) as above: Pr(survive|neg) = 0.2059. Then we have the following decision tree: live (0.95) U(12) die (0.05) op no op pos (0.66) live take test U(0) U(3) live (0.21) neg (0.34) op no op die die (0.79) U(0) don’t take test U(12) U(0) U(3) EU(no test) 6. Suppose that the probability of death during the test is 0.005 for the patient. Should Dr. No advise the patient to have the test prior to deciding on the operation? Solution: Yes, the test should be taken. The evaluated decision tree is as follows: live (0.95) 0.95 op 1 die (0.05) 0 pos (0.66) live (0.995) take test 0.896 0.896 don’t take test 0.90 0.95 no op 0.8 neg (0.34) 0.8 die (0.005) op live (0.21) 0.21 no op 0 1 die (0.79) 0 0.8 0.8 2 2 Decision Theory II You go to the racetrack. You can: • Decline to place any bets at all. • Bet on Belle. It costs $1 to place a bet; you will be paid $2 if she wins (for a net profit of $1). • Bet on Jeb. It costs $1 to place a bet; you will be paid $11 if he wins (for a net profit of $10). You believe that Belle has probability 0.7 of winning and that Jeb has probability 0.1 of winning. 1. Your goal is to maximize the expected value of your actions. What, if any, bet should you place, and what is your expected value? Draw the decision tree that supports your conclusion. Assume that you are risk-neutral. Solution: You should bet on Belle. The expected utility is U ($0.40). 2. Someone comes and offers you gambler’s anti-insurance. If you agree to it, • they pay you $2 up front • you agree to pay them 50% of any winnings (that is, $.50 if Belle wins, and $5 if Jeb wins). How would it affect the expected value of each of your courses of action? What would be the best action to take now, again assuming risk-neutrality? Draw the new decision tree. Solution: You should take the $2 insurance and bet on Belle. The expected utility is U ($2.05) (see figure 1). win (0.7) $2.05 lose (0.3) $2.50 1.00 Belle win (0.1) Take insurance $2.05 $2.05 Jeb $1.60 $7 lose (0.9) $1 Don’t take insurance win (0.7) $0.40 Belle $0.40 $1 lose (0.3) Jeb $-1 $0.1 win (0.1) lose (0.9) $10 $-1 Figure 1: Decision tree for the betting problem 3 Decision Theory III You’re an olympic skier. In practice today, you fell down and hurt your ankle. Based on the results of an x-ray, the doctor thinks that it’s broken with probability 0.2. So, the question is, should you ski in the race tomorrow? 3 If you ski, you think you’ll win with probability 0.1. If your leg is broken and you ski on it, then you’ll damage it further. So, your utilities are as follows: if you win the race and your leg isn’t broken, +100; if you win and your leg is broken, +50; if you lose and your leg isn’t broken 0; if you lose and your leg is broken -50. If you don’t ski, then if your leg is broken your utility is -10, and if it isn’t, it’s 0. 1. Draw the decision tree for this problem. 2. Evaluate the tree, indicating the best action choice and its expected utility. Solution: U(ski) = 0 and U(not ski) = -2, so we ski (see figure 2). broken (0.2) win (0.1) ski 0 lose (0.9) 90 fine (0.8) -10 broken (0.2) fine (0.8) 50 100 -50 don’t ski -2 broken (0.2) fine (0.8) -10 0 0 Figure 2: Decision tree for the skier’s choice. You might be able to gather some more information about the state of your leg by having more tests. You might be able to gather more information about whether you’ll win the race by talking to your coach or the TV sports commentators. 3. Compute the expected value of perfect information about the state of your leg. Solution: Given perfect information about the leg, we have the tree in figure 3, so the expected value of the information is E(Uinf o ) − E(Unoinf o ) = 6 − 0 = 6. 4. Compute the expected value of perfect information about whether you’ll win the race. Solution: Given perfect information about winning, we have the tree in figure 4, so the expected value of the information is E(Uinf o ) − E(Unoinf o ) = 7.2 − 0 = 7.2. In the original statement of the problem, the probability that your leg is broken and the probability that you’ll win the race are independent. That’s a pretty unreasonable assumption. 6. Is it possible to use a decision tree in the case that the probability that you’ll win the race depends on whether your leg is broken? 4 win (0.1) -40 ski 50 lose (0.9) -50 broken (0.2) 6 -10 don’t ski -10 fine (0.8) 10 win (0.1) ski don’t ski 10 100 lose (0.9) 0 0 Figure 3: Decision tree given perfect information about the leg. broken (0.2) 90 fine (0.8) 50 100 ski broken (0.2) win (0.1) 7.2 90 don’t ski -2 -10 fine (0.8) 0 lose (0.9) -2 ski broken (0.2) -10 -50 fine (0.8) don’t ski 0 -2 broken (0.2) fine (0.8) -10 0 Figure 4: Decision tree given perfect information about winning. Solution: Yes. Just put the win branch after the broken branch and use the conditional probabilities for the given state of the leg. 5