Lecture 10: Decision Trees and Expected Utility 1 Introduction 2

advertisement
Lecture 10: Decision Trees and Expected
Utility
A cynic is a man who knows the price of everything and the value of nothing
– Oscar Wilde (Lady Windermere’s Farm)
1
Introduction
In the previous lecture, we discussed the basics of probability theory. In this
lecture, we want to apply probability theory to analyze decision making. We
will focus on two concepts: the value of information and the value of control
over a decision. Both of these are covered in great detail by Lave and March.
So, if you do not understand these examples, do some of the problems in
Lave and March.
After the decision tree analysis, we will consider the subject of risk. In
doing so we will differentiate between trees with monetary payoffs and trees
with utility payoffs.
Our discussion of decision trees seems a bit removed from our core motivations of predict, explain, design, and think. Though trees prove helpful for
all four, they might be better thought of as a tool for helping people make
better decisions. The key insight is this: if we assume that people make good
decisions (using trees or not), then trees will be helpful for explaining why
people did what they did.
2
Decision Trees And Expected Value
We begin with some simple examples to show how we can use trees to get at
expected value. In the first example, we use a tree to compute a probability.
Example 1: There is a 40% chance that you will get a call back from a firm.
If called back there is a 50% chance that you get a job offer. To compute the
probability that you get the job, you draw a tree beginning with 100 chances,
40 times you will get a call back. And of those 40 call backs, you will get the
job 50% of the time or 20 times.
In the second example, we compute an expected value.
1
Example 2: You lose the extension cord for your laptop during a time when
you are contemplating buying a new laptop for $1500. You know that there
is a 50% chance that the same company will introduce a new model of the
laptop. If so, there is a 60% chance it will be $200 cheaper and a 40% chance
it will cost the same amount. You can buy a new extension cord for $50. Do
you do it? Out of 100 chances, 50 times the company will introduce a new
computer. Of those 50 times, 30 times (60%) the new model will cost less.
Therefore, 30% of the time, you save $200. That means that if you wait, you
save $60.
In our last example, we are not given probabilities. Instead, we back out
the appropriate probabilities for each choice.
Example 3: As President, you want to get rid of a dictator who has destabilized a region. If you attack without cause, you anticipate a payoff of 5.
(Ignore the units of the payoffs.) However, you think that this person might
have violated a treaty. You can detect violations 80% of the time. If you find
violations and then attack, your payoff is 9. If you do not find violations,
then you cannot attack and your payoff is 0. How sure do you have to be of
violations in order to want to investigate?
To determine this probability, we draw the tree with 100 chances on top.
Let p be the probability that they are cheating. This means that of the
100 chances, that 100p times they will be cheating and you will detect those
violations 80p times. Therefore, the probability of finding violations is 0.8p
and the payoff is 7.2p. If you do not inspect the payoff is five. You inspect
when 7.2p > 5, or p > 5/7.2 = 69.44%.
3
The Value of Information and Control
The simple tree models are easy. Computing the value of information and
the value of control are more difficult. We first construct an example in which
we compute the value of having information.
Example 4: Suppose that I am forced into an agreement with someone
whom I don’t know. I have two options: To trade in good faith or to renege.
Suppose that I have reason to think that there is a 75% chance this person
is honest. If the person is honest and I trade I make $100. If the person is
2
dishonest and I trade I lose $200. In either case, if I bail out I lose $100.
How much would I pay for information whether the person is honest?
To answer questions like this we draw two trees. The first tree we use to
compute the expected value without the information. In the second tree,
we compute the expected value with the information. First, we compute my
expected value. If I do not trade, I lose $100. If I do, I gain $100 75% of the
time and lose $200 25% of the time for a net gain of $25. Therefore, I should
trade. If I had perfect information, then I would trade when he is honest and
get $100 75% of the time. And, when he is dishonest, I would not trade and
I would only lose $100. This happens 25% of the time. So, my net would be
$50. Therefore, I would pay $25 for the information.
Here is another example
Example 5: You can buy either a Dell or a Mac computer. You are indifferent between the two but the Dell is $50 cheaper. You read in the paper
that Apple is going to send free IPODs to 20% of their customers. You were
just about to spend $300 on an IPOD. A friend tells you that she has learned
that Apple is using a simple random number generator applied to your social
security number to determine who gets the IPODs. She says that for $20,
she will find out if you would get the IPOD or not. Should you take her up
on her offer?
Example 6: Party A is considering moving its convention until after party
B’s. The reason for this is that they would like to have second mover advantage in the selection of their vice president. Party B will choose a governor
from either California or New York. Party A will choose a senator from one
of three states: California, Virginia, or Ohio. The payoffs to party A for each
pair of choices are as follows:
Virginia
Ohio
California
California New York
5
7
6
2
2
8
Suppose that the probability that part B picks the governor from New
York is 20%. The expected payoff from choosing Virginia is 4 + 1.4 = 5.4.
3
The expected payoff from choosing Ohio is 4.8+0.4 = 5.2, and from choosing
California equals 1.6+1.6 = 3.2. Therefore, if A goes first, they pick Virginia
and get a payoff of 5.4. If A picks second and B picks California, A picks
Ohio and gets 6. This happens 80% of the time. If B picks New York, then A
chooses California and gets 8. This happens 20% of the time. The expected
payoff equals 4.8 + 1.6 = 6.4. The gain from full information equals 1.
Note, that the senator from Virginia hates this plan. More information
can mean that a choice that was previously made is now never made. Lave
and March also discuss the value of control. The value of control is computed in the same way as the value of information. The difference between
the two concepts is that the value of control is not buying information but
implementing a plan so that there is no longer any uncertainty.
4
What’s it all about?
This is a class about building models. Why did we just study these trees
in such detail when these trees seem to be prescriptive: here’s how to make
good decisions. To repeat what I said at the beginning, we typically assume
that if people do something often enough than they probably don’t do it
too badly. If we’re modeling it from the outside, we can use these trees to
boil a decision down to its essence. Look for example at Lave and March’s
discussion of voting.
To see the power of trees, let’s consider an example that looks at the effect
of unpredictability. Congress has passed a law that forces states to pay for
some environmental damage. Everyone agrees that the law is constitutional
under current law. The states can challenge this law at a cost C. The only
way that the states can win in the courts is if the courts forge new law. The
probability of this is p. The loss to the states of the new law is L. If we draw
out the tree we see that the states challenge only if C + (1 − p) ∗ L ≤ L,
which reduces to C ≤ p ∗ L. A whole bunch of obvious and cool implications
pop out of this.
1. the lower C, the more likely the states challenge (obv)
2. the higher L, the more likely the states challenge (obv)
3. the higher p, the more likely the states challenge (obv)
4
4. congress wants L to be big but not too big
5. the court wants p to be large enough that laws get challenged
6. which gets set first L or p makes a huge difference
5
Expected Utility
We’ve been using monetary terms and utility terms interchangeably so far.
There is a problem with using money if we’re talking about individuals and
not firms or organizations: risk. Most people would rather get $50 for sure
than $100 half of the time and nothing the other half of the time. (Three
caveats: we like small risks, hence all of this gambling. People tend to be risk
loving over losses. And, for really big risks we hate risk.) Therefore, when
we cut branches off the trees and compute expected values, we are lopping
of risk as well. That’s a mistake.
This begs the question. What about utilities? Is the utility of 5 the same
as half of the time getting a utility of 10 and half of the time getting a utility
of 0? The answer is, it can be if you want it to provided that two assumptions
hold.
Substitution Axiom If x is preferred to y, then a lottery that gives x with
probability p and z with probability (1 − p) is preferred to lottery that gives y
with probability p and z with probability (1 − p).
Archimedean Axiom If x is preferred to y is preferred to z, then there exist
probabilities p and q in (0, 1) such that a lottery that gives x with probability
p and z with probability (1 − p) is preferred to y which is in turn preferred to
lottery that gives x with probability q and z with probability (1 − q).
The first assumption tells us that substituting something better always
makes us happier.. The second says that we can take combinations of something better than y and something worse than y to give us something both
better and worse than y. This assumption rules out extreme risk aversion or
risk loving behavior. I couldn’t have a y that I preferred to any lottery of
something better and worse than it.
5
5.1
Does this make sense
As in the case of creating utility functions, we again have these innocuous
looking axioms that give a powerful result. This happens because axioms
build upon themselves quickly. As you might expect, whether people satisfy
expected utility theory is complicated. It is easy to construct examples where
people violate it.
Example 6: Lottery 1: would you rather have $4 million for sure or a 90%
chance at $6 million? Lottery 2: would you rather have a ten in a million
chance at $4 million or a nine in a million chance at $6 million?
If you choose the former and the latter in this example, you violated the
substitution axiom. The only difference between the two lotteries is that the
second includes a big chance at getting nothing.
6
Practice Question
You have two options: law school or a government internship. However, your
government contacts are all republican and if the Republicans lose the House,
then your internship won’t be as much fun. Your utility from the internship
if they maintain the House majority equals 8. But your utility if they lose is
only 2. You figure that there is a 75% chance that they will maintain it. Your
utility from going to Law School is 6. You have to decide whether to accept
at the law school by September 1 but the internship deadline is December
1. Suppose that the internship deadline were to be moved to September 1.
Suppose that on September 1 that you figure that you will know either that
the republicans have a 100% chance of keeping the House or that they have
a 50% chance. If each utility unit is worth about $10,000, how much would
you pay to get an extension to September 1st for the internship deadline?
6
Download