The value of information

advertisement
Judgments and Decisions
Psych 253
• Using Decision Analysis to Answer Questions
about the Value of Getting Information
• Perfect vs. Imperfect Information
• Signal Detection Theory
• Medical Example
• Jury Decision Making
You can select A, B, or C. Below are the payoffs for each
depending on the throw of a die. Which one do you
want to have?
A
1
$1
2
$2
Die Roll
3
4
$3
$4
B
$6
$2
$5
$2
$2
$2
C
$7
$5
$4
$4
$2
$1
What should you choose?
5
$5
6
$6
A
1
$1
2
$2
Die Roll
3
4
$3 $4
B
$6
$2
$5
$2
$2
$2
$3.17
C
$7
$5
$4
$4
$2
$1
$3.83
5
$5
6
$6
EV
$3.50
Which one would you select if you knew what
number would come up on the die?
A
1
$1
2
$2
Die Roll
3
4
$3 $4
B
$6
$2
$5
$2
$2
$2
C
$7
$5
$4
$4
$2
$1
5
$5
6
$6
What is the expected value of knowing the number
that would come up?
1/6*[ $7 + $5 + $5 + $4 + $5 + $6 ] = $5.33
A
1
$1
2
$2
Die Roll
3
4
$3
$4
B
$6
$2
$5
$2
$2
$2
C
$7
$5
$4
$4
$2
$1
5
$5
6
$6
The value of the information is
$5.33 - $3.83 = $1.50
Value of the
decision
WITH
perfect
information
Value of the
decision
WITHOUT
perfect
information
Miracle Movers rents out trucks with a crew of 2
people. One day, Miracle discovers it is a truck
short. When this happens, Miracle rents a truck
from a local firm. Small trucks cost $130 per day,
and large ones cost $200. The advantage of the
small truck may vanish if the crew has to make 2
trips. The extra cost of a second trip is $150, and the
probability of 2 trips is 40%.
Large Truck
$200 $200
P(NL) =.40
Small Truck
Need Large
Truck $280$280
P(NS) =.60 Need Small
Truck $130$130
What would it be worth to Michelle to know for
sure whether she would need a small truck?
Large Truck
$200 $200
P(NL) =.40
$280$280
Small Truck
P(NS) =.60 $130$130
What would it be worth to Michelle to know for
sure whether she would need a small truck?
Large Truck
$200 $200
Small Truck
EV= $190280
What would it be worth to Michelle to know for
sure whether she would need a small truck?
Don’t Collect
Information
Small Truck $190
“NL” P(NL) = .40
Collect
Information “NS”
(that happens
to be perfect)
P(NS) =.60
Need Large Truck
Need Small Truck
Need Small Truck
Need Large Truck
What does Perfect Information Mean?
Collect Perfect
Information
“NL”
P(NL)=.40
Need Large
Truck $200
“NS”
P(NS) =.60
P (“NL”|NL) = 1 and P (“NL”|NS) = 0
P(“NS”|NS) = 1 and P (“NS”|NL ) = 0
Need Small
Truck $130
Value of the decision with perfect information
= p(“NL” and NL)*$200 + p(“NS” and NS)*$130
+ p(“NL” and NS)*$200 + p(“NS” and NL)*$280
But because the information is perfect,
p(“NL” and NS) and p(“NS” and NL) never occur.
So the value of the decision with perfect
information is
= p(“NL” and NL)*$200 + p(“NS” and NS)*$130
What is P(“NL” and NL) and P(“NS” and NS)?
P(“NL” and NL) = P(“L”|NL)*P(NL) = 1.0*(.4) =.4
P(“NS” and NS) = P(“S”|NS)*P(NS) = 1.0*(.6)=.6
Value of the decision with perfect info
= p(“NL” and NL)*$200 + p(“NS” and NS)*$130
= .4*$200 + .6*$130 = $158
Value of the Decision – Value of the Decision
with perfect info = Value of the Information
Value of the information = $190 - $158 = $32
Perfect information means this…
Information
“S”
“L”
NS P(“S”|NS) =1 P(“L”|NS)=0
Actual
NL P(“S”|NL) =0 P(“L”|NL) =1
If the information is imperfect, we need to
quantify that uncertainty.
Information
“S”
“L”
NS P(“S”|NS) =.9 P(“L”|NS)=.1
Actual
NL P(“S”|NL)=.2 P(“L”|NL) =.8
Signal Detection Theory:
A Theory of Repeated Decision Making
when Information is Fuzzy
-deciding which customers would default on
a loan
-deciding whether to hire an employee
-determining whether someone is lying or
not
-and any other decision that is made
repeatedly using ambiguous evidence
Imagine a doctor who treats patients for a
disease. A test for the number of goodies in the
patient’s blood is the only predictor or clue he
has to predict whether or not the patient has
the disease.
From past research, the doctor knows that
goodies are normally distributed in the blood
of Healthy and Sick people. The average # of
goodies in the blood of Healthy people is 100,
and the average number of goodies in the
blood of Sick people is 115.
0.45
Healthy
0.4
Sick
0.3
0.25
0.2
0.15
0.1
0.05
# Goodies in the Blood
14
5
13
0
12
5
12
0
11
5
10
0
85
70
0
55
Probability of Disease
0.35
We can analyze the doctor’s ability to diagnose a
patients with Signal Detection Theory. This theory
has two parameters:
d´ = d prime - the ability to distinguish Healthy people
from Sick people with the blood test
B = beta (also called the threshold or cutoff) - the
doctor’s tendency to call a patient “Healthy” or
“Sick”
0.45
d’
0.4
Healthy
Sick
0.3
0.25
0.2
0.15
0.1
0.05
# Goodies in the Blood
14
5
13
0
12
5
12
0
11
5
10
0
85
70
0
55
Probability of Disease
0.35
“Healthy”
B
“Sick”
0.45
0.4
Healthy
Sick
0.3
0.25
0.2
0.15
0.1
0.05
# Goodies in the Blood
14
5
13
0
12
5
12
0
11
5
10
0
85
70
0
55
Probability of Disease
0.35
No matter where the doctor sets his cutoff,
there are two errors and two correct decisions.
Decision
“Healthy” “Sick”
State of Healthy
P(“H”|H) P(“S”|H)
the World
Sick P(“H”|S) P(“S”|S)
Each outcome has a name.
Decision
“Healthy” “Sick”
Correct
State of Healthy Rejection
the World
Sick
Miss
False
Alarm
Hit
“Healthy”
B
“Sick”
0.45
0.4
Healthy
Sick
0.3
0.25
0.2
0.15
0.1
0.05
# Goodies in the Blood
14
5
13
0
12
5
12
0
11
5
10
0
85
70
0
55
Probability of Disease
0.35
Which error is worse? It depends.
If the decision is whether to treat a patient
who may have cancer, a miss may be worse
than a false alarm.
If the decision is whether to attack an
unidentified plane, a false alarm may be
worse than a miss.
Where should the doctor set his threshold?
B
“Healthy”
“Sick”
0.45
0.45
0.4
0.4
0.35
0.35
0.2
0.15
# Goodies in the Blood
14
5
55
14
5
13
0
12
5
12
0
11
5
0
10
0
0
85
0.05
70
0.05
13
0
0.1
12
5
0.1
0.25
12
0
0.15
Sick
11
5
0.2
“Sick”
Healthy
10
0
0.25
0.3
85
Sick
70
Healthy
Probability of Disease
0.3
55
Probability of Disease
“Healthy”
B
# Goodies in the Blood
As we move B from the left to the right, we decrease P(“S”|H)
(false alarms), but increase P(“H”|S) (misses)
Where shouldthe doctor to put his cutoff (B)?
SDT tells the decision maker where to locate B,
the cutoff, for a given objective. In this sense,
it is a normative theory.
But SDT can also be used to estimate where
someone puts his or her B. In this sense, it is
a descriptive theory.
The optimal location for B depends on the base
rate for the event (in this case, the disease)
and the utilities associated with the four
possible outcomes.
d’ d’
Healthy
Sick
70
55
85
10
0
11
5
13
0
14
5
16
0
17
5
Sick
Healthy
40
25
Probability
0.45
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
Values of Y hat
Increasing accuracy requires one to find
variables that strengthen the predictability
of the criterion.
Using those variables in a consistent, statistical
fashion improves predictability.
Using a fixed (and optimal) response threshold
reduces noise further improves
predictability.
Another Example: Jury Decision Making
Probability
“Acquit”
“Convict”
Guilty
Innocent
Perceived Culpability
What are acceptable probabilities of errors?
Convicting the innocent - False Alarm
Acquitting the guilty - Miss
Within this framework, how good must the
juror be to achieve reasonable rates of error ?
Assuming normal distributions and equal variances,
If false alarms and misses are 1%, d’ = 4.7.
If false alarm and miss rates are 5%, d’ = 3.3.
If false alarm and miss rates are 10%, d’= 2.6.
What are common values of d’?
Legal Settings
d’
Detecting liars in a mock crime
Detecting liars with polygraphs
0
0.5 - 1.0
Personnel Selection
d’
Job placement with the Armed
Forces Qualification Test
0.6 - 0.8
Weather Forecasting
d’
Fog-risk in Canberra, Australia
24 hours earlier
18 hours earlier
12 hours earlier
Tornados near Kansas City
Rain in Chicago
Minimum temp in Albuquerque
0.8
1.0
1.2
1.0
1.5
1.7
Medical Settings
d’
Detecting breast cancer w mammograms
Detecting brain lesions with RN Scans
Experts detecting cervical cancer
Detecting cervical cancer with computer
based systems
Detecting prostate cancer with PSA tests
Detecting brain lesions with CT Scans
1.3
1.4 - 1.7
1.6
1.8
2.0
2.4 - 2.9
Signal detection theory and empirical estimates
of d’ (with all possible technological advances)
tells us we are not achieving errors that are even
remotely desirable.
What might this imply for policy?
Download