Constraints on credences in two not mutually exclusive propositions:

advertisement
Constraints on credences in two
not mutually exclusive propositions:
the search for the best belief update function. . .
Leszek Wroński
Institute of Philosophy
Jagiellonian University
Kraków, Poland
’Full and Partial Belief’ Workshop
The Tilburg Center for Logic, General Ethics, and Philosophy of Science
20 X 2014
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
1 / 23
. . . or . . .
Leszek Wroński
Institute of Philosophy
Jagiellonian University
Kraków, Poland
’Full and Partial Belief’ Workshop
The Tilburg Center for Logic, General Ethics, and Philosophy of Science
20 X 2014
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
1 / 23
Some problems with
minimizing expected epistemic inaccuracy
Leszek Wroński
Institute of Philosophy
Jagiellonian University
Kraków, Poland
’Full and Partial Belief’ Workshop
The Tilburg Center for Logic, General Ethics, and Philosophy of Science
20 X 2014
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
1 / 23
Some problems with
minimizing expected epistemic inaccuracy:
logarithmic measures to the rescue
Leszek Wroński
Institute of Philosophy
Jagiellonian University
Kraków, Poland
’Full and Partial Belief’ Workshop
The Tilburg Center for Logic, General Ethics, and Philosophy of Science
20 X 2014
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
1 / 23
Outline:
the accuracy-centered perspective in formal epistemology: the basic
notions;
a problem from Leitgeb & Pettigrew (2010b) regarding updating one’s
belief function in response to a certain type of evidence;
the solution in the simplest non-trivial case and its intransferability to
different cases;
a better (?) solution obtained using a different measure of inaccuracy
from the one suggested by Leitgeb & Pettigrew.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
2 / 23
Alethic inaccuracy
An important notion in formal epistemology: Inaccuracy as distance
from truth.
Properly fleshed out, this notion has been used to justify numerous
epistemic norms.
In this talk assume that belief functions are probability functions on the
power sets of finite sets of epistemically possible worlds. (It follows that
we assume probabilism; with an agent at a given time we associate a
finite probability space hW , P(W ), bi.)
An inaccuracy measure should be a function giving us a real number for a
possible world and a belief function:
I (w , b) = r
(“If w is the actual world, then the inaccuracy of the belief function b is r ”).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
3 / 23
I (w , b) = r
(“If w is the actual world, then the inaccuracy of the belief function b is r ”).
The actual shape of an inaccuracy measure follows from which scoring
rule is used. Two widely used types of scoring rules lead to the so called
quadratic and logarithmic inaccuracy measures.
We can also consider the expected inaccuracy of a belief function b0
given a belief function b:
X
b({w }) · I (w , b0 )
w ∈W
and require from an ideally rational agent to, in response to the incoming
evidence, aim to minimise expected inaccuracy from the perspective of
their current belief function.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
4 / 23
Updating your credences
Suppose, starting with a belief function b, you learn (only) that a
proposition E is true. How should your new belief function b0 look like?
The traditional Bayesian answer: conditionalise. That is, for any A,
b0 (A) = b(A|E ).
This way of updating is sometimes called full conditionalisation. It only
makes sense if b(E ) 6= 0. And it will never raise your credence from 0 or
lower it from 1.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
5 / 23
But what if our evidence is not certain? E.g. we learn our new credences in
disjoint propositions E1 , · · · , Em , which form a partition of W ?
The traditional answer: use Jeffrey Conditionalisation. That is, for any
A,
m
X
0
b (A) =
b0 (Ei ) · b(A|Ei ).
i=1
Full conditionalisation is a special case of Jeffrey conditionalisation (JC).
JC will also never make your extreme credences non-extreme. It is also
order-dependent: two sets of constraints applied in reverse order may
result in a different belief function.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
6 / 23
The updating problem
What if the constraints implied by your evidence are (only) your new
credences in some two propositions A and B such that A ∩ B 6= ∅?
This is an open problem in Leitgeb & Pettigrew (2010b).
I will show a partial answer in the L & P approach and a general answer in
a rival framework.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
7 / 23
A sketch of (a part) of Leitgeb & Pettigrew (2010):
a series of arguments for using quadratic inaccuracy measures;
a proof that full conditionalisation minimizes expected inaccuracy;
an example showing that Jeffrey conditionalisation does not!
a proof that an Alternative Jeffrey Conditionalisation (AJC) rule does.
AJC sometimes dictates that you raise your credence in a proposition from
0. The way of updating depends not only on your new credences in the
elements of a partition, but also on the cardinality of the elements.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
8 / 23
Levinstein (2012):
the AJC rule leads to unintuitive updating behaviour;
if the logarithmic inaccuracy measure is used, Jeffrey
Conditionalisation follows as the way to minimise one’s expected
inaccuracy in cases of the type discussed above.
So: what happens in our case? To repeat: you learn just the new credences
in some two propositions A and B such that A ∩ B 6= ∅.
Let us tackle the simplest non-trivial—four-world—setup first. Assume
then that W = {w1 , w2 , w3 , w4 }, A = {w1 , w2 } and B = {w2 , w3 } and you
start with a belief function b; you learn your new credences b0 (A) = p and
b0 (B) = q. Notice that the task reduces to figuring out b0 (A ∩ B).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
9 / 23
Levinstein (2012):
the AJC rule leads to unintuitive updating behaviour;
if the logarithmic inaccuracy measure is used, Jeffrey
Conditionalisation follows as the way to minimise one’s expected
inaccuracy in cases of the type discussed above.
So: what happens in our case? To repeat: you learn just the new credences
in some two propositions A and B such that A ∩ B 6= ∅.
Let us tackle the simplest non-trivial—four-world—setup first. Assume
then that W = {w1 , w2 , w3 , w4 }, A = {w1 , w2 } and B = {w2 , w3 } and you
start with a belief function b; you learn your new credences b0 (A) = p and
b0 (B) = q. Notice that the task reduces to figuring out b0 (A ∩ B).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
9 / 23
Fact 1
Let A be an agent with the belief function b at time t. Let
W = {w1 , w2 , w3 , w4 } be the set of epistemically possible worlds for A.
Let A be the proposition {w1 , w2 }; let B be {w2 , w3 }. Suppose that
between times t and t 0 the agent learns (only) the following constraints on
her belief function b0 at t 0 : b0 (A) = p and b0 (B) = q. Let us label the
as K . The belief update function which
number b(A ∩ B) + p−b(A)+q−b(B)
2
minimizes global expected inaccuracy in the sense of Leitgeb & Pettigrew
(2010) is fully determined by the two constraints and the following
condition:

if K < 0

0
0
b (A ∩ B) = min {p, q} if min {p, q} < K


K
otherwise.
So: the change in your credence in A ∩ B is the average of the changes in
your credences in A and in B (whenever it makes mathematical sense).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
10 / 23
At this point a prima facie reasonable way to proceed would be the
following:
given any similar situation with an arbitrary finite W , calculate
b0 (A ∩ B) using the above formula;
this, together with the two constraints, gives us the credences at time
t 0 in all propositions in the set {A ∩ B, A ∩ ¬B, ¬A ∩ B, ¬A ∩ ¬B},
which is a partition of W ;
We can (it would seem) now use AJC to derive the full shape of b0 .
But no. It turns out the answer depends on the cardinality of the
propositions. (Not just the AJC does; but also the method of calculating
the input for AJC.)
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
11 / 23
At this point a prima facie reasonable way to proceed would be the
following:
given any similar situation with an arbitrary finite W , calculate
b0 (A ∩ B) using the above formula;
this, together with the two constraints, gives us the credences at time
t 0 in all propositions in the set {A ∩ B, A ∩ ¬B, ¬A ∩ B, ¬A ∩ ¬B},
which is a partition of W ;
We can (it would seem) now use AJC to derive the full shape of b0 .
But no. It turns out the answer depends on the cardinality of the
propositions. (Not just the AJC does; but also the method of calculating
the input for AJC.)
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
11 / 23
Consider the following example (adapted from Osherson 2002): “Listening
to the radio I hear a forecast for rain but I’m not sure whether it comes
from the chief meteorologist or from his unreliable deputy.” At that
moment I have the following belief function b1 , where R is “It rains today”
and C is “The chief was speaking”;
b1
¬RC
.2
RC
.4
R¬C
.1
¬R¬C
.3
Now, a glance at the sky raises my credence in R to .7. Assume that the
forecast is rebroadcasted and even though I strain my ears, I conclude I
should not change my credence in C . The following is my new credence
dictated by minimizing my expected epistemic inaccuracy in the sense of
Leitgeb & Pettigrew (i.e. by Fact 1):
0
b1
Leszek Wroński (JU, Poland)
¬RC
.1
RC
.5
R¬C
.2
¬R¬C
.2
Problems with minimizing GExpInacc
20 X 2014
12 / 23
I’d like to raise two problems in this situation.
a problem in the 4-world case, with reacting to a certain type of
evidence;
a problem with extending it to a 5-world case and reacting to the
same evidence as on the previous slide.
Let us start with the second problem.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
13 / 23
Suppose I owe the chief meteorologist money and I don’t want to meet
him. I know that whenever it doesn’t rain, he always walks home through a
park he never visits otherwise and in which I walk my dog (which I do
regardless of the weather); not wanting to disturb my dog’s routine, I will
also go to the park today. (If it rains, the chief meteorologist takes a bus
home.) He can traverse the park via one of two paths, call them “1” and
“2”. Let B be the proposition “the chief meteorologist will walk via path 1
today”. My four epistemic possible worlds
¬RC
RC
R¬C
¬R¬C
become
¬RCB
¬RC ¬B
RC ¬B
R¬C ¬B
¬R¬C ¬B
and my initial credence becomes, say,
b2
¬RCB
.1
¬RC ¬B
.1
RC ¬B
.4
R¬C ¬B
.1
¬R¬C ¬B
.3
The ¬RC world has effectively split in two. (Think of this as of two agents
with a similar space of epistemically possible worlds.)
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
14 / 23
Consider the same evidential situation: a glance at the sky raises my
credence in R to .7; the forecast is rebroadcasted and even though I strain
my ears I conclude I should not change my credence in C . The following is
my new credence dictated by minimizing the expected inaccuracy in the
L & P sense:
0
b2
¬RCB
.043
¬RC ¬B
.043
RC ¬B
.514
R¬C ¬B
.186
¬R¬C ¬B
.214
Let us bring together the initial and updated credences for the two agents:
b1
0
b1
b2
0
b2
¬RCB
.1
.043
Leszek Wroński (JU, Poland)
¬RC
.2
.1
¬RC ¬B
.1
.043
RC
.4
.5
R¬C
.1
.2
RC ¬B
.4
.514
¬R¬C
.3
.2
R¬C ¬B
.1
.186
Problems with minimizing GExpInacc
¬R¬C ¬B
.3
.214
20 X 2014
15 / 23
Notice that the two agents
started with the same credences in R, C , and RC ;
faced the same evidence implying (only) that they should increase their
credence in R in the same way and not to change their credence in C ;
ended with a different credence in RC ;
and it seems the only difference was whether the park was considered
or not.
Note: there are cases in which not only the credence in a suitably chosen
proposition ends up being different, but the two agents end up with
different probability rankings regarding two propositions which they
initially believe to the same degree: that is, cases in which
b1 (φ) = b2 (φ) and b1 (ψ) = b2 (ψ)
but
0
0
0
0
b2 (φ) < b2 (ψ) while b1 (φ) > b1 (ψ).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
16 / 23
Now, for a problematic example in the 4-world case. Start again with the
same credence function:
b1
¬RC
.2
RC
.4
R¬C
.1
¬R¬C
.3
Consider now two pieces of evidence:
a glance at the sky lowers my credence in R to .4 (by .1);
a rebroadcasting of the forecast increases my credence in C to .7 (by
.1).
The following is my new credence dictated by Fact 1:
00
b1
¬RC
.3
RC
.4
R¬C
.0
¬R¬C
.3
Notice that somehow I ended up with credence 0 in R¬C .
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
17 / 23
What happened is I updated my credences in R and C as if they were
perfectly anticorrelated. I interpreted an increase in the credence in C as
an increase in the credence in ¬RC ; I interpreted a decrease in the
credence in R as a decrease in the credence in R¬C .
In general:
increase in R=increase in C : update as if they are perfectly correlated
(increase RC by the same amount);
increase in C =−increase in R: update as if they are perfectly
anticorrelated (leave RC as it was, increase ¬RC and decrease R¬C
by the same amount).
Even if some would like to protest against using the word “correlation”
here, I think we should find e.g. the example involving the decrease of
credence in R¬C to 0 troubling.
This behaviour extends beyond the 4-world case.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
18 / 23
Logarithmic measures to the rescue
Levinstein (2012): JC does minimise expected inaccuracy if a logarithmic
inaccuracy measure is used.
The measure used by Levinstein is Ilog (w , b) = − ln(b({w }).
What happens if we use this measure in our case?
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
19 / 23
Fact 2
Let A be an agent with the belief function b at time t. Let A and B be
two propositions such that A ∩ B 6= ∅. Suppose that the evidence received
by A between times t and t 0 entails (only) her credences b0 (A) and b0 (B).
To minimise expected inaccuracy calculated by the logarithmic measure A
should calculate her new credences in the logical combinations of A and B
so that the following is true:
b(¬A ∩ B)
b(A ∩ B)
b(¬A ∩ ¬B)
b(A ∩ ¬B)
+ 0
= 0
+ 0
0
b (A ∩ ¬B) b (¬A ∩ B)
b (A ∩ B) b (¬A ∩ ¬B)
and then use Jeffrey Conditionalisation.
This does not depend on the cardinality of the propositions.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
20 / 23
...or, in other words:
Fact 2
Let A be an agent with the belief function b at time t. Let A and B be
two propositions such that A ∩ B 6= ∅. Suppose that the evidence received
by A between times t and t 0 entails (only) the following constraints on her
belief function b0 at t 0 : b0 (A) = p and b0 (B) = q. To minimise expected
inaccuracy calculated by the logarithmic measure A should calculate her
new credences in the logical combinations of A and B so that the following
is true:
b(¬A ∩ B)
b(A ∩ B)
b(¬A ∩ ¬B)
b(A ∩ ¬B)
+
= 0
+ 0
0
0
p − b (A ∩ B) q − b (A ∩ B)
b (A ∩ B) b (A ∩ B) + 1 − p − q
and then use Jeffrey Conditionalisation.
This does not depend on the cardinality of the propositions.
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
20 / 23
Conclusions
I have shown:
the solution to Leitgeb & Pettigrew’s problem in the simplest
non-trivial case;
that the solution in their framework is not transferable to different
cases;
a general solution using the logarithmic inaccuracy measure.
This might be an argument against using quadratic inaccuracy measures.
But what to do with L & P’s powerful arguments in their 2010 paper?
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
21 / 23
Thank you!
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
22 / 23
References:
Leitgeb, H., and Pettigrew, R. (2010), “An Objective Justification of
Bayesianism I: Measuring Inaccuracy”, Philosophy of Science 77(2);
Leitgeb, H., and Pettigrew, R. (2010), “An Objective Justification of
Bayesianism II: The Consequences of Minimizing Inaccuracy”,
Philosophy of Science 77(2);
Levinstein, B. (2012), “Leitgeb and Pettigrew on Accuracy and
Updating”, Philosophy of Science 79(3).
Leszek Wroński (JU, Poland)
Problems with minimizing GExpInacc
20 X 2014
23 / 23
Download