Document 11044647

advertisement
LIBRARY
OF THE
MASSACHUSETTS INSTITUTE
OF TECHNOLOGY
CONSTRUCTION OF OPTIMAL
SEQUENTIAL DECISION TREES
by
CZrfhu^r
William A. Martin
April 1971
MASSACHUSETTS
^UTE OF TECHNOLOGY
50 MEMORIAL DRIVE
[BRIDGE. MASSACHUSETTS
(
I
MASS. INST. TECH.
JUL
12
1971
DEWEr LIBRARY
CONSTRUCTION OF OPTIMAL
SEQUENTIAL DECISION TREES
by
Cirthu.r
William A. Martin
April 1971
528/71
JUL
19
^^]l^,
ABSTRACT
The goal of a sequential decision problem is to identify one object
selected from
object.
a
knovm set of objects by applying a series of tests to the
The outcomes of tests already performed can be used to select the
next test.
Each test may be applied at most once.
Given the a priori probabi-
lities of the objects, the cost of each test, and the probabilities of the
outcomes of each test given the object to vjhlch it is applied, an algorithm
is given for constructing the tree of tests v;hich minimizes the average
cost of identifying an object.
It is assumed that the test results do not
depend on the order in which the tests are applied.
The algorithm uses a
branch and bound procedure to construct partial decision trees.
A lov7er
bound on the cost of completing a partial tree is taken as the minimum
cost of a set of tests V7hich will provide information about the objects
equal to their entropy, under the assum.ption that the tests are independent.
632653
A Lower Bound
We are given a set X of objects and a series of K tests.
has outcomes in the set Y
.
We are also given the cost C
the joint probability distribution p(>:, y
and tests 1 to k produce outcomes y
to y
,
.
.
.
The k
test
of each test and
that object X occurs
,y, )
•
The entropy H(X) of the set of objects before any tests are performed
is defined to be
Since each term in the sum is > 0, it is clear that the entropy is
zero if and only if the probability of some object is equal to 1.
There-
fore, if the entropy is not zero after several tests have been applied it
is not possible that the object has been identified.
Similarly, if the
average decrease in entropy produced from several tests is not at least
equal to the initial entropy then it is not possible for the objects to
be identified in every case.
The decrease in entropy from applying test Y, is
1(X;Y
)
= H(X) - H(xIy
)
=
-V
^'^1
S
p(>:,y,
,
•
•
•
,y^) log p(x)
\
^
X.Y-..,Y^
S-^
X,Y^....,\^
p^'^-'^'i
V
^°2 p(^'Iyi)
PCxIy
)
- 2
Similarly,
the decrease in entropy from applying
H(X) - HCXJY.Y
)
=
1 J
-Z2
Y Y
tests Y
and Y
p(x,y^,...,y^,) log p(x)
Y
"*'2_/
=
tv.'O
p(x,y ,...,y
)
log p(x|y y
)
2^
^
pU.yi>--.>yK)
1^2-:^—
2.
.
p(>-'»yi.---.yK>
^^^-YuWir
+
I(X;YJ + I(X;Y^|y^)
But I(X;Y.|y.) < I(X;Y.)
since
EpCy.)p(y.)
V7hich
using the inequality log
l(X;Y.|Y^) - l(X;Y.) <
(
Y^
2.
V7
<
(u - 1)
becomes
pCy^pCyJ
p(x,y,....y,)
-J^^^r
)
- 1 =
is
i
(Note that equality occurs when pCy.jy.) = p(y.!)p(y.).)
We use the above vrell-knovm result to compute a lower bound on the cost
of testing the objects.
The result says that each test yields at most an
entropy equal to that Which results when it is applied first.
tests the benefit of the doubt and compute these entropies.
costs of the tests
vje
We give the
Using the
then compute the cost per unit of entropy hopefully
to be received from each test and rank the tests in the order of increasing
cost per unit.
of tests
v.-'hich
objects.
Finally, using this ranking, we form the
lox-?est
cost set
could possibly yield entropy equal to that of the set of
A fractional part of one test may be included in this set in order
to get the exact amount of entropy needed.
then include the
VJe
sair.e
fraction
of its cost in the cost of this set, which is the lower bound in question.
Branch and Bound
Armed V7ith this lower bound we can employ a branch and bound algorithm
like that described by Reinwald and Soland (2)
decision trees.
.
We form a tree of partial
The root node of the tree of partial decision trees
corresponds to no tests applied.
A node is a complete node of
a decision
tree if no tests remain which have not been applied in reaching that node,
if one object has probability one at that node, or if branches extend from
that node.
A node of the tree of partial decision trees
is a terminal node
if the decision tree at that node contains no incomplete nodes.
The ave-
rage cost of the decision tree of each non-terminal node is bounded by the
costs of the tests applied so far plus the lovjer bounds on the cost of
I
- 4
each of the sub-decision trees v^hich has
xiot
been completed.
When the cost of a terminal node is less. than the lower bounds on all
the non-terninal nodes, this terminal node represents the optimum tree.
When this is not the case the non-terminal node V7hich contains the partial
decision tree V7ith the lovrest cost bound is extended by adding a branch
for each test not yet applied.
It is possible that some uncertainty vjill remain after all of the tests
are applied.
In this case, more than one order of applying the tests vill
yield the same results.
This situation can be detected by evaluatlns the
result of applying all of the tests in a given order.
If the reinaining un-
certainty is small it may be desirable to find the best tree which reduces
the entropy below some threshold.
.
References
1.
Gallager, H.G., "Information Tlieory and Reliable Communication",
Wiley (1968).
2.
Reinwaldj
L_.
and Soland R.
Tables to Opt'
Processing
2.
Reinv7ald, L.
Tj
;!
.
,
"Conversion of Limited-Entry Decision
Computer Programs I:
and Soland R.
,
"Conversion of Limited-Entr>' Decision
Tables to Optimum Computer Programs II:
Requirement".
Minimum Average
JACM (July 1966)
JACM (October 1967).
Minimum Storage
Date Due
N**"
^2-?-
3
TDfiO 0D3
fl75
3b3
5A^3
TOflO
0D3 TOb
40fi
HD28
M.I.T. Alfred P. Sloan
Mklk
school of Management
Nos.523-71-Working Papers.
Nos. 532-71 Nos. 523-71 to Nos. 532-71
||||||||ll|||||||((||||||n
5'z5,ff
lllllillllllllllllllll
3T0flDD03TDb3fl5
3 TOflO
DD3
fl7S
3T7
I
1527-7^
3
TOflD
DD3 TDb 3T0
3
TDflD
0D3 TOb 3bb
3
TDfiD
D03
BASFMENF
^
fl7S
SIGN
this
card and present
it
with book
J,
413
at
the
CIRCULATION DESK
5>9'7|
3
^OfiD DD3
TDb
MASSACHUSETTS INSTITUTE OF
fi4b
TECHNOLOGY
LIBRARIES
^M-71
531
3
S060 003 TOb
ttbJ.
DEWEY LIBRARY
f?2-7'
3 TDfiO
003 TOb 671
Download