Document 11070408

advertisement
DEWET
^0
HD28
.M414
o t
MIT
Sloan School of Management
Working Paper 4340-01
December 2001
PRICING AMERICAN OPTIONS:
A DUALITY APPROACH
Leonid Kogan and Martin Haugh
© 2001
by Leonid Kogan and Martin Haugh. All
Short sections of text, not to exceed two paragraphs,
permission provided that
full credit
including
rights reserved.
may be quoted
© notice
is
without explicit
given to the source.
This paper also can be downloaded without charge from the
Social Science Research
Network Electronic Paper Collection:
http://papers.ssm.com/abstract_id=294821
MASSACHucFlTS INSTITUTE
OF TECHNQLOGY_
J
UN
3 2002
LIBRARIES
Pricing American Options:
A
Duality Approach^
Martin B. Haugh^ and Leonid Kogan*
Abstract
We develop a new method
for pricing
American options. The main
practical contribution
a general algorithm for constructing upper and lower bounds on the true
price of the option using any approximation to the option price. We show that our
bounds are tight, so that if the initial approximation is close to the true price of the
of this paper
option, the
is
bounds are
also guaranteed to
be
close.
worst-case performance of the pricing bounds.
is
We
also explicitly characterize the
The computation
of the lower
bound
straightforward and relies on simulating the suboptimal exercise strategy implied by
the approximate option price.
The upper bound
is
also
computed using Monte Carlo
by the representation of the American option price as a
solution of a properly defined dual minimization problem, which is the main theoretical
result of this paper. Our algorithm proves to be accurate on a set of sample problems
where we price call options on the maximum and the geometric mean of a collection
of stocks. These numerical results suggest that our pricing method can be successfully
simulation. This
is
made
feasible
applied to problems of practical interest.
*An
earlier draft of this
Approach
.
We thank Mark
paper was
titled
A
Duality
Wang and
seminar
Pricing High- Dimensional American Options:
Broadie, Domenico Cuoco,
Andrew
Lo, Jacob Sagi, Jiang
Columbia University, INSEAD, London Business School, MIT, Princeton, Stanford, and
Wharton for valuable comments. We thank an anonymous referee for pointing out the alternative proof
of Theorem 1 in Section 3.1. Financial support from the MIT Laboratory for Financial Engineering, the
Rodney L. White Center for Financial Research at the Wharton School, and TIAA-CREF is gratefully
participants at
acknowledged.
We
are grateful to Anargyros Papageorgiou for providing the
generate the low discrepancy sequences used
in this
^Operations Research Center, MIT, Cambridge,
FINDER
software to
paper.
MA
02139; email; mhaugh@alum.mit.edu.
'Corresponding author: Sloan School of Management, MIT, 50 Memorial Drive, E52-455, Cambridge,
MA
02142; email: lkogan@mit.edu.
Introduction
1
Valuation and optimal exercise of American options remains one of the most challenging practical problems in option pricing theory.
The computational
cost of traditional
valuation methods, such as lattice and tree-based techniques, increases rapidly with the
number
known
of underlying securities
and other payoff-related
Due
variables.
to this well-
curse of dimensionality, practical applications of such methods are limited to
problems of low dimension.
In recent years, several
sionality.
methods have been proposed
to address this curse of dimen-
Instead of using traditional deterministic approaches, these methods use Monte
Carlo simulation to estimate option prices. Tilley (1993) was the
first
Other important
that American options could be priced using simulation techniques.
work
in this literature includes
mar and Zwecher
to demonstrate
Barraquand and Martineau (1995), Carriere (1996), Ray-
(1997), Ibanez
and Zapatero (1999) and Garcia (1999). Longstaff and
Schwartz (2001) and Tsitsiklis and Van Roy (1999, 2000) have proposed an approximate
dynamic programming approach that can compute good
in practice. Tsitsiklis
and Van Roy
success of approximate
price estimates
and
very fast
is
also provide theoretical results that help explain the
dynamic programming methods. The
price estimates these tech-
niques construct, however, typically only give rise to lower bounds on the true option
price.
As a
result, there
is
usually no formal
method
for evaluating the
accuracy of the
price estimates.
In an important contribution to the literature, Broadie
develop two stochastic mesh methods
for
American option
and Glasserman
pricing.
tages of their procedure over the previously proposed methods
is
One
that
it
(
1997a, b)
of the advan-
allows
them
to
generate both lower and upper bounds on the option price that converge asymptotically
to the true option price. Their
bounds are based on an application of Jensen's inequal-
and can be evaluated by Monte Carlo simulations.
ity
necessarily generalize to other pricing methods.
drawback but nonetheless appears
of Broadie
method
of their first
suffer
demanding. In an
to be computationally
is
from
effort
and Tan (2001) generalize the stochastic
to address this drawback, Boyle, Kolkiewicz
mesh method
The complexity
The second approach does not
exponential in the number of exercise periods.
this
However, such bounds do not
and Glasserman (1997b) using low discrepancy sequences to
improve the efficiency of the approach.
The main
practical contribution of this paper
is
a general algorithm for constructing
upper and lower bounds on the true price of the option using any approximation to the
option price.
is
We
show that our bounds are
close to the true price of the option, the
addition,
tight, so that
bounds are
the
if
of the lower
bound
is
obtained by simulating a different stochastic process that
appropriate supermartingale.
We
In
of the pricing bounds.
straightforward and relies on simulating the
suboptimal exercise strategy implied by the approximate option
is
approximation
also guaranteed to be close.
we exphcitly characterize the worst-case performance
The computation
initial
justify this
is
price.
The upper bound
determined by choosing an
procedure by representing the American
option price as a solution of a dual minimization problem, which
is
the
main
theoretical
result of this paper.
In order to determine the option price approximation underlying the estimation of
bounds, we also implement a
dynamic programming
fast
and accurate valuation method based on approximate
(see Bertsekas
and
Tsitsiklis 1996)
where we use non-linear
gression techniques to approximate the value function. Unlike
Monte Carlo simulation
most procedures that use
to estimate the continuation value of the option, our
deterministic and relies on low discrepancy sequences as an alternative to
simulation.
For the examples considered
in this paper,
we
re-
method
is
Monte Carlo
find that low discrepancy
sequences provide a significant computational improvement over simulation.
While the duality-based approach to portfolio optimization problems has proved
successful
and
is
now widely used
1998), the dual approach to the
in finance,
(see, for
example, Karatzas and Shreve
American option pricing problem does not seem
been previously developed other than
in recent
to have
independent work by Rogers (2001).
Rogers establishes a dual representation of American option prices similar to ours and
applies the
new
options using
approach
representation to compute upper bounds on several types of American
Monte Carlo
simulation. However, he does not provide a formal systematic
for generating tight
upper bounds and
his
computational approach
is
problem
specific.
Andersen and Broadie (2001) use the methodology developed
in this
paper to
for-
A
dis-
malulate another computational algorithm based on Monte Carlo simulation.
tinguishing feature of their approach
They
the use of an approximate exercise policy, as
price, to estimate the
opposed to an approximate option
option.
is
bounds on the true
price of the
also observe that a straightforward modification (taking the martingale
part of the supermartingale) of the stochastic process used to estimate the upper
bound
also applies to the algorithm
we use
leads to
in this
more accurate estimates. This observation
paper where we begin with an
led to a significant
improvement
initial
in the
approximation to the option
price.
computational results that were presented
This
in
an
earlier draft of this paper.
The algorithm
riods, given
of
Andersen and Broadie
is
quadratic in the number of exercise pe-
knowledge of the approximate exercise
policy, while
given knowledge of the approximate option price. While
results are not currently available,
it
it is
our approach
is
linear
true that formal complexity
does seem to be the case that accurate approxi-
mations to the option price can be computed very quickly.
(See, for
example, Longstaff
and Schwartz 2001,
gramming methods such
that pricing
and Van Roy 2000, and other approximate dynamic pro-
Tsitsiklis
as the
method used
methods based on an
initial
in this paper.)
As a
approximation to the option
we beheve
result,
price, rather
than
the exercise policy, will be more efficient in practice. This should hold in particular for
problems with many exercise problems. For a particular
exercise dates,
it is
The
compare the
difficult to
paper
rest of the
In Section 3,
we
is
derive the
We
problems with more than 10 exercise periods and
for
practical tradeoffs
between our alternative approaches.
new
duality result for
price. In Section 4
American options and use
it
to derive
we describe the implementation
report numerical results in Section 5 and
we conclude
of the
in Section 6.
Problem Formulation
2
In this section
we formulate the American option
Information Set.
pricing problem.
We consider an economy with a set of dynamically
markets, described by the underlying probability space,
risk-neutral valuation
if
ours.
organized as follows. In Section 2 we formulate the problem.
an upper bound on the option
algorithm.
sample problems with 10
Andersen and Broadie compute pricing bounds that are similar to
However, they do not report results
so
set of
financial
measure Q.
It is
well
known
il,
complete financial
the sigma algebra T, and the
(see Harrison
and Kreps 1979) that
markets are dynamically complete, then under mild regularity assumptions
there exists a unique risk-neutral measure, allowing one to
compute
prices of
contingent claims as the expected value of their discounted cash flows.
economy
More
we assume that Tt
specifically,
is
is
^
-.t
€ [0,r]}.
'
t
€
[0.7"]}.
generated by Z(, a d-dimensional standard Brow-
nian motion, and the state of the economy
process {Xt e
state-
The information
represented by the augmented filtration {Tt
structure in this
all
is
represented by an .F(-adapted Markovian
Option Payoff.
Let
/i,
=
/j(A',)
be a nonnegative adapted process representing the
payoff of the option, so that the holder of the option receives
at time
t.
We
also define a riskless account process Bt
We
denotes the instantaneously risk-free rate of return.
/i,
=
if
the option
exp
(
j^
/-^c/s-
is
j,
exercised
where
r^
assume that the discounted
payoff processes satisfies the following integrability condition
max
(1)
(=0,1,. ..,T
where E,
ditional
[]
denotes the expected value under the risk-neutral probability measure, con-
on the time
t
information,
JF,.
Exercise Dates. The American feature
at
any of the pre-specified exercise dates
of the option allows its holder to exercise
T = {0,
in
and T. Equal spacing of the exercise dates
is
1,
.
,
.
.
assumed
it
T}, equally spaced between
to simplify the notation
and
mapped
not restrictive. Moreover, a unit time increment within the model can be
is
into
any period of calendar time.
Option Price. The
value process of the American option,
the option conditional on
it
not having been exercised before
Vt
=
Vt, is
t.
the price process of
It satisfies
Bthr
sup Et
(2)
Br
where r
is
any stopping time with values
in the set
Tn[t,T].
This
is
a well-known
characterization of American options (see, for example, Bensoussan 1984, Karatzas 1988
and Pliska 1997).
3
Theory
Our approach
Step
1.
to the
American option pricing problem consists of the following
Compute an approximation
to the
steps.
market price of the option as a function of the
time and state. Specifically, we use an approximate dynamic programming algorithm to
determine the continuation value of the option,
on not exercising
Step
2.
it
i.e.,
the value of the option conditional
at the current time period.
Estimate the lower bound on the option price by simulating the approximate
exercise strategy based on the option price approximation from Step
Step
it
3.
Based on the option price approximation, define a martingale process and use
to estimate the
new dual
the
We
of the
upper bound by Monte Carlo simulation. This
last step is
based on
representation of the option price presented below.
begin this section with our main theoretical result on the dual representation
American option
price.
compute bounds on the option
We
then show
how
to use this price characterization to
and study the properties
price,
of these bounds.
The Dual Problem
3.1
The problem
an American option, the primal problem,
of pricing
Vo
=
Fit.^)._^^
Then the dual problem
gales,
TTj.
is
iTt,
max
se{\t,T]nT}
Bt
that of computing
we
I
define the dual function F{t,
—
+t7r
^TT
it)
as
(3)
\Bs
to minimize the dual function at time
over
all
supermartin-
Let Uq denote the optimal value of the dual problem, so that
Uo
=
inf F(0,
tt)
=
inf
Eq
max
—
—
TT,
+^
following theorem shows that the optimal values of the dual
coincide.
is
sup Eo
rer
For an arbitrary adapted supermartingale,
The
1.
(D)
and primal problems
Theorem
(Duality Relation) The optimal values of
1
the dual
problem (D) are
problem
is
given by
=
tTj*
equal,
Vq
i.e.,
=
Vt
is
the value process for the
Vq
=
sup
Br
—
teT
where the
TTf,
Vt/Bt
is
<
sup Eo
4-(yr
+(yr
(TT
(4)
\Bt
inequahty follows from the optional sampling theorem
first
gales (see Billingsley 1995)
gales,
tt;,
hr
<En max —
on the
right
hand
option,
Et
sup Eo
r6T
sup Eo
T&T
American
'hrB,'
=
re{[t,T]nT}
Proof For any supermartingale
and
Moreover, an optimal solution of the dual
f/o-
Vt/Bt, where Vt
the primal problem (P)
and condition
(1).
for
supermartin-
all
supermartin-
Taking the infimum over
side of (4) implies Vq
<
f/o.
On
the other hand, the process
a supermartingale, which implies
Uo<Eo max
{ht/Bt
teT
Since Vt
>
attained
when
ht for all
Theorem
1
=
tt^*
t,
-Vt/Bt)
we conclude that Uq <
Vq.
+Vo.
Therefore, Vq
=
Uo,
and equality
is
V/Bt.
shows that an upper bound on the price of the American option can be
constructed simply by evaluating the dual function over an arbitrary supermartingale
TTt-
In particular,
bounded above by
if
such a supermartingale
tto.
Theorem
1
satisfies nt
ht/Bt, the option price Vq
is
therefore implies the following well-known character-
ization of the
American option
Proposition
1
cess Vt/Bt
the smallest supermartingale that
is
>
price (see, for example,
Phska 1997).
(Option Price Characterization) The discounted option
price pro-
dominates the discounted payoff of the
all exercise periods.
option at
The
reverse
is
also true,
i.e.,
one can use Proposition
1
to prove
Theorem
1.
that since both processes on the right-hand side of (3) are supermartingales, so
discounted dual function F{t,n)/Bt. Clearly F(i,
-^(0,
tt)
>
Vo by Proposition
1.
When
tt^
is
>
tt)
ht/Bt for any
Theorem
1
The
pricing problem
American option, which
we use
is
is
to evaluate the
closely related to the
method
in
in a
new
price.
While various methods
in the literature,
for
computing
remained a challenging problem. The result of Theorem
for super-replicating the
American option.
Since the discounted dual function F{t, 7r)/Bt
markets
1
problem of dynamic replication of the
equally important in practice.
reliable replication strategies has
suggests a
and the dual problems
upper bound on the option
approximating American option prices have been suggested
1
implying that
therefore expresses the well-known result of Proposition
constructive form, which
the
is
chosen to be the discounted option price, the
reverse inequality holds, implying that the values of the primal
coincide.
t,
Note
is
a supermartingale and financial
our model are dynamically complete, there exists a self-financing trading
strategy with an initial cost F(0, n) which almost surely dominates F{t,
1996, Section
2.1).
tt)
(see Duffie
Such a trading strategy super-replicates the payoff of the American
option, since F{t, n) dominates the price of the option
Using an approximation to the option
price,
we can
and hence
define
tt,
its
payoff at exercise.
and F(^7r)
explicitly, so
that super-replicating the option can be a relatively straightforward task. In particular,
Boyle et
al.
(1997) and Garcia et
al.
(2000) describe
Monte Carlo methods
for estimating
the portfolio strategies replicating the present value process of a state contingent claim.
This claim could correspond to a derivative security or some consumption process. Their
results are directly applicable to (3).
The Upper Bound on the Option Price
3.2
When
the supermartingale ni in (3) coincides with the discounted option value process,
Vt/Bt, the upper
bound F(0,
TTf
in
the exact price,
equals the true price of the American option.
such a way that when the approximate option price,
V(,
tt^
equals the discounted option price, Vt/ Bt-
to use either of the following
Tfo
This
bound can be obtained from an accurate approximation,
suggests that a tight upper
by defining
tt)
=
two definitions of
It
V,,
Vt,
coincides with
seems natural then
ttj:
Vo
(5)
(6)
(7)
where
TT,,
Vt
(x)"*"
implying
=
Vf,
Tit
:= max(x,0).
tt^
=
is
By
construction, E^
an adapted supermartingale
[ttj+i
for
Vt/Bt, because the latter process
is
— (Tt
bound corresponding
the properties of the
to (5,6)
bound under the
is
first
<
for either definition of
any approximation,
Also,
Vj.
when
a supermartingale and the positive
part of expectations in (5,6) and (5,7) equals zero. While
the upper
]
we cannot claim a
superior to the
priori that
bound determined by
(5,7),
definition are considerably easier to analyze.
Note also that the upper bound can be tightened further by omitting the positive
part in the definition of
of
TTf)
The
resulting process
and therefore a supermartingale, so that
bound.
It
tt;.
It
coincides with nt at
t
=
and
is
we used the formulation
a martingale (the martingale part
too can be used to construct an upper
always greater than or equal to
therefore leads to a lower value of the upper
of this paper,
it
is
bound defined by
(5,6) to
(3).
(In
an
nt for
t
>
0.
earlier draft
compute upper bounds on the option
price.
Andersen and Broadie (2001) point out, however, that
in general tighter
upper
bounds can be obtained by using the martingale component of the supermartingale,
component
In our framework, the martingale
tt^.
obtained by simply omitting the positive
is
in (5,6)).
For the remainder of the paper we
TTo
T^t
+l
=
the upper
bound
Fn =
Vn
+
=
TT,
+
max
Vt
Bt
(9)
to (8)
and
(9).
Then
it is
easy to see that
-^
-;:;
\Bt
+
y
(10)
tjj.
Bj
Bt
Bj_i
bound
deter-
Upper Bound) Consider any approximation
to the
following theorem relates the worst-case performance of the upper
mined by
Theorem
(8)
and
(9) to
the accuracy of the original approximation,
2 (Tightness of the
American option
price, Vt, satisfying Vt
Ko <
Proof
Vt+i
Bt+i
E(
-Df
by
(
t€T
The
-
-5-
bound corresponding
En
to be defined by
(8)
V„
explicitly given
is
ttj
Vo
•D(+l
Let V'o denote the upper
take
will therefore
K)
+
>
hf.
Vf.
Then
2 Eo
;ii)
Bt
See Appendix A.l.
Theorem
2 suggests
in practice. First,
it
two possible reasons
suggests that the
exercise periods. However, this
Eo
max
\
Tif'
\
is
for
why
the upper
bound may deteriorate
bound may be
linearly in the
limited
number
of
a worst case bound. Indeed, the quantity of interest
——
TIT
Bt
Bt
+
2_]
10
Ej-
V
VJ-1
5,
Bj-i
(12)
and while we would expect
clear that
options
it
should increase
when we
typically
of exercise times in
[0,
The second reason
Van Roy
to increase with the
it
(2000) have
This
linearly.
is
number
of exercise periods,
particularly true for pricing
want to keep the horizon, T,
fixed, while
we
let
it
not
is
American
the
number
T] increase.
is
due
to the
approximation
shown that under
dynamic programming algorithms,
certain conditions,
it
assumes that
—
\Vt
and
and
Tsitsiklis
Vt\/Bt.
for certain
approximate
can be bounded above by a constant,
this error
dependent of the number of exercise periods. This
to perpetual options since
error,
T
result,
however,
is
in-
only applicable
-^ oo while the interval between exercise
times remains constant. As mentioned in the previous paragraph, however, we are typically interested in
decreases.
problems where
T
is
In this case, Tsitsiklis and
fixed
and the
interval
Van Roy show
bounded above by a constant times \/N, where
A'' is
between exercise periods
that the approximation error
the
number
is
of exercise periods.
These two observations suggest that the quality of the upper bound should deteriorate
with
A^,
when we
3.3
but not in a linear fashion. In Section 5 we shall see evidence to support this
successfully price options with as
many
as 100 exercise periods.
The Lower Bound on the Option Price
To construct a lower bound on the
true option price,
we
define the Q-value function to
be
"
Qt(Xt) := Et
The Q-value
at time
t
is
being exercised at time
available, iov
t
=
1, ...
Bt
R
Dt+l
Vt+AXt+i)
(13)
simply the expected value of the option, conditional on
t.
,T —
it
not
Suppose that an approximation to the Q-value, Qt{Xt),
1.
Then, to compute the lower bound on the option
we simulate a number of sample paths originating
11
at A'q. For each
is
price,
sample path, we
find
the
first
exercise period
exists, in
if it
t,
which h(Xt)
>
exercised at this time and the discounted payoff' of the path
this
is
a feasible Tt
-
adapted exercise poHcy,
is
it
The option
Qt{Xt).
is
is
then
given by h{Xt)/ Bt- Since
clear that the expected discounted
payoff from following this policy defines a lower bound, Vq, on the true option price, Vq.
The
T
f = min{f E
Formally,
:
Qt
<
ht}
and
following theorem characterizes the worst-case performance of the lower bound.
Theorem
3 (Tightness of the
Lower Bound) The
lower bound on the option price
satisfies
Vo
>
Vo
-
Eo
y-
\Qt-
'
(14)
Proof See Appendix A. 2.
While
theorem suggests that the performance of the lower bound may deteriorate
this
linearly in the
Theorem
not the case.
in
of exercise periods, numerical experiments indicate that this
exercised,
it is
i.e.,
is
3 describes the worst case performance of the bound. However,
order for the exercise strategy that defines the lower
performance,
it
number
bound
to achieve the worst case
necessary that at each exercise period the option
the condition Qt{Xt)
must be the case that
<
h{Xt)
< Qt{Xt)
is
is
satisfied.
mistakenly
left
un-
For this to happen,
at each exercise period, the underlying state variables are close
to the optimal exercise boundary. In addition, Qt
true value Qt so that the option
is
must systematically overestimate the
not exercised while
it is
optimal to do
so. In practice,
the variability of the underlying state variables, Xt, might suggest that Xt spends
time near the optimal exercise boundary.
This suggests that as long as Qt
approximation to Qt near the optimal exercise
12
frontier, the lower
is
little
a good
bound should be a
good estimate
of the true price, regardless of the
of exercise periods.
Implementation
4
we describe
In this section
that
we use
algorithm we describe
it
initial
now standard
example, Bertsekas and
algorithms,
some
in
obtaining the
for
of this kind are
for
number
is
is
detail
approximate Q-value
approximation to the value function,
in the
Vj.
Algorithms
approximate dynamic programming literature
Tsitsiklis,
An
1996).
that, in contrast to
deterministic.
iteration, the algorithm
(see
interesting feature of the particular
most approximate dynamic programming
This deterministic property
is
achieved through the
use of low discrepancy sequences. While such sequences are used in the
same
spirit as
independent and identically distributed sequences of random numbers, we found that
their application significantly
They
4.1
As
improved the computational efficiency of the algorithm.
are described in further detail in
Appendix B.
Q- Value Iteration
before, the
problem
is
to
compute
Vo
In theory this
problem
is
=
sup Eo
T6T
easily solved using value iteration
where we solve
for the value
functions, Vt, recursively so that
Vt
Vt
The
price of the option
economy.
= HXt)
= max
is
(15)
h{X,),E
(16)
then given by V'o(Xo) where
In practice, however,
if
d
is
large so that
13
Xt
Xq
is
is
the initial state of the
high dimensional, then the
'curse of dimensionality' implies that value iteration
As an
is
not practical.
alternative to value iteration consider again the Q-value function, which was
defined earlier to be the continuation value of the option
=
QtiXt)
E,
Vt+i{Xt+,]
(17)
Bt+i
The
value of the option at time
t
+
Vt+i{Xt+,)
so that
we can
1 is
=
then
miix{h(Xt+i).Qt+AXt+,))
(18)
also write
Bt
msLx{h{Xt+i),Qt+i{Xt+i))
QtiXt
(19)
Bt+i
Equation (19) clearly gives a natural extension of value iteration to Q-value
The algorithm we
iteration.
use in this section consists of performing an approximate Q-value
iteration.
why
it is
preferable to concentrate on approximat-
ing Qt rather than approximating Vt directly.
Letting Qt and Vt denote our estimates
There are a number of reasons
of Qt
and
and value
Vt respectively,
for
we can write the
defining equations for approximate Q-value
iteration as follows:
Bt
QtiXt)
=
Vt{Xt)
= nmx{h(X,),E
max{hiXt+,),Qt+i{Xt+:,
E,
(20)
Bt
(21)
Vt+dXt^i]
Bt+i
The
functional forms of (20) and (21) suggest that Qt
more
easily
is
smoother than
14,
and therefore
approximated.
More importantly, however, Qt
is
the
unknown quantity
of interest,
and the decision
to exercise or continue at a particular point will require a comparison of Qt{Xt)
14
and
h{Xt).
we only have
If
difficult to
However,
if
make. For example,
Vt{Xt)
us then such a comparison will often be very
Vt available to
if
Vt{Xi)
>
h{Xt) then we do not exercise the option.
only marginally greater than h{Xt), then
is
> Qt{Xt) and
ation,
we misinterpret Vi{Xt) and assume that
is
is
the case that
it is
optimal to continue when
in fact
it
optimal to exercise. This problem can be quite severe when there are relatively few
exercise periods because in this case, there
value of exercising
now and
often a significant difference between the
When we
have a direct estimate of
arise.
Approximate Q- Value Iteration
4.2
The
is
the continuation value.
Qt{Xt) available, this problem does not
first
\ Qt{-',
is
may be
actually attempting to approximate h{Xt). In this situ-
h{Xt)
Vt{Xi)
it
step in approximate Q-value iteration
Pt)
:
/3t
e 5R^
[,
which
is
is
to select an approximation architecture,
a class of functions from which
we
select Qt. This class
parametrized by the vector Pt £ 5?^ so that the problem of determining Qt
to the problem of selecting Pt where pt
The
choice of architecture does not
sufficiently 'rich' to accurately
is
is
reduced
chosen to minimize some approximation error.
seem to be particularly important as long
as
it
is
approximate the true value value.
Possible architectures are the linearly parametrized architectures of Longstaff and
Schwartz (2000) and Tsitsiklis and Van Roy (2000), or non-linearly parametrized architectures such as neural networks or support vector machines (see
paper we use a multi-layered perceptron with a single hidden
of neural networks.
Vapnik 1999). In
layer,
a particular class
Multi-layered perceptrons with a single hidden layer are
to possess the universal approximation property so that they are able to
any continuous function over a compact
number
set arbitrarily well,
of neurons are used (see Hornick, Stinchcombe
15
this
known
approximate
provided that a sufficient
and White 1989).
The second
step in the procedure
is
to select for each
t
=
I,
.
.
.
,
K—
I,
a set
Sr.= {Pl.--.Ph,}
St in
do
=
makes sense to choose the
sets
such a way that they are representative of the probabihty distribution of Xt-
We
of training points
this using
points, then
where
G
P,[
n
for
JR*^
1,
.
.
.
low discrepancy sequences so that
we simply take Nt points from a
the technique described in Appendix B,
it is
,
A^j.
if
Nt
It
is
of training
then straightforward to convert these points
Of
course,
also possible to
it is
by simply simulating from the distribution of the state vector,
select the training points
Our Umited experience shows
work very
number
particular low discrepancy sequence. Using
into training points that represent the distribution of Xt-
A'(.
the desired
that both simulation and low discrepancy sequences
The performace
well in practice.
of the low discrepancy sequences, however,
appeared to be marginally superior when applied to the problems we consider in
this
paper.
The
with
t
third step
=
K—
1
is
and
to perform a training point evaluation. Defining
for
n =
Qt{Pl)
1,
.
= Et
.
.
,Nt,
_Bt
-
we use Qt{Pn)
max
1
The operator
This
is
E[
.
]
in (22)
necessary as
it
is
is
Qt =
0,
we begin
to estimate Qt{Pi) where
(h{Xt+i), Qt+i{Xt+i)
(22)
^
intended to approximate the expectation operator, E[
usually not possible to
.
].
compute the expectation exactly on
account of the high dimensionality of the state space. For example, E[
.
],
could corre-
spond to Monte Carlo simulation where we simulate from the distribution oi Xt+i given
that Xt
= PI Then
E[
.
]
is
defined by
M
Bt
m&x(^h{Xt+i),Qt+i{Xt+i))
:=^^^|^5]max(M:i--(),4+i(-i^())
Bt+\
16
(23)
where
M xis are drawn randomly from the conditional distribution of Xt+\-
lem with
this
method
which can be too slow
generate the
5,
x;'s,
with
M was set equal to
that the rate of convergence to the true expectation
is
for
M chosen
in
0(-^)
advance. For the 5-dimensional problems of Section
we
1000. For the 10-dimensional problems,
set
sinuilation, then in order to achieve a
accuracy, a substantially larger value of
we estimate Qt with
Q,(x;
is
prob-
our purposes. Instead, we use a low discrepancy sequence to
Had we used Monte Carlo
Finally,
The
Qt{,
M equal to 2000.
comparable
level of
M would have been required.
where
\t3t)
A) = Y.r,{j)e U(j) + J2 (^'O'-O^IO)
(24)
)
and
K
^=
n=
The parameters
9{.)
fe((j),
rt{j)
_
and
(25)
l
constitute the parameter vector Pt, while
r,(j, /) in (24)
denotes the logistic function so that
e{x)
K
2
&vgmmY,[Qt{K)-Q>{Pn-,Pt))
refers to the
number
of 'neurons',
—^
=
and d
is
.
the dimensionality of the input vector,
While the input vector x may simply correspond to a sample state vector,
to
augment x with certain
state vector.
features, that
From an informational
is,
easily
computed functions
point of view, features add no
the neural network. However, they often
make
it
it is
common
of the current
new information
easier for the neural
x.
network
to
(or other
approximation architecture) to approximate the true value function.
In practice,
we
usually minimize a variant of the quantity in (25) in order to avoid the
difficulties associated
with overfitting. This
is
17
done using cross validation, an approach
that requires the training points to be divided into three sets, namely training, vahdation
and
test sets. Initially, only the training
and validation
sets are used in the minimization
so that the quantity
Y,{Qt{P!)-Qt{P!;Pt)y'
is
minimized where the sum
tion
is
in (26)
is
taken over points in the training
performed using the Levenberg Marquardt method
(see Bertsekas
and
the validation set
Tsitsiklis 1996).
is
also
(26)
for least
squares optimization
At each iteration of the minimization, the error
computed and
as long as overfitting
is
validation error starts increasing at any point then
The algorithm then terminates
number
of iterations,
and
pt
is
if
it is
in
not taking place, then
However,
the validation error should decrease along with the training set error.
place.
The minimiza-
set.
likely that overfitting
is
if
the
taking
the validation error increases for a prespecified
then set equal to the value of pt in the
last iteration of
the minimization before the validation error began to increase.
There
addressed.
is
one further
difficulty
The neural network
with the neural network architecture that needs to be
have
will typically
case that the algorithm will terminate at a local
minimum.
In this case,
it is
a different starting value for
many
minima and
local
minimum
that
is
far
pt.
This
may be
repeated until a satisfactory local
minimum
set.
use this training algorithm for finding Qt-\- For the remaining Q- value functions,
however, the problem
Qt-\-
from the global
necessary to repeat the minimization again, this time using
has been found, where 'satisfactory' refers to performance on the test
We
often the
it is
We
network.
is
now somewhat
can therefore use
In practice, this
quickly and that
simplified since
it is
usually the case that Qt
pt as the initial solution for training the
time
t
—
1
~
neural
means that the other neural networks can be trained very
we only need
to perform the minimization once.
18
It also
means that we
can dispense with the need
for
having a test set
Once Qt has been found, we then
We
have found Qq.
for all
iterate in the
but the terminal neural network.
manner
we
of value iteration until
could then use
= max{h{Xo),Qo{X,
Vo{Xo)
While such an estimate may be quite accurate,
to estimate the value of the option.
of limited value since
(27)
we can say very
little
about the estimation
it is
error. In addition,
we
do not have at hand an exercise strategy that has expected value equal to V'o(Xo) nor
do we know
if
such a strategy even
little
information with
for the price of the
American option
Finally,
exists.
it
provides
regards to hedging the option.
Computing the Upper and Lower Bounds
4.3
In Section 3.1
is
we showed that an upper bound
given by (10).
However,
estimated lower bound.
and so Theorem
in (10)
(Since Vq
we can
>
alternatively set Vq
ho, this
new
Vn
+
max
En
ter
7^ - 1^ + >
\Bt
Bt
^
I
upper bound
K-
Kj-i
B,
Bj-i
cjj'
This then gives a natural decomposition of the upper bound into a
nents.
in
The
first
component
is
is
where V^
%
>
is
the
ht for all
t,
given by
(28)
sum
of
the estimated lower bound, while the second
some sense measures the extent
fails
Vq,
definition satisfies
2 continues to hold.) In this case, the
Vn =
=
two compo-
component
to which the discounted approximate value function
to be a supermartingale.
We
estimate
Vq by simulating sample paths
max
ter
Vt
I
-;:::
T^
I
Bt
Bt
+ V2_^ ^J"
j=i
19
of the state variables, evaluating
(29)
Bj
Bj-i
along each path and taking the average over
all
Evaluating (29) numerically
paths.
is
time consuming since at each point, {t,Xt), on the path, we need to compute
Any unbiased
V,
V,j-i
Bj
Bj-i
(30)
-j-i
noisy estimate of the expectation in (30), however, will result in an upwards
biased estimate of the upper bound, due to the application of the
It
maximum
operator.
therefore important that accurate estimates of the expectation in (30) can be
is
computed. As before, we again use low discrepancy sequences to estimate these highdimensional integrals.
bound,
it
is
we wish
Since
to
compute an accurate estimate
of the
upper
important to use a good stopping criterion to determine the number of
points that will be used to estimate the expectation in (30).
Let E(A^) denote the estimate of (30)
some
fixed value of
M, we
when
A^ low discrepancy points are used. For
then examine E{Mi) for
i
=
1,
.
.
,
.
L and
terminate either
when
|E(Af(z
or
when
i
=
L,
if
+ l))-E(Mz)| <e
the condition in (31)
In the numerical results of Section
and 10-dimensional problems,
L
—
is
5,
(31)
not satsified for any
we
respectively.
set
For
M
all
i
problems, we set
48000. These values typically result in estimates of the upper
experiments where
M
and
e
is
One
of,
e
=
.2
cents and
bound that appear
to
based upon further numerical
were increased and decreased respectively. The results of
these experiments invariably resulted in estimates of the upper
or 2 cents
L.
equal to 2000 and 4000 for the 5
be very accurate for practical purposes. This observation
1
<
and often considerably
bound that were within
closer to, the original estimate.
possible concern about the use of low discrepancy sequences for estimating
20
the upper
bound
might result
many
that even a small bias in the estimate of the expectation in (30)
is
This
exercise periods.
In particular, as the
is
number
not a problem, however, for reasons mentioned before.
of exercise periods increases, the time interval between
exercise periods decreases which
means that the
As a
next period also decreases.
result,
An Automated
variability of the value function in the
the conditional expectation in (30) can be
estimated more accurately for fixed values of
4.4
upper bound when there are
in a considerable bias in the estimate of the
M and
e.
Pricing Algorithm
Perhaps the obvious way to compute the lower and upper bounds
fashion so that after estimating the Q-value functions,
is
a sequential
in
we simulate a number
of sample
paths to compute the lower bound, and simulate another set to compute the upper
bound. One
Vq might be
When
difficulty
with this strategy, however,
significant so that there
this occurs,
ing points or a
more
we
is
is
that the difference between V^ and
a large duality gap.
are forced to re-estimate the Q/s, possibly using
flexible
approximation architecture.
This process
more
train-
may need
be repeated a number of times before we obtain a sufficiently small duality gap.
American options that need
to be priced regularly, this
is
to
For
unlikely to be a problem since
experience should allow us to determine in advance the appropriate parameter settings.
However,
for pricing
more exotic American options, we might need
ad hoc approach. This might be very
method
We
inefficient in practice, so
we now
to use such
an
briefly outline a
to address this problem.
have already mentioned that while the upper bound should work very well in
practice, the lower
Section 5
when we
bound should
still
be superior.
price call options on the geometric
these problems, where
it
is
actually possible to
21
This observation
mean
of a
compute option
is
number
borne out
in
of stocks. For
prices exactly,
we
see
that the lower
bound
is
closer than the
This then suggests that when there
upper bound
is
a large duality gap
is
not sufficiently close to Vq.
given in (28), suggests that
upper bound
upper bound to the true
if
E(_i
The expression
V(/5(— 14-i/5(-i
it
price.
is
usually because the
upper bound, as
for the
tends to be large, then the
not be very tight. Based on this observation, we propose the following
will
modification to the approximate Q-value iteration.
After Qt has been computed, we do not proceed directly to computing Qt-\- Instead,
we simulate a number
compute
E(
threshold,
Vt+\/ Bt+\
e^,
of points from the distribution of
—
V't/Bt
.
If
Xt and
the average value of these quantities
then believing that we have a good estimate of
Qt-i- Otherwise
we re-estimate
each point, we
for
l^,
we proceed
The
We
to estimate
by increasing the number of time
Qt, either
points or by refining the approximation architecture, depending on the
seems more appropriate.
below some
is
then repeat the process until the average
training
t
remedy that
than
is less
e(.
and upper
resulting estimates of the Q-value functions should lead to tight lower
bounds.
A
further advantage of this proposal
computational
effort
termine online how
is
is
many
how much
allows us to determine
training points are needed or
how complex
we can now
de-
the approximation
good estimates of the option
price.
Numerical Results
In this section
of 5
it
required to obtain a good solution. In particular,
architecture needs to be in order to obtain
5
that
we
and 10 stocks
illustrate our
respectively,
results for call options
easier to solve
methodology by pricing
call
and the geometric mean of
options on the
5 stocks.
on the geometric mean of 10 stocks since
than the 5 stock
case.
this
We
do not present
problem
While somewhat counterintuitive,
22
maximum
this
is
is
in fact
explained
by noting that the
volatility of the
geometric
mean
decreases as the
number
of stocks
increases.
In the problems that follow,
we assume that the market has
A'^
traded securities with
price processes given by
dSl
where Z[
Z/
is
is
Each security pays dividends
pij.
interval [0,T].
The
first
T
-
5,)dt
+,a dZl\
(32)
at a continuous rate of
We
(5j.
assume that
and that there are n equally spaced exercise dates
date occurs at
the n"' exercise date occurs at
let r
Sl[{r
a standard Brownian motion and the instantaneous correlation of Z] and
the option expires at time
and
=
=
i
T.
i
=
We
which we
call
is
the P' exercise period and
use k to denote the strike price of the option
be the annual continuously compounded interest
constant though this assumption
in the
rate. It is
assumed that
r
is
easily relaxed.
In order to generate the initial approximation to the option price, certain problem
specific information
extraction,
network.
was
also used.
For example, we have already mentioned feature
where functions of the current state vector are used as inputs
In
all
the problems of this section, we found
stock prices before using
them
it
advantageous to order the
as inputs to the neural network.
In addition,
the current intrinsic value of the option as a feature for options on the
for
we used
maximum,
while
options on the geometric mean, the corresponding European option value was used.
Though
will
for the neural
it is
true in general that the exact
not be available, this
is
European value of a high-dimensional option
not important since
it is
known
(see
Hutchinson
et al (1994)
that European options prices can be quickly and accurately approximated using learning
networks
.
(This observation also suggests another
dynamic programming algorithm. Instead
way
of
implementing the approximate
of using the approximation architecture to
23
estimate the Q- value, we could use
to estimate the early exercise
it
premium, conditional
on not exercising the option at the current exercise period. This conditional early exercise
premium
plus the estimated
European option
Q- value function.)
of the
Another method by which problem
Broadie and Glasserman 1997).
is
more accurate estimate
price might give a
We
specific information
know,
for
was used
European option. Such
easily be incorporated to the
We
by simply redefining the estimated Q-value to be the
maximum
timated value function one time period ahead, and a European option that
bound on the Q-value.
infor-
approximate dynamic programming algorithm.
mation can
this
policy fixing (see
example, that the American option price
greater than or equal to the price of the corresponding
do
is
of the es-
a lower
is
For options on the geometric mean, we again use the corre-
sponding European option value to bound the Q-value from below. For options on the
maximum, we
The
use the European option on the stock that
most
'in
the money'.
approximation to the option price was obtained using 2000 and 2500
initial
training points per period for the 5
assigned 65%,
is
25% and 10%
and 10-dimensional problems,
of the training points for the time
respectively.
T-
I
cent.
and stopped as soon
The remainder
of the time
t
+
these networks,
1
as the
network as the starting point
70%
lower
test set
maximum
was
less
than
of
1
of the neural networks were trained only once, using the solution
for training the
time
t
network.
For
of the training points were assigned to the training set with the
remainder assigned to the validation
The
mean-squared error on the
also
neural network to
the training, validation and test sets in turn. This network was trained a
5 times
We
set.
bound was obtained by simulating the approximate
8 million sample paths.
exercise strategy along
The upper bound was computed using 1,000 sample paths
estimate the expectation in (28).
24
to
We
(7,
on the
Call
5.1
Maximum
assume that there are
=
The
0.2
and
p,j
=
for
5 assets, r
^
i
of 5 Assets
=
T=
0.05,
assumed
All stocks are
j.
results are given in Table
to have the
problems where there are
for
1
=
and k
3 years
We
100.
same
=
let 5,
0.1,
initial price Sq.
50 and 100 time
10, 25,
periods.
It
can be seen that the estimated lower and upper bounds are typically very
thereby providing very accurate estimates of the true price.
duality gap widens with the
number
of exercise periods
it
even for the problems with 100 exercise periods, we can
As we argued
before, this widening of the duality
deterioration of the upper
bound
as the
see further evidence to support this
In Table
allow us to
we
1
is
typically
less.
due to the gradual
We
of exercise periods increases.
will
when we examine options on the geometric mean.
early exercise premia of the
approximately
exercise periods or
duality gap
is
obtain very good estimates.
also report the prices of the corresponding
compute the
the duality gap
number
true that the
is
it
does so quite gradually so that
still
gap
While
close,
1%
American options.
of the early exercise
For problems with as
many
European options which
premium
for
We
see that
options with 25
as 50 or 100 exercise periods, the
between 1% and 3%. Using the midpoint of the lower and upper bounds
is
should therefore enable us to price the early exercise premium of these options to within
1%
or 2%.
lower
a,
bound
is
Call
5.2
We
Even more accurate
usually closer to the true price than the upper bound.
on the Geometric Mean of
assume that there are
=
0.4
price So-
and
The
price estimates could be obtained by noting that the
pij
=
for
5 assets, r
i
^
j.
=
0.03,
T=
1
5
years and k
All stocks are again
results are given in Table
1
for
25
Assets
=
100.
We
let J,
=
assumed to have the same
problems where there are
10, 25,
0.05,
initial
50 and
100 time periods.
For
tlie
American
call
price of the option can be
option on the geometric
mean
computed using a standard binomial
mean
process that describes the evolution of the geometric
We
motion.
mean
therefore report the true price of the
together with our niunerical results in Table
The
results are similar to those in Table
percentage of the early exercise premium,
for the options that start
number
5.3
1,
tree, since
is itself
the stochastic
a geometric Brownian
American options on the geometric
2.
though the duality gap, measured as a
now tends
to be
somewhat wider than before
out of the money. Again, this duality gap increases with the
of exercise periods though at a very gradual pace.
on the Mciximuni of
Call
We make
the same assumptions for the
did for the 5 asset case except
3.
of a collection of stocks the true
Measured
now
10 Assets
call
T =
option on the
year.
I
in absolute terms, the duality
gap
is
The
maximum
of 10 assets as
we
results are displayed in Table
again very small for these problems.
However, measured as a percentage of the early exercise premium, the duality gap,
though
still
quite small, can be as large as 20%.
This could be due to a number of
reasons.
First, the early exercise
overall option vahie
represents a
much
smaller component of the
than before, so that using the duality gap (measured as a percentage
of the early exercise
errors.
premium now
premium)
to
measure performance
is
likely to
accentuate pricing
This problem could possibly be overcome by approximating the early exercise
premium
rather than the continuation value of the option, as mentioned earlier.
The more
likely
reason for the wider duality gap
training points or neurons were used.
the state space,
Indeed, while
we made only a moderate
26
number
of
we doubled the dimensionality
of
is
that an insufficient
increase in the
number
of training points
and did not increase
at all the
number
As a
of neurons.
estimated lower bound for this problem, $15,181,
bound, $15,178,
for the
corresponding 50 period problem,
100 period problem
is
is
it is
inferior.
is
greater than the true option price for the
clear that the initial option price
Since the upper bound
decreased to $15,228.
is
quite sensitive to the
for the
100 period
The
gap
resulting duality
is
upper bound
to $15,196, while the
only 3 cents which, for
all
practical
very tight.
Conclusions
In this paper
options.
we have developed a new method
Our approach
is
tion of the
for pricing
and exercising American
based on approximate dynamic programming using nonlinear
regression to approximate the option price.
on
is
approximation
resolved the 100 period problem using 4,000 training points per period
and 25 neurons, the estimated lower bound increased
6
The
considerably wider than the duality gap for the 50 period problem.
When we
purposes,
$90.
greater than the estimated lower
is
approximation, we are not surprised to see that the duality gap
problem
=
corresponding 100 period problem. Since we know that the true
option price for the 100 period problem
initial
we expect performance
Consider, for example, the 50 exercise period problem with So
to suffer.
for the
result,
American option
Our main
theoretical result
is
a representa-
price as a solution of a dual minimization problem.
this dual characterization of the price function,
Based
we use Monte Carlo simulation
to
construct tight upper and lower bounds on the option price. These bounds do not rely on
a specific approximation algorithm and can be used in conjunction with other methods
for pricing
American options.
We characterize
the theoretical worst-case performance of
the pricing bounds and show that they are very accurate on a set of sample problems
where we price
call
options on the
maximum and
27
the geometric
mean
of a collection
of stocks.
These numerical
results suggest that our pricing
applied to problems of practical interest.
28
method can be
successfully
A
Proofs
Proof of Theorem 2
A.l
Simplifying (8) and
Vo
as
=
V'o
an upper bound on
Vo
=
K)
+
and using Theorem
(9),
+
^;
I
—
price of the
max EEj-.
j=i
<
V^o
<
+
Eo
American option.
At time
Vf,
then have
5,_i
S,_,
B,_i
-j-i
5j
5j-l
-^J
-^J-l
K)+|Vb-K)|+EoJ](E,_i
option price process,
A. 2
B,
B,
We
V,
E
where the second inequaUty
Theorem
(Al)
Ej-
j=i
is
5,
5.
j=i
of
>
Bt
B,
ter
we obtain
+
-^
\Bt
teT
tlie
Eo
max
Eo
1,
Bj-i
Bt-.,
due to the supermartingale property of the discounted
and the
last step follows
from the triangle
inequality.
The
result
2 then follows.
Proof of Theorem 3
t,
the following six mutually exclusive events are possible:
(ii)
Qt<Qt<
We
define
f,
=
ht,
(iii)
min{s €
Qt <
[f,
T]
ht
n
<
T
(iv)
Qt.
:Q~,
Vt
<
=
Qt
<
ht
<
Qt,
(v) ht
<
(i)
Qt
Qt
<
<Qt <
Qt,
ht,
(vi) Ih
<
/ij and
BtEt
For each of the six scenarios above, we establish a relation between the lower bound and
the true option price.
(i),(ii)
The algorithm
for
estimating the lower bound correctly prescribes immediate
29
Qt
<
Qt-
.
=
exercise of the option so that Vt— y^
(ill)
In this case the option
Vt-V<
is
0.
exercised incorrectly.
=
]4
ht
and
=
Qt implying
\Qt-Qt\.
(iv) In this case the option
is
not exercised though
it is
optimal to do
Bt
Vt
= -^Et \VtJ
while
Vt
= ht<Qt+ {Qt-
Bt
Qt)
= -^Et
+[Qt-Q
[Vt+,]
This implies
Vt-V<\Qt-Qt\
(v),(vi) In this case the option
is
+^E,
correctly
Vt-Vt =
left
-^Et
\Vt+x
-
Vt+i]
unexercised so that
\Vt+i
-
Vt+i]
.
Therefore by considering the four possible scenarios, we find that
Vt-Vt<
Iterating
Vt
and using the
\Qt
fact that
-
Qt\
Vr= Vt
+^Ef
\Vt+i
-
Vt+x]
implies the result.
30
.
so.
Therefore
B
Low Discrepancy Sequences
A
low discrepancy sequence
in
some
fixed
unit cube
is
a deterministic sequence of points that
domain. Often, and without
Because the points
[0, 1]''.
in
loss of generality,
we take
this
where
{y,
quence.
=
:i
An
1,
.
,
.
.
A^},
is
evenly dispersed
domain
to be the
a low discrepancy sequence are evenly dispersed,
they are often used to numerically integrate some function /(.) over
/
is
[0, Ij'^so
that
f{x)dx^^^^l^
(Bl)
a set of A^ consecutive terms from the low discrepancy se-
important property that low discrepancy sequences possess
terms are added, the sequence remains evenly dispersed.
is
that as
This property implies that,
in contrast to
other numerical integration schemes, the term
determined
advance and can therefore be chosen according to some termination
terion.
in
Because these sequences are evenly dispersed, their use
often results in a rate of convergence that
where the convergence rate
is
is
much
faster
0(-^). For the technical
new
A'^
in
in
(Bl) need not be
cri-
numerical integration
than Monte Carlo simulation
definition of a low discrepancy
sequence and a more detailed introduction to their properties and financial applications,
see Boyle, Broadie
and Glasserman (1997). See Birge (1994),
and Paskov and Traub (1995)
In this paper
for
some
and Tan (1996)
of these applications.
we use low discrepancy sequences
ing point evaluation.
Joy, Boyle
for training point selection
The low discrepancy sequences
and
train-
are of particular value for training
point evaluation since a good estimate of
_Bt
E,
^max(M.Yt+i),Q,4-i(A-,+i))
B
can usually be computed much faster by using a low discrepancy sequence
Monte Carlo
simulation.
31
in place of
Even though a low discrepancy point y G
interpret
it is
it
as being
[0, 1]"^ is
sampled from a uniform distribution
then straightforward to convert y into a point,
variable
Z
deterministic
in
[0, 1]''.
it
can be useful to
With
this in
x, that is representative of
with cumulative distribution function F(.).
mind,
a random
For example, suppose
Z
is
a
d-dimensional standard normal random variable with correlation matrix equal to the
identity.
We
can then construct a point
2
a multivariate normal
is
5R'^
that
is
representative of
by setting
(B2)
taken componentwise in (B2).
random
Z
= F-\y)
z
where the operation F~^
6
variable,
X
If
instead
we wish
to 'simulate'
with covariance matrix E, then we simply
premultiply z by the Cholesky decomposition of S.
In financial applications
tributed.
able
is
random
variables are often
Since transforming a normal
easy, however,
we can do
random
likewise with
assumed
to be lognormally dis-
variable into a lognormal
z.
We
we use
in this
paper are Sobol sequences.
32
vari-
can therefore easily convert a
d-dimensional low discrepancy sequence into a sequence of points that
of a d-dimensional multivariate log-normal distribution.
random
is
representative
The low discrepancy sequences
References
Andersen, L. and M. Bioadie, 2001, A Primal-Dial Simulation Algorithm for Pricing
Multi-Dimensional American Options, Working Paper, Columbia University.
J. and D. Martineau, 1995, Numerical Valuation of High-Dimensional
Multivariate American Securities, Journal of Financial and Quantitative Analysis
30, 383-405.
Barraquand,
Bensoussan, A., 1984,
icae 2, 139-158.
Bertsekas, D. and
Belmont,
On
the Theory of Option Pricing, Acta
J. Tsitsiklis,
AppHcandae Mathemat-
1996, Ncilto- Dynamic Programming,
Athena
Scientific,
MA.
Billingsley, P., 1995, Probability
and Measure, Third Edition, John Wiley
&
Sons.
J. R., 1994, Quasi-Monte Carlo Approaches to Option Pricing, Technical Report 94-19. Department of Industrial and Operations Engineering, University of
Michigan, Ann Arbor.
Birge,
M. Broadie, and P. Glasserman, 1997, Monte Carlo Methods
Economic Dynamics and Control 21, 1267-1321.
Boyle, P.,
for Security
Pricing. Journal of
Boyle,
P.,
A.W. Kolkiewicz and
K.S. Tan, 2001, Pricing American Style Options Using
Paper, University of Waterloo, Water-
Low Discrepancy Mesh Methods, Working
loo.
Broadie, M., P. Glasserman, 1997a, Pricing American-style Securities Using Simulation.
Journal of Economic Dynamics and Control 21, 1323-1352.
Broadie, M., P. Glasserman, 1997b, A Stochastic Mesh Method for Pricing HighDimensional American Options, Working Paper, Columbia Business School, Columbia
University, New York.
Broadie, M., P. Glasserman, and G. Jain, 1997c, Enhanced Monte Carlo Estimates for
American Option Prices, Journal of Derivatives, 5, 25-44.
J., 1996, Valuation of Early-Exercise Price of Options using Simulations and
Npnparametric Regression, Insurance: Mathematics and Economics 19, 19-30.
Carriere,
Duffie,
Darrell,
Princeton,
1996.
New
Dynamic Asset Pricing Theory, Princeton
University Press,
Jersey.
Garcia, D., 1999, A Monte Carlo Method for Pricing American Options, Working
Paper, University of California at Berkeley.
Garcia, R.,
J.
Detemple and M. Rindisbacher, 2000,
Working Paper.
A Monte-Carlo Method for Optimal
Portfolios,
Geske, R. and H.E. Johnson, 1984, The American Put Options Valued Analytically,
Journal of Finance 39, 1511-1524.
Goonatilake, S. and P. Treleaven, Eds., 1995, Intelligent Systems for Finance and
Business, John Wiley & Sons.
33
J. M. and D. Kreps, 1979, Martingales and Arbitrage in Multiperiod SecuriMarkets, Journal of Economic Theory 20, 381-408.
Harrison,
ties
J. M. and S. R. Pliska 1981, Martingales and Stochastic Integrals in the
Theory of Continuous Trading, Stochastic Processes and their Applications 11,
Harrison,
215-260.
Hornick, K., M. Stinchcombe and H. White, 1989, Multilayer feedforward Networks
are Universal Approximators, Neural Networks 2, 359-366.
A Nonparametric Approach to the Pricing
Via Learning Networks, Journal of Finance
Hutchinson, A., A. Lo and T. Poggio, 1994,
and Hedging
of Derivative Securities
49, 851-889.
Ibanez, A. and F. Zapatero, 1998, Monte Carlo Valuation of American Options through
Computation of the Optimal Exercise Frontier, Institute Tecnologico Autonomo
de Mexico
.
Joy, C, P.P. Boyle, and K.S. Tan, 1996, Quasi-Monte Carlo
Finance, Management Science 42, 926-938.
Methods
in
Numerical
Karatzas, I, 1998, On the Pricing of American Options, Applied Mathematics and
Optimization 17, 37-60.
Karatzas,
New
and
I.
S.
Shreve, 1998, Methods of Mathematical Finance, Springer- Verlag
York.
and E. Schwartz, Valuing American Options by Simulation:
Least-Squares Approach, Review of Financial Studies 14, 113-148.
Longstaff, F.
Niederreiter, H., 1992,
CBMS-NSF
Paskov,
S.
and
Portfolio
Pliska, S.,
63,
J.
Random Number
SIAM,
A
Generation and Quasi-Monte Carlo Methods.
Philadelphia, PA.
Traub, 1995, Faster Valuation of Financial Derivatives, Journal of
22, Fall, 113-120.
Management
1997, Introduction to Mathematical Finance, Blackwell PubUshers
Maiden,
Simple
Inc.,
MA.
W.H., S.A. Teukolsky, S.A. Vetterhng and W.T. Flannery, 1992, Numerical
2nd Ed, Cambridge University
Recipes in C: The Art of Scientific Computing
Press,
,
Press, Cambridge.
Raymar, S. and M. Zwecher, 1997, A Monte Carlo Valuation of American Call Options
on the Maximum of Several Stocks, Journal of Derivatives 5, 7-23.
Rogers, L.C.G., 2001, Monte Carlo Valuation of American Options, Working Paper,
University of Bath.
S., 1995, Uniform
Publishers, Boston.
Tezuka,
Tilley, J. A., 1993,
Random Numbers: Theory and
Valuing American Options
in a
Practice,
Kluwer Academic
Path Simulation Model, Transactions
of the Society of Actuaries 45, 83-104.
and B. Van Roy, 1999, Optimal Stopping of Markov Processes; Hilbert
Space Theory, Approximation Algorithms, and an Application to Pricing Financial
Derivatives, IEEE Transactions on Automatic Control 44, 1840-1851.
Tsitsiklis, J.,
34
and B. Van Roy, 2000, Regression Methods for Pricing Complex AmericanStyle Options, Working Paper, Laboratory for Information and Decision Systems,
Tsitsiklis, J.,
MIT, Cambridge.
Vapnik, V.N., 1999, The Nature of Statistical Learning Theory (Statistics for Engineering and Information Science), 2nd Ed, Springer Verlag.
35
Table
Table
We
1
1:
Call on the
mciximum
contains estimates of the price of an American
use the following set of parameter values: r
=
call
0.05,
of 5 assets
option on the
T —
3,
k
—
maximum
100,
6^
—
of 5 assets.
0.1, a^
—
0.2
and pij =
for i,j = 1, ...,5. All stocks are assumed to have the same initial price Sq. The
columns "Lower Bound" and "Upper Bound" contain estimates of the lower and upper bounds,
respectively.
for
and
The standard
problems with
10, 25,
errors of these estimates are given in brackets. Results are displayed
50 and 100 time periods.
We
report estimated options prices in $'s
their standard errors in cents.
So
Lower Bound
Upper Bound
10 exercise periods
90
100
110
90
100
no
90
100
no
90
100
no
16.640
European Price
Table
2:
Call on the geometric
Table 2 contains estimates of the price of an American
5 assets.
We
mean
call
use the following set of parameter values: r
—
of 5 assets
option on the geometric
0.03,
T=
1,
k
=
100,
6i
mean
—
of
0.05,
= 0.4 and pij — for i,j — 1, ..., 5. All stocks are assumed to have the same initial price Sq.
The columns "Lower Bound" and "Upper Bound" contain estimates of the lower and upper
bounds, respectively. The standard errors of these estimates are given in brackets. Results
CT(
are displayed for problems with 10, 25, 50
prices in $'s
and
Sq
and 100 time periods.
Lower Boundd
Upper Bound
True Price
10 exercise periods
90
We
report estimated options
their standard errors in cents.
European Price
Table
3:
Call on the
maximum
Table 3 contains estimates of the price of an American
assets.
We
use the following set of parameter values: r
of 10 assets
call
=
option on the
0.05,
T =
I,
k
maximum
—
100,
Sj,
of 10
=
0.1,
and pij = for i, j = 1, ..., 5. All stocks are assumed to have the same initial price SoThe columns "Lower Bound" and "Upper Bound" contain estimates of the lower and upper
ai
=
0.2
bounds, respectively.
ai'e
The standard
errors of these estimates are given in brackets.
displayed for problems with 10, 25, 50 and 100 time periods.
prices in $'s
and
We
their standard errors in cents.
5o
Luwer Bound
Upper Bound
European Price
10 exercise periods
90
100
110
15.082
15.092
(0.40)
(0.37)
26.860
26.863
(0.46)
0.60)
39.076
39.076
(0.51)
(0.48)
14.747
26.403
38.522
25 exercise periods
90
100
110
15.158
15.172
(0.39)
(0.51)
26.960
26.970
(0.45)
(0.74)
39.182
39.193
(0.50)
(0.64)
14.747
26.403
38.522
50 exercise periods
90
100
110
15.181
15.216
(0.38)
(0.79)
26.978
27.092
(0.45)
(3.46)
39.225
39.261
(0.49)
(0.96)
14.747
26.403
38.522
100 exercise periods
90
100
110
15.178
15.283
(0.39)
(1.36)
26.996
27.070
(0.45)
(1.17)
39.223
39.306
Results
report estimated options
14.747
26.403
38.522
3
Date Due
3 9080 02246 0726
Download