Faculty of Sciences and Mathematics, University of Niš, Serbia

advertisement
Faculty of Sciences and Mathematics, University of Niš, Serbia
Available at: http://www.pmf.ni.ac.yu/filomat
Filomat 21:2 (2007), 285–290 _____________________________________
A TRUST REGION ALGORITHM FOR LC1 UNCONSTRAINED
OPTIMIZATION
Nada Đuranović-Miličić
Abstract: In this paper a trust region algorithm for minimization of locally
Lipschitzian functions, which uses the second order Dini upper directional derivative is
considered. A convergence proof is given, as well as an estimate of the rate of convergence.
Keywords and phrases: Directional derivative, second order Dini upper directional
derivative, uniformly convex functions, trust region algorithm.
2000 Mathematics Subject Classification: 65K05, 90C30.
Received: November 21, 2006.
1. INTRODUCTION
We shall consider the following LC1 problem of unconstrained optimization
min  f ( x) | x  D  R n  ,
(1)
where f : D  R n  R is a LC1 function on the open convex set D , that means the
objective function we want to minimize is continuously differentiable and its gradient
f is locally Lipschitzian, i.e.
f ( x)  f ( y)  L x  y
for x, y  D
for some L  0 .
We shall present an iterative algorithm which is based on the results from [2] and [3]
for finding an optimal solution to problem (1) generating the sequence of points {xk } of
the
following
form:
xk 1  xk  dk , k  0,1,...,
dk  0,
where the directional vector dk is defined by the particular algorithm.
285
(2)
The algorithm which we are going to present is a trust region algorithm, which uses
the second order Dini upper directional derivative instead of the Hessian in the
quadratic approximatiom model.
The convergence of trust region algorithms has been usually proved for C 2 functions,
i.e. for twice continuously differentiable functions ( see for example [2]). The aim of
this paper is to establish convergence under weaker assumptions, if f  LC1 .
2. PRELIMINARIES
We shall give some preliminaries that will be used for the remainder of the paper.
Definition (see [3]) The second order Dini upper directional derivative of the function
n
n
f  LC 1 at xk  R in the direction d  R is defined to be
f ( xk   d )  f ( xk )
f ( x ; d )  lim sup
T
D
k
d

 0
If f is directionally differentiable at xk , we have
f  xk   d   f  xk   d
f  xk ; d   f  xk ; d   lim 
T
"
D
"

 0
for all d  R n .
Lemma 1 (See [3]) Let f : D  R n  R be a LC1 function on D , where D  R n is an
open subset. If x is a solution of LC1 optimization problem (1), then:
f ( x; d )  0
and f D( x; d )  0, d  R n .
Lemma 2 (See [3]) Let f : D  R n  R be a LC1 function on D , where D  R n is an
open subset. If x satisfies
f ( x; d )  0
and f D( x; d )  0, d  0, d  R n , then x is a strict local minimizer of (1).
3. THE OPTIMIZATION ALGORITHM
286
At the k-th iteration we want to find a solution dk of the following direction finding
subproblem
min  k  d   f  xk  d 
T
1 "
f D  xk , d 
2
(3)
s.t. d   k
for some bound  k  0 , where the norm  is arbitrary, but is usually chosen as the l2 norm. We choose the radius  k to be as large as possible depending on the agreement
between  k  dk  and f  xk  dk  . We can measure this agreement by comparing the
actual reduction in f on the k-th step
Aredk  f k  f  xk   f  xk  dk 
and the corresponding predicted reduction
Pr edk  k  k  0  k  dk  .
Hence, if the ratio k 
Ared k
is closer to unity, the better is the agreement beteween
Pr ed k
the quadratic model k  d  and f  xk  dk  .
The trust region algorithm for LC1 optimization
Step 0. Given x0  D,   0,  0  0, 0  1   2  1, k=0;
Step 1. If f  xk    , STOP; otherwise find a solution dk to the subproblem (3);
Step 2. Compute f  xk  dk  and  k ;
Step 3.  k 1

1 d k if  k  1 ;

 2 k if  k   2 and d k   k ;

 k , otherwise.

Step 4. If k  0, set xk 1  xk ; otherwise xk 1  xk  dk ;
Step 5. set k  1  k , go to step 1.
Proposition. If the function f  LC1 satisfies the condition
287
c1 d
 f D( x; d )  c2 d
2
2
(4)
then: 1) the function f is uniformly and, hence, strictly convex, and, consequently; 2)
the level set L( x0 )  x  D : f ( x)  f ( x0 ) is a compact convex set for any x0  D ; 3)
there exists a unique point x* such that f ( x* )  min f ( x) .
xL ( x0 )
.
Proof. See [1].
Lemma 3.
Pr ed k   k  0    k  d k  
f  xk 
2

f  xk 

min  k ,
c2




.


Proof is analogous to the proof of Lemma 5.2 in [3].
Convergence theorem. If the function f  LC1 satisfies the condition (4) and if any
level set
x  D | f  x   f  x 
is bounded, then for any initial point xo  D,
k
k  1, inf  k  0 and xk  x as k   , where x is a unique minimal point.
Proof. The algorithm generates a subsequence xk  , k  K for which either
a) k  1 and  k 1  0 ( and hence  k  0 ), or
b) k  1 and inf  k  0 .
From the boundness of level sets
x  D | f  x   f  x 
k
it follows that
xk  , k  K
belongs to a bounded set. Hence there exist accumulation points of xk  , k  K . Let x
be any accumulation points of such a subsequence. We only have to prove that
f  x   0 since, according to the Proposition, the stationary point is a unique
minimizer.
We shall show that x can not arise from case a) but only from case b). Namely, since
f  LC 1 from (3) it follows that there exists a sequence  k  (see [3]) converging to
zero such that
 f x   f x
k
k

 dk    k  0  k  dk     k  k , i.e.
Aredk  Pr edk   k  k .
(5)
Dividing (5) by Pr ed k and taking the limit as k  , k  K gives k  1. Hence
k  1 and inf  k  0 ; so x can arise only from case b). In case b)
288
f  x1   f  x    f k
implies
kK
f k  0 since
f  x1   f  x 
is constant.
Hence
 k  0 since k  1 .
From Lemma3 it follows that
Pr ed k   k  0    k  d k  
f  xk 
2

f  xk 

min  k ,
c2




.


Since k  0 and inf  k  0 , it follows that f  xk   0. From continuity of f we
have that f  x   0. Following Theorem 4.4 from [3] it can be analogously proved that
xk 
converges Q-superlinearly to x .
4. CONCLUSION
In this paper we presented a trust region algorithm which is based on the results from
[2] and [3]. We proved the convergence of the trust region method for LC1 class of
functions using the second order Dini upper directional derivative instead of the
Hessian in the quadratic approximatiom model, аs it has been done in [2]. In [2] the
convergence proof of the trust region algorithm is given for f  C 2 , hence under
stronger assumptions. In [3] a trust region algorithm is also defined for LC1 functions.
It uses the exact solution dk and an inexact solution d k of the subproblem (3) and it
requires that the decrease in  k d k  must be at least a fraction of the optimal decrease in
 k d k  . In [3] the convergence proof is given under a little stronger assumptions than
it has been done in this paper.
5. ACKNOWLEDGMENT
This research was supported by Science Fund of Serbia, grant number 144015G ,
through
Institute
of
Mathematics,
SANU.
REFERENCES
[1] Djuranovic-Milicic, N., On minimization of locally Lipschitzian functions,
Proceedings of SYM-OP-IS 2003, H.Novi, 347-350, 2003.
[2] Fletcher, R., Practical methods of optimization, Vol.1, J. Wiley, New York,1980.
[3] Sun, V.J., Sampaio, R.J.B., and Yuan, Y.J., Two algorithms for LC1 unconstrained
optimization, Journal of Computational Mathematics 6 , 621-632, 2000.
289
[4] Vujčić, V., Ašić, M., and Miličić, N., Mathematical programming (in Serbian),
Institute of Mathematics, Belgrade, 1980.
Department of Mathematics, Faculty of Technology and Metallurgy, University of
Belgrade, Belgrade, Serbia
290
Download