Research Article Global Convergence of a New Nonmonotone Filter Method for

advertisement
Hindawi Publishing Corporation
Journal of Applied Mathematics
Volume 2013, Article ID 976509, 9 pages
http://dx.doi.org/10.1155/2013/976509
Research Article
Global Convergence of a New Nonmonotone Filter Method for
Equality Constrained Optimization
Ke Su, Wei Liu, and Xiaoli Lu
College of Mathematics and Computer Science, Hebei University, Baoding 071002, China
Correspondence should be addressed to Ke Su; pigeonsk@163.com
Received 6 December 2012; Accepted 5 March 2013
Academic Editor: Hadi Nasseri
Copyright © 2013 Ke Su et al. This is an open access article distributed under the Creative Commons Attribution License, which
permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
A new nonmonotone filter trust region method is introduced for solving optimization problems with equality constraints. This
method directly uses the dominated area of the filter as an acceptability criterion for trial points and allows the dominated area
decreasing nonmonotonically. Compared with the filter-type method, our method has more flexible criteria and can avoid Maratos
effect in a certain degree. Under reasonable assumptions, we prove that the given algorithm is globally convergent to a first order
stationary point for all possible choices of the starting point. Numerical tests are presented to show the effectiveness of the proposed
algorithm.
1. Introduction
We analyze an algorithm for solving optimization problems
where a smooth objective function is to be minimized subject
to smooth nonlinear equality constraints. More formally, we
consider the problem,
min
𝑓 (π‘₯) ,
s.t. 𝑐𝑖 (π‘₯) = 0,
𝑖 ∈ 𝐼 = {1, 2, . . . , π‘š} ,
(𝑃)
where π‘₯ ∈ 𝑅𝑛 , the functions 𝑓 : 𝑅𝑛 → 𝑅 and 𝑐𝑖 (𝑖 ∈ 𝐼) : 𝑅𝑛 →
𝑅 are all twice continuously differentiable. For convenience,
let 𝑔(π‘₯) = ∇𝑓(π‘₯), 𝑐(π‘₯) = (𝑐1 (π‘₯), 𝑐2 (π‘₯), . . . , π‘π‘š (π‘₯))𝑇 and
𝐴(π‘₯) = (∇𝑐1 (π‘₯), ∇𝑐2 (π‘₯), . . . , ∇π‘π‘š (π‘₯)), and π‘“π‘˜ refers to 𝑓(π‘₯π‘˜ ),
π‘π‘˜ to 𝑐(π‘₯π‘˜ ), π‘”π‘˜ to 𝑔(π‘₯π‘˜ ) and 𝐴 π‘˜ to 𝐴(π‘₯π‘˜ ), and so forth.
There are many trust region methods for equality constrained nonlinear programming (𝑃), for example, Byrd et
al. [1], Dennis Jr. et al. [2] and Powell and Yuan [3], but
in these works, a penalty or augmented Lagrange function is always used to test the acceptability of the iterates.
However, there are several difficulties associated with the
use of penalty function, and in particular the choice of the
penalty parameter. Hence, in 2002, Fletcher and Leyffer [4]
proposed a class of filter method, which does not require
any penalty parameter and has promising numerical results.
Consequently, filter technique has been employed to many
approaches, for instance, SLP methods [5], SQP methods [6–
8], interior point approaches [9], bundle techniques [10], and
so on.
Filter technique, in fact, exhibits a certain degree of nonmonotonicity. The nonmonotone technique was proposed by
Grippo et al. in 1986 [11] and combined with many other
methods. M. Ulbrich and S. Ulbrich [12] proposed a class
of penalty-function-free nonmonotone trust region methods
for nonlinear equality constrained optimization without filter
technique. Su and Pu [13] introduced a nonmonotone trust
region method which used the nonmonotone technique in
the traditional filter criteria. Su and Yu [14] presented a
nonmonotone method without penalty function or filter.
Gould and Toint [15] directly used the dominated area of the
filter as an acceptability criteria for trial points and obtained
the global convergence properties. We refer the reader [16–18]
for some works about this issue.
Motivated by the ideas and methods above, we propose
a modified nonmonotone filter trust region method for
solving problem (𝑃). Similar to the Byrd-Omojokun class of
algorithms, each step is decomposed into the sum of two
distinct components, a quasi-normal step and a tangential
step. The main contribution of our paper is to employ the
nonmonotone idea to the dominated area of the filter so
that the new and more flexible criteria is given, which is
different from that of Gould and Toint [15] and Su and Pu [13].
2
Journal of Applied Mathematics
Under usual assumptions, we prove that the given algorithm
is globally convergent to first order stationary points.
This paper is organized as follows. In Section 2, we
introduce the fraction of Cauchy decrease and the composite
SQP step. The new nonmonotone filter technique is given in
Section 3. In Section 4, we propose the new nonmonotone
filter method and present the global convergence properties
in Section 5. Some numerical results are reported in the last
section.
π‘‘π‘˜π‘› is the solution to the subproblem
min
s.t.
1 σ΅„©σ΅„©
σ΅„©2
󡄩󡄩𝑐 + π΄π‘‡π‘˜ 𝑑𝑛 σ΅„©σ΅„©σ΅„© +
σ΅„©
2σ΅„© π‘˜
σ΅„©σ΅„© 𝑛 σ΅„©σ΅„©
󡄩󡄩𝑑 σ΅„©σ΅„© ≤ Δ π‘˜ ,
min
s.t.
Consider the following unconstraint minimization optimization problem:
(1)
π‘₯∈𝑅
𝑛
where 𝑓 : 𝑅 → 𝑅 is a continuously differentiable function.
A trust region algorithm for solving the above problem is an
iterate procedure that computes a trial step as an approximate
solution to the following subproblems:
min
s.t.
1
π‘ž (𝑑) = 𝑔𝑇 𝑑 + 𝑑𝑇 𝐻𝑑,
2
where 𝐻 is the Hessian matrix ∇2 𝑓(π‘₯) or an approximate to
it and Δ > 0 is a given trust region radius.
To assure the global convergence, the step is required only
to satisfy a fraction of Cauchy decrease condition. This means
that 𝑑 must predict via the quadratic model function π‘ž(𝑑)
at least as much as a fraction of the decreased given by the
Cauchy step on π‘ž(𝑑); that is, there exists a constant 𝜎 > 0
fixed across all iterations, such that
𝑐𝑝
π‘ž (0) − π‘ž (𝑑) ≥ 𝜎 (π‘ž (0) − π‘ž (𝑑 )) ,
(3)
where 𝑑𝑐𝑝 is the steepest descent step for π‘ž(𝑑) inside the trust
region.
Lemma 1. If the trial step 𝑑 satisfies a fraction of Cauchy
decrease condition, then
π‘ž (0) − π‘ž (𝑑) ≥
σ΅„©
σ΅„©σ΅„©
σ΅„©∇𝑓 (π‘₯)σ΅„©σ΅„©σ΅„©
𝜎 σ΅„©σ΅„©
σ΅„©
}.
σ΅„©σ΅„©∇𝑓 (π‘₯)σ΅„©σ΅„©σ΅„© min {Δ, σ΅„©
2
‖𝐻‖
(4)
Proof (see Powell [19] for the proof). Now, we turn to explain
the composite SQP step. Given an approximate estimate of
the solution π‘₯π‘˜ at π‘˜th iteration, following Dennis Jr. et al. [2]
and M. Ulbrich and S. Ulbrich [12], we obtain the trial step
π‘‘π‘˜ = π‘‘π‘˜π‘› + π‘‘π‘˜π‘‘ by computing a quasi-normal step π‘‘π‘˜π‘› and a
tangential step π‘‘π‘˜π‘‘ . The purpose of the quasi-normal step π‘‘π‘˜π‘›
is to improve feasibility. To improve optimality, we seek π‘‘π‘˜π‘‘
in the tangential space of the linearized constraints in such a
way that it provides sufficient decrease for a quadratic model
of the objective function 𝑓(π‘₯). Let π‘žπ‘˜ (𝑑) = π‘”π‘˜π‘‡ 𝑑+(1/2)𝑑𝑇 π»π‘˜ 𝑑,
where π»π‘˜ is a symmetric approximation of ∇2 𝑓(π‘₯).
π‘žπ‘˜ (π‘‘π‘˜π‘› + 𝑑𝑑 ) ,
π΄π‘‡π‘˜ 𝑑𝑑 = 0
(6)
σ΅„©σ΅„© 𝑑 σ΅„©σ΅„©
󡄩󡄩𝑑 σ΅„©σ΅„© ≤ Δ π‘˜ .
σ΅„© σ΅„©
Then we get the current trial step π‘‘π‘˜ = π‘‘π‘˜π‘› + π‘‘π‘˜π‘‘ . Let π‘‘π‘˜π‘‘ =
𝑑
𝑑
π‘Šπ‘˜ π‘‘π‘˜ , where π‘‘π‘˜ ∈ 𝑅𝑛−π‘š and π‘Šπ‘˜ ∈ 𝑅𝑛×(𝑛−π‘š) denote a matrix
whose columns form a basis of the null space of π΄π‘‡π‘˜ . We refer
to [2] for a more detailed discussion of this issue.
In usual way that impose a trust region in stepdecomposition methods, the quasi-normal step π‘‘π‘˜π‘› and the
tangential step π‘‘π‘˜π‘‘ are required to satisfy
σ΅„©σ΅„© 𝑛 σ΅„©σ΅„©
󡄩󡄩𝑑 σ΅„©σ΅„© ≤ πœ…Δ π‘˜ ,
(2)
‖𝑑‖ ≤ Δ,
(5)
where Δ π‘˜ is a trust region radius and 𝐴 π‘˜ = ∇𝑐(π‘₯π‘˜ ) ∈
𝑅𝑛×π‘š , πœ‰ > 0. In order to improve the value of the objective
function, we solve the following subproblem to get π‘‘π‘˜π‘‘ :
2. The Fraction of Cauchy Decrease and
the Composite SQP Step
min𝑛 𝑓 (π‘₯) ,
πœ‰ σ΅„©σ΅„© 𝑛 σ΅„©σ΅„©2
󡄩𝑑 σ΅„© ,
2σ΅„© σ΅„©
σ΅„©
σ΅„©σ΅„© 𝑛
σ΅„©σ΅„©π‘‘π‘˜ + 𝑑𝑑 σ΅„©σ΅„©σ΅„© ≤ Δ π‘˜ ,
σ΅„©
σ΅„©
(7)
where 0 < πœ… < 1. Here, to simplify the proof, we only impose
a trust region on β€– 𝑑𝑛 β€–≤ Δ π‘˜ and β€– 𝑑𝑑 β€–≤ Δ π‘˜ , which is natural.
Note that π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (𝑑𝑑 ) is the reduced gradient of π‘žπ‘˜ in
terms of the representation 𝑑𝑑 = π‘Šπ‘˜ 𝑠 of the tangential step:
∇𝑠 (π‘žπ‘˜ (π‘Šπ‘˜ 𝑠)) = π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (π‘Šπ‘˜ 𝑠) = π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (𝑑𝑑 ) .
(8)
Define
𝑔̂ (π‘₯) = π‘Š(π‘₯)𝑇 𝑔 (π‘₯) .
(9)
Then the first order necessary optimality conditions (KarushKuhn-Tucker or KKT conditions) at a local solution π‘₯ ∈ 𝑅𝑛
of problem (𝑃) can be written as
𝑐 (π‘₯) = 0,
𝑔̂ (π‘₯) = 0.
(10)
3. A New Nonmonotone Filter Technique
In filter method, originally proposed by Fletcher and Leyffer
[4], the acceptability of iterates is determined by comparing
the value of constraint violation and the objective function
with previous iterates collected in a filter. Define the violation
function β„Ž(π‘₯) by β„Ž(π‘₯) =β€– 𝑐(π‘₯)β€–22 , it is easy to see that β„Ž(π‘₯) = 0
if and only if π‘₯ is a feasible point, so a trial point should reduce
either the value of constraint violation or that of the objective
function 𝑓.
In the process of the algorithm, we need to decide whether
the trial point π‘₯π‘˜+ is any better than π‘₯π‘˜ as an approximate
solution to the problem (𝑃). If we decide that this is the case,
we say that the iteration π‘˜ is successful and choose π‘₯π‘˜+ as
Journal of Applied Mathematics
3
𝑓(π‘₯)
𝑓(π‘₯)
π‘π‘Š(β„±π‘˜ )
π‘˜β„±
β„±π‘˜
+ π‘˜β„±
𝑓max
β„±π‘˜
𝑓max
π’Ÿ(β„±π‘˜ )
β„±
β„±π‘˜
𝑓max
β„±
π‘˜
β„Žmin
π‘˜
β„Žmax
(β„Žπ‘˜+ , π‘“π‘˜+ )
(β„Žπ‘˜+ , π‘“π‘˜+ )
π’«π‘˜
𝑓max
β„Ž (π‘₯)
(β„Žπ‘˜+ , π‘“π‘˜+ )
π‘†π‘Š(β„±π‘˜ )
β„±
β„±
π‘˜
β„Žmin
β„±π‘˜
𝑓min
β„±π‘˜
β„Žmax
+ π‘˜β„±
π‘˜
β„Žmax
π’«π‘˜
𝑓min
𝑆𝐸(β„±π‘˜ )
β„Ž (π‘₯)
π‘˜β„±
β„±π‘˜
𝑓min
Figure 1
(β„Žπ‘˜+ , π‘“π‘˜+ )
the next iterate. Let us denote by S the set of all successful
iterations, that is,
󡄨
(11)
S = {π‘˜ 󡄨󡄨󡄨π‘₯π‘˜+1 = π‘₯π‘˜+ } .
In traditional filter method, a point π‘₯ is called acceptable
to the filter if and only if
β„Ž (π‘₯) ≤ π›½β„Žπ‘— or 𝑓 (π‘₯) ≤ 𝑓𝑗 − π›Ύβ„Žπ‘— ,
∀ (β„Žπ‘— , 𝑓𝑗 ) ∈ F,
where 0 < 𝛾 < 𝛽 < 1, F denotes the filter set. Define
󡄨
D (F) = {(β„Ž, 𝑓) σ΅„¨σ΅„¨σ΅„¨σ΅„¨β„Ž > β„Žπ‘— and 𝑓 > 𝑓𝑗 , ∃𝑗 ∈ F } .
F
Fπ‘˜
β„Žmax
= max β„Žπ‘— ,
F def
𝑓minπ‘˜ =
Fπ‘˜ def
𝑓max
=
def
𝑗∈Fπ‘˜
min 𝑓𝑗 ,
𝑗∈Fπ‘˜
Consider the trial point π‘₯π‘˜+ , if the filter is empty, then define
its contribution to the area of the filter by
𝛼 (π‘₯π‘˜+ , Fπ‘˜ ) = πœ…πΉ 2 ,
def
(12)
(13)
A trial point π‘₯π‘˜+ is accepted if and only if (β„Žπ‘˜+ , π‘“π‘˜+ ) ∉ D(Fπ‘˜ ).
Now, similar to the idea of Gould and Toint [15], we give a
new modified nonmonotone filter technique. For any (β„Ž, 𝑓)pair, define an area that represents its contribution to the area
of D(F), we hope this contribution is positive; that is, the
area of D(F) is increasing. For convenience, we partition
the right half-plane [0, +∞] × [−∞, +∞] into four different
regions (see Figure 1). Define D(Fπ‘˜ )𝑐 to be the complement
of D(Fπ‘˜ ). Let
π‘˜
= min β„Žπ‘— ,
β„Žmin
Figure 2
where πœ…πΉ > 0 is a constant. If the filter is not empty, then
define the contribution of π‘₯π‘˜+ to the area of the filter by four
different formulae.
If (β„Žπ‘˜+ , π‘“π‘˜+ ) ∈ π‘†π‘Š(Fπ‘˜ ), assume
𝑐
Fπ‘˜
𝛼 (π‘₯π‘˜+ , Fπ‘˜ ) = area (D(Fπ‘˜ ) ∩ [β„Žπ‘˜+ , β„Žmax
+ πœ…πΉ ]
def
Fπ‘˜
× [π‘“π‘˜+ , 𝑓max
(17)
+ πœ…πΉ ]) .
If (β„Žπ‘˜+ , π‘“π‘˜+ ) ∈ π‘π‘Š(Fπ‘˜ ), assume
F
(18)
F
(19)
π‘˜
𝛼 (π‘₯π‘˜+ , Fπ‘˜ ) = πœ…πΉ (β„Žmin
− β„Žπ‘˜+ ) .
def
If (β„Žπ‘˜+ , π‘“π‘˜+ ) ∈ 𝑆𝐸(Fπ‘˜ ), assume
def
𝑗∈Fπ‘˜
(14)
𝛼 (π‘₯π‘˜+ , Fπ‘˜ ) = πœ…πΉ (𝑓minπ‘˜ − π‘“π‘˜+ ) .
def
max 𝑓𝑗 .
𝑗∈Fπ‘˜
If (β„Žπ‘˜+ , π‘“π‘˜+ ) ∈ 𝑁𝐸(Fπ‘˜ ) = D(Fπ‘˜ ), assume
These four parts are
def
def
(2) the undominated part of lower left corner of the half
plane:
𝑐
Fπ‘˜
Fπ‘˜
π‘†π‘Š (Fπ‘˜ ) = D(Fπ‘˜ ) ∩ [0, β„Žmax
] × [−∞, 𝑓max
],
(15)
P
π‘˜
𝛼 (π‘₯π‘˜+ , Fπ‘˜ ) = −area (D (Fπ‘˜ ) ∩ [β„Žπ‘˜+ − β„Žmin
]
(1) the dominated part of the filter: 𝑁𝐸(Fπ‘˜ ) = D(Fπ‘˜ );
def
(16)
(20)
P
× [π‘“π‘˜+ − 𝑓minπ‘˜ ]) ,
P
def
π‘˜
where Pπ‘˜ = {(β„Ž, 𝑓) ∈ Fπ‘˜ |β„Žπ‘— < β„Žπ‘˜+ , 𝑓𝑗 < π‘“π‘˜+ }, and β„Žmin
=
P
def
min β„Žπ‘— , 𝑓minπ‘˜ = min𝑓𝑗 .
(3) the undominated upper left corner: π‘π‘Š(Fπ‘˜ )
Fπ‘˜
Fπ‘˜
) × (𝑓max
, +∞];
[0, β„Žmin
def
=
𝑗∈Pπ‘˜
(4) the undominated lower right corner: 𝑆𝐸(Fπ‘˜ )
F
Fπ‘˜
(β„Žmax
, +∞] × [−∞, 𝑓minπ‘˜ ).
def
Figure 2 illustrate the corresponding areas in the filter.
Horizontally dashed surfaces indicate a positive contribution
and vertically dashed ones a negative contribution. Note that
𝛼(π‘₯, F) is a continuous function of (β„Ž(π‘₯), 𝑓(π‘₯)).
=
𝑗∈Pπ‘˜
4
Journal of Applied Mathematics
Next, we should consider the updating of the filter. If
(β„Žπ‘˜ , π‘“π‘˜ ) ∉ D(Fπ‘˜ ), then
Fπ‘˜+1 ←󳨀 Fπ‘˜ ∪ (β„Žπ‘˜ , π‘“π‘˜ ) .
is true or not, where πœ‚3 and πœ‡ are all positive constants, if it
is not true, then turn to the feasibility restoration phase and
define
(21)
𝑗
π‘Ÿπ‘˜
If (β„Žπ‘˜ , π‘“π‘˜ ) ∈ D(Fπ‘˜ ), then
P
P
π‘˜
Fπ‘˜+1 ←󳨀 (Fπ‘˜ \ Pπ‘˜ ) ∪ (β„Žmin
, π‘“π‘˜ ) ∪ (β„Žπ‘˜ , 𝑓minπ‘˜ ) .
(22)
2
(23)
def
where π›Όπ‘˜ = 𝛼(π‘₯π‘˜+ , Fπ‘˜ ), 𝛾F ∈ (0, 1) is a constant. Under the
nonmonotonic situation, we relax condition (23) to be
}
{
π‘˜
}
{
}
{
2
max {π›Όπ‘˜ , ∑ πœ† 𝑝(𝑗) 𝛼𝑝(𝑗) } ≥ 𝛾F (β„Žπ‘˜+ ) ,
}
{
}
{ 𝑗=π‘Ÿ(π‘˜)+1
𝑗∈U
}
{
(24)
def
where 𝛼𝑝(𝑗) = 𝛼(π‘₯𝑗 , F𝑝(𝑗) ), U = {π‘˜| filter is updated for
(β„Žπ‘˜ , π‘“π‘˜ )}, π‘Ÿ(π‘˜) ≤ π‘˜ is some reference iteration for U, π‘Ÿ(π‘˜) ∈ U,
U ⊆ S, πœ† 𝑝(𝑗) ∈ [0, 1], ∑π‘˜π‘—=π‘Ÿ(π‘˜)+1, 𝑗∈U πœ† 𝑝(𝑗) = 1.
Compared to condition (2.21) [15], our condition (24) is
more flexible if π›Όπ‘˜ is negative.
According to condition (24), it is possible to accept π‘₯π‘˜+
even though it may be dominated. Then π‘₯π‘˜+ will be accepted
if either (23) or (24) holds.
4. The New Nonmonotone Filter Trust
Region Algorithm
Our algorithm is based on the usual trust region technique;
define the predict reduction for the function π‘žπ‘˜ (π‘₯) to be
pred (π‘‘π‘˜ ) = π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ )
(28)
A formal description of the algorithm is given as follows.
We now return to the question of deciding whether a trial
point π‘₯π‘˜+ is acceptable for the filter or not. We will insist that
this is a necessary condition for the iteration π‘˜ to be successful
in the sense that π‘₯π‘˜+1 = π‘₯π‘˜+ . If we consider an iterate π‘₯π‘˜ , there
+
= π‘₯𝑝(π‘˜)+1 =
must exist a predecessor iteration such that π‘₯𝑝(π‘˜)
π‘₯π‘˜ . Under the monotonic situation, a trial point π‘₯π‘˜+ would be
accepted whenever it results in an sufficient increase in the
dominated area of the filter, that means π‘₯π‘˜+ would be accepted
whenever
π›Όπ‘˜ ≥ 𝛾F (β„Žπ‘˜+ ) ,
σ΅„©σ΅„© 𝑗 σ΅„©σ΅„©2 σ΅„©σ΅„© 𝑗
σ΅„©2
󡄩󡄩𝑐(π‘₯π‘˜ )σ΅„©σ΅„© − 󡄩󡄩𝑐(π‘₯π‘˜ + π‘‘π‘˜π‘— )σ΅„©σ΅„©σ΅„©
σ΅„©
σ΅„©
σ΅„©
σ΅„©
=σ΅„©
.
󡄩󡄩𝑐(π‘₯𝑗 )σ΅„©σ΅„©σ΅„©2 − 󡄩󡄩󡄩𝑐(π‘₯𝑗 ) + 𝐴(π‘₯𝑗 )𝑇 𝑑𝑗 σ΅„©σ΅„©σ΅„©2
σ΅„©σ΅„© π‘˜ σ΅„©σ΅„© σ΅„©σ΅„© π‘˜
π‘˜
π‘˜σ΅„©
σ΅„©
(25)
Algorithm A 𝑆𝑑𝑒𝑝 0. Choose an initial point π‘₯0 ∈ 𝑅𝑛 , a
symmetric matrix 𝐻0 ∈ 𝑅𝑛×𝑛 , let Δ 0 > 0, πœ–π‘‘ > 0, πœ‚2 , πœ‚3 ∈
(0, 1), 𝛼1 , 𝛼2 ∈ [0, 1], πœ‚1 > 0, 0 < 𝛾0 < 𝛾1 < 1 ≤ 𝛾2 < 𝛾3 ≤
2, 0 < πœ‰ < 1, 𝛾F ∈ (0, 1), πœ‡ > 0, compute 𝑓(π‘₯0 ), β„Ž(π‘₯0 ), let
π‘˜ = 0, F0 = 0.
Step 1. Compute π‘π‘˜ , 𝐴 π‘˜ , π‘Šπ‘˜ , π‘“π‘˜ , β„Žπ‘˜ , π‘”π‘˜ , π‘”Μ‚π‘˜ = π‘Šπ‘˜π‘‡ π‘”π‘˜ .
Step 2. If β€– π‘”Μ‚π‘˜ β€– +β„Žπ‘˜ ≤ πœ–π‘‘ , stop.
Step 3. Solve the subproblems (5) and (6) to get the quasinormal step π‘‘π‘˜π‘› and the tangential step π‘‘π‘˜π‘‘ . Let π‘‘π‘˜ = π‘‘π‘˜π‘› +
π‘‘π‘˜π‘‘ , π‘₯π‘˜+ = π‘₯π‘˜ + π‘‘π‘˜ .
Step 4. If π‘Ÿπ‘˜ < πœ‚1 , let π‘₯π‘˜+1 = π‘₯π‘˜ , then go to Step 8.
Step 5. If π‘₯π‘˜ + π‘‘π‘˜ is not acceptable to the filter, go to Step 8.
Otherwise π‘₯π‘˜+1 = π‘₯π‘˜+ and update the filter according to
(21) and (22), the trust region radius and π»π‘˜ , then get the
corresponding β„Žπ‘˜+1 , π‘“π‘˜+1 .
2+𝛼2
Step 6. If β„Žπ‘˜+1 ≤ πœ‚3 min{πœ‡, 𝛼1 Δ π‘˜
Step 1, otherwise go to Step 7.
}, π‘˜ = π‘˜ + 1 and go to
Step 7. By restoration Algorithm B to get π‘‘π‘˜π‘Ÿ , then the trial
point π‘₯π‘˜π‘Ÿ = π‘₯π‘˜ + π‘‘π‘˜π‘Ÿ .
Step 8. Update the trust region radius by Algorithm C, let π‘˜ =
π‘˜ + 1 and go to Step 3.
We aim to reduce the value of β„Ž(π‘₯) in the restoration
Algorithm B, that is to get 𝑐(π‘₯π‘˜π‘Ÿ ) = 0 by Newton-type method.
Algorithm B 𝑆𝑑𝑒𝑝 0. Let π‘₯π‘˜0 = π‘₯π‘˜ , Δ0π‘˜ = Δ π‘˜ , 𝑗 = 0, πœ‚2 , πœ‚3 ∈
(0, 1), 𝛼1 , 𝛼2 ∈ [0, 1], πœ‡ > 0.
𝑗
F
2+𝛼2
Step 1. If β„Ž(π‘₯π‘˜ ) ≤ πœ‚3 min{β„Žπ‘˜ π‘˜ , 𝛼1 Δ π‘˜
𝑗
by the filter, then π‘₯π‘˜π‘Ÿ = π‘₯π‘˜ , stop.
𝑗
} and π‘₯π‘˜ is acceptable
Step 2. Compute
and the actual reduction
ared (π‘‘π‘˜ ) = 𝑓 (π‘₯π‘˜ ) − 𝑓 (π‘₯π‘˜+ ) .
Moreover, let π‘Ÿπ‘˜ = ared(π‘‘π‘˜ )/pred(π‘‘π‘˜ ), if there exists a nonzero
constant πœ‚1 such that π‘Ÿπ‘˜ ≥ πœ‚1 and condition (23) and (24)
hold, the trial point π‘₯π‘˜+ will be called acceptable. Then the next
trial point π‘₯π‘˜+1 is obtained, and for its feasibility, we consider
the condition
2+𝛼2
β„Žπ‘˜+1 ≤ πœ‚3 min {πœ‡, 𝛼1 Δ π‘˜
}
min
(26)
(27)
s.t.
𝑗
σ΅„©2
σ΅„©σ΅„© 𝑗
󡄩󡄩𝑐(π‘₯π‘˜ ) + 𝐴(π‘₯π‘˜π‘— )𝑇 π‘‘π‘˜π‘— σ΅„©σ΅„©σ΅„© ,
σ΅„©
σ΅„©
σ΅„©σ΅„© 𝑗 σ΅„©σ΅„©
𝑗
σ΅„©σ΅„©σ΅„©π‘‘π‘˜ σ΅„©σ΅„©σ΅„© ≤ Δ π‘˜
(29)
𝑗
to get π‘‘π‘˜ , then compute π‘Ÿπ‘˜ .
𝑗
𝑗+1
Step 3. If π‘Ÿπ‘˜ ≤ πœ‚2 , π‘₯π‘˜
to Step 2.
𝑗
𝑗+1
𝑗
= π‘₯π‘˜ , Δ π‘˜ = Δ π‘˜ /2, 𝑗 = 𝑗 + 1 and go
Journal of Applied Mathematics
𝑗
𝑗+1
𝑗
5
𝑗
𝑗+1
𝑗
Step 4. If π‘Ÿπ‘˜ > πœ‚2 , π‘₯π‘˜ = π‘₯π‘˜ + π‘‘π‘˜ , Δ π‘˜ = 2Δ π‘˜ , 𝑗 = 𝑗 + 1,
𝑗+1
F
compute 𝐴 π‘˜ and go to Step 1, where β„Žπ‘˜ π‘˜ = min{min𝑗∈Fπ‘˜
{β„Žπ‘— |β„Žπ‘— > 0}, πœ‡}.
Lemma 2. At the current iterate π‘₯π‘˜ , let the trial point component π‘‘π‘˜π‘› actually be normal to the tangential space. Under
the problem assumptions, there exists a constant π‘˜1 > 0
independent of the iterates such that
Algorithm C (updating the trust region radius). Given πœ‚1 >
0, 0 < 𝛾0 < 𝛾1 < 1 ≤ 𝛾2 < 𝛾3 ≤ 2, we have the following.
σ΅„© σ΅„©
σ΅„©σ΅„© 𝑛 σ΅„©σ΅„©
σ΅„©σ΅„©π‘‘π‘˜ σ΅„©σ΅„© ≤ 𝛼1 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© .
(1) If π‘Ÿπ‘˜ < πœ‚1 or π‘₯π‘˜+ is not acceptable to the filter, Δ π‘˜+1 ∈
[𝛾0 Δ π‘˜ , 𝛾1 Δ π‘˜ ].
(2) If π‘₯π‘˜+ is acceptable to the filter but does not satisfiy
condition (27), Δ π‘˜+1 ∈ (Δ π‘˜ , 𝛾2 Δ π‘˜ ).
(3) If π‘₯π‘˜+ is acceptable to the filter and satisfies
(27), Δ π‘˜+1 ∈ [𝛾2 Δ π‘˜ , 𝛾3 Δ π‘˜ ].
From the description above and the idea of the algorithm,
we can see that our algorithm is more flexible. Every successful iterate must be any better than the predecessor one in
some degree according to the traditional filter method. But
our algorithm relaxes this demand by using the nonmonotone technique and also avoids Maratos effect in a certain
degree. Moreover, Algorithm C allows a relatively wide choice
of the trust region.
5. The Convergence Properties
Assumption
(A1) The objective function 𝑓 and the constraint functions
𝑐𝑖 (𝑖 ∈ 𝐼 = {1, 2, . . . , π‘š}) are twice continuously
differentiable.
(A2) For all π‘˜, π‘₯π‘˜ , and π‘₯π‘˜ + π‘‘π‘˜ all remain in a closed,
bounded convex subset Ω ⊂ 𝑅𝑛 .
(A3) The matrix 𝐴(π‘₯) = ∇𝑐(π‘₯) is nonsingular matrix for all
π‘₯ ∈ Ω.
−1
𝑇
(A5) The matrix π»π‘˜ is uniformly bounded.
By the assumptions, we can suppose there exist constants
V0 , V1 , V2 , V3 such that β€– 𝑓(π‘₯) β€–≤ V0 , β€– 𝑔(π‘₯) β€–≤ V0 , β€– 𝑐(π‘₯) β€–≤
V0 , β€– 𝐴(π‘₯) β€–≤ V0 , β€– 𝐴(π‘₯)𝑇 𝐴(π‘₯)−1 β€–≤ V1 , β€– π‘Š(π‘₯) β€–≤ V2 , β€–
π»π‘˜ β€–≤ V3 , β€– π‘Šπ‘˜π‘‡ π»π‘˜ β€–≤ V3 , β€– π‘Šπ‘˜π‘‡ π»π‘˜ π‘Šπ‘˜ β€–≤ V3 .
By (A1) and (A2), it holds
0 ≤ β„Žπ‘˜ ≤ β„Žmax
Lemma 3. Under Assumptions, there exist positive constants
π‘˜2 , π‘˜3 , π‘˜4 independent of the iterates such that
σ΅„©σ΅„© σ΅„©σ΅„©2 σ΅„©σ΅„©σ΅„©
σ΅„© σ΅„©
σ΅„© σ΅„©
𝑇 𝑛 σ΅„©2
σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„© − σ΅„©σ΅„©π‘π‘˜ + 𝐴 π‘˜ π‘‘π‘˜ σ΅„©σ΅„©σ΅„©σ΅„© ≥ π‘˜2 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© min {π‘˜3 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© , Δ π‘˜ } ,
π‘žπ‘˜ (π‘‘π‘˜π‘› ) − π‘žπ‘˜ (π‘‘π‘˜ )
≥
𝜎 σ΅„©σ΅„© 𝑇
σ΅„©
σ΅„©
σ΅„©
σ΅„©σ΅„©π‘Šπ‘˜ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„© min {π‘˜4 σ΅„©σ΅„©σ΅„©π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„© , Δ π‘˜ } .
σ΅„©
σ΅„©
σ΅„©
σ΅„©
2
(33)
Proof. The proof is an application of Lemma 1 to the two
subproblems (5) and (6).
Lemma 4. Suppose that Assumptions hold, then restoration
Algorithm B is well defined.
∀π‘˜,
(30)
where 𝑓min , β„Žmax > 0, hence in the (β„Ž, 𝑓)-plane, the (β„Ž, 𝑓)pair lies in the area [0, β„Žmax ] × [𝑓min , +∞].
From (A1), (A2), and (A3), it exists a constant ] such that
󡄨󡄨
󡄨
2
󡄨󡄨𝑓 (π‘₯π‘˜ + π‘‘π‘˜ ) − π‘žπ‘˜ (π‘‘π‘˜ )󡄨󡄨󡄨 ≤ ]Δ π‘˜ .
Proof. The conclusion is obvious, if β„Žπ‘˜ → 0. Otherwise it
𝑗
exists πœ– > 0 such that for all 𝑗, it holds β„Žπ‘˜ > πœ–. Consider the
set
󡄨󡄨
σ΅„©σ΅„© 𝑗 σ΅„©σ΅„©2 σ΅„©σ΅„©
σ΅„©2
σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„© − 󡄩󡄩𝑐 (π‘₯π‘˜π‘— + π‘‘π‘˜π‘— )σ΅„©σ΅„©σ΅„©
{ 󡄨󡄨󡄨 𝑗
σ΅„©
σ΅„©
σ΅„©
σ΅„© > πœ‚ > 0} , (34)
󡄨
𝐾 = {𝑗 σ΅„¨σ΅„¨π‘Ÿπ‘˜ = σ΅„© σ΅„©2 σ΅„©
2
}
𝑗
𝑗
𝑗
𝑗
󡄨󡄨
󡄩󡄩𝑐 σ΅„©σ΅„© − 󡄩󡄩𝑐 + (𝐴 )𝑇 𝑑 σ΅„©σ΅„©σ΅„©2
σ΅„©σ΅„© π‘˜ σ΅„©σ΅„© σ΅„©σ΅„© π‘˜
π‘˜
π‘˜σ΅„©
}
{ 󡄨󡄨
σ΅„©
𝑗
𝑗
𝑗
(31)
𝑗
where π‘π‘˜ = 𝑐(π‘₯π‘˜ ), 𝐴 π‘˜ = 𝐴(π‘₯π‘˜ ). By Lemma 3 and the
definition of β„Žπ‘˜ , we have
∞
σ΅„©2
σ΅„© 𝑗 σ΅„©2 σ΅„©σ΅„© 𝑗
𝑗−1
𝑗
𝑗 𝑇 𝑗󡄩
+∞ > ∑ (β„Žπ‘˜ − β„Žπ‘˜ ) ≥ ∑ (σ΅„©σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„©σ΅„© − σ΅„©σ΅„©σ΅„©σ΅„©π‘π‘˜ + (𝐴 π‘˜ ) π‘‘π‘˜ σ΅„©σ΅„©σ΅„©σ΅„© )
σ΅„©
σ΅„©
𝑗=1
𝑗∈𝐾
σ΅„© σ΅„©
σ΅„© σ΅„© 𝑗
≥ πœ‚2 π‘˜2 ∑ σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© min {π‘˜3 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© , Δ π‘˜ } .
−1
(A4) The matrices (𝐴(π‘₯) 𝐴(π‘₯)) , π‘Š(π‘₯), (π‘Š π‘Š) are
uniformly bounded in Ω, where π‘Š(π‘₯) denotes a
matrix whose columns form a basis of the null space
of 𝐴(π‘₯)𝑇 .
𝑓min ≤ π‘“π‘˜ ,
Proof. It is similar to the proof of Lemma 2 in [13].
𝑗
In this section, to present a proof of global convergence of
algorithm, we always assume that the following conditions
hold.
𝑇
(32)
𝑗∈𝐾
(35)
𝑗
𝑗
By β„Žπ‘˜ > πœ–, it holds Δ π‘˜ → 0 for 𝑗 ∈ 𝐾. From Algorithm B, we
𝑗
can obtain that Δ π‘˜ → 0 for all 𝑗.
On the other side,
𝑇 σ΅„©
σ΅„©2
σ΅„©σ΅„© 𝑗 σ΅„©σ΅„©2 σ΅„©σ΅„©
σ΅„©2 σ΅„© σ΅„©2 σ΅„©σ΅„©
σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„© − 󡄩󡄩𝑐 (π‘₯π‘˜π‘— + π‘‘π‘˜π‘— )σ΅„©σ΅„©σ΅„© = σ΅„©σ΅„©σ΅„©π‘π‘˜π‘— σ΅„©σ΅„©σ΅„© − σ΅„©σ΅„©σ΅„©π‘π‘˜π‘— + (π΄π‘—π‘˜ ) π‘‘π‘˜π‘— σ΅„©σ΅„©σ΅„© + π‘œ (Δπ‘—π‘˜ )
σ΅„©σ΅„©
σ΅„© σ΅„© σ΅„©
σ΅„© σ΅„© σ΅„©σ΅„©
σ΅„©
(36)
𝑗
𝑗
for Δ π‘˜ → 0. By the algorithm, the radius Δ π‘˜ should be
𝑗+1
𝑗
𝑗
satisfied Δ π‘˜ > Δ π‘˜ , that is contradicted to Δ π‘˜ → 0. The
proof is complete.
Now, we analyze the impact of the criteria (23) and (24).
Once a trial point is accepted as a new iterate, it must be
provided some improvement, and we formalize this by saying
that iterate π‘₯π‘˜ = π‘₯𝑝(π‘˜)+1 improves on iterate π‘₯𝑖(π‘˜) . That is
6
Journal of Applied Mathematics
the trial point π‘₯π‘˜ is accepted at iterate 𝑝(π‘˜); it happens under
two situations, one is by the criteria (23), that is,
𝑖 (π‘˜) = 𝑝 (π‘˜)
if 𝑝 (π‘˜) ∉ A
(37)
the other is by the criteria (24), that is,
𝑖 (π‘˜) = 𝑝 (π‘˜)
if 𝑝 (π‘˜) ∉ A.
(38)
Now consider any iterate π‘₯π‘˜ , it improved on π‘₯𝑖(π‘˜) , which
was itself accepted because it improved on π‘₯𝑖(𝑖(π‘˜)) , and so on,
until back to π‘₯0 . Hence we may construct a chain of successful
iterations indexed by Cπ‘˜ = {𝑙1 , 𝑙2 , . . . , π‘™π‘ž } for each π‘˜, such that
Lemma 6. Suppose that Assumptions hold. If Algorithm A
does not terminate finitely and the filter contains infinite
iterates, then limπ‘˜ → ∞ β„Žπ‘˜ = 0.
Proof. Suppose by contradiction that there exists a constant
πœ– > 0 and infinite sequence {π‘˜π‘– } ⊆ S such that β„Žπ‘˜π‘– ≥ πœ– for
all 𝑖. Because there are infinite iterations in the filter, we have
|S| = ∞, then β„Žπ‘™π‘ž ≥ πœ– for ∀π‘ž.
area (D (Fπ‘˜ )) ≥ 𝛾F ⋅ π‘ž ⋅ πœ–2 .
(46)
𝑗 = 1, 2, . . . , π‘ž − 1,
(39)
Then by (31), area(D(Fπ‘˜ )) is upper bounded for each π‘˜. That
max
max
≥ 0 such that area(D(Fπ‘˜ )) ≤ πœ…F
, so
means it exists πœ…F
max
2
𝑖 ≤ (πœ…F /𝛾F πœ– ). Hence 𝑖 must be finite, it contradicts to the
infinity of {π‘˜π‘– }. The proof is complete.
where 𝑙1 is the smallest index in the chain of successful
iterations.
Lemma 7. Suppose that Assumptions hold and Algorithm A
terminate finitely, then β„Žπ‘˜ = 0.
Lemma 5. Suppose that Assumptions hold and Algorithm A
does not terminate finitely, apply Algorithm A to the problem
(𝑃), then for all π‘˜ and Cπ‘˜ = {𝑙1 , 𝑙2 , . . . , π‘™π‘ž }, it holds
Proof. From the Algorithm A and the definition of filter, the
conclusion follows.
π‘₯𝑙1 = π‘₯0 ,
π‘₯π‘™π‘ž = π‘₯π‘˜ ,
π‘₯𝑙𝑗 = π‘₯𝑖(𝑙𝑗+1 ) ,
π‘ž
area (D (Fπ‘˜ )) ≥ 𝛾F ∑β„Žπ‘™2𝑗 .
(40)
𝑗=1
Proof. For ∀𝑙𝑗 ∈ Cπ‘˜ , if 𝑝(𝑙𝑗 ) ∈ A, by (24), 𝑖(𝑙𝑗 ) = π‘Ÿ(𝑝(𝑙𝑗 )) =
𝑙𝑗−1 , then
}
{
𝑙𝑗
}
{
}
{
max {𝛼𝑝(𝑙𝑗 ) , ∑ πœ† 𝑝(𝑖) 𝛼𝑝(𝑖) } ≥ 𝛾F β„Žπ‘™2𝑗 .
}
{
}
{
𝑖=𝑙𝑗−1 +1
𝑖∈U
}
{
𝑙
𝑗
If max{𝛼𝑝(𝑙𝑗 ) , ∑ 𝑖=𝑙
𝑗−1 +1, 𝑖∈U
(41)
Lemma 8. For any trial point π‘₯π‘˜+1 =ΜΈ π‘₯π‘˜ , there must be one
accepted by the filter.
Lemma 9. Suppose that Assumptions hold, there exists π‘˜5 > 0
independent of the iterates such that
σ΅„© σ΅„©
π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› ) ≥ −π‘˜5 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© .
σ΅„© σ΅„©
Proof. By (32), the assumptions and σ΅„©σ΅„©σ΅„©π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„© ≤ Δ max , it is obvious
that
1
𝑇
π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› ) = −π‘”π‘˜π‘‡ π‘‘π‘˜π‘› − (π‘‘π‘˜π‘› ) π»π‘˜ π‘‘π‘˜π‘›
2
πœ† 𝑝(𝑖) 𝛼𝑝(𝑖) } = 𝛼𝑝(𝑙𝑗 ) ,
𝛼𝑝(𝑙𝑗 ) ≥ 𝛾F β„Žπ‘™2𝑗 .
𝑙
𝑗
πœ† 𝑝(𝑖) 𝛼𝑝(𝑖) }
If max{𝛼𝑝(𝑙𝑗 ) , ∑ 𝑖=𝑙
𝑗−1 +1, 𝑖∈U
𝛼𝑝(𝑖) ,
σ΅„© σ΅„© σ΅„© σ΅„© 1 σ΅„© σ΅„©σ΅„© σ΅„©
≥ − σ΅„©σ΅„©σ΅„©π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„© (σ΅„©σ΅„©σ΅„©π‘”π‘˜ σ΅„©σ΅„©σ΅„© + σ΅„©σ΅„©σ΅„©π»π‘˜ σ΅„©σ΅„©σ΅„© σ΅„©σ΅„©σ΅„©π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„©)
2
σ΅„©σ΅„© σ΅„©σ΅„©
≥ −π‘˜1 σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„© (]0 + ]3 Δ max )
(42)
=
𝑙
𝑗
∑ 𝑖=𝑙
𝑗−1 +1, 𝑖∈U
(47)
(48)
σ΅„© σ΅„©
= −π‘˜5 σ΅„©σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©σ΅„© .
πœ† 𝑝(𝑖)
def
The proof is complete.
𝑙𝑗
∑ 𝛼𝑝(𝑖) ≥ 𝛾F β„Žπ‘™2𝑗 .
(43)
𝑖=𝑙𝑗−1 +1
𝑖∈U
σ΅„© σ΅„©
Lemma 10. Suppose that Assumptions hold and σ΅„©σ΅„©σ΅„©π‘”Μ‚π‘˜ σ΅„©σ΅„©σ΅„© ≥ πœ–π‘‘ , if
Δ π‘˜ ≤ min{(
If 𝑝(𝑙𝑗 ) ∉ A, 𝑙𝑗−1 = 𝑝(𝑙𝑗 ). By U ⊆ S, it holds
{𝑙𝑗−1 + 1, . . . , 𝑙𝑗 } ∩ U ⊆ {𝑙𝑗−1 + 1, . . . , 𝑙𝑗 } ∩ S = {𝑙𝑗 } . (44)
(
Then from (23), 𝛼𝑝(𝑖) ≥ 𝛾F β„Žπ‘™2𝑗 . It implies (43). Moreover
π‘˜
π‘ž
𝑖=0
𝑖∈U
𝑗=0
area (D (Fπ‘˜ )) ≥ ∑ 𝛼𝑝(𝑖) = ∑ ( ∑ 𝛼𝑝(𝑖) ) .
𝑖=𝑙𝑗−1 +1
𝑖∈U
Together with (42) and (43), the result follows.
]0
)
π‘˜1 ]3 √πœ‚3 𝛼1
(
𝑙𝑗
(45)
πœ–π‘‘
)
2π‘˜1 ]3 √πœ‚3 𝛼1
2/(2+𝛼2 )
2/(2+𝛼2 )
2/𝛼2
πœŽπœ–π‘‘
)
16π‘˜1 ]0 √πœ‚3 𝛼1
,
π‘˜4 πœ–π‘‘
,
2
,
(49)
def
} = 𝛿1
one can deduce
π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ ) ≥
πœŽπœ–π‘‘
def
Δ π‘˜ = πœŽΔ π‘˜ .
8
(50)
Journal of Applied Mathematics
7
Proof. By the assumptions and the definition of π‘‘π‘˜π‘› , it holds
σ΅„©
σ΅„©
σ΅„©σ΅„© 𝑇
σ΅„© σ΅„©
σ΅„©
σ΅„©σ΅„©π‘Šπ‘˜ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„© = σ΅„©σ΅„©σ΅„©π‘Šπ‘˜π‘‡ (π‘”π‘˜ + π»π‘˜ π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„© ≥ σ΅„©σ΅„©σ΅„©σ΅„©π‘”Μ‚π‘˜ σ΅„©σ΅„©σ΅„©σ΅„© − σ΅„©σ΅„©σ΅„©π‘Šπ‘˜π‘‡ π»π‘˜ π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„©
σ΅„©
σ΅„©
σ΅„©
σ΅„© σ΅„©
σ΅„©
σ΅„©σ΅„© σ΅„©σ΅„©
≥ πœ–π‘‘ − ]3 π‘˜1 σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„©
πœ–
1+(𝛼 /2)
≥ πœ–π‘‘ − ]3 π‘˜1 √πœ‚3 𝛼1 Δ π‘˜ 2 ≥ 𝑑 .
2
(51)
From Lemma 3, π‘žπ‘˜ (π‘‘π‘˜π‘› ) − π‘žπ‘˜ (π‘‘π‘˜ ) ≥ (𝜎/4)πœ–π‘‘ min{Δ π‘˜ , π‘˜4 πœ–π‘‘ /2} ≥
(πœŽπœ–π‘‘ /4)Δ π‘˜ . Together with
1
𝑇
π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› ) = − π‘”π‘˜π‘‡ π‘‘π‘˜π‘› − (π‘‘π‘˜π‘› ) π»π‘˜ π‘‘π‘˜π‘›
2
σ΅„© σ΅„© σ΅„© σ΅„© 1 σ΅„© σ΅„© σ΅„© σ΅„©2
≥ − σ΅„©σ΅„©σ΅„©π‘”π‘˜ σ΅„©σ΅„©σ΅„© σ΅„©σ΅„©σ΅„©π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„© − σ΅„©σ΅„©σ΅„©π»π‘˜ σ΅„©σ΅„©σ΅„© σ΅„©σ΅„©σ΅„©π‘‘π‘˜π‘› σ΅„©σ΅„©σ΅„©
2
1+(𝛼2 /2)
≥ − π‘˜1 ]0 √πœ‚3 𝛼1 Δ π‘˜
Theorem 13. Suppose the assumptions hold, there must exist
Δ min > 0 such that for each π‘˜, it holds
Δ π‘˜ ≥ Δ min .
Proof. Let π‘˜1 be large enough such that β„Žπ‘˜1 ≤ πœ–π‘‘ , it is true by
Lemmas 6 and 7. Suppose by contradiction that the index 𝑗 is
the first one after π‘˜1 , which satisfies
{
(1 − √𝛾F ) β„ŽπΉ
)
Δ π‘— ≤ 𝛾0 min {𝛿2 , (
πœ‚3 𝛼1
{
2+𝛼2
1+(𝛼2 /2)
we know 𝑗 ≥ π‘˜1 + 1, that is 𝑗 − 1 ≥ π‘˜1 . From the Algorithm
and (60), it concludes
Δ π‘—−1 ≤
(52)
then π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ ) ≥ −(πœŽπœ–π‘‘ /8)Δ π‘˜ + (πœŽπœ–π‘‘ /4)Δ π‘˜ = (πœŽπœ–π‘‘ /8)Δ π‘˜ =
πœŽΔ π‘˜ . It is the conclusion.
1
Δ ≤ 𝛿3 .
𝛾0 𝑗
(61)
By (60) and (61), (53) can be obtained. In Lemma 11, let 𝑗 − 1
instead of π‘˜, it deduces
π‘Ÿπ‘—−1 ≥ πœ‚1 .
(62)
Based on Lemma 12, together with (60), (61), and the algorithm, we can see
Lemma 11. Suppose the conditions of Lemma 10 hold, if
(1 − πœ‚1 ) 𝜎 def
} = 𝛿2
]
} def
, Δ π‘˜1 } = 𝛾0 𝛿3 ,
}
(60)
where β„ŽπΉ = min𝑖∈U β„Žπ‘– is the smallest value of violation
function in filter. Then Δ π‘— ≤ 𝛾0 Δ π‘˜1 . By the above analysis,
≥ − 2π‘˜1 ]0 √πœ‚3 𝛼1 Δ π‘˜
Δ π‘˜ ≤ min {𝛿1 ,
1/(2+𝛼2 )
def
− π‘˜12 ]3 πœ‚3 𝛼1 Δ π‘˜
πœŽπœ–
≥ − 𝑑 Δπ‘˜
8
(59)
(53)
2+𝛼
+
β„Žπ‘—−1
≤ πœ‚3 𝛼1 Δ π‘—−12 ≤ (1 − √𝛾F ) β„ŽπΉ .
(63)
It can be seen that (53) is true for 𝑗 − 1 ≥ π‘˜, with (55), we can
deduce
then π‘Ÿπ‘˜ ≥ πœ‚1 .
Proof. From the definition of π‘Ÿπ‘˜ and Lemma 10, together with
(31), we have
󡄨
󡄨
2
󡄨 󡄨󡄨𝑓 (π‘₯π‘˜ + π‘‘π‘˜ ) − π‘žπ‘˜ (π‘‘π‘˜ )󡄨󡄨󡄨 ]Δ π‘˜
󡄨󡄨
≤
≤ 1 − πœ‚1 . (54)
σ΅„¨σ΅„¨π‘Ÿπ‘˜ − 1󡄨󡄨󡄨 = 󡄨 󡄨󡄨
󡄨
πœŽΔ π‘˜
σ΅„¨σ΅„¨π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ )󡄨󡄨󡄨
It is obvious that π‘Ÿπ‘˜ ≥ πœ‚1 .
Lemma 12. Suppose the conditions of Lemmas 10 and 11 hold,
if
−1/(1+𝛼2 )
β„Žπ‘˜ ≤ (πœ‚3 𝛼1 )
(
πœ‚1 𝜎
)
√𝛾F
(2+𝛼2 )/(1+𝛼2 )
(55)
then
𝑓 (π‘₯π‘˜+ ) ≤ 𝑓 (π‘₯π‘˜ ) − √𝛾F β„Žπ‘˜ .
2+𝛼2
From the Algorithm, β„Žπ‘˜ ≤ πœ‚3 𝛼1 Δ π‘˜
𝑓 (π‘₯π‘˜ ) − 𝑓 (π‘₯π‘˜+ ) ≥ πœ‚1 𝜎(
Hence 𝑓(π‘₯π‘˜ ) − 𝑓(π‘₯π‘˜+ ) ≥ √𝛾F β„Žπ‘˜ .
(57)
, then
β„Žπ‘˜ 1/(2+𝛼2 )
)
.
πœ‚3 𝛼1
(64)
+
That means π‘₯𝑗−1
can be accepted by the filter. From above and
(55), we know Δ π‘— ≥ Δ π‘—−1 . Hence the index 𝑗 is not the first
one after π‘˜1 which satisfied (60), that is a contradiction. So,
for any π‘˜ > π‘˜1 , it holds Δ π‘˜ ≥ 𝛾0 𝛿3 . Define
Δ min = min {Δ 0 , . . . , Δ π‘˜1 , 𝛾0 𝛿3 }
(65)
we can see that
Δ π‘˜ ≥ Δ min
(66)
holds for each π‘˜. The proof is complete.
(56)
Proof. By Lemmas 3, 10, and 11, together with β€– π‘”Μ‚π‘˜ β€–≥ πœ–π‘‘ , it
holds
𝑓 (π‘₯π‘˜ ) − 𝑓 (π‘₯π‘˜+ ) ≥ πœ‚1 (π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ )) ≥ πœ‚1 πœŽΔ π‘˜ .
+
≤ 𝑓𝑗−1 − √𝛾F β„Žπ‘—−1 .
𝑓𝑗−1
(58)
Lemma 14. Suppose that Assumptions hold and Algorithm A
σ΅„© σ΅„©
does not terminate finitely, then lim inf π‘˜ → ∞ σ΅„©σ΅„©σ΅„©π‘”Μ‚π‘˜ σ΅„©σ΅„©σ΅„© = 0.
Proof. Suppose by contradiction that for πœ–π‘‘ , there exists a
constant π‘˜ > 0 such that β€– π‘”Μ‚π‘˜ β€–≥ πœ–π‘‘ .
σ΅„©
σ΅„©
σ΅„© σ΅„©
By Assumption (A3) and (A4), σ΅„©σ΅„©σ΅„©σ΅„©π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„©σ΅„© ≥ σ΅„©σ΅„©σ΅„©π‘”Μ‚π‘˜ σ΅„©σ΅„©σ΅„© −
]3 π‘˜1 β€– π‘π‘˜ β€–. From Lemma 6, we know β„Žπ‘˜ → 0. Hence there
exists Μƒπ‘˜ > 0 such that
2πœ–π‘‘
σ΅„©σ΅„© σ΅„©σ΅„©
σ΅„©σ΅„©π‘π‘˜ σ΅„©σ΅„© ≤
3]3 π‘˜1
(67)
8
Journal of Applied Mathematics
for π‘˜ > Μƒπ‘˜. Then β€– π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› ) β€–≥ (1/3) β€– π‘”Μ‚π‘˜ β€–≥ (1/3)πœ–π‘‘ for
def
π‘˜ > Μ‚π‘˜ = max{π‘˜, Μƒπ‘˜}.
It is obvious that
π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ ) = π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› ) + π‘žπ‘˜ (π‘‘π‘˜π‘› ) − π‘žπ‘˜ (π‘‘π‘˜ ) .
(68)
By the proof of Lemma 9, it holds |π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› )| ≤
]0 β€– π‘‘π‘˜π‘› β€– +(1/2)]3 β€– π‘‘π‘˜π‘› β€–2 . Together with Lemma 6 and the
definition of π‘‘π‘˜ , we have
lim (π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜π‘› )) = 0.
π‘˜→∞
(69)
By the Algorithm, we can get
∞
∞
π‘˜=0
π‘˜=0
+∞ > ∑ (π‘“π‘˜ − π‘“π‘˜+1 ) ≥ πœ‚1 ∑ (π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ )) .
(70)
Table 1
Problem
HS6
HS7
HS8
HS9
HS26
HS39
HS40
HS42
HS78
𝑛
2
2
2
2
3
4
4
4
5
π‘š
1
1
2
1
1
2
3
2
3
NF
11
9
7
6
24
15
7
8
6
NG
11
3
4
6
24
9
5
8
6
L’s NF
20
15
—
6
42
26
7
11
8
L’s NG
12
11
—
6
24
19
5
11
7
Proof. By Assumption (A1), there exist a subsequence {π‘˜π‘— }
and π‘₯∗ , such that lim𝑗 → ∞ π‘₯π‘˜π‘— = π‘₯∗ . Together with Assump-
tion (A3) and (A4), it hods lim𝑗 → ∞ π‘Šπ‘˜π‘‡π‘— π‘”π‘˜π‘— = 0, which means
for large enough 𝑗, π‘”π‘˜π‘— lies in the space spaned by the columns
of π΄π‘‡π‘˜π‘— . That is there exists πœ† π‘˜π‘— such that
Then
lim (π‘žπ‘˜ (0) − π‘žπ‘˜ (π‘‘π‘˜ )) = 0.
π‘˜→∞
(71)
By (68), (69), and (71), it deduces
lim (π‘žπ‘˜ (π‘‘π‘˜π‘› ) − π‘žπ‘˜ (π‘‘π‘˜ )) = 0.
π‘˜→∞
The conclusion follows.
π‘žπ‘˜ (π‘‘π‘˜π‘› ) − π‘žπ‘˜ (π‘‘π‘˜ )
≥
𝜎 σ΅„©σ΅„© 𝑇
σ΅„©
σ΅„©
σ΅„©
σ΅„©σ΅„©π‘Š ∇π‘ž (𝑑𝑛 )σ΅„©σ΅„©σ΅„© min {Δ π‘˜ , π‘˜4 σ΅„©σ΅„©σ΅„©π‘Šπ‘˜π‘‡ ∇π‘žπ‘˜ (π‘‘π‘˜π‘› )σ΅„©σ΅„©σ΅„©}
σ΅„©
σ΅„© (73)
2σ΅„© π‘˜ π‘˜ π‘˜ σ΅„©
≥
πœŽπœ–π‘‘
π‘˜πœ–
min {Δ min , 4 𝑑 } > 0,
6
6
Theorem 15. Suppose the assumptions hold, and apply the
algorithm to problem (𝑃), then
(74)
where π‘”Μ‚π‘˜ = π‘Šπ‘˜π‘‡ π‘”π‘˜ , π‘”π‘˜ = ∇𝑓(π‘₯π‘˜ ), π‘Š(π‘₯) denotes a matrix
whose columns form a basis of the null space of 𝐴(π‘₯)𝑇 .
Proof. If the algorithm terminates finitely, it is obvious that it
holds. Otherwise, by Lemmas 6 and 14, the conclusion also
can be obtained.
Theorem 16. Suppose the assumptions hold, and {π‘₯π‘˜ } is the
infinite sequence obtained by the algorithm, then there must
exist a subsequence such that
lim π‘₯π‘˜π‘— = π‘₯∗
and π‘₯∗ satisfies the one order KKT condition of (𝑃).
6. Some Numerical Experiments
(1) Updating of π»π‘˜ is done by π»π‘˜+1 = π»π‘˜ +(π‘¦π‘˜π‘‡ π‘¦π‘˜ /π‘¦π‘˜π‘‡ π‘ π‘˜ )−
(π»π‘˜ π‘ π‘˜ π‘ π‘˜π‘‡ π»π‘˜ /π‘ π‘˜π‘‡ π»π‘˜ π‘ π‘˜ ), where π‘¦π‘˜ = πœƒπ‘˜ π‘¦Μ‚π‘˜ + (1 − πœƒπ‘˜ )π»π‘˜ π‘ π‘˜
1
{
{
πœƒπ‘˜ = { 0.8π‘ π‘˜π‘‡ π»π‘˜ π‘ π‘˜
{ 𝑇
𝑇
{ π‘ π‘˜ π»π‘˜ π‘ π‘˜ − π‘ π‘˜ π‘¦Μ‚π‘˜
π‘ π‘˜π‘‡ π‘¦Μ‚π‘˜ ≥ 0.2π‘ π‘˜π‘‡ π»π‘˜ π‘ π‘˜
(77)
and π‘¦Μ‚π‘˜ = π‘”π‘˜+1 − π‘”π‘˜ + (𝐴 π‘˜+1 − 𝐴 π‘˜ )πœŒπ‘˜ , π‘ π‘˜ = π‘₯π‘˜+1 −
π‘₯π‘˜ πœŒπ‘˜ is the multipluser of corresponding quadratic
subproblems.
which contradicts (72). The conclusion follows.
𝑗→∞
(76)
(72)
Based on the assumptions, Lemma 3 and Theorem 13, for π‘˜ >
Μ‚π‘˜, it holds
σ΅„© σ΅„©
lim inf (β„Žπ‘˜ + σ΅„©σ΅„©σ΅„©π‘”Μ‚π‘˜ σ΅„©σ΅„©σ΅„©) = 0,
π‘˜→∞
lim π‘”π‘˜π‘— + π΄π‘‡π‘˜π‘— πœ† π‘˜π‘— = 0.
𝑗→∞
(75)
(2) We assume the error toleration is 10−5 .
(3) The algorithm parameters were set as follows: 𝐻0 =
𝐼 ∈ 𝑅𝑛×𝑛 , πœ‚1 = 0.25, πœ‚2 = 0.25, πœ‚3 = 0.1, 𝛼1 = 𝛼2 = 0.5,
𝛾 = 0.02, 𝜌 = 0.5, πœ‰ = 10−6 , 𝛾0 = 0.1, 𝛾1 = 0.5, 𝛾2 = 2,
𝛾𝐹 = 10−4 , Δ 0 = 1. The program is written in Matlab.
The numerical results for the test problems are listed in
Table 1.
In Table 1, the problems are numbered in the same way
as in Schittkowski [20] and Hock and Schittkowski [21]. For
example, “S216” is the problem (216) in Schittkowski [20]
and “HS6” is the problem (6) in Hock and Schittkowski
[21]. NF, NG represent the numbers of function and gradient
calculations and “L’s” is the solution in [22]. The numerical
results show that the our algorithm is more effective than the
L’s for most test examples. Moreover, the higher the level of
nonmonotonic, the better the numerical results. The results
show that the new algorithm is robust and effective, and more
flexible for the acceptance of the the trial iterate.
Journal of Applied Mathematics
Acknowledgments
This research is supported by the National Natural Science
Foundation of China (no. 11101115) and the Natural Science
Foundation of Hebei Province (nos. A2010000191, Y2012021).
References
[1] R. H. Byrd, R. B. Schnabel, and G. A. Shultz, “A trust region
algorithm for nonlinearly constrained optimization,” SIAM
Journal on Numerical Analysis, vol. 24, no. 5, pp. 1152–1170, 1987.
[2] J. E. Dennis Jr., M. El-Alem, and M. C. Maciel, “A global
convergence theory for general trust-region-based algorithms
for equality constrained optimization,” SIAM Journal on Optimization, vol. 7, no. 1, pp. 177–207, 1997.
[3] M. J. D. Powell and Y. Yuan, “A trust region algorithm for
equality constrained optimization,” Mathematical Programming
A, vol. 49, no. 2, pp. 189–211, 1991.
[4] R. Fletcher and S. Leyffer, “Nonlinear programming without a
penalty function,” Mathematical Programming A, vol. 91, no. 2,
pp. 239–269, 2002.
[5] C. M. Chin and R. Fletcher, “On the global convergence of
an SLP-filter algorithm that takes EQP steps,” Mathematical
Programming A, vol. 96, no. 1, pp. 161–177, 2003.
[6] R. Fletcher, N. I. M. Gould, S. Leyffer, P. L. Toint, and A.
Wächter, “Global convergence of a trust-region SQP-filter
algorithm for general nonlinear programming,” SIAM Journal
on Optimization, vol. 13, no. 3, pp. 635–659, 2002.
[7] R. Fletcher, S. Leyffer, and P. L. Toint, “On the global convergence of a filter-SQP algorithm,” SIAM Journal on Optimization,
vol. 13, no. 1, pp. 44–59, 2002.
[8] S. Ulbrich, “On the superlinear local convergence of a filter-SQP
method,” Mathematical Programming B, vol. 100, no. 1, pp. 217–
245, 2004.
[9] M. Ulbrich, S. Ulbrich, and L. N. Vicente, “A globally convergent primal-dual interior-point filter method for nonlinear
programming,” Mathematical Programming A, vol. 100, no. 2,
pp. 379–410, 2004.
[10] R. Fletcher and S. Leyffer, “A bundle filter method for nonsmooth nonlinear optimization,” Tech. Rep. NA/195, Department of Mathematics, University of Dundee, Dundee, Scotland,
December 1999.
[11] L. Grippo, F. Lampariello, and S. Lucidi, “A nonmonotone
line search technique for Newton’s method,” SIAM Journal on
Numerical Analysis, vol. 23, no. 4, pp. 707–716, 1986.
[12] M. Ulbrich and S. Ulbrich, “Non-monotone trust region methods for nonlinear equality constrained optimization without a
penalty function,” Mathematical Programming B, vol. 95, no. 1,
pp. 103–135, 2003.
[13] K. Su and D. Pu, “A nonmonotone filter trust region method for
nonlinear constrained optimization,” Journal of Computational
and Applied Mathematics, vol. 223, no. 1, pp. 230–239, 2009.
[14] K. Su and Z. Yu, “A modified SQP method with nonmonotone
technique and its global convergence,” Computers & Mathematics with Applications, vol. 57, no. 2, pp. 240–247, 2009.
[15] N. I. M. Gould and P. L. Toint, “Global convergence of a nonmonotone trust-region filter algorithm for nonlinear programming,” in Multiscale optimization methods and applications, vol.
82, pp. 125–150, Springer, New York, NY, USA, 2006.
9
[16] Z. Chen and X. Zhang, “A nonmonotone trust-region algorithm
with nonmonotone penalty parameters for constrained optimization,” Journal of Computational and Applied Mathematics,
vol. 172, no. 1, pp. 7–39, 2004.
[17] Z. W. Chen, “A penalty-free-type nonmonotone trust-region
method for nonlinear constrained optimization,” Applied Mathematics and Computation, vol. 173, no. 2, pp. 1014–1046, 2006.
[18] F. A. M. Gomes, M. C. Maciel, and J. M. Martı́nez, “Nonlinear
programming algorithms using trust regions and augmented
Lagrangians with nonmonotone penalty parameters,” Mathematical Programming A, vol. 84, no. 1, pp. 161–200, 1999.
[19] M. J. D. Powell, “Convergence properties of a class of minimization algorithms,” in Nonlinear Programming, pp. 1–27, Academic
Press, New York, NY, USA, 1974.
[20] K. Schittkowski, More Test Examples for Nonlinear Programming
Codes, vol. 282, Springer, Berlin, Germany, 1987.
[21] W. Hock and K. Schittkowski, Test Examples for Nonlinear Programming Codes, vol. 187, Springer, Berlin, Germany, 1981.
[22] M. Lalee, J. Nocedal, and T. Plantenga, “On the implementation
of an algorithm for large-scale equality constrained optimization,” SIAM Journal on Optimization, vol. 8, no. 3, pp. 682–706,
1998.
Advances in
Operations Research
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Advances in
Decision Sciences
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Mathematical Problems
in Engineering
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Journal of
Algebra
Hindawi Publishing Corporation
http://www.hindawi.com
Probability and Statistics
Volume 2014
The Scientific
World Journal
Hindawi Publishing Corporation
http://www.hindawi.com
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
International Journal of
Differential Equations
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Volume 2014
Submit your manuscripts at
http://www.hindawi.com
International Journal of
Advances in
Combinatorics
Hindawi Publishing Corporation
http://www.hindawi.com
Mathematical Physics
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Journal of
Complex Analysis
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
International
Journal of
Mathematics and
Mathematical
Sciences
Journal of
Hindawi Publishing Corporation
http://www.hindawi.com
Stochastic Analysis
Abstract and
Applied Analysis
Hindawi Publishing Corporation
http://www.hindawi.com
Hindawi Publishing Corporation
http://www.hindawi.com
International Journal of
Mathematics
Volume 2014
Volume 2014
Discrete Dynamics in
Nature and Society
Volume 2014
Volume 2014
Journal of
Journal of
Discrete Mathematics
Journal of
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Applied Mathematics
Journal of
Function Spaces
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Optimization
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Hindawi Publishing Corporation
http://www.hindawi.com
Volume 2014
Download