Hindawi Publishing Corporation Abstract and Applied Analysis Volume 2013, Article ID 163487, 9 pages http://dx.doi.org/10.1155/2013/163487 Research Article An Improved Nonmonotone Filter Trust Region Method for Equality Constrained Optimization Zhong Jin Department of Mathematics, Shanghai Maritime University, Shanghai 201306, China Correspondence should be addressed to Zhong Jin; asdjz1234@163.com Received 22 October 2012; Accepted 11 January 2013 Academic Editor: Nikolaos Papageorgiou Copyright © 2013 Zhong Jin. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Motivated by the method of Su and Pu (2009), we present an improved nonmonotone filter trust region algorithm for solving nonlinear equality constrained optimization. In our algorithm a modified nonmonotone filter technique is proposed and the restoration phase is not needed. At every iteration, in common with the composite-step SQP methods, the step is viewed as the sum of two distinct components, a quasinormal step and a tangential step. A more relaxed accepted condition for trial step is given and a crucial criterion is weakened. Under some suitable conditions, the global convergence is established. In the end, numerical results show our method is effective. 1. Introduction We consider the problem of minimizing a nonlinear function subject to a set of nonlinear equality constraints as follows: (đ) min đ (đĽ) s.t. đđ (đĽ) = 0, đ ∈ đź = {1, 2, . . . , đ} , (1) where the objective function đ : đ đ → đ and the equality constraints đđ (đ ∈ đź) : đ đ → đ are all twice continuously differentiable. The algorithm that we discuss belongs to the class of trust region methods and, more specifically, to that filter methods suggested first by Fletcher and Leyffer [1], in which the use of a penalty function, a common feature of the large majority of the algorithms for constrained optimization, is replaced by the technique so-called “filter.” Subsequently, global convergences of the trust region filter SQP methods were established by Fletcher et al. [2, 3], and Ulbrich [4] got its superlinear local convergence. Especially, the step framework in [3] is similar in spirit to the composite-step SQP methods pioneered by Vardi [5], Byrd et al. [6], and Omojokun [7]. Consequently, filter method has been actually applied in many optimization techniques, for instance, the pattern search method [8], the SLP method [9], the interior method [10], the bundle approaches [11, 12], the system of nonlinear equations and nonlinear least squares [13], multidimensional filter method [14], line search filter methods [15, 16], and so on. In fact, filter method exhibits a certain degree of nonmonotonicity. The idea of nonmonotone technique can be traced back to Grippo et al. [17] in 1986, combined with the line search strategy. Due to its excellent numerical exhibition, over the last decades, the nonmonotone technique has been used in trust region method to deal with unconstrained and constrained optimization problems [18–25]. More recently, the nonmonotone trust region method without penalty function has also been developed for constrained optimization [26–29]. Especially, in [29] the nonmonotone idea is employed to the filter technique and restoration phase, a common feature of the large majority of filter methods, is not needed. Based on [29], motivated by above ideas and methods, in this paper we present an improved filter algorithm that combines the nonmonotone and trust region techniques for solving nonlinear equality constrained optimization. Our method improves previous results and gives a more flexible mechanism, weakens some needed conditions, and has the merits as follows. (i) An improved nonnomotone filter technique is presented. Filter technique is viewed as a biobjective 2 Abstract and Applied Analysis optimization problem that minimizes objective function đ and violation function â. In [29], the same nonmonotone measure is utilized for both đ and â at the đth iteration. In the paper we improve it and define a new measure for â, which may make the discussion and consideration of the nonmonotone properties of đ and â more freely. (ii) The restoration phase is not needed. The restoration procedure has to be considered in most of filter methods. We employ the nonmonotone idea to the filter technique, so as [29] the restoration phase is not needed. (iii) A more relaxed accepted condition for trial step is considered. Compared to general trust region methods, in [29] the condition for the accepted trial step is relaxed. In the paper we improve it and then the accepted condition for trial step is more relaxed. (iv) A crucial criterion is weakened in the algorithm. By introducing a new parameter đ ∈ (0, 1], we improve the crucial criterion for implementation of the method in [29]. It is obvious that our criterion becomes same with [29] if đ = 1, but new criterion can be easier to satisfy, so this crucial criterion has been weakened when setting đ < 1 in the initialization of our algorithm. The presentation is organized as follows. Section 2 introduces some preliminary results and improvements on filter technique and some conditions. Section 3 develops a modified nonmonotone filter trust region algorithm, whose global convergence is shown in Section 4. The results of numerical experience with the proposed method are discussed in the last section. 2. Preliminary and Improvements In this section, we first recall some definitions and preliminary results about the techniques of composite-step SQP type trust region method and fraction of Cauchy decrease condition. And then the improvements on filter technique and some conditions are given. 2.1. Fraction of Cauchy Decrease Condition. To introduce the corresponding results, we consider the unconstrained optimization problem: min{đ(đĽ) | đĽ ∈ đ đ }, where đ : đ đ → đ is continuously differentiable. At iteration point đĽđ , it obtains a trial step đđ by solving the following quadratic program subproblem: 1 min đ (đ) = ∇đ (đĽđ ) đ + đđ đťđ đ 2 s.t. âđâ ≤ Δ đ , (2) where đťđ is a symmetric matrix which is either the Hessian matrix of đ at đĽđ or an approximation to it, and Δ đ > 0 is a trust region radius. To assure the global convergence, the step is only required to satisfy a fraction of Cauchy decrease condition. It means đ must predict via the quadratic model function đ(đ) at least as much as a fraction of the decreased given by the Cauchy step on đ(đ), that is, there exists a constant đ > 0 fixed across all iterations, such that đ (0) − đ (đ) ≥ đ (đ (0) − đ (đcp )) , (3) where đcp is the steepest descent step for đ(đ) inside the trust region. And we have the following lemma. Lemma 1. If the trial step đ satisfies a fraction of Cauchy decrease condition, then đ (0) − đ (đ) ≥ óľŠ óľŠóľŠ óľŠ∇đ (đĽ)óľŠóľŠóľŠ đ óľŠóľŠ óľŠ }. óľŠóľŠ∇đ (đĽ)óľŠóľŠóľŠ min {Δ, óľŠ 2 âđťâ (4) Proof. See Powell [30] for the proof. 2.2. The Subproblems of Composite-Step SQP Type Trust Region Method. In the composite-step SQP type trust region methods, at current iterate đĽđ we obtain the trial step đđ = đđđ + đđđĄ by computing a quasi-normal step đđđ and a tangential step đđđĄ . The purpose of the quasi-normal step đđđ is to improve feasibility and đđđĄ in the tangential space of the linearized constraints can provide sufficient decrease for a quadratic model of the objective function đ(đĽ) to improve optimality. For problem (đ), đđđ is the solution to the subproblem 1 óľŠóľŠ óľŠ2 óľŠóľŠđ + đ´đ đđ óľŠóľŠóľŠ 2 óľŠóľŠ đđ óľŠ đ óľŠ óľŠóľŠđ óľŠóľŠ ≤ Δ đ , s.t. óľŠ óľŠ min (5) where Δ đ is a trust region radius, đ1 (đĽ) đ2 (đĽ) đ (đĽ) = ( .. ) , . (6) đđ (đĽ) đ´(đĽ) = ∇đ(đĽ), đđ = đ(đĽđ ) and đ´ đ = đ´(đĽđ ) ∈ đ đ×đ . In order to improve the value of the objective function, đđđĄ can be obtained by the following subproblem min đđ (đđđ + đđĄ ) s.t. đ´đđ đđĄ = 0 óľŠóľŠ đĄ óľŠóľŠ óľŠóľŠđ óľŠóľŠ ≤ Δ đ , (7) where đđ (đ) = đđđ đ + (1/2)đđ đťđ đ and đđ = ∇đ(đĽđ ). Then we get the current trial step đđ = đđđ + đđđĄ . In usual ways that impose a trust region in step-decomposition methods, the quasi-normal step đđđ and the tangential step đđđĄ are required to satisfy óľŠóľŠ đ óľŠóľŠ óľŠóľŠđ óľŠóľŠ ≤ 0.8Δ đ , óľŠ óľŠóľŠ đ óľŠóľŠđđ + đđĄ óľŠóľŠóľŠ ≤ Δ đ . óľŠ óľŠ (8) Here, to simplify the proof, we only impose a trust region on âđđ â ≤ Δ đ and âđđĄ â ≤ Δ đ , which is natural. Abstract and Applied Analysis 3 2.3. The Improved Accepted Condition for đđ . Borrowed from the usual trust region idea, we also need to define the following predicted reduction for the violation function â(đĽ) = âđ(đĽ)â2 a feasible point. So in the traditional filter method, a point đĽ is called acceptance to the filter if and only if óľŠ2 óľŠ predđđ = â (đĽđ ) − óľŠóľŠóľŠóľŠđđ + đ´đđ đđđ óľŠóľŠóľŠóľŠ where 0 < đž < đ˝ < 1, F denotes the filter set. Different from above criteria of filter idea, with nonmonotone technique, in [29] a point đĽ is called to be acceptable to the filter if and only if (9) and the actual reduction óľŠ óľŠ2 óľŠ óľŠ2 aredđđ = â (đĽđ ) − â (đĽđ + đđ ) = óľŠóľŠóľŠđđ óľŠóľŠóľŠ − óľŠóľŠóľŠđ (đĽđ + đđ )óľŠóľŠóľŠ . (10) To evaluate the descent properties of the step for the objective function, we use the predicted reduction of đ(đĽ) đ predđ = đđ (0) − đđ (đđ ) = −đđ (đđ ) đ (12) In general trust region method, the step đđ will be accepted if đ đ aredđ ≥ đ predđ , (13) where đ ∈ (0, 1) is a fixed constant. In [29], considering nonmonotone technique, the condition (13) is replaced by đ đ raredđ ≥ đ predđ , where đ raredđ (14) is the relaxed actual reduction of đ(đĽ), that is, đ(đ)−1 đ raredđ = max {đ (đĽđ ) , ∑ đ đđ đ (đĽđ−đ )} − đ (đĽđ + đđ ) . đ=0 (15) But in this paper, we consider more relaxed accepted conđ đ đ dition by increasing aredđ to raredđ and reducing predđ to đ min{predđ , predđđ +đ0 } simultaneously, then the condition (14) is replaced by đ đ raredđ ≥ đ min {predđ , predđđ + đ0 } , (16) where đ0 is a small positive number. 2.4. The Improved Nonmonotone Filter Technique. In order to obtain next iterate, it needs to determine the step; this procedure that decides which trial step is accepted is “filter method.” For the optimization problem with equality constraints min {đ (đĽ) | đ (đĽ) = 0} , orđ (đĽ) ≤ đđ − đžâđ ∀ (âđ , đđ ) ∈ F, (18) either â (đĽ) ≤ đ˝ max âđ−đ 0≤đ≤đ(đ)−1 đ(đ)−1 (19) or đ (đĽ) ≤ max [đđ , ∑ đ đđ đđ−đ ] − đžâ (đĽ) , đ=0 (11) and the actual reduction of đ(đĽ) aredđ = đ (đĽđ ) − đ (đĽđ + đđ ) . â (đĽ) ≤ đ˝âđ (17) a promising trial step should either reduce the constraint violation â or the objective function value đ. Since â(đĽ) = âđ(đĽ)â2 , it is easy to see that â(đĽ) = 0 if and only if đĽ is where (âđ−đ , đđ−đ ) ∈ Fđ for 0 ≤ đ ≤ đ(đ) − 1, and 1 ≤ đ(đ) ≤ min{đ(đ − 1) + 1, đ}, đ ≥ 1 is a given positive integer, đ đđ = 1, đ đđ ∈ (0, 1), and there exists a positive ∑đ(đ)−1 đ=0 constant đ such that đ đđ ≥ đ. Observe that đ(đ) is used in both conditions (19), while in this paper we wish to reduce the relationship of the nonmonotone properties of đ and â. We consider the nonmonotone properties of đ and â, respectively, and call a trial point đĽ is acceptable to the filter if and only if either â (đĽ) ≤ đ˝ max 0≤đ≤đ(đ)−1 đđ+đ−đ (20) đ(đ)−1 or đ (đĽ) ≤ max [đđ , ∑ đ đđ đđ−đ ] − đžâ (đĽ) , đ=0 where đđ+đ−đ comes from {đđ }+∞ đ=1 such that đ1 = đ2 = ⋅ ⋅ ⋅ = +∞ đđ−1 = 0 and {đđ }đ=đ is a subsequence of {â(đĽđ )}, and 1 ≤ đ(đ) ≤ min{đ(đ − 1) + 1, đ}. Similar to the traditional filter methods, if (20) is satisfied, it is called an â-type iteration and the filter set Fđ needs to be updated at each successful â-type iteration. 2.5. The Weakened Criterion for Implementation of Algorithm. đ We replace the crucial criterion predđ ≥ predđđ in [29] by đ predđ ≥ đ, predđđ (21) where đ ∈ (0, 1]. It is obvious that this new criterion becomes the same with [29] if đ = 1, but it can be easier to satisfy, so the criterion has been weakened when setting đ < 1 in the initialization of our algorithm. 3. Description of the Algorithm Ě = It will be convenient to introduce the reduced gradient đ(đĽ) đ(đĽ)đ đ(đĽ) where đ(đĽ) = ∇đ(đĽ) and đ(đĽ) denotes a matrix whose columns are from a basis of the null space of đ´(đĽ)đ . The first order necessary optimality conditions (KarushKuhn-Tucher or KKT conditions) at a local solution đĽ ∈ đ đ of (đ) can be written as đ (đĽ) = 0, đĚ (đĽ) = 0. (22) 4 Abstract and Applied Analysis For brevity, at the current iterate đĽđ , we use đđ , đđ , đ´ đ , đđ , âđ , đđ , đťđ to denote ∇đ(đĽđ ), đ(đĽđ ), đ´(đĽđ ), đ(đĽđ ), â(đĽđ ), đ(đĽđ ), đť(đĽđ ). We are now in a position to present a formal statement of our algorithm. Algorithm 2. Step 0. Initialization: choose an initial guess đĽ0 , a symmetric matrix đť0 ∈ đ đ×đ , an initial region radius Δ 0 ≥ Δ min > 0, 0 < đž < đ˝ < 1, 0 < đ ≤ 1 and F0 = {(â0 , đ0 )}. Set đ ∈ (0, 1], đ0 > 0, 0 < đ < 1, 0 < đž0 < đž1 < 1 < đž2 and đ > 0. Let đ(0) = 1, đ1 = đ2 = ⋅ ⋅ ⋅ = đđ−1 = 0, đđ = â(đĽ0 ), đ = 0 and đ = 0. Step 1. Compute đđ , đđ , đđ , âđ , đ´ đ , đđ . If âđĚđ â + âđ = 0, then stop. Step 2. Compute (5) and (7) to obtain đđđ and đđđĄ , and set đđ = đđđ + đđđĄ . Step 3. If đĽđ + đđ is acceptable to the filter, go to Step 4, otherwise go to Step 5. đ đ đ Step 4. If raredđ < đ min{predđ , predđđ + đ0 } and predđ / predđđ ≥ đ, then go to Step 5, otherwise go to Step 6. Step 5. Δ đ ∈ [đž0 Δ đ , đž1 Δ đ ], go to Step 2. Step 6. đĽđ+1 = đĽđ + đđ , Δ đ+1 ∈ [Δ đ , đž2 Δ đ ] ≥ Δ min . Update đťđ to đťđ+1 , đ(đ+1) = min{đ(đ)+1, đ}. If (20) holds, update the filter set; let đđ+đ+1 = â(đĽđ+1 ), đ(đ+1) = min{đ(đ)+1, đ} and đ = đ + 1. Let đ = đ + 1 and go to Step 1. Remark 3. At the beginning of each iteration, we always set Δ đ ≥ Δ min , which will avoid too small trust region radius. (A2 ) All points đĽđ that are sampled by algorithm lie in a nonempty closed and bounded set đ. (A3 ) The matrix sequence {đťđ } is uniformly bounded. −1 (A4 ) đ´(đĽ) = ∇đ(đĽ), (đ´đ (đĽ)đ´(đĽ)) , đ(đĽ) and −1 đ (đ (đĽ)đ(đĽ)) are uniformly bounded on đ. From the above assumptions, it is easier to suppose that there exist five constants đŁ1 , đŁ2 , đŁ3 , đŁ4 , đŁ5 such that âđ(đĽ)â ≤ đŁ1 , â∇đ(đĽ)â ≤ đŁ1 , â∇2 đ(đĽ)â ≤ đŁ1 , âđ(đĽ)â ≤ đŁ2 , −1 âđ´(đĽ)â ≤ đŁ2 , â∇2 đ(đĽ)â ≤ đŁ2 , â(đ´đ (đĽ)đ´(đĽ)) â ≤ đŁ3 , âđ(đĽ)â ≤ đŁ4 , âđđ (đĽ)đť(đĽ)â ≤ đŁ5 ; this can make our following results convenient to prove. Lemma 6. At the current iterate đĽđ , let the trial point component đđđ actually be normal to the tangential space. Under the above assumptions, there exists a constant đź1 > 0 independent of the iterates such that óľŠóľŠ óľŠóľŠ óľŠóľŠ đ óľŠóľŠ (25) óľŠóľŠđđ óľŠóľŠ ≤ đź1 óľŠóľŠđđ óľŠóľŠ . đ Proof. Since đđ is actually normal to the tangential space, it follows that óľŠ óľŠ óľŠóľŠóľŠđđ óľŠóľŠóľŠ = óľŠóľŠóľŠđ´ đ (đ´đ đ´ đ )−1 đ´đ đđ óľŠóľŠóľŠ đ đ óľŠ óľŠ đ óľŠ óľŠóľŠ óľŠ −1 óľŠóľŠóľŠ óľŠóľŠ đ (26) = óľŠóľŠđ´ đ (đ´ đ đ´ đ ) (đđ + đ´đđ đđ − đđ )óľŠóľŠóľŠ óľŠ óľŠ óľŠóľŠ −1 óľŠ óľŠ óľŠ óľŠ óľŠ óľŠ = óľŠóľŠóľŠđ´ đ (đ´đđ đ´ đ ) óľŠóľŠóľŠ + (óľŠóľŠóľŠóľŠđđ + đ´đđ đđ óľŠóľŠóľŠóľŠ + óľŠóľŠóľŠđđ óľŠóľŠóľŠ) , óľŠ óľŠ together with the fact that âđđ + đ´đđ đđ â ≤ âđđ â, we have −1 óľŠ óľŠóľŠ óľŠóľŠ óľŠ óľŠóľŠ đ óľŠóľŠ óľŠ óľŠ đ (27) óľŠóľŠđđ óľŠóľŠ ≤ 2 óľŠóľŠóľŠđ´ đ (đ´ đ đ´ đ ) óľŠóľŠóľŠ óľŠóľŠóľŠđđ óľŠóľŠóľŠ â đź1 óľŠóľŠóľŠđđ óľŠóľŠóľŠ . óľŠ óľŠ Remark 4. If â-type iteration is satisfied, then đđ+đ+1 is generated, so if the number of â-type iterations is infinite, we obtain a sequence {đđ }+∞ đ=1 such that đ1 = đ2 = ⋅ ⋅ ⋅ = đđ−1 = 0 and {đđ }+∞ is a subsequence of {â(đĽđ )}. Specially, (20) always đ=đ holds on the whole obtained sequence {đđ }+∞ đ=1 . Lemma 7. Suppose that the assumptions hold, there exist positive đź2 , đź3 independent of the iterates such that óľŠ óľŠ óľŠ óľŠ óľŠóľŠ óľŠóľŠ2 óľŠóľŠóľŠ đ óľŠ2 (28) óľŠóľŠđđ óľŠóľŠ − óľŠóľŠđđ + đ´ đ đđ óľŠóľŠóľŠóľŠ ≥ đź2 óľŠóľŠóľŠđđ óľŠóľŠóľŠ min {óľŠóľŠóľŠđđ óľŠóľŠóľŠ , Δ đ } , Remark 5. In the above algorithm, let đ be a positive integer. For each đ, let đ(đ) satisfying óľŠ óľŠ óľŠ óľŠ ≥ đź3 óľŠóľŠóľŠóľŠđđđ ∇đđ (đđđ )óľŠóľŠóľŠóľŠ min {óľŠóľŠóľŠóľŠđđđ ∇đđ (đđđ )óľŠóľŠóľŠóľŠ , Δ đ } . đ (0) = 1, 1 ≤ đ (đ) ≤ min {đ (đ − 1) + 1, đ} , for đ ≥ 1, (23) 1 ≤ đ (đ) ≤ min {đ (đ − 1) + 1, đ} , for đ ≥ 1, (29) Proof. The statement is established by an application of Lemma 1 to two subproblems (5) and (7). Lemma 8. Under problem assumptions, Algorithm 2 is well defined. and we also find đ(đ) satisfy đ (0) = 1, đđ (đđđ ) − đđ (đđ ) (24) so the nonmonotonicity is showed as đ > 1. 4. Global Convergence of Algorithm In this section, we will prove the global convergence properties of Algorithm 2 under the following assumptions. (A1 ) The objective đ(đĽ) and the constraint functions đđ (đĽ), đ = 1, 2, . . . đ are all twice continuously differentiable on an open set containing đ. Proof. We will show that there exists đż > 0 such that step đđ is accepted whenever Δ đ ≤ đż. In fact, óľ¨ óľ¨óľ¨ óľ¨óľ¨aredđ − predđ óľ¨óľ¨óľ¨ óľ¨óľ¨ đ đ óľ¨óľ¨ óľŠóľŠ óľŠóľŠ 1 = óľŠóľŠóľŠóľŠđ (đĽđ ) − đ (đĽđ + đđ ) + đđđ đđ + đđđ đťđ đđ óľŠóľŠóľŠóľŠ 2 óľŠ óľŠ óľŠóľŠ đ óľŠóľŠ (30) 1 đ 2 1 đ đ óľŠ óľŠ = óľŠóľŠ−đđ đđ − đđ ∇ đ (đŚđ ) đđ + đđ đđ + đđ đťđ đđ óľŠóľŠóľŠóľŠ 2 2 óľŠ óľŠ 1óľŠ óľŠ = 4đΔ2đ óľŠóľŠóľŠóľŠ∇2 đ (đŚđ ) − đťđ óľŠóľŠóľŠóľŠ ≤ 4đΔ2đ đ, 2 Abstract and Applied Analysis 5 where đŚđ = đĽđ + đđđ , đ ∈ (0, 1) denotes some point on the line segment from đĽđ to đĽđ + đđ , đ = (1/2)(sup âđťđ â + maxđĽ∈đ â∇2 đ(đĽ)â), âđđđĄ â ≤ √đΔ đ , âđđđ â ≤ √đΔ đ and âđđ â ≤ 2√đΔ đ . We consider two cases in the following. Case 1 (â(đĽđ ) ≠ 0). To prove the implementation of Algođ rithm 2, we need only to show that if predđ /predđđ ≥ đ > 0, đ đ it follows raredđ ≥ đ min{predđ , predđđ + đ0 }. Without loss of generality, we can assume that âđđ â ≥ đ. Then we start with đż ∈ (0, đ] such that the closed đż-ball about đĽđ lies in đ. It is obvious from đż ≤ đ that Δ đ ≤ đż ≤ đ. And from (28), we have predđđ ≥ đź2 đΔ đ . Observe that óľ¨ óľ¨ óľ¨ óľ¨ óľ¨ óľ¨óľ¨ óľ¨óľ¨ aredđđ − predđđ óľ¨óľ¨óľ¨ óľ¨óľ¨óľ¨ predđđ óľ¨óľ¨óľ¨ óľ¨óľ¨óľ¨ aredđđ − predđđ óľ¨óľ¨óľ¨ óľ¨óľ¨ ⋅ óľ¨óľ¨ óľ¨óľ¨ = óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ óľ¨ óľ¨ óľ¨óľ¨ óľ¨óľ¨ đ predđđ óľ¨óľ¨ óľ¨óľ¨ predđđ óľ¨óľ¨óľ¨ óľ¨óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ pred óľ¨ óľ¨ óľ¨ óľ¨ óľ¨ óľ¨ đ 4đΔ2đ đ ≤ ół¨→ 0 as Δ đ ół¨→ 0, đź2 đΔ đ Lemma 8 has provided the implementation of Algorithm 2. By mechanism of Algorithm 2, it is obvious that there exists a constant Δ > 0, such that Δ đ ≥ Δ for sufficiently large đ. In the remainder of this paper we denote the set of indices of those iterations in which the filter has been augmented by A, that is to say, if the đth iteration is â-type iteration, we have đ + 1 ∈ A. By this definition, if (20) holds for infinite iterations, then |A| = ∞; otherwise, we can say |A| ≠ ∞. Lemma 9. Let {đĽđ } be an infinite sequence generated by Algorithm 2. If |A| ≠ ∞, then limđ → ∞ â(đĽđ ) = 0. Proof. By mechanism of Algorithm 2, we can assume that for all đ, it follows đ(đ)−1 đ (đĽđ+1 ) ≤ max {đ (đĽđ ) , ∑ đ đđ đ (đĽđ−đ )} − đžâ (đĽđ ) . đ=0 (37) đ predđ ≥ đ > 0. predđđ Hence we have óľ¨óľ¨ đ đ óľ¨óľ¨ óľ¨óľ¨óľ¨ aredđ − predđ óľ¨óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ ół¨→ 0 đ óľ¨óľ¨ óľ¨óľ¨ predđ óľ¨óľ¨ óľ¨óľ¨ Then we first show that for all đ, it holds (31) đ=0 as Δ đ ół¨→ 0, đ aredđ (32) đ đ đ which implies ≥ đ predđ ≥ 0, then raredđ ≥ aredđ đ đ đ predđ ≥ đ min{predđ , predđđ + đ0 } for some đ ∈ (0, 1). ≥ đ − đ5 Δ đ . And it holds from â(đĽđ ) = 0 that đđđ = 0, then, combining with (29), we find that đź đź đ đ min { , Δ đ } = đΔ đ . 2 2 2 Hence, óľ¨óľ¨ óľ¨ 2 óľ¨óľ¨ aredđđ − predđđ óľ¨óľ¨óľ¨ óľ¨óľ¨ óľ¨óľ¨ ≤ 4đΔ đ đ ół¨→ 0 as Δ ół¨→ 0, đ đ óľ¨óľ¨óľ¨ óľ¨óľ¨óľ¨ (đź/2) đΔ đ predđ óľ¨óľ¨ óľ¨óľ¨ đ đ ≤ đ (đĽ0 ) − đđž ∑ â (đĽđ ) . đ=0 Next we prove (38) by induction. If đ = 1, we have đ(đĽ1 ) ≤ đ(đĽ0 ) − đžâ(đĽ0 ) ≤ đ(đĽ0 ) − đđžâ(đĽ0 ). Assume that (38) holds for 1, 2, . . . , đ, then we consider that (38) holds for đ + 1 in the following two cases. đ đđ đ(đĽđ−đ )} = đ(đĽđ ), then Case 1. If max{đ(đĽđ ), ∑đ(đ)−1 đ=0 đ−1 đ (đĽđ+1 ) ≤ đ (đĽđ ) − đžâ (đĽđ ) ≤ đ (đĽ0 ) − đđž ∑ â (đĽđ ) − đžâ (đĽđ ) đ=0 Next, if we reduce đż such that đż ≤ (1/2đ5 )đ, then for all Δ đ ≤ đż, we have Δ đ ≤ (1/2đ5 )đ. Therefore óľŠóľŠóľŠđđ ∇đ (đđ )óľŠóľŠóľŠ ≥ đ . (34) óľŠóľŠ đ đ đ óľŠóľŠ 2 đ (38) đ−1 ≥ Case 2 (â(đĽđ ) = 0). If âđĚđ â = 0, then đĽđ is a KKT-point of (1) by Algorithm 2. Thus, we assume that there exists a constant đ such that âđĚđ â ≥ đ. Then óľŠóľŠóľŠđđ ∇đ (đđ )óľŠóľŠóľŠ = óľŠóľŠóľŠđđ (đ + đť đđ )óľŠóľŠóľŠ óľŠóľŠ đ đ đ óľŠóľŠ óľŠóľŠ đ đ đ đ óľŠ óľŠ óľŠóľŠ đ óľŠóľŠ óľŠóľŠ đ óľŠ (33) ≥ óľŠóľŠóľŠđđ đđ óľŠóľŠóľŠ − óľŠóľŠóľŠđđ đťđ đđđ óľŠóľŠóľŠóľŠ predđ = −đđ (đđ ) ≥ đ−2 đ (đĽđ ) ≤ đ (đĽ0 ) − đđž ∑ â (đĽđ ) − đžâ (đĽđ−1 ) đ (35) đ ≤ đ (đĽ0 ) − đđž∑â (đĽđ ) . đ=0 (39) Case 2. If max{đ(đĽđ ), ∑đ(đ)−1 đ đđ đ(đĽđ−đ )} = ∑đ(đ)−1 đ đđ đ=0 đ=0 đ(đĽđ−đ ), let đ = đ(đ) − 1, then đ đ (đĽđ+1 ) ≤ ∑đ đđĄ đ (đĽđ−đĄ ) − đžâ (đĽđ ) đĄ=0 (36) đ which also implies raredđ ≥ aredđ ≥ đ predđ ≥ đ min{predđ , predđđ + đ0 } for some đ ∈ (0, 1). Thus, the trial step is accepted for all Δ đ ≤ đż. đ đ−đĄ−2 ≤ ∑đ đđĄ (đ (đĽ0 ) − đđž ∑ â (đĽđ ) đĄ=0 đ=0 −đžâ (đĽđ−đĄ−1 ) ) − đžâ (đĽđ ) 6 Abstract and Applied Analysis đ−đ−2 Proof. Suppose to contrary that there exist constant đ > 0 and đ > 0 such that âđĚđ â > 2đ for all đ > đ. Then similar to the proof of Lemma 8, we have = đ đ0 (đ (đĽ0 ) − đđž ∑ â (đĽđ ) − đđž đ=0 đ−2 óľŠ óľŠóľŠ đ óľŠ óľŠ óľŠóľŠđđ ∇đđ (đđđ )óľŠóľŠóľŠ ≥ óľŠóľŠóľŠóľŠđĚđ óľŠóľŠóľŠóľŠ − óľŠóľŠóľŠđđđ đťđ đđđ óľŠóľŠóľŠ óľŠ óľŠ óľŠ óľŠ óľŠóľŠ đ óľŠóľŠ óľŠ óľŠ ≥ 2đ − đ5 óľŠóľŠđđ óľŠóľŠ ≥ 2đ − đ5 đź1 óľŠóľŠóľŠđđ óľŠóľŠóľŠ , × ∑ â (đĽđ ) − đžâ (đĽđ−1 )) đ=đ−đ−1 đ−đ−2 + đ đ1 (đ (đĽ0 ) − đđž ∑ â (đĽđ ) for all đ > đ. Since â(đĽđ ) → 0, then âđđ â → 0; there exists Ěđ > đ such that óľŠóľŠ đ óľŠ óľŠóľŠđđ ∇đđ (đđđ )óľŠóľŠóľŠ ≥ đ ∀đ > Ěđ. (44) óľŠ óľŠ đ=0 đ−3 −đđž ∑ â (đĽđ ) − đžâ (đĽđ−2 )) đ=đ−đ−1 From â(đĽđ ) → 0, we also have đđđ → 0, then đđ (đđđ ) → 0. Hence đ−đ−2 + ⋅ ⋅ ⋅ + đ đđ (đ0 (đĽ0 ) − đđž ∑ â (đĽđ ) đ predđ = −đđ (đđ ) = −đđ (đđđ ) + đđ (đđđ ) − đđ (đđ ) đ=0 ≥ −đđ (đđđ ) + đźđ min {Δ đ , đ} ≥ −đžâ (đĽđ−đ−1 ) ) − đžâ (đĽđ ) đ đ−đ−2 đ=0 đĄ=0 đ(đ)−1 đ đ (đĽđ+1 ) ≤ max {đ (đĽđ ) , ∑ đ đđ đ (đĽđ−đ )} − ∑đ đđĄ đžâ (đĽđ−đĄ−1 ) − đžâ (đĽđ ) . đ=0 đĄ=0 đ By the fact that ∑đĄ=0 đ đđĄ = 1, đ đđĄ ≥ đ and â(đĽđ ) ≥ 0, we have đ−đ−2 đ In common with the proof of Lemma 9, we have ∞ đ đ ∑ min {predđ , predđđ + đ0 } ół¨→ 0, (47) đ=0 đ (đĽđ+1 ) ≤ đ (đĽ0 ) − đđž ∑ â (đĽđ ) đ đ=0 which implies predđ → 0. It contradicts (45). Hence the result follows. đ−1 − đđž ∑ â (đĽđ ) − đžâ (đĽđ ) (41) đ−1 = đ (đĽ0 ) − đđž ∑ â (đĽđ ) − đžâ (đĽđ ) Lemma 11. Let {đĽđ } be an infinite sequence generated by Algorithm 2. If |A| = ∞, then limđ∈A, đ → ∞ â(đĽđ ) = 0. Proof. In view of convenience, denote đ=0 đđ(đ) = đ ≤ đ (đĽ0 ) − đđž∑â (đĽđ ) . đ=0 Then for all đ, (38) holds. Moreover, since {đ(đĽđ )} is bounded below, let đ → ∞, we can get that max 0≤đ≤đ(đ)−1 đđ+đ−đ , (48) where đ + đ − đ(đ) + 1 ≤ đ(đ) ≤ đ + đ. From the algorithm, we know đđ+đ+1 = â(đĽđ+1 ) and it holds đđ+đ+1 ≤ đ˝đđ(đ) . (49) Since đ(đ + 1) ≤ đ(đ) + 1, we have ∞ đđž ∑ â (đĽđ ) < ∞, (46) − đ min {predđ , predđđ + đ0 } . (40) đ=đ−đ−1 (45) đź đ min {Δ, đ} . 2 As in the proof of Lemma 8, there exists đ ∈ (0, 1) such đ đ that raredđ ≥ đ min{predđ , predđđ + đ0 }. That is, đ ≤ ∑đ đđĄ đ (đĽ0 ) − đđž ∑ (∑đ đđĄ ) â (đĽđ ) đĄ=0 (43) (42) đ=0 which implies â(đĽđ ) → 0. Hence the result follows. Lemma 10. Suppose that the assumptions hold. If Algorithm 2 does not terminate finitely, let {đĽđ } be an infinite sequence generated by Algorithm 2; if |A| ≠ ∞, then limđ → ∞ inf âđĚđ â = 0. đđ(đ+1) = max 0≤đ≤đ(đ+1)−1 đđ+đ+1−đ ≤ max đđ+đ+1−đ 0≤đ≤đ(đ) = max {đđ+đ+1 , đđ(đ) } = đđ(đ) . (50) Abstract and Applied Analysis 7 This implies that {đđ(đ) } converges. Moreover, by đđ+đ+1 ≤ đ˝đđ(đ) , we obtain đđ(đ) ≤ đ˝đđ(đ(đ)−đ−1) . (51) And since đ˝ ∈ (0, 1), eventually đđ(đ) → 0 (đ → ∞). Thus đđ+đ+1 ≤ đ˝đđ(đ) ół¨→ 0 (52) holds by Algorithm 2, then, with the facts đđ+đ+1 = â(đĽđ+1 ) and đ + 1 ∈ A, we have limđ∈A, đ → ∞ â(đĽđ ) = 0. Lemma 12. Suppose that the assumptions hold. If Algorithm 2 does not terminate finitely, let {đĽđ } be an infinite sequence generated by Algorithm 2; if |A| = ∞, then limđ∈A,đ → ∞ inf âđĚđ â = 0. Proof. Suppose to the contrary that there exist constant đ > 0 and đ > 0 such that âđĚđ â > 2đ for all đ > đ and đ ∈ A. Then similar to the proof of Lemma 8, we have óľŠ óľŠ óľŠóľŠ đ óľŠ óľŠóľŠđđ ∇đđ (đđđ )óľŠóľŠóľŠ ≥ óľŠóľŠóľŠóľŠđĚđ óľŠóľŠóľŠóľŠ − óľŠóľŠóľŠđđđ đťđ đđđ óľŠóľŠóľŠ óľŠ óľŠ óľŠ óľŠ óľŠóľŠ đ óľŠóľŠ óľŠ óľŠ ≥ 2đ − đ5 óľŠóľŠđđ óľŠóľŠ ≥ 2đ − đ5 đź1 óľŠóľŠóľŠđđ óľŠóľŠóľŠ , (53) for all đ > đ and đ ∈ A. Since limđ∈A, đ → ∞ â(đĽđ ) → 0, that is, âđđ â → 0 for đ ∈ A, then there exists Ěđ > đ such that óľŠóľŠ đ óľŠ óľŠóľŠđđ ∇đđ (đđđ )óľŠóľŠóľŠ ≥ đ ∀đ > Ěđ, đ ∈ A. óľŠ óľŠ (54) From limđ∈A, đ → ∞ â(đĽđ ) = 0, we also have đđđ → 0 for đ ∈ A, then đđ (đđđ ) → 0 for đ ∈ A. Thus, for đ ∈ A đ predđ = −đđ (đđ ) = −đđ (đđđ ) + đđ (đđđ ) ≥ −đđ (đđđ ) + đźđ min {Δ đ , đ} ≥ − đđ (đđ ) đź đ min {Δ, đ} . 2 đ đ ∑ min {predđ , predđđ + đ0 } ół¨→ 0, (55) (56) đ Theorem 13. Suppose {đĽđ } is an infinite sequence generated by Algorithm 2, one has óľŠ óľŠ óľŠ óľŠ lim inf [óľŠóľŠóľŠđđ óľŠóľŠóľŠ + óľŠóľŠóľŠđĚđ óľŠóľŠóľŠ] = 0. đťđ đ đ (đ đ ) đťđ đ (đ đ ) đťđ đ đ đ + đđ (đđ ) đ (đ đ ) đđ , (58) where đ đ = đĽđ+1 − đĽđ , đđ = đđ đŚđ + (1 − đđ )đťđ đ đ , đŚđ = đđ+1 − đđ and 1 { { { { { đđ = { đ { (đ đ ) đťđ đ đ { { 0.8 { đ đ (đ đ ) đťđ đ đ − (đ đ ) đŚđ { đ if (đ đ ) đŚđ đ ≥ 0.2(đ đ ) đťđ đ đ , otherwise. (59) In the following tables, the notations mean as follows: (i) SC: the problems from Schittkowski [32]; (ii) HS: the problems from Hock and Schittkowski [31]; (iii) n: the number of variables; (iv) m: the number of inequality constraints; (v) NIT: the number of iterations; (vii) NG: the number of evaluations of scalar constraint functions; (ix) Algorithm 2: the method proposed in [29]; which implies predđ → 0 for all đ. But predđ → 0 for đ ∈ A is also true, which contradicts (55). Hence the result follows. đ→∞ đ đťđ+1 = đťđ − (viii) Algorithm 1: our method in this paper; đ=0 đ In this section, we carry out some numerical experiments based on the algorithm over a set of problems from [31, 32]. The program is written in MATLAB and we test and give the performance of Algorithm 2 on these problems. For each test example, we choose the initial matrix đť0 = đź. As to those constants, we set đž = 0.02, đ˝ = 0.98, đ = 0.5, đ = 0.75, đ0 = 0.05, đž0 = 0.1, đž1 = 0.5, đž2 = 2, Δ min = 10−6 , Δ 0 = 1, đ = 3. We use 10−6 as the stopping criterion. During the numerical experiments, updating of đťđ is done by (vi) NF: the number of evaluations of the objective functions; Similar to the proof of Lemma 10, it is obtained that ∞ 5. Numerical Experiments (57) Namely, there exists at least one cluster point of {đĽđ } that is a KKT point of problem (đ). Proof. The conclusion follows immediately by Lemmas 9, 10, 11, and 12. Thus, the whole proof is completed. (x) Algorithm 3: the method proposed in [33]. The detailed numerical results are summarized in Tables 1 and 2. Now we give a brief analysis for numerical test results. From Table 1, we can see that our algorithm executes well for these typical problems taken from [31, 32]. From the computation efficiency in Table 2, by solving same problems in [31, 32] we should point out our algorithm is competitive with the some existed nonmonotone filter type methods to solve equality constrained optimization, for example, [29, 33]. However, in our method an improved nonmonotone filter technique is proposed, which may make the discussion and consideration of the nonmonotone properties of đ and â more freely. Furthermore, we consider a more relaxed accepted condition for trial step and a weakened crucial criterion is presented. It is easier to meet those new criteria, so our method is much more flexible. All results summarized show that our algorithm is promising and numerically effective. 8 Abstract and Applied Analysis Table 1: Detailed numerical results. đ 2 4 3 3 3 4 5 3 2 2 2 2 3 3 4 5 5 5 5 5 Problem SC216 SC219 SC235 SC252 SC254 SC265 SC269 SC344 HS006 HS007 HS008 HS009 HS026 HS028 HS042 HS047 HS050 HS051 HS077 HS079 đ 1 2 1 1 2 2 3 1 1 1 2 1 1 1 2 3 3 3 2 3 NIT 5 28 20 22 18 3 5 12 4 8 3 5 35 7 8 25 12 7 12 11 NF 16 52 48 46 45 7 10 24 7 14 11 12 87 29 31 86 39 18 22 23 NG 7 36 30 36 37 3 8 16 5 8 5 6 45 11 10 72 14 9 16 13 Table 2: Comparison among three algorithms for same problems in [31, 32]. Problem SC216 SC235 SC252 SC265 HS006 HS007 HS047 HS050 HS051 HS077 (đ, đ) (2, 1) (3, 1) (3, 1) (4, 2) (2, 1) (2, 1) (5, 3) (5, 3) (5, 3) (5, 2) NG-NF (Algorithm 1) 7-16 30-48 36-46 3-7 5-7 8-14 72-86 14-39 9-18 16-22 NG-NF (Algorithm 2) 6-15 39-50 99-122 3-3 7-7 10-13 95-95 15-15 16-16 22-28 Acknowledgments The author would like to thank the anonymous referees for the careful reading and helpful comments, which led to an improved version of the original paper. The research is supported by the Science & Technology Program of Shanghai Maritime University (no. 20120060). References [1] R. Fletcher and S. Leyffer, “Nonlinear programming without a penalty function,” Mathematical Programming, vol. 91, no. 2, pp. 239–269, 2002. [2] R. Fletcher, S. Leyffer, and P. L. Toint, “On the global convergence of a filter-SQP algorithm,” SIAM Journal on Optimization, vol. 13, no. 1, pp. 44–59, 2002. [3] R. Fletcher, N. I. M. Gould, S. Leyffer, P. L. Toint, and A. WaĚchter, “Global convergence of a trust-region SQP-filter [4] [5] [6] [7] [8] NG-NF (Algorithm 3) 11-13 39-50 26-35 3-3 9-9 10-13 95-95 15-15 16-16 22-28 algorithm for general nonlinear programming,” SIAM Journal on Optimization, vol. 13, no. 3, pp. 635–659, 2002. S. Ulbrich, “On the superlinear local convergence of a filter-SQP method,” Mathematical Programming, vol. 100, no. 1, pp. 217– 245, 2004. A. Vardi, “A trust region algorithm for equality constrained minimization: convergence properties and implementation,” SIAM Journal on Numerical Analysis, vol. 22, no. 3, pp. 575–591, 1985. R. H. Byrd, R. B. Schnabel, and G. A. Shultz, “A trust region algorithm for nonlinearly constrained optimization,” SIAM Journal on Numerical Analysis, vol. 24, no. 5, pp. 1152–1170, 1987. E. O. Omojokun, Trust region algorithms for optimization with nonlinear equality and inequality constraints [Ph.D. thesis], University of Colorado, Boulder, Colo, USA, 1989. C. Audet and J. E. Dennis,, “A pattern search filter method for nonlinear programming without derivatives,” SIAM Journal on Optimization, vol. 14, no. 4, pp. 980–1010, 2004. Abstract and Applied Analysis [9] C. M. Chin and R. Fletcher, “On the global convergence of an SLP-filter algorithm that takes EQP steps,” Mathematical Programming, vol. 96, no. 1, pp. 161–177, 2003. [10] M. Ulbrich, S. Ulbrich, and L. N. Vicente, “A globally convergent primal-dual interior-point filter method for nonlinear programming,” Mathematical Programming, vol. 100, no. 2, pp. 379–410, 2004. [11] R. Fletcher and S. Leyffer, “A bundle filter method for nonsmooth nonlinear optimization,” Tech. Rep. NA/195, Department of Mathematics, University of Dundee, Dundee, Scotland, 1999. [12] E. Karas, A. Ribeiro, C. SagastizaĚbal, and M. Solodov, “A bundlefilter method for nonsmooth convex constrained optimization,” Mathematical Programming, vol. 116, no. 1-2, pp. 297–320, 2009. [13] N. I. M. Gould, S. Leyffer, and P. L. Toint, “A multidimensional filter algorithm for nonlinear equations and nonlinear leastsquares,” SIAM Journal on Optimization, vol. 15, no. 1, pp. 17–38, 2004. [14] N. I. M. Gould, C. Sainvitu, and P. L. Toint, “A filter-trustregion method for unconstrained optimization,” SIAM Journal on Optimization, vol. 16, no. 2, pp. 341–357, 2006. [15] A. WaĚchter and L. T. Biegler, “Line search filter methods for nonlinear programming: motivation and global convergence,” SIAM Journal on Optimization, vol. 16, no. 1, pp. 1–31, 2005. [16] A. WaĚchter and L. T. Biegler, “Line search filter methods for nonlinear programming: local convergence,” SIAM Journal on Optimization, vol. 16, no. 1, pp. 32–48, 2005. [17] L. Grippo, F. Lampariello, and S. Lucidi, “A nonmonotone line search technique for Newton’s method,” SIAM Journal on Numerical Analysis, vol. 23, no. 4, pp. 707–716, 1986. [18] Z. W. Chen and X. S. Zhang, “A nonmonotone trust-region algorithm with nonmonotone penalty parameters for constrained optimization,” Journal of Computational and Applied Mathematics, vol. 172, no. 1, pp. 7–39, 2004. [19] N. Y. Deng, Y. Xiao, and F. J. Zhou, “Nonmonotonic trust region algorithm,” Journal of Optimization Theory and Applications, vol. 76, no. 2, pp. 259–285, 1993. [20] P. L. Toint, “Non-monotone trust-region algorithms for nonlinear optimization subject to convex constraints,” Mathematical Programming, vol. 77, no. 3, pp. 69–94, 1997. [21] M. Ulbrich, “Nonmonotone trust-region methods for boundconstrained semismooth equations with applications to nonlinear mixed complementarity problems,” SIAM Journal on Optimization, vol. 11, no. 4, pp. 889–917, 2001. [22] D. C. Xu, J. Y. Han, and Z. W. Chen, “Nonmonotone trust-region method for nonlinear programming with general constraints and simple bounds,” Journal of Optimization Theory and Applications, vol. 122, no. 1, pp. 185–206, 2004. [23] D. T. Zhu, “A nonmonotonic trust region technique for nonlinear constrained optimization,” Journal of Computational Mathematics, vol. 13, no. 1, pp. 20–31, 1995. [24] Z. S. Yu and D. G. Pu, “A new nonmonotone line search technique for unconstrained optimization,” Journal of Computational and Applied Mathematics, vol. 219, no. 1, pp. 134–144, 2008. [25] Z. S. Yu, C. X. He, and Y. Tian, “Global and local convergence of a nonmonotone trust region algorithm for equality constrained optimization,” Applied Mathematical Modelling, vol. 34, no. 5, pp. 1194–1202, 2010. [26] M. Ulbrich and S. Ulbrich, “Non-monotone trust region methods for nonlinear equality constrained optimization without a 9 [27] [28] [29] [30] [31] [32] [33] penalty function,” Mathematical Programming, vol. 95, no. 1, pp. 103–135, 2003. Z. W. Chen, “A penalty-free-type nonmonotone trust-region method for nonlinear constrained optimization,” Applied Mathematics and Computation, vol. 173, no. 2, pp. 1014–1046, 2006. N. Gould and P. L. Toint, “Global convergence of a nonmonotone trust-region filter algorithm for nonlinear programming,” in Proceedings of Gainesville Conference on Multilevel Optimization, W. Hager, Ed., pp. 129–154, Kluwer, Dordrecht, The Netherlands, 2005. K. Su and D. G. Pu, “A nonmonotone filter trust region method for nonlinear constrained optimization,” Journal of Computational and Applied Mathematics, vol. 223, no. 1, pp. 230–239, 2009. M. J. D. Powell, “Convergence properties of a class of minimization algorithm,” in Nonlinear Programming 2, O. Margasarian, R. Meyer, and S. Robinson, Eds., pp. 1–27, Academic Press, New York, NY, USA, 1975. W. Hock and K. Schittkowski, Test Examples for Nonlinear Programming Codes, vol. 187 of Lecture Notes in Economics and Mathematical Systems, Springer, 1981. K. Schittkowski, More Test Examples for Nonlinear Programming Codes, vol. 282 of Lecture Notes in Economics and Mathematical Systems, Springer, Berlin, Germany, 1987. K. Su and H. An, “Global convergence of a nonmonotone filter method for equality constrained optimization,” Applied Mathematics and Computation, vol. 218, no. 18, pp. 9396–9404, 2012. Advances in Operations Research Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Advances in Decision Sciences Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Mathematical Problems in Engineering Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Algebra Hindawi Publishing Corporation http://www.hindawi.com Probability and Statistics Volume 2014 The Scientific World Journal Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Differential Equations Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Volume 2014 Submit your manuscripts at http://www.hindawi.com International Journal of Advances in Combinatorics Hindawi Publishing Corporation http://www.hindawi.com Mathematical Physics Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Complex Analysis Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Mathematics and Mathematical Sciences Journal of Hindawi Publishing Corporation http://www.hindawi.com Stochastic Analysis Abstract and Applied Analysis Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com International Journal of Mathematics Volume 2014 Volume 2014 Discrete Dynamics in Nature and Society Volume 2014 Volume 2014 Journal of Journal of Discrete Mathematics Journal of Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Applied Mathematics Journal of Function Spaces Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Optimization Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014