Appl. Math. J. Chinese Univ. Ser. B 2006, 21(3): 289-301 PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE METHOD Pu Dingguo 1 Zhou Yan1,2 Abstract. In this paper, a QP-free feasible method with piecewise NCP functions is proposed for nonlinear inequality constrained optimization problems. The new NCP functions are piecewise linear-rational, regular pseudo-smooth and have nice properties. This method is based on the solutions of linear systems of equation reformulation of KKT optimality conditions, by using the piecewise NCP functions. This method is implementable and globally convergent without assuming the strict complementarity condition, the isolatedness of accumulation points. Furthermore, the gradients of active constraints are not requested to be linearly independent. The submatrix which may be obtained by quasi-Newton methods, is not requested to be uniformly positive definite. Preliminary numerical results indicate that this new QP-free method is quite promising. §1 Introduction Consider the constrained nonlinear optimization problem (NLP) min f (x), x ∈ Rn , s.t. G(x) ≤ 0, (1) where f : Rn → R and G(x) = (g1 (x), g2 (x), . . . , gm (x))T : Rn → Rm are Lipchitz continuously differentiable functions. We denote by D = {x ∈ Rn |G(x) < 0} and D̄ = cl(D) the strictly feasible set and the feasible set of Problem (NLP), respectively. The Lagrangian function associated with Problem (NLP) is the function L(x, λ) = f (x) + λT G(x), (2) where λ = (λ1 , λ2 , · · · , λm )T ∈ Rm is the multiplier vector. For simplicity, we use (x, λ) to denote the column vector (xT , λT )T . Received:2005-10-20. MR Subject Classification: 90C30,65K10. Keywords: constrained optimization, semismooth, nonlinear complementarity, convergence. Supported by the Natural Science Foundation of China (10371089, 10571137). 290 Appl. Math. J. Chinese Univ. Ser. B Vol. 21, No. 3 A Karush-Kuhn-Tucker (KKT) point (x̄, λ̄) ∈ Rn ×Rm is a point that satisfies the necessary optimality conditions for Problem (NLP), ∇x L(x̄, λ̄) = 0, G(x̄) ≤ 0, λ̄ ≥ 0, λ̄i gi (x̄) = 0, (3) where 1 ≤ i ≤ m. We also say x̄ is a KKT point if there exists a λ̄ such that (x̄, λ̄) satisfy (3). Finding KKT points for Problem (NLP) can be equivalently reformulated as solving the mixed nonlinear complementarity of Problem (NCP) in (3), Problem (NCP) has attracted much attention due to its various applications, see [1, 2]. One method to solve the nonlinear complementarity problem is to construct a Newton method for solving a system of nonlinear equations,see [3, 4]. Qi H D and Qi L Q[5] proposed a new QP-free method which ensures the strict feasibility of all iterates. Their work is based on the Fischer-Burmeister NCP function. They proved the global convergence without isolatedness of accumulation point and the strict complementarity condition. They also proved the superlinear convergence under mild conditions. Stimulated by the progress in the above aspects, in this paper, a QP free feasible method with piecewise NCP functions is proposed for the minimization of a smooth function subject to smooth inequality constraints. The new NCP functions are piecewise linear-rational, regular pseudo-smooth and have nice properties. This method is based on the solutions of linear systems of equation reformulation of KKT optimality conditions, by using the piecewise NCP functions. The method is an iterative method in which, locally, the iteration can be viewed as a perturbation of a Newton-quasi Newton iteration on the primal variables for the solution of equalities in KKT first order conditions of optimality. This method ensures the feasibility of all iterations. We modify the Qi H D and Qi L Q’s[5] method slightly to obtain the local convergence under some weaker conditions. In particular, this method is implementable and globally convergent without assuming the strict complementarity condition, the isolatedness of accumulation points. Furthermore, the gradients of active constraints are not requested to be linearly independent. The submatrix which may be obtained by quasi Newton methods, is not requested to be uniformly positive definite. We also prove that the method has superlinear convergence rate under some mild conditions. Preliminary numerical results indicate that this smoothing QP-free infeasible method is quite promising. Definition 1.1. NCP pair and SNCP pair We call a pair (a, b) ∈ R2 to be an NCP pair if a ≥ 0, b ≥ 0 and ab = 0; and call (a, b) to be an SNCP pair if (a, b) is an NCP pair and a2 + b2 6= 0. Definition 1.2. NCP function. A function φ : R2 → R is called an NCP function if φ(a, b) = 0, if and only if (a, b) is an NCP pair. The following definition is presented and its properties are discussed by Qi L Q[5] . Definition 1.3. Pseudo-smooth function. Let ψ : R2 → R be a strongly semismooth function. Denote Pψ as the set of points where ψ takes zero. Let Eψ be the extreme point set of Pψ . We say that ψ is a Pseudo-smooth function if it is smooth everywhere in R2 \Eψ . Two most famous NCP functions are the min function and the Fischer-Burmeister NCP function[6] . However, the min function may not be differentiable at infinite points, and the Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... 291 inexact generalized Newton direction with this NCP function may not be a descent direction. The Fischer-Burmeister NCP function is irrational. This makes some originally linear terms, such as the Lagrangian multipliers in KKT system for the variational inequality problem or the constrained nonlinear programming problems, to become nonlinear after reformulation, see [7]. In this paper, we present a new class of interesting piecewise linear NCP functions ψ. The 3-piecewise linear NCP function is as follows. 2 3a − a /b if b ≥ a > 0, or 3b > −a ≥ 0, 2 ψ(a, b) = (4) 3b − b /a if a > b > 0 or 3a > −b ≥ 0, 9a + 9b if 0 ≥ a and − a ≥ 3b, or − 3a ≤ b ≤ 0. If (a, b) 6= (0, 0), then à ! 3 − 2a/b a2 /b2 à ! b2 /a2 ∇ψ(a, b) = 3 − 2b/a à ! 9 9 if b ≥ a > 0, or 3b > −a ≥ 0, if a > b > 0 or 3a > −b ≥ 0, (5) if 0 ≥ a and − a ≥ 3b, or − 3a ≤ b ≤ 0, and (à Aψ = ∂B ψ(0, 0) = 3 − 2t t2 ! ) (à : −3 ≤ t ≤ 1 ∪ t2 3 − 2t ! ) : −3 ≤ t ≤ 1 . (6) The 4-piecewise linear NCP function ψ is as follows. ψ(0, 0) = 0 and 2 k a 2kb − b2 /a ψ(a, b) = 2k 2 a + 2kb + b2 /a 2 k a + 4kb if if if if b > k|a|, a ≥ |b|/k and a > 0, a ≤ −|b|/k and a < 0, b < −k|a|. (7) If (a, b) 6= (0, 0), then à ! k2 0 à ! 2 2 b /a 2k − 2b/a à ! ∇ψ(a, b) = 2k − b2 /a2 2k + 2b/a à ! 2 k 4k and if b > k|a|, if a ≥ |b|/k and a > 0, (8) if a ≤ −|b|/k and a < 0, if b < −k|a|, 292 Appl. Math. J. Chinese Univ. Ser. B (à Aψ = ∂B ψ(0, 0) = k 2 t2 2k − 2t ! ) (à : |t| ≤ 1 ∪ 2 − k 2 t2 2k − 2t Vol. 21, No. 3 ! ) : |t| ≤ 1 . (9) In this paper, we use the 3-piecewise linear NCP function ψ(a, b). It is easy to check the following Proposition. Proposition 1.1. For the new function ψ(a, b) the following holds. 1. ψ(a, b) = 0 ⇐⇒ a ≥ 0, b ≥ 0, ab = 0; 2. the square of ψ is continuously differentiable; 3. ψ is twice continuously differentiable everywhere except at the origin, but it is strongly semismooth at the origin and is a pseudo-smooth NCP function; 4. for any (α, β) ∈ ∂B ψ(a, b), (a, b) 6= (0, 0), or any (α, β) ∈ ∂B ψ(0, 0), α2 + β 2 ≥ 1 > 0. To construct the semismooth equation Φ(x, Λ) = 0 , we use the form Φ(x, Λ) = 0 which is T equivalently reformulated as KKT point conditions. Let Φ(x, Λ) = (φ1 (x, Λ) · · · φn+m (x, Λ)) , where φi (x, Λ) = ∇xi L(x, Λ), 1 ≤ i ≤ n, and φi (x, Λ) = ψ((−gj (x)), λj ), n + 1 ≤ i = n + j ≤ n + m. If 1 ≤ i ≤ n, or if (gj (x), λj ) 6= (0, 0) and n + 1 ≤ i = j + n ≤ n + m, then φi (x, Λ) is continuously differentiable at (x, Λ) ∈ Rn+m . If (gj (x), λj ) = (0, 0) and n + 1 ≤ i = j + n ≤ n + m, then φi (x, Λ) is strongly semismooth at (x, Λ) ∈ Rn+m . Given any s ∈ Rn+m , s 6= 0, there is a directional derivative φ0i ((x, Λ); s) of φi (x, Λ) at (x, Λ) in the direction s ∈ Rn+m such that, for α > 0, φi ((x, Λ) + αs) = φi (x, Λ) + αφ0i ((x, Λ); s) + o(α), (10) If we use piecewise linear NCP functions, the generalized Newton direction dk is a descent direction of Θk = kΦ(z k )k2 at xk (Φ(x, Λ). We consider that ψ(a, b) is the new NCP function (4). In this case, for n + 1 ≤ i = j + n ≤ n + m, we havef or( − g(x), 0) 6= (0, 0) à ! −(3 + 2g (x))∇g (x)/λ j i j (λ /g (x))2 ej à j j ! −(λj /gj (x))2 ∇gi (x) ∇φi = (3 + 2λj /gj (x))ej à ! −9∇g (x) i 9e j if λj ≥ −gj (x) > 0 or 3λj > gj (x) ≥ 0, if − gj (x) > λj > 0 or − 3gj (x) > −λj ≥ 0, (11) if 0 ≥ −gj (x) and gj (x) ≥ 3λj , or 3gj (x) ≤ λj ≤ 0, where ej = (0, · · · , 0, 1, 0 · · · , 0)T ∈ Rm is the jth column of the unit matrix, its jth element is 1, and other elements are 0. If gj (x) = 0, λj = 0, and n + 1 ≤ i = j + n ≤ n + m, then φi (x, Λ) is semismooth and directionally differentiable at (x, Λ). We have Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... (à Aψ = ∂B ψ(0, 0) ! ) −(3 − 2t)∇gi (x) = : −3 ≤ t ≤ 1 ∪ t2 ej (à ! ) −t2 ∇gi (x) : −3 ≤ t ≤ 1 . (3 − 2t)ej 293 (12) We define the index sets I0 and I1 as follows: I1 (x, Λ) = {i|(gj (x), λj ) 6= (0, 0); I0 (x, Λ) = {i|(λj , gj (x)) = (0, 0)}. (13) The paper is organized as follows. In Section 2 we propose a QP free feasible method. In Section 3 we show that the algorithm is well defined. In Section 4 and Section 5 we discuss the conditions of global convergence and superlinear convergence of the algorithm, respectively. In Section 6 we give a brief discussion and some numerical tests. §2 Algorithm Let (ξj (x, µ), γj (x, µ)) = (−1, 1) if j ∈ I0 , otherwise let (ξj (x, µ), γj (x, µ)) = ∇ψ(a, b)|a=−gj (x),b=µj . We have ξj (x, µ) < 0 γj (x, µ)) > 0, −(3 + 2gj (x)/µj ), if µj ≥ −gj (x) > 0 or 3µj > gj (x) ≥ 0, ξj (x, µ) = −(µj /gj (x))2 , if − gj (x) > µj > 0 or − 3gj (x) > −µj ≥ 0, −9, if 0 ≥ −gj (x) and gj (x) ≥ 3µj or 3gj (x) ≤ µj ≤ 0, (14) and 2 (µj /gj (x)) ej , if µj ≥ −gj (x) > 0 or 3µj > gj (x) ≥ 0, γ(x, µ) = (3 + 2µj /gj (x)), if − gj (x) > µj > 0 or − 3gj (x) > −µj ≥ 0, 9, if 0 ≥ −gj (x) and gj (x) ≥ 3µj or 3gj (x) ≤ µj ≤ 0, q In the following algorithm, let ξjk = ξj (xk , µk ) and γjk = γj (xk , µk ), ηjk = 2γjk , à k V = k V11 k V21 k V12 k V22 ! à = H k + ck1 In diag(ξ k )(∇Gk )T ∇Gk diag(η k ) (15) ! , (16) where In is the n order unit matrix, ck1 = c1 min{1, kΦ̄k kν }, Φ̄k = Φ(xk , λ̄k ), λ̄k is obtained in Algorithm 1, c1 ∈ (0, 1), diag(ξ k ) or diag(η k ) denotes the diagonal matrix whose jth diagonal element is ξj (xk , µk ) or ηj (xk , µk ), respectively. Algorithm 2.1 Step 0. Initialization. Give an initial guess (x0 , µ0 ), x0 ∈ D, µ0 ≥ 0, λ̄0 = µ0 , c1 ∈ (0, 1), τ ∈ (0, 1), ν > 1, µ̄ ≥ µ0 > 0, κ ∈ (0, 1), θ ∈ (0, 1). Give a symmetric positive definite matrix H 0 . Denote ∇Gk = ∇G(xk ), V k = V (xk , µk ), f k = f (xk ) and so on. Step 1. Compute. 294 Appl. Math. J. Chinese Univ. Ser. B Compute dk0 and λk0 by solving the following linear system in (d, λ): à ! à ! k d −∇f Vk = . λ 0 Vol. 21, No. 3 (17) If dk0 = 0, then stop. Otherwise, compute dk1 and λk1 by solving the following linear system in (d, λ): ! à ! à d −∇f k k . (18) V = 3 diag(ξ k )(λk0 λ −) k0 k2 where λk0 and λk2 by solving the following linear system in (d, λ): − = min{λ , 0}. Compute d à ! à ! d −∇f k k V = , (19) k0 3 λ diag(ξ k )(λ− ) + kdk1 kν diag(ξ k )e where e = (1, . . . , 1)T ∈ Rm . Let à ! à ! à ! dk dk1 dk2 k k =b +ρ , λk λk1 λk2 (20) where bk = (1 − ρk ) and ρk = (θ − 1) (dk1 )T ∇f k ¯P ¯ ¯ m ¯ k1 ν . 1 + ¯ j=1 λk0 j ¯ kd k (21) ˆ the solution of the least square problem in d. Compute a correction d, min dT Hd, s.t. gi (xk + dk ) + dT ∇gik = −ψ k for any i ∈ Ik , (22) where Ik = {i|gik ≥ −λki } and ½ ½¯ ¯ k ν ψ = max kd k , max ¯¯ k i∈Ik ¯κ ¾ ¾ ¯ ξik k 2 ¯ − 1¯ kd k . −ηik λki (23) If (22) has no solution or if kdˆk k ≥ kdk k, set kdˆk k = 0. Step 2. Line search. Let tk = τ j , where j is the smallest non-negative integer such that f (xk + tk dk + (tk )2 dˆk ) ≤ f k + θtk (dk )T ∇f k (24) gi (xk + tk dk + (tk )2 dˆk ) < 0, 1 ≤ i ≤ m. (25) and Step 3. Update. Set xk+1 = xk + tk dk + (tk )2 dˆk . Set λ̄k+1 = min{λk0 , µ̄e}, Φ̄k+1 = Φ(xk+1 , λ̄k+1 ) and µk+1 = min{max{λk0 , kdk ke}, µ̄e}. If Φ̄k+1 = 0 or Φ(xk+1 , µk+1 ) = 0 then stop; otherwise update H k and obtain a symmetric positive definite matrix H k+1 . Set k = k + 1. Go to Step 1. Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... 295 §3 Implement of algorithm In this section, we assume that the following assumptions A1-A3 hold. A1 The strictly feasible set D is nonempty. The level set S = {x|f (x) ≤ f (x0 ) and x ∈ D̄} is bounded. A2 f and gi are Lipschitz continuously differentiable, and for all y, z ∈ Rn+m , kL(y) − L(z)k ≤ c2 ky − zk. A3 H k is positive definite and there exists a positive number m1 such that 0 < dT H k d ≤ m1 kdk2 for all d ∈ Rn , d 6= 0. Lemma 3.1. If Φ̄k 6= 0, then V k is nonsingular. Furthermore, assume that (x∗ , µ∗ ) is an accumulation point of {(xk , µk )}, (xk(i) , µk(i) ) → (x∗ , µ∗ ), Φ̄k(i) → Φ∗ and V k(i) → V ∗ . If k(i) −1 ∗ Φ∗ 6= 0, then k(V à !) k is bounded and V is nonsingular. u Proof. If V k = 0 for some (u, v) ∈ Rn+m , where u = (u1 . . . , un )T , v = (v1 . . . , vm )T , v then we have (H k + ck1 In )u + ∇Gk v = 0, (26) and diag(ξ k )(∇Gk )T u + (diag(η k ))v = 0. (27) Assume Φ̄k 6= 0. Obviously, ck1 6= 0 and, by the definitions of ξjk and ηjk , ξjk < 0 and ηjk > 0, j = 1, 2, · · · , m. Thus, diag(η k ) is nonsingular. We have v = −(diag(η k ))−1 diag(ξ k )(∇Gk )T u. (28) Putting (28) into (26), we have uT (H k + ck1 In )u − uT ∇Gk diag(ξ k )(diag(η k ))−1 (∇Gk )T u = 0. (29) uT (H k + ck1 In )u = 0 and u = 0 are implied by the fact that H k + ck1 In is positive definite and −∇Gk diag(ξ k )(diag(η k ))−1 (∇Gk )T is positive semi-definite, then v = 0 by (27). The first part of this lemma holds. k(i) On the other hand, without loss of generality we may assume that c1 → c∗1 6= 0, diag(ξ k(i) ) → diag(ξ ∗ ), diag(η k(i) ) → diag(η ∗ ) and H k(i) → H ∗ . We know ηj∗ > 0 for all j = 1, 2, · · · , m. H k(i) → H ∗ implies that H ∗ is positive semi-definite and H ∗ + c∗j In is positive definite. By replacing index k by ∗ in the above proof, it is easy to check that V ∗ is nonsingular. Assumption V k(i) → V ∗ implies that k(V k(i) )−1 k is bounded. This lemma holds. If Φ̄k = 0 or Φ(xk , µk ) = 0 , then (xk , λ̄k ) or (xk , µk ) is a KKT point of Problem (NLP). Without loss of generality, in the sequel we may assume that Φ̄k 6= 0 and Φ(xk , µk ) 6= 0 for all k. Because V k is nonsingular, (17) or (18) always has a unique solution. V k is nonsingular and Ak = (V k )−1 exists. Let à k A = H k + ck1 In diag(ξ k )(∇Gk )T ∇Gk diag(η k ) !−1 à = Ak11 Ak21 Ak12 Ak22 ! . (30) 296 Appl. Math. J. Chinese Univ. Ser. B Vol. 21, No. 3 By calculating directly we have Ak11 = (H k + ck1 In )−1 + (H k + ck1 In )−1 (∇Gk )(Qk )−1 diag(ξ k )(∇Gk )T (H k + ck1 In )−1 ; Ak12 = −(H k + ck1 In )−1 (∇Gk )(Qk )−1 ; Ak21 = −(Qk )−1 diag(ξ k )(∇Gk )T (H k + ck1 In )−1 ; Ak22 = (Qk )−1 . where Qk = diag(η k ) − diag(ξ k )(∇Gk )T (H k + ck1 In )−1 (∇Gk ). Lemma 3.2. If Φ̄k 6= 0, then dk0 = 0 if and only if ∇f (xk ) = 0, and dk0 = 0 implies λk0 = 0 and (xk , λk0 ) is a KKT point of Problem (NLP). Proof. If ∇f (xk ) = 0, then dk0 = 0 and λk0 = 0 by (17). If dk0 = 0, then (17) implies ∇Gk λk0 = −∇f (xk ) and diag(η k )λk0 = 0. So, λk0 = 0 and ∇f (xk ) = 0. Without loss of generality, we assume that the algorithm never terminates at any k, i.e., dk0 6= 0 for all k in the remainder part of this paper. Lemma 3.3. If dk0 6= 0, then 1. ck1 kdk0 k2 ≤ (dk0 )T (H k + c̄k1 In )dk0 ≤ −(dk0 )T ∇f k . P 4 2. (dk1 )T ∇f k = (dk0 )T ∇f k − i:λk0 <0 (λk0 i ) . i k T k k1 T k 3. (d ) ∇f ≤ θ(d ) ∇f . Proof. (17) implies (H k + ck1 In )dk0 + ∇Gk λk0 = −∇f k (31) diag(ξ k )(∇Gk )T dk0 + (diag(η k ))λk0 = 0. (32) λk0 = −(diag(η k ))−1 diag(ξ k )(∇Gk )T dk0 . (33) and We have Putting (33) into (31), we have −(dk0 )T ∇f k = (dk0 )T ((H k dk0 + ck1 In ) + ∇Gk λk0 ) = (dk0 )T (H k + ck1 In )dk0 −(dk0 )T ∇Gk (diag(η k ))−1 diag(ξ k )(∇Gk )T dk0 . (34) (dk0 )T ∇Gk (diag(η k ))−1 diag(ξ k )(∇Gk )T dk0 ≤ 0 implies ck1 kdk0 k2 ≤ (dk0 )T (H k dk0 + ck1 In )dk0 ≤ −(dk0 )T ∇f k . (35) The first part of the lemma holds. (17) and (30) imply (dk0 )T = −Ak11 ∇f k , λk0 = −Ak21 ∇f k . (36) Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... 297 The property of the matrix implies (Qk )−1 diag(ξ k ) = ((diag(ξ k ))−1 Qk )−1 = {(diag(ξ k ))−1 [diag(η k ) − diag(ξ k )(∇Gk )T (H k + ck1 In )−1 (∇Gk )]}−1 = {[diag(η k ) − (∇Gk )T (H k + ck1 In )−1 (∇Gk )diag(ξ k )](diag(ξ k ))−1 }−1 = [(Qk )T (diag(ξ k ))−1 ]−1 = diag(ξ k )((Qk )T )−1 . (37) (18), (30) and (37) imply Ak12 diag(ξ k ) = (Ak21 )T and 3 (dk1 )T ∇f k = −(Ak11 ∇f k )T ∇f k − [(Ak12 )T diag(ξ k )]T (∇f k )(λk0 −) X 4 = (dk0 )T ∇f k − (λk0 i ) . (38) i:λk0 <0 i The second part of this lemma holds. Finally, (19)-(21) and (38) imply (dk2 − dk1 )T ∇f k = kdk1 kν [Ak12 diag(ξ k )e]T ∇f k = kdk1 kν m X λk0 j=1 and (dk )T ∇f k = (1 − ρk )(dk1 )T ∇f k + ρk (dk2 )T ∇f k ≤ θ(dk1 )T ∇f k . (39) This lemma holds. It is easy to prove the following (see [5]) Lemma 3.4. If dk0 6= 0, then there is a t̄ such that, for all t ∈ (0, t̄), (24) and (25) are satisfied. Proof. If dk0 6= 0, then by the continuous differentiability of f , we have f k − f (xk + tdk + t2 dˆk ) ≥ −t(dk )T ∇f k + O(t2 ). (40) (40), gik < 0 and the continuous differentiability of gi imply that there is a t̄ > 0 such that, for any 0 < t ≤ t̄, (24) and (25) are satisfied. This lemma holds. Lemmas 3.1-3.4 show that Algorithm 2.1 can be implemented. §4 Convergence In this section we assume that assumptions A1-A3 hold. Lemma 4.1. Assume xk(i) → x∗ and Φ̄k(i) > ε > 0 for some ε, then the sequences of {(dk(i)0 , λk(i)0 )}, {(dk(i)1 , λk(i)1 )} and {(dk(i)2 , λk(i)2 )} are all bounded on k = 0, 1, · · ·. Proof. If xk(i) → x∗ and Φ̄k(i) > ε > 0 for some ε, then the matrix sequence {(V k(i) )−1 } is proved to be uniformly bounded in Lemma 3.1. {xk(i) } is bounded due to the assumption A3. The solvability of system (17) means that {(dk(i)0 , λk(i)0 )} is bounded, which implies the boundedness of {dk(i)1 } of the right-hand side of (18). Hence {(dk(i)1 , λk(i)1 )} is also bounded. 298 Appl. Math. J. Chinese Univ. Ser. B Vol. 21, No. 3 Finally, the boundedness of {dk(i)1 , λk(i)1 )} implies the boundedness of the right-hand side of (19). Hence {(dk(i)2 , λk(i)2 )} is also bounded. Lemma 4.2. Assume xk(i) → x∗ and Φ̄k(i) > ε > 0 for some ε. There is a c3 > 0 such that, for all k = 1, 2, · · ·, kdk(i) − dk(i)1 k ≤ c3 kdk(i)0 k. Proof. It is from the Lemma 3.1 that there exists a c3 > 0 such that, for all k = 0, 1, · · ·, c3 ≥ 2mρk k(V k(i) )−1 k. Let ∆dk(i) = dk(i) − dk(i)1 and ∆λk(i) = λk(i) − λk(i)1 , then by (18)-(20), (∆xk(i) , ∆λk(i) ) is the solution of à ! à ! k(i) ∆d 0 V k(i) = . ∆λk(i) −ρk kdk(i)0 kν diag(ξ k )e (41) It is easy to see that k(∆xk(i) , ∆λk(i) )k ≤ c3 kdk(i)0 kν The lemma holds. Lemma 4.3. Assume xk(i) → x∗ and λk(i) → λ∗ , 1. if dk(i) → 0, then λ∗j ≥ 0 for any 1 ≤ j ≤ m; 2. if dk(i)0 → 0, then x∗ is a KKT point of Problem (NLP); 3. if dk(i) → 0 and Φ̄k(i) > ε > 0 for some ε, then x∗ is a KKT point of Problem (NLP). Proof. It follows from Lemma 3.3 that X k(i)0 4 k(i) (λi ) . (42) (dk(i) )T ∇f k(i) ≤ −c1 θkdk(i)0 k2 − θ k(i)0 i:λi <0 Hence {dk(i) } → 0 implies X k(i)0 i:λi k(i)0 4 (λi ) → 0 and λ∗j ≥ 0, 1 ≤ j ≤ m. (43) <0 The first part of this lemma holds. Because {λ̄k(i) } ≤ µ̄ and µk(i) are bounded, there is an accumulation point λ̄∗ of {λ̄k(i) }. Without loss of generality, we may assume that ck(i) → c∗ , µk(i) → µ∗ and λ̄k(i) → λ̄∗ . (42) implies that, for any accumulation point λ∗ of {λk(i) }, λ∗i ≥ 0, 1 ≤ i ≤ m. Taking limitation in both sides of (17), and noting dk(i)0 → 0, we obtain (λ∗ )T ∇G∗ = −∇f ∗ and diag(η ∗ )λ∗ = 0. If −gi∗ > 0, for some 1 ≤ i ≤ n, then −ηi∗ + c∗i ≥ δ > 0 and λ∗i = 0, that is, for any 1 ≤ i ≤ m, gi∗ λ∗i = 0. The second part of this lemma holds. If dk(i) → 0 and Φ̄k(i) > ε > 0 for some ε, then (42) implies dk(i)0 → 0, so x∗ is a KKT point of the Problem (NLP). This lemma holds. By using Lemma 4.3, the proof of the following lemma is the same as that of Lemma 3.8 in [3]. Lemma 4.4. Assume xk(i) → x∗ and Φ̄k(i) > ε > 0 for some ε. If dk(i)−1 → 0, then (x∗ , λ∗ ) is a KKT point of Problem (NLP), where λ∗ is an accumulation point of {λk(i) }. Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... 299 The following result is the same as that in Lemma 3.6 of [5]. Lemma 4.5. Assume xk(i) → x∗ and Φ̄k(i) > ε > 0 for some ε. If lim inf{kdk(i)−1k k} > 0, then (x∗ , λ∗ ) is a KKT point of Problem (NLP), where λ∗ is an accumulation point of {λk(i) }. The following global convergence theorem holds. Theorem 4.1. If x∗ is a limit point of {xk }, then x∗ is a KKT point of Problem (NLP). Proof. Assume (xk(i) , λ̄k(i) ) → (x∗ , λ̄∗ ). If Φ̄∗ = 0, then (x∗ , λ̄∗ ) is a KKT point of Problem (NLP). If Φ̄∗ 6= 0, then λk(i) is bounded by Lemma 4.1. It follows from Lemmas 4.2-4.5 that (x∗ , λ∗ ) is a KKT point, where λ∗ is an accumulation point of {λk(i) }. The proof of this theorem is completed. §5 Superlinear convergence Let I2 (x, λ) = {i|gi (x) = 0, λi > 0} and X(x, λ) = {d|dT ∇gi (x) ≥ 0, i ∈ I2 (x, λ); }. We need the following conception and assumptions for the superlinear convergence. Definition 5.1 A point (x, λ) is said to satisfy the strong second-order sufficient condition for Problem(NLP) if it satisfies the first-order KKT conditions (3) and if dT V d > 0 for all d ∈ X(x, λ), d 6= 0 and any V ∈ ∂B Φ(x, λ). Notice that the strong second-order sufficient condition implies that x is a strict local minimum of Problem (NLP) (see [8]). A4 {∇gi (x∗ )} are linear independent, where i ∈ I(x∗ ) = {i|gi (x∗ ) = 0}, x∗ is an accumulation point of {xk } and a KKT point of Problem (NLP) as well. A5 H k is uniformly positive definite and there exist two positive numbers m1 and m2 such that 0 < m2 kdk2 ≤ dT H k d ≤ m1 kdk2 for all d ∈ Rn and all k. A6 The strong second-order sufficient condition for Problem(NLP) holds at each KKT point (x∗ , λ∗ ). A7 The strict complementarity condition holds at each KKT point (x∗ , µ∗ ). A8 The sequence of {H k } satisfies kP k (H k − ∇2x L(xk , µk ))dk1 k → 0, kdk1 k (44) where P k = I − N k ((N k )T N k )−1 N k )T and N k = (∇gik ), i ∈ I k = {i|gik = 0}. Assumption A7 implies that Φ is continuously differentiable at each KKT point (x∗ , µ∗ ). Like the proof of Lemma 3.1, we get Lemma 5.1. Assume A4 and A5 hold, then {k(V k )−1 k} and {k(V̂ k )−1 k} are bounded. Furthermore, if V ∗ is an accumulation matrix of {V k }, then V ∗ is nonsingular. Lemma 5.1 and the proof of Lemma 4.1 imply Lemma 5.2. Assume A4 and A5 hold. The sequences of {(dk(i)0 , λk(i)0 )}, {(dk(i)1 , λk(i)1 )} and {(dk(i)2 , λk(i)2 )} are all bounded on k = 0, 1, . . .. A9 λk0 < µ̄ and λ̄k = λk0 for sufficiently large k. It is easy to check that limk→∞ Φ(xk , λ̄k ) = limk→∞ Φ(xk , λk0 ) = 0. (44) is equivalent to the following k(P k (H k + ck1 ) − ∇2x L(xk , µk ))dk1 k → 0. (45) kdk1 k 300 Appl. Math. J. Chinese Univ. Ser. B Vol. 21, No. 3 We have kckj k/kΦ̄k k → 0 and c̄k1 /kΦ̄k k → 0 as k → ∞, respectively. Theorem 5.1. Assume A1-A3 and A6 hold. If (x∗ , λ∗ ) is an accumulation point of {(xk , λk0 )}, then 1. (x∗ , λ∗ ) is a KKT point of Problem (NLP), 2.(xk , λk0 ) → (x∗ , λ∗ ), 3. dk0 → 0, dk1 → 0 and dk2 → 0. The proof of Theorem 5.1 is as the same as that of Theorem 3.7, Lemma 3.1 and Corollary 3.3 in [5]. Assume A1 and A3-A9 hold. The proof of Lemma 5.2 and Theorem 5.1 is the same as that of Lemma 4.6 and Theorem 4.9 in [5]. Lemma 5.3. For k large enough the step tk = 1 is accepted by the line search. Theorem 5.2. Let Algorithm 2.1 be implemented to generate a sequence {(xk , λk )}, and (x∗ , λ∗ ) be an accumulation point of {(xk , λk )}, then (x∗ , λ∗ ) is an KKT point of Problem (NLP), and (xk , λk ) converges to (x∗ , λ∗ ) superlinearly. §6 Discussion and numerical tests In this section, we give some numerical results of Algorithm 2.1 for constrained optimization problems, the details about implementation are described as follows. 1. Termination criterion. We use the termination condition on Φ and ε. Because Φ(x, µ) = 0 if and only if (x, µ) is a KKT point, the termination criterion is kΦk k ≤ 10−5 . 2. Update H k . H k is updated by BFGS method. In particular, we set H k+1 = H k − where ( k y = and H k sk (sk )T H k y k (y k )T + , (sk )T H k sk (sk )T y k ŷ k θk ŷ k + (1 − θk )H k sk (sk )T ŷ k ≥ 0.2(sk )T H k sk , otherwise, k k+1 − xk , s =x k ŷ = ∇f (xk ) − ∇f (xk+1 ) + (∇g(xk+1 ) − ∇g(xk ))µk , k θ = 0.8(sk )T H k sk /((sk )T H k sk − (sk )T y k ). 3. The test problems are chosen from [9]. 4. The parameters are set as follows: c1 = 0.1, τ = 0.5, ν > 2, µ̄ = 105 , κ = 0.9 and θ = 0.8. In Table 1 which present the results of the numerical experiments, we use the following notations: Feasible: the method in this paper, Problem: the number of problems in [9], x0 : the starting vector, It: the number of iterations, kΦk: the value of kΦ(·)k at the final iterate (xk , µk ), Pu Dingguo,et al. PIECEWISE LINEAR NCP FUNCTION FOR QP FREE FEASIBLE... 301 ε: the value of ε at the final iteration (xk , µk ), FV: the objective function value at the final iteration. Table 1 Problem 4 5 12 24 30 35 0 x (1.125,0.125) (0,0) (0,0) (1,0.5) (1,1,1) (0,1.05,2.9) Feasible It kΦk 4 1.0e-08 8 6.25e-06 21 4.25e-06 26 2.25e-06 21 9.05e-06 20 1.35e-06 FV 2.6667e+00 -1.9132e+00 1.1132e+00 -1.1543e+00 1.0000e+00 1.1111e-01 From Table 1 we can see that the numerical results indicate that this method is quite promising. References 1 Ferris M C, Pang J S. Engineering and economic applications of complementarity problems, SIAM Review, 39 (1997), 669-713. 2 Harker P T, Pang J S. Finite-dimensional variational and nonlinear complementarity problems: a survey of theory, algorithm and applications, Mathematical Programming, 48 (1990), 161-220. 3 Panier E R, Tits A L, Herskovits J N. A QP-free, globally, locally superlinear convergent method for the inequality constrained optimization problems, SIAM Journal on Control and Optimization, 36 (1988) 788-811. 4 Pu D G, Zhou Y, Zhang H Y. A QP free feasible method, Journal of Computational Mathematics, 2004, 22: 651-660. 5 Qi H D, Qi L Q. A New QP-free, globally V, locally superlinear convergent feasible method for the solution of inequality constrained optimization problems, SIAM Journal on Optimization, 11 (2000) 113-132. 6 Fischer A. A special Newton-type optimization method, Optimization, 24 (1992), 269-284. 7 Qi L Q, Jiang H. Semismooth Karush-Kuhn-Tuchker equations and convergence anaylsie of Newton and quasi-Newton methods for solving these equations, Mathematics of Operations Research, 22(1997), 301-325. 8 Qi L Q. Convergence analysis of some method for solving nonsmooth equations, Mathematics of Operations Research, 18(1993), 227-243. 9 Hock W, Schittkowski K. Test Example for Nonlinear Programming Codes, Lecture Notes in Econom. and Math. Systems 187, Berlin: Springer-Verlag, 1981. 1 Department of Applied Mathematics, Tongji University, Shanghai, 200092, China. 2 Department of Management Science and Engineering, Qingdao University, Qingdao, 266071, China.