Hindawi Publishing Corporation Journal of Applied Mathematics Volume 2013, Article ID 295147, 7 pages http://dx.doi.org/10.1155/2013/295147 Research Article Parallel Variable Distribution Algorithm for Constrained Optimization with Nonmonotone Technique Congying Han,1,2 Tingting Feng,2 Guoping He,2 and Tiande Guo1 1 School of Mathematical Sciences, University of the Chinese Academy of Sciences, No. 19(A), Yuquan Road, Shijingshan District, Beijing 100049, China 2 College of Information Science and Engineering, Shandong University of Science and Technology, Qingdao 266510, China Correspondence should be addressed to Congying Han; hancy@ucas.ac.cn Received 6 September 2012; Accepted 28 February 2013 Academic Editor: Ching-Jong Liao Copyright © 2013 Congying Han et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. A modified parallel variable distribution (PVD) algorithm for solving large-scale constrained optimization problems is developed, which modifies quadratic subproblem πππ at each iteration instead of the πππ0 of the SQP-type PVD algorithm proposed by C. A. SagastizaΜbal and M. V. Solodov in 2002. The algorithm can circumvent the difficulties associated with the possible inconsistency of πππ0 subproblem of the original SQP method. Moreover, we introduce a nonmonotone technique instead of the penalty function to carry out the line search procedure with more flexibly. Under appropriate conditions, the global convergence of the method is established. In the final part, parallel numerical experiments are implemented on CUDA based on GPU (Graphics Processing unit). 1. Introduction In this paper, we consider the following nonlinear programming problem: min π (π₯) , (1) s.t. π (π₯) ≤ 0, where π : π π → π and π(π₯) : π π → π π are all continuously differentiable. Suppose the feasible set π in (1) is described by a system of inequality constraints: π = {π₯ ∈ π π | ππ (π₯) ≤ 0, πΌ = {π | π = 1, 2, . . . , π}} . (2) In this paper, we give a new algorithm based on the method in [1], which partitions the problem variables π₯ ∈ π π into π blocks π₯1 , . . . , π₯π , such that π π₯ = (π₯1 , . . . , π₯π ) , π₯π ∈ π ππ , π = 1, . . . , π, ∑ππ = π. π=1 (3) And assume that π has block-separable structure. Specifically, π (π₯) = {π1 (π₯1 ) , π2 (π₯2 ) , . . . , ππ (π₯π )} , ππ (π₯π ) : π ππ σ³¨→ π ππ , π π ∈ {1, 2, . . . , π} , ∑ππ = π. (4) π=1 Parallel variable distribution (PVD) algorithm for solving optimization problems was first proposed by Ferris and Mangasarian [2], in 1994, in which the variables are distributed among π processors. Each processor has the primary responsibility for updating its block of variables while allowing the remaining secondary variables to change in a restricted fashion along some easily computable directions. The distinctive novel feature of this algorithm is the presence of the “forget-me-not” term which allows for a change in “secondary” variables. This makes PVD fundamentally different from the block Jacobi [3] and coordinate descent [4] methods. The forget-me-not approach improves robustness and accelerates convergence of the algorithm and is the key to its success. In 1997, Solodov [5] proposed useful generalizations that consist, for the general unconstrained case, of replacing exact global solution of the subproblems by 2 a certain natural sufficient descent condition, and, for the convex case, of inexact subproblem solution in the PVD algorithm. Solodov [6] proposed an algorithm applied the PVD approach to problems with general convex constraints directly use projected gradient residual function as the direction and show that the algorithm converges, provided that certain conditions are imposed on the change of secondary variables. In 2002, SagastizaΜbal and Solodov [1] proposed two new variants of PVD for the constrained case. Without assuming convexity of constraints, but assuming blockseparable structure, they showed that PVD subproblems can be solved inexactly by solving their quadratic programming approximations. This extends PVD to nonconvex (separable) feasible sets and provided a constructive practical way of solving the parallel subproblems. For inseparable constraints, but assuming convexity, they developed a PVD method based on suitable approximate projected gradient directions. The approximation criterion was based on a certain error bound result, and it was readily implementable. In 2011, Zheng et al. [7] gave a parallel SSLE algorithm, in which the PVD subproblems are solved inexactly by serial sequential linear equations, for solving large-scale constrained optimization with block-separable structure. Without assuming the convexity of constraints, the algorithm is proved to be globally convergent to a KKT point. Han et al. [8] proposed an asynchronous PVT algorithm for solving large-scale linearly constrained convex minimization problems with the idea of [9] in 2009, which based on the idea that a constrained optimization problem is equivalent to a differentiable unconstrained optimization problem by introducing the Fischer Function. And in particular, different from [9] the linear rate of convergence does not depend on the number of processors. In this paper, we use [1] as our main reference on SQPtype PVD method for problems (1) and (4). Firstly, we introduce the algorithm in [1] simply. The original problem is distributed into π parallel subproblems which treatment among π parallel processors. The algorithm of [1] may result in that the linear constraints in quadratic programming subproblems are inconsistent or the solution of quadratic programming subproblems is unbounded, so that the algorithm may fail. This drawback has been overcome by many researchers, such as [10]. In [1], exact penalty function is used as merit function to carry out the line-search. For the penalty function, as pointed out by Fletcher and Leyffer [11], the biggest drawback is that the penalty parameter estimates could be problematic to obtain. To overcome this drawback, we use a nonmonotone technique to carry out the line search instead of penalty function. Recent research [12–14] indicates that the monotone line search technique may have some drawbacks. In particular, enforcing monotonicity may considerably reduce the rate of convergence when the iteration is trapped near a narrow curved valley, which can result in very short steps or zigzagging. Therefore, it might be advantageous to allow the iterative sequence to occasionally generate points with nonmonotone objective values. Grippo et al. [12] generalized the Armijo rule and proposed a nonmonotone line search technique for Newton’s method which permits increase in function value, while retaining global convergence of the Journal of Applied Mathematics minimization algorithm. In [15], Sun et al. give several nonmonotone line search techniques, such as nonmonotone Armijo rule, nonmonotone Wolfe rule, nonmonotone F-rule, and so on. Several numerical tests show that the nonmonotone line search technique for unconstrained optimization and constrained optimization is efficient and competitive [12– 14, 16]. Recently, [17] gives a method to overcome the drawback of zigzagging with nonmonotone line search technique to determine the step length, which makes the algorithm more flexible. In this paper, we combine [1] with the ideas of [10, 17] and propose an infeasible SQP-type Parallel Variable Distribution algorithm for constrained optimization with nonmonotone technique. The paper is organized as follows. The algorithm is presented in Section 2. In Section 3, under mild assumptions, some global convergence results are proved. The numerical results are shown in Section 4. And conclusions are given in the last section. 2. A Modified SQP-Type PVD Method with Nonmonotone Technique To describe our algorithm, we first give the modified quadratic subproblems of SQP method. Given π₯π ∈ π π , we distribute it into π blocks, as the iterate π₯ππ ∈ π ππ , π = 1 . . . , π, π»π ∈ π ππ ×ππ , π = 1, 2, . . . , π denote an approximate Hessian. We use π§π ∈ π to perturb the subproblem of πππ0 of [1] and give a new subproblem πππ instead of it. Consider πππ (π₯ππ , π»ππ ) min { { { (ππ ,π§π )∈π ππ +1 { { { {s.t. { ={ { { { { { { { { 1 π§π + πlπ π»ππ ππ 2 π (5) (πππ ) ππ ≤ π§π , π ππ (π₯ππ ) + ∇ππ (π₯ππ ) ππ ≤ π§π , π ∈ πΌπ = {1, . . . , ππ } . For convenience, given π₯, the πth block of ∇π(π₯) will be denoted by ππ = (πΌππ ×ππ 0ππ ×ππ ) ∇π (π₯) . (6) In (5), πππ is ππ in step π. Note that (5) is always feasible for ππ = 0, π§π = maxπ∈πΌπ {ππ (π₯ππ ), 0}. To test whether constraints are satisfied or not, we denote the violation function β(π₯) as follows: σ΅© σ΅© β (π₯) = σ΅©σ΅©σ΅©π(π₯)+ σ΅©σ΅©σ΅© , (7) where ππ (π₯)+ = max{ππ (π₯), 0}, π ∈ πΌ, β ⋅ β denotes the Euclidean norm on π π . It is easy to see that β(π₯) = 0 if and only if π₯ is a feasible point (β(π₯) > 0 if and only if π₯ is infeasible). We describe our modified PVD algorithm as follows. Algorithm A. Consider the following. Journal of Applied Mathematics 3 Step 0. Start with any π₯0 ∈ π π . Choose parameters π’ ∈ (0, 1/2), π½ ∈ (1/2, 1), πΎ ∈ (0, 1), positive integer π > 0, and positive definite matrices π»π0 ∈ π ππ ×ππ , π ∈ {1, 2, . . . , π}. Set π = 0, π(π) = 0. Having π₯π , check a stopping criterion. If it is not satisfied, compute π₯π+1 as follows. Step 1. Parallelization. For each processor π ∈ {1, 2, . . . , π}, solve subproblem (5) to get (πππ , π§ππ ). If πππ = 0, π = 1, . . . , π stop, otherwise return to Step 2. π π»π = ( If πππ ππ π πππ = ((π1π ) , . . . , (πππ ) ) , π»1π Step 2. Compute Ψππ (π) max s.t. βπβ ≤ Δπ (12) to get πππ . Calculate πππ . Step 4. If πππ > π, then let π₯ππ+1 = π₯ππ + πππ , Δπ+1 = 2Δπ , π = π + 1, π(π) = min{π(π − 1) + 1, π} and go to Step 1. (8) .. ). . 3. The Convergence Properties π»ππ −(1/2)πππ π»π ππ , return to Step ≤ 1 and return to Step 4. 3; otherwise, set π π = π (π₯π + πππ ) ≤ max [π (π₯π−π )] − πππππ π»π ππ . 0≤π≤π(π) (9) Step 4. Set π₯π+1 = π₯π + π π ππ . Let β = max0≤π≤π(π) [β(π₯π−π )]. If β(π₯π+1 ) ≤ π½β, then set π(π + 1) = min{π(π) + 1, π}, update π»ππ to π»ππ+1 , π = 1, . . . , π, set π = π + 1, and go to Step 1. Otherwise, let π = π + 1, call Restoration Algorithm of [17] to obtain π₯ππ , let π₯π = π₯ππ , π(π) = π(π), and go to Step 1. Remark 1. In Step 3 of Algorithm A, the nonmonotone parameter π(π) satisfies 0 ≤ π (π) ≤ min {π (π − 1) + 1, π} , π ≥ 1. (10) For the convenience, we denote π(π₯π(π))=max0≤π≤π(π) [π(π₯π−π)], where π − π(π) ≤ π(π) ≤ k. In a restoration algorithm, we aim to decrease the value of β(π₯) more precisely, we will use a trust region type method to obtain β(π₯π ) → 0, π → ∞ by the help of the nonmonotone technique. Let πππ (π) = β (π₯ππ ) − β (π₯ππ + π) , +σ΅© σ΅©σ΅© π σ΅© Ψππ (π) = β (π₯ππ ) − σ΅©σ΅©σ΅©σ΅©(π (π₯ππ ) + ∇π(π₯ππ ) π) σ΅©σ΅©σ΅©σ΅© , σ΅© σ΅© (11) and πππ = πππ (π)/Ψππ (π). Algorithm B is similar to the restoration phase given by Su and Yu [17], we describe the Restoration Algorithm as follows. Algorithm B. Consider the following. To prove the global convergence of Algorithm A, we make the following assumptions. Assumptions Step 3. Choose π π which is the largest one in the sequence {1, πΎ, πΎ2 , . . .} satisfying: π (0) = 0, Step 1. If β(π₯ππ ) ≤ π½β, then π₯ππ = π₯ππ stop. Step 3. If πππ ≤ π, then let π₯ππ+1 = π₯ππ , Δπ+1 = Δπ /2, π = π + 1, π(π) = min{π(π − 1) + 1, π} and go to Step 2. Step 2. Synchronization. Set ππ = (π1π , . . . , πππ ) , Step 0. Assume π₯π0 = π₯π , Δ0 > 0, π = 0, π ∈ (0, 1), π(π) = 0. (A1) The iterate {π₯π } remains in compacted subset π ⊂ π π . (A2) The objective function π and the constraint functions ππ (π ∈ πΌ) are twice continuously differentiable on π π . (A3) For all π₯π ∈ π ππ , the set {∇ππ (π₯π ) : π ∈ πΌπ (π₯π )} is linearly independent, where πΌπ (π₯π ) = {π ∈ πΌπ : ππ (π₯π ) = Φπ (π₯π ), Φπ (π₯π ) = maxπ∈πΌπ {ππ (π₯π ), 0}}. (A4) There exist two constants 0 < π ≤ π such that πβππ β2 ≤ πππ π»ππ ππ ≤ πβππ β2 , π = 1, . . . , π for all iteration π and ππ ∈ π ππ . (A5) The solution of problem (12) satisfies +σ΅© σ΅© π σ΅©σ΅© Ψππ (π) = β (π₯ππ ) − σ΅©σ΅©σ΅©σ΅©(π (π₯ππ ) + ∇π(π₯ππ ) π) σ΅©σ΅©σ΅©σ΅© σ΅© σ΅© (13) π π π ≥ π½2 Δ min {β (π₯π ) , Δ } , where π½2 is a constant. Remark 2. Assumptions (A1) and (A2) are the standard assumptions. (A3) is the LICQ constraint qualification. (A4) plays an important role in obtaining the convergence results. (A5) is the sufficient reduction condition which guarantees the global convergence in a trust region method. Under the assumptions, π is bounded below and the gradient function ππ , π = 1, . . . , π is uniformly continuous in ππ , where π = (π1 , . . . , ππ ). Remark 3. We can use quasi-Newton BFGS methods to update π»ππ , different from [18], we can use a small modification of π¦ππ to make π»ππ reserve positive definite according to the formula (33) of [17]. 4 Journal of Applied Mathematics Similar to Lemma 1 in [19], we can get the following conclusions. Lemma 4. Suppose (A1)–(A4) hold, π»ππ , π = 1, . . . , π is a symmetric positive definite, then πππ is well defined and (πππ , π§ππ ) is the unique KKT point of (5). Furthermore, πlπ is bounded over compact subsets of ππ × ππ , where ππ is the set of symmetric positive definite ππ × ππ matrices. Together with the definination of Φπ (π₯ππ ), implies that (18) holds. (B2) Since (πππ , π§ππ ) is the solution of (5), there exists V ∈ π and π ∈ π ππ such that the KKT condition of (5), while πππ = 0, we have π Vπππ + ∇ππ (π₯ππ ) πππ = 0, ππ 1 − V − ∑π’ππ = 0, Proof. First note that the feasible set for (5) is nonempty, since (ππ , π§π ) = (0, maxπ∈πΌπ {ππ (π₯π ), 0}) is always feasible. It is clear that, if (πππ , π§ππ ) is a solution to (5) if an only if πππ solves the unconstrained problem 1 min πππ π»ππ ππ ππ ∈π ππ 2 π π (14) + max {(πππ ) ππ , max {ππ (π₯ππ ) + ∇ππ (π₯ππ ) ππ }} , π π π∈πΌπ (15) Since the function being minimized in (14) is strictly convex and radially unbounded, it follows that πππ is well defined and unique as a global minimizer for the convex problem (5) and thus unique as a KKT point for that problem. So we have π π»ππ πππ + Vπππ + ∇ππ (π₯ππ ) πππ = 0, (16) ππ (17) π=1 due to (17) and π’π ≥ 0, V ≥ 0, we have π’π , π ∈ πΌπ is bounded. Since π, ππ is twice continuously differentiable and assumption (A1) holds, for all π₯π ∈ ππ , β − (Vππ + ∇ππ (π₯ππ )π πππ β is bounded, set the maximum is π. For (16), βπ»ππ πππ β ≤ π and combining with Assumption (A4), we have πβπππ β2 ≤ ππππ π»ππ πππ ≤ βπππ ββπ»ππ πππ β ≤ πβπππ β, so that βπππ β ≤ π/π; therefore, πππ is bounded on ππ × ππ . Lemma 5. Suppose that π₯ππ ∈ π ππ , π = 1, . . . , π, definite matrix, and (πππ , π§ππ ) is the solution of have the results (B1) and (B2). π»ππ is positive πππ . Then we (B1) The following inequality holds: π§ππ ≤ Φπ (π₯ππ ) π 1 − (πππ ) π»ππ πππ , 2 −π§ππ ≤ 0, ππ (π₯ππ ) − π§ππ ≤ 0, Vπ§ππ = 0, π = 1, . . . , π, (18) where Φπ (π₯π ) = maxπ∈πΌπ {ππ (π₯π ), 0}. (B2) If πππ = 0 then π§ππ = 0, π = 1, . . . , π, and π₯π = (π₯1π , . . . , π₯ππ ) is a KKT point of problem (1). Proof. (B1) Since ππ = 0, π§π = maxπ∈πΌπ {ππ (π₯π ), 0} is a feasible point of ππl , from the optimality of (πππ , π§ππ ), we have π 1 π§ππ + (πππ ) π»ππ πππ ≤ π§π + 0 = max {ππ (π₯ππ ) , 0} . π∈πΌπ 2 π ∈ πΌπ , V ≥ 0, π’ππ [ππ (π₯ππ ) − π§ππ ] = 0, (22) (23) π’ππ ≥ 0, π’ππ = 0, π ∈ πΌπ . ∀π ∉ πΌπ0 = {π ∈ πΌπ : ππ (π₯ππ ) = Φπ (π₯ππ )} . (24) (19) (25) Hence, it follows from Assumption (A3) Together with (20) and (21) that V > 0. For (23) we have π§ππ = 0. From (22) and (24), we have π’ππ ππ (π₯ππ ) = 0, π’ππ ≥ 0, ππ (π₯ππ ) ≤ 0, 1 − V − ∑π’π = 0, (21) π=1 By the definition of Φπ (π₯ππ ) and (22), Φπ (π₯ππ ) ≤ π§ππ . Then, from (18) and πππ = 0, Φπ (π₯ππ ) ≥ π§ππ . From (24), it follows that π∈πΌπ π§π = max {(πππ ) ππ , max {ππ (π₯ππ ) + ∇ππ (π₯ππ ) ππ }} . (20) π ∈ πΌπ , π ∈ πΌπ , (26) due to (20) and set πππ = π’ππ /V, we have π πππ + ∇ππ (π₯ππ ) πππ = 0. (27) Taking into account separability of constraints π(π₯), and π let πΌ = ∪π=1 πΌπ , then πππ ππ (π₯π ) = 0, πππ ≥ 0, ππ (π₯π ) ≤ 0, π ∈ πΌ, π ∈ πΌ, (28) π ππ + ∇ππ (π₯π ) ππ = 0; that is, π₯π is a KKT point of (1). Lemma 6 (see [17, Lemma 3]). In Step 3, the line search procedure is well defined. Lemma 7 (see [17, Lemma 4]). Under Assumption (A5), the Restoration Algorithm (Algorithm B) terminates finitely. From Lemmas 6 and 7, we know that the Algorithm A and Algorithm B are well implemented. The following Lemma implies that every cluster point of {π₯π } which generated by Algorithm A is a feasible point of problem (1). Lemma 8. Let {π₯π } be an infinite sequence generated by Algorithm A, then β(π₯π ) → 0 (π → ∞). Journal of Applied Mathematics 5 Theorem 9. Suppose {π₯π } is an infinite sequence generated by Algorithm A, πππ , π = 1, . . . , π is the solution of (5), ππ = (π1π , . . . , πππ ), then limπ → ∞ βπππ β = 0, π = 1, . . . , π. ∗ Proof. By Assumption (A1), there exists a point π₯ such that π₯π → π₯∗ for π ∈ πΎ, where πΎ is an infinite index set. By Algorithm A and Lemma 8, we consider the following two possible cases. If π = 1, since {π(π)} ⊂ {π(π)}, (35) and (36) follow from (33) and (34). For a given number π > 1, we have π (π₯π(π)−π ) π ≤ π (π₯π(π(π)−π−1) ) − π π(π)−π−1 πππ(π)−π−1 π»π(π)−π−1 ππ(π)−π−1 . (37) Using the same arguments above, we obtain σ΅©σ΅© π(π)−π−1 σ΅©σ΅© σ΅©σ΅© = 0, lim σ΅©σ΅©σ΅©σ΅©ππ σ΅©σ΅© π→∞ σ΅© σ΅© Case I. πΎ0 = {π ∈ πΎ | πππ ππ ≤ −(1/2)πππ π»π ππ } is an infinite index set. In this case, we have π (π₯π + π π ππ ) ≤ π (π₯π(π) ) − π π→∞ (29) ≤ max max 0≤π≤π(π)+1 [π (π₯π+1−π )] [π (π₯π+1−π )] π ≤ π(π) − π − 1. Therefore, lim π (π₯π ) = lim π (π₯π(π)−1 ) = lim π (π₯π(π) ) . Hence for π(π) ≤ π, it holds π→∞ π π (π₯π(π) ) ≤ π (π₯π(π(π)−1) ) − π π(π)−1 πππ(π)−1 π»π(π)−1 ππ(π)−1 π π π=1 (31) Since π is bounded below, {π(π₯π(π) )} converges. Due to (31), we can obtain π lim π π(π)−1 π∑(πππ(π)−1 ) π»ππ(π)−1 πππ(π)−1 = 0. (32) π=1 By Lemma 6, there exists π > 0 such that π π(π)−1 ≥ π. By Assumption (A4), we obtain π = 1, . . . , π. (33) From the uniform continuity of π(π₯), this implies that lim π (π₯π(π)−1 ) = lim π (π₯π(π) ) . π→∞ π→∞ (34) Let π(π) = π(π + π + 2), for any given π ≥ 1, we can have σ΅©σ΅© π(π)−π σ΅©σ΅© σ΅©σ΅© = 0, lim σ΅©σ΅©σ΅©σ΅©ππ σ΅©σ΅© π→∞ σ΅© σ΅© (35) lim π (π₯π(π)−π ) = lim π (π₯π(π) ) . (36) π→∞ π→∞ π→∞ (40) π lim ∑π π πππππ π»ππ πππ = 0. π→∞ (41) π=1 Using the same arguments for deriving (33), we obtain that σ΅© σ΅© lim σ΅©σ΅©σ΅©πππ σ΅©σ΅©σ΅©σ΅© = 0, π→∞ σ΅© π = 1, . . . , π. (42) Case II. πΎ0 is a finite index set, which implies πΎ1 = {π ∈ πΎ | πππ ππ > −(1/2)πππ π»π ππ } is an infinite index set. If limπ → ∞ βπππ β = 0, π = 1, . . . , π does not hold, then there exists a positive number π > 0 and an infinite index set πΎ2 , such that βπππ β > π, for all π ∈ πΎ2 ⊂ πΎ1 . Since (πππ , π§ππ ) is the solution of πππ . By the KKT condition of (5), we have π π»ππ πππ + Vπππ + ∇ππ (π₯ππ ) πππ = 0, (43) ππ 1 − V − ∑π’π = 0, (44) ππππ πππ − π§ππ ≤ 0, (45) π=1 π π = 1, . . . , π, π→∞ (39) Then by the relation (29), we have = π (π₯π(π(π)−1) ) − π π(π)−1 π∑(πππ(π)−1 ) π»ππ(π)−1 πππ(π)−1 . σ΅© σ΅© lim σ΅©σ΅©σ΅©σ΅©πππ(π)−1 σ΅©σ΅©σ΅©σ΅© = 0, π→∞ π→∞ σ΅© Since {π(π₯π(π) )} admits a limit, by the uniform continuity of π on π, it holds = π (π₯π(π) ) . π→∞ σ΅© σ΅© lim σ΅©σ΅©σ΅©π₯π+1 − π₯π(π) σ΅©σ΅©σ΅©σ΅© = 0. (30) = max {π (π₯π(π) ) , π (π₯π+1 )} π Therefore (35) and (36) hold for any given π ≥ 1. Remark 1, we obtain π(π) = π(π + π + 2) < π + π + 2, that l(π)−π π(π)−π is π(π) − π − 1 ≤ π + 1. Since ππ(π)−π = {π1 , . . . , ππ } together with (35), we can obtain limπ → ∞ βππ(π)−π β = 0, 1 ≤ Since π(π + 1) ≤ π(π) + 1, we obtain 0≤π≤π(π+1) π→∞ Now for any π, π₯π+1 = π₯π(π) − ∑π(π)−π−1 π π(π)−π ππ(π)−π , from π=1 π=1 π (π₯π(π+1) ) = (38) lim π (π₯π(π)−π−1 ) = lim π (π₯π(π) ) . π π ππππ π»π ππ = π (π₯π(π) ) − π π π∑ππππ π»ππ πππ . π = 1, . . . , π, π’π [ππ (π₯ππ ) + ∇ππ (π₯ππ ) πππ − π§ππ ] = 0, π’π ≥ 0, π ∈ Iπ . (46) Since V ≥ 0, π’π ≥ 0 and (44), we obtain 0 ≤ π’π ≤ 1, therefore there exists ππ satisfied βπ’ππ β ≤ ππ . 6 Journal of Applied Mathematics By Lemma 8, β(π₯π ) → 0 implies βπ (π₯ππ ) → 0. Hence there exists π0 , such that for π > π0 , π ∈ πΎ2 , it holds σ΅© σ΅©2 πσ΅©σ΅©σ΅©σ΅©πππ σ΅©σ΅©σ΅©σ΅© πππ π»π ππ ππ2 π βπ (π₯π ) ≤ ≤ ≤ π π π, 2ππ 2ππ 2ππ π = 1, . . . , π. (47) Consequently, from (43), we deduce Vππππ πππ = −ππππ π»ππ πππ − ππππ ∇ππ (π₯ππ ) πππ ππ = −ππππ π»ππ πππ + ∑π’π (ππ (π₯ππ ) − π§ππ ) Table 1: Results for Algorithm A. Problem 1 π, ] ]=3 π π=6 π = 12 π = 24 π = 48 π = 96 π = 192 2 4 8 16 32 64 Iteration number Algorithm A Iteration number of πππ subproblem Running times 5 10 15 9 6 7 52 66 78 46 50 24 0.5 0.5 1 2 9 13 π=1 ππ ≤ −ππππ π»ππ πππ + ππ βπ (π₯ππ ) − ∑π’π π§ππ (48) π=1 Taking into account separability of constraints π(π₯) and π₯, we conclude π∗ + π∗ ∇π∗ = 0, ππ ≤ −ππππ π»ππ πππ + ππ βπ (π₯ππ ) − ∑π’π ππππ πππ . ππ∗π ππ (π₯∗ ) = 0, π=1 ππ ππππ πππ = (∑π’π + V) ππππ πππ ≤ 1 + ππππ π»ππ πππ 2 (49) π−1 π V ∑ ∑ ∑ π₯(π−1)V+π‘ π₯(π−1)V+π‘ π=1 π=π+1 π‘=1 V Taking into account separability of constraints π(π₯), then π πππ ππ = ∑ππππ πππ π=1 π (51) To get some insight into computational properties of our approach in Section 2, we considered the same test problems taken from [1]. Choose Problem 1 of [1] as follows: min 1 = − ππππ π»ππ πππ . 2 1 ≤ −∑ ππππ π»ππ πππ 2 π=1 π ∈ πΌ. 4. Numerical Results π=1 −πlππ π»ππ πππ ππ∗ ≥ 0, Therefore π₯∗ is a KKT point of problem (1). Together with (44), (47), and transpose ≤ −ππππ π»ππ πππ + ππ βπ (π₯ππ ) π (π₯∗ ) ≤ 0, (50) 1 = − πππ π»π ππ , 2 which contradicts the definition of πΎ1 . The proof is complete. Theorem 10. Suppose {π₯π } is an infinite sequence generated by Algorithm A and the assumptions in Theorem 9 hold. Then every cluster point of {π₯π } is a KKT point of problem (1). Proof. By Assumption (A2), there exists π₯∗ , such that π₯π → π₯∗ , π ∈ πΎ. From Lemma 8, we have β(π₯π ) → 0, π ∈ πΎ, which means that π₯∗ is a feasible point. From Theorem 9, we have limπ → ∞ βπππ β = 0, π = 1, . . . , π. By Lemma 5, πππ → 0 implies that π§ππ → 0, π = 1, . . . , π, that is, ππ∗ = 0, π = 1, . . . , π, π§π∗ = 0, π = 1, . . . , π. Compare the KKT condition of (5) with the KKT condition of problem (1), and let π∗ = (π1∗ , . . . ππ∗ ), ππ∗ = (π’π∗ /V, π ∈ π πΌπ ), πΌ = ∪π=1 πΌπ . s.t. 2 − 1 = 0, ∑π₯(π−1)V+π‘ (52) π = 1, . . . , π. π‘=1 In our test, we only verify our convergence theory without comparing with serial algorithms. In the future, we will propose the convergent rate of the parallel algorithm. Not having access to a parallel computer, we have carried out on Graphics Processing Unit (GPU) to solve them and implement the algorithm on Compute Unified Device Architecture (CUDA) [20]. However, the input and output are implemented by CPU. And all the subproblems are solved on blocks which are constructed into many threads. If there are π blocks in GPU, which is equivalent to π processors in a parallel computer. Many threads of the block can achieve a large number of vector calculations in current block. We have implemented Algorithm A in C language. All codes were run under Visual Studio 2008 and Cuda 4.0 on the DELL Precision T7400. In Table 1, we report the number of iterations and the running times for Algorithms A on Problem (52). The results in Table 1 confirm that the proposed approach certainly makes sense. We solve the subproblem (5) which is a positive semidefinite quadratic programs by the modified interior-point methods. And the Restoration Algorithm of [17] is solved by Trust-region Approach combined with local pattern search methods. All algorithms are coded by C language without using any function from optimization Toolbox; hence the results are not the best. Journal of Applied Mathematics In Table 1, π denotes the dimensions of test problem and ] denotes the variable number of one block. There is less iteration number of πππ subproblem with more numbers of blocks for high-dimensional problem, which shows the algorithm is reliable. 7 [9] [10] 5. Conclusion The PVD algorithm which was proposed in 1994 is used to solve unconstrained optimization problems or has a special case of convex block-separable constraints. In 1997, M. V. Solodov proposed useful generalizations that consist, for the general unconstrained case, of replacing exact global solution of the subproblems by a certain natural sufficient descent condition, and, for the convex case, of inexact subproblem solution in the PVD algorithm. M. V. Solodov proposed a PVD approach in 1998 which applied to problems with general convex constraints directly and show that the algorithm converges. In 2002, C. A. SagastizaΜbal et al. propose two variants of PVD for the constrained case: without assuming convexity of constraints, but assuming block-separable structure and for inseparable constraints, but assuming convexity. In this paper, we propose the modified algorithm of [1] used to the general constraints but with block-separable structure. In a word, the algorithm above is a special structure of the objective function or constraints with a special structure. In the further, we will study the parallel algorithm with general inseparable constraints. Acknowledgments This paper was supported by National Natural Science Foundation of China (11101420, 71271204) and other foundations (2010BSE06047). The authors would like to thank Dr. Jiang Zhipeng for valuable work on numerical experiments. References [1] C. A. SagastizaΜbal and M. V. Solodov, “Parallel variable distribution for constrained optimization,” Computational Optimization and Applications, vol. 22, no. 1, pp. 111–131, 2002. [2] M. C. Ferris and O. L. Mangasarian, “Parallel constraint distribution,” SIAM Journal on Optimization, vol. 1, no. 1, pp. 487–500, 1994. [3] D. P. Bertsekas and J. N. Tsitsiklis, Parallel and Distributed Computation, Prentice Hall, Englewood Cliffs, NJ, USA, 1989. [4] P. Tseng, “Dual coordinate ascent methods for non-strictly convex minimization,” Mathematical Programming, vol. 59, no. 2, pp. 231–247, 1993. [5] M. V. Solodov, “New inexact parallel variable distribution algorithms,” Computational Optimization and Applications, vol. 7, no. 2, pp. 165–182, 1997. [6] M. V. Solodov, “On the convergence of constrained parallel variable distribution algorithms,” SIAM Journal on Optimization, vol. 8, no. 1, pp. 187–196, 1998. [7] F. Zheng, C. Han, and Y. Wang, “Parallel SSLE algorithm for large scale constrained optimization,” Applied Mathematics and Computation, vol. 217, no. 12, pp. 5377–5384, 2011. [8] C. Han, Y. Wang, and G. He, “On the convergence of asynchronous parallel algorithm for large-scale linearly constrained [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] minimization problem,” Applied Mathematics and Computation, vol. 211, no. 2, pp. 434–441, 2009. M. Fukushima, “Parallel variable transformation in unconstrained optimization,” SIAM Journal on Optimization, vol. 8, no. 3, pp. 658–672, 1998. J. Mo, K. Zhang, and Z. Wei, “A variant of SQP method for inequality constrained optimization and its global convergence,” Journal of Computational and Applied Mathematics, vol. 197, no. 1, pp. 270–281, 2006. R. Fletcher and S. Leyffer, “Nonlinear programming without a penalty function,” Mathematical Programming, vol. 91, pp. 239– 269, 2002. L. Grippo, F. Lampariello, and S. Lucidi, “A nonmonotone line search technique for Newton’s method,” SIAM Journal on Numerical Analysis, vol. 23, no. 4, pp. 707–716, 1986. L. Grippo, F. Lampariello, and S. Lucidi, “A truncated Newton method with nonmonotone line search for unconstrained optimization,” Journal of Optimization Theory and Applications, vol. 60, no. 3, pp. 401–419, 1989. G. Liu, J. Han, and D. Sun, “Global convergence of the BFGS algorithm with nonmonotone linesearch,” Optimization, vol. 34, no. 2, pp. 147–159, 1995. W. Sun, J. Han, and J. Sun, “Global convergence of nonmonotone descent methods for unconstrained optimization problems,” Journal of Computational and Applied Mathematics, vol. 146, no. 1, pp. 89–98, 2002. P. L. Toint, “An assessment of nonmonotone linesearch techniques for unconstrained optimization,” SIAM Journal on Scientific Computing, vol. 17, no. 3, pp. 725–739, 1996. K. Su and Z. Yu, “A modified SQP method with nonmonotone technique and its global convergence,” Computers & Mathematics with Applications, vol. 57, no. 2, pp. 240–247, 2009. Y.-X. Yuan and W. Sun, Optimization Theory and Methods, Springer, New York, NY, USA, 2006. C. T. Lawrence and A. L. Tits, “A computationally efficient feasible sequential quadratic programming algorithm,” SIAM Journal on Optimization, vol. 11, no. 4, pp. 1092–1118, 2001. I. Buck, Parallel Programming with CUDA, 2008, http://gpgpu .org/sc2008. Advances in Operations Research Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Advances in Decision Sciences Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Mathematical Problems in Engineering Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Algebra Hindawi Publishing Corporation http://www.hindawi.com Probability and Statistics Volume 2014 The Scientific World Journal Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Differential Equations Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Volume 2014 Submit your manuscripts at http://www.hindawi.com International Journal of Advances in Combinatorics Hindawi Publishing Corporation http://www.hindawi.com Mathematical Physics Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Complex Analysis Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Mathematics and Mathematical Sciences Journal of Hindawi Publishing Corporation http://www.hindawi.com Stochastic Analysis Abstract and Applied Analysis Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com International Journal of Mathematics Volume 2014 Volume 2014 Discrete Dynamics in Nature and Society Volume 2014 Volume 2014 Journal of Journal of Discrete Mathematics Journal of Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Applied Mathematics Journal of Function Spaces Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Optimization Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014