Hindawi Publishing Corporation Mathematical Problems in Engineering Volume 2011, Article ID 463087, 22 pages doi:10.1155/2011/463087 Research Article Global Convergence of a Nonlinear Conjugate Gradient Method Liu Jin-kui, Zou Li-min, and Song Xiao-qian School of Mathematics and Statistics, Chongqing Three Gorges University, Chongqing 404000, China Correspondence should be addressed to Liu Jin-kui, liujinkui2006@126.com Received 5 March 2011; Revised 21 May 2011; Accepted 4 June 2011 Academic Editor: Piermarco Cannarsa Copyright q 2011 Liu Jin-kui et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. A modified PRP nonlinear conjugate gradient method to solve unconstrained optimization problems is proposed. The important property of the proposed method is that the sufficient descent property is guaranteed independent of any line search. By the use of the Wolfe line search, the global convergence of the proposed method is established for nonconvex minimization. Numerical results show that the proposed method is effective and promising by comparing with the VPRP, CG-DESCENT, and DL methods. 1. Introduction The nonlinear conjugate gradient method is one of the most efficient methods in solving unconstrained optimization problems. It comprises a class of unconstrained optimization algorithms which is characterized by low memory requirements and simplicity. Consider the unconstrained optimization problem minn fx, x∈R 1.1 where f : Rn → R is continuously differentiable, and its gradient g is available. The iterates of the conjugate gradient method for solving 1.1 are given by xk1 xk αk dk , 1.2 2 Mathematical Problems in Engineering where stepsize αk is positive and computed by certain line search, and the search direction dk is defined by dk ⎧ ⎨−gk , for k 1, ⎩−g β d , k k k−1 for k ≥ 2, 1.3 where gk ∇fxk , and βk is a scalar. Some well-known conjugate gradient methods include Polak-Ribière-Polyak PRP method 1, 2 , Hestenes-Stiefel HS method 3 , Hager-Zhang HZ method 4 , and Dai-Liao DL method 5 . The parameters βk of these methods are specified as follows: βkPRP βkHS gkT gk − gk−1 , gk−1 2 gkT gk − gk−1 T , dk−1 gk − gk−1 βkHZ βkDL yk−1 − 2dk−1 gkT yk−1 T dk−1 yk−1 −t yk−1 2 T T dk−1 yk−1 gkT sk−1 T dk−1 yk−1 1.4 gk T dk−1 yk−1 , , t ≥ 0, where || · || is the Euclidean norm and yk−1 gk − gk−1 . We know that if f is a strictly convex quadratic function, the above methods are equivalent in the case that an exact line search is used. If f is nonconvex, their behaviors may be further different. In the past few years, the PRP method has been regarded as the most efficient conjugate gradient method in practical computation. One remarkable property of the PRP method is that it essentially performs a restart if a bad direction occurs see 6 . Powell 7 constructed an example which showed that the PRP method can cycle infinitely without approaching any stationary point even if an exact line search is used. This counterexample also indicates that the PRP method has a drawback that it may not globally be convergent when the objective function is nonconvex. Powell 8 suggested that the parameter βk is negative in the PRP method and defined βk as βk max 0, βkPRP . 1.5 Gilbert and Nocedal 9 considered Powell’s suggestion and proved the global convergence of the modified PRP method for nonconvex functions under the appropriate line search. In addition, there are many researches on convergence properties of the PRP method see 10– 12 . In recent years, much effort has been investigated to create new methods, which not only possess global convergence properties for general functions but also are superior to original methods from the computation point of view. For example, Yu et al. 13 proposed a Mathematical Problems in Engineering 3 new nonlinear conjugate gradient method in which the parameter βk is defined on the basic of βkPRP such as βkVPRP ⎧ 2 T ⎪ ⎪ ⎨ gk − gk gk−1 ν g T dk−1 gk−1 2 k ⎪ ⎪ ⎩0, 2 if gk > gkT gk−1 , 1.6 otherwise, where ν > 1 in this paper, we call this method as VPRP method. And they proved the global convergence of the VPRP method with the Wolfe line search. Hager and Zhang 4 discussed the global convergence of the HZ method for strong convex functions under the Wolfe line search and Goldstein line search. In order to prove the global convergence for general functions, Hager and Zhang modified the parameter βkHZ as βkMHZ max βkHZ , ηk , 1.7 −1 , dk min η, gk 1.8 where ηk η > 0. The corresponding method of 1.7 is the famous CG-DESCENT method. Dai and Liao 5 proposed a new conjugate condition, that is, dkT yk−1 −tgkT sk−1 , t ≥ 0. 1.9 Under the new conjugate condition, they proved global convergence of the DL conjugate gradient method for uniformly convex functions. According to Powell’s suggestion, Dai and Liao gave a modified parameter βk max gkT yk−1 T dk−1 yk−1 ,0 − t gkT sk−1 T dk−1 yk−1 , t ≥ 0. 1.10 The corresponding method of 1.10 is the famous DL method. Under the strong Wolfe line search, they researched the global convergence of the DL method for general functions. Zhang et al. 14 proposed a modified DL conjugate gradient method and proved its global convergence. Moreover, some researchers have been studying a new type of method called the spectral conjugate gradient method see 15–17 . This paper is organized as follows: in the next section, we propose a modified PRP method and prove its sufficient descent property. In Section 3, the global convergence of the method with the Wolfe line search is given. In Section 4, numerical results are reported. We have a conclusion in the last section. 4 Mathematical Problems in Engineering 2. Modified PRP Method In this section, we propose a modified PRP conjugate gradient method in which the parameter βk is defined on the basic of βkPRP as follows: βkMPRP ⎧ ⎪ ⎪ ⎨ 2 gk − g T gk−1 k max 0, g T dk−1 gk−1 2 k ⎪ ⎪ ⎩0, 2 2 if gk ≥ gkT gk−1 ≥ mgk , 2.1 otherwise, in which m ∈ 0, 1. We introduce the modified PRP method as follows. 2.1. Modified PRP (MPRP) Method Step 1. Set x1 ∈ Rn , ε ≥ 0, and d1 −g1 , if g1 ≤ ε, then stop. Step 2. Compute αk by some inexact line search. Step 3. Let xk1 xk αk dk , gk1 gxk1 , if ||gk1 || ≤ ε, then stop. Step 4. Compute βk1 by 2.1, and generate dk1 by 1.3. Step 5. Set k k 1, and go to Step 2. In the convergence analyses and implementations of conjugate gradient methods, one often requires the inexact line search to satisfy the Wolfe line search or the strong Wolfe line search. The Wolfe line search is to find αk such that fxk αk dk ≤ fxk δαk gkT dk , 2.2 gxk αk dk T dk ≥ σgkT dk , 2.3 where 0 < δ < σ < 1. The strong Wolfe line search consists of 2.2 and the following strengthened version of 2.3: gxk αk dk T dk ≤ −σgkT dk . 2.4 Moreover, in most references, we can see that the sufficient descent condition 2 gkT dk ≤ −cgk , c>0 2.5 is always given which plays a vital role in guaranteeing the global convergence properties of conjugate gradient methods. But, in this paper, dk can satisfy 2.5 without any line search. Theorem 2.1. Consider any method 1.2-1.3, where βk βkMPRP . If gk / 0 for all k ≥ 1, then 2 gkT dk < −gk , ∀k ≥ 1. 2.6 Mathematical Problems in Engineering 5 Proof. Multiplying 1.3 by gkT , we get 2 gkT dk −gk βkMPRP gkT dk−1 . 2.7 If βkMPRP 0, from 2.7, we know that the conclusion 2.6 holds. If βkMPRP / 0, the proof is divided into two cases in the following. Firstly, if gkT dk−1 ≤ 0, then from 2.1 and 2.7, one has gkT dk 2 gk − g T gk−1 k T 2 · gk dk−1 T max 0, gk dk−1 gk−1 2 2 gk − gkT gk−1 − gk · gkT dk−1 gk−1 2 2 2 gkT gk−1 /gk · gkT dk−1 − gkT dk−1 gk−1 2 −gk · gk−1 2 2 2 gk−1 − gkT dk−1 1 − gkT gk−1 /gk 2 −gk · gk−1 2 gk−1 2 2 2 ≤ −gk · 2 −gk < 0. gk−1 2 −gk 2.8 Secondly, if gkT dk−1 > 0, then from 2.7, we also have gkT dk 2 2 2 gk − gkT gk−1 < − gk · gkT dk−1 − gkT gk−1 ≤ −mgk . T gk dk−1 2.9 From the above, the conclusion 2.6 holds under any line search. 3. Global Convergences of the Modified PRP Method In order to prove the global convergence of the modified PRP method, we assume that the objective function fx satisfies the following assumption. Assumption H i The level set Ω {x ∈ Rn | fx ≤ fx1 } is bounded, that is, there exists a positive constant ξ > 0 such that for all x ∈ Ω, ||x|| ≤ ξ. ii In a neighborhood V of Ω, f is continuously differentiable and its gradient g is Lipchitz continuous, namely, there exists a constant L > 0 such that gx − g y ≤ Lx − y, ∀x, y ∈ V. 3.1 6 Mathematical Problems in Engineering Under these assumptions on f, there exists a constant γ > 0 such that gx ≤ γ ∀x ∈ Ω. 3.2 The conclusion of the following lemma, often called the Zoutendijk condition, is used to prove the global convergence properties of nonlinear conjugate gradient methods. It was originally given by Zoutendijk 18 . Lemma 3.1. Suppose that, Assumption H holds. Consider any iteration of 1.2-1.3, where dk satisfies gkT dk < 0 for k ∈ N and αk satisfies the Wolfe line search, then gkT dk 2 k≥1 < ∞. dk 2 3.3 Lemma 3.2. Suppose that Assumption H holds. Consider the method 1.2-1.3, where βk βkMPRP , and αk satisfies the Wolfe line search and 2.6. If there exists a constant r > 0, such that gk ≥ r, ∀k ≥ 1, 3.4 then one has uk − uk−1 2 < ∞, 3.5 k≥2 where uk dk /||dk ||. Proof. From 2.1 and 3.4, we get gkT gk−1 / 0. 3.6 By 2.6 and 3.6, we know that dk / 0 for each k. Define the quantities rk −gk , dk δk βkMPRP dk−1 dk . 3.7 By 1.3, one has uk −gk βkMPRP dk−1 dk rk δk uk−1 . dk dk 3.8 Since uk is unit vector, we get rk uk − δk uk−1 δk uk − uk−1 . 3.9 Mathematical Problems in Engineering 7 1 0.9 0.8 0.7 0.6 Y 0.5 0.4 0.3 0.2 0.1 0 0 2 4 6 8 10 X MPRP method VPRP method CG-DESCENT method DL+ method Figure 1: Performance profiles of the conjugate gradient methods with respect to CPU time. 1 0.9 0.8 0.7 0.6 Y 0.5 0.4 0.3 0.2 0.1 0 0 1 2 3 4 5 6 7 8 9 10 X MPRP method VPRP method CG-DESCENT method DL+ method Figure 2: Performance profiles with respect to the number of iterations. From δk ≥ 0 and the above equation, one has uk − uk−1 ≤ 1 δk uk − uk−1 1 δk uk − 1 δk uk−1 8 Mathematical Problems in Engineering ≤ uk − δk uk−1 δk uk − uk−1 2rk . 3.10 By 2.1, 3.4, and 3.6, one has g T gk−1 1 ≥ k 2 gk 2 3.11 > m. From 3.3, 2.6, 3.4, and 3.11, one has ⎞ 2 T g g k−1 k ⎝rk · ⎠ m2 rk ≤ gk 2 0 0 k≥1,dk / k≥1,dk / 2 ⎛ 2 gkT gk−1 0 k≥1,dk / dk 2 ≤ 0 k≥1,dk / gkT dk 2 3.12 2 dk 2 < ∞, so rk 2 < ∞. 3.13 k≥1,dk /0 By 3.10 and the above inequality, one has uk − uk−1 2 < ∞. 3.14 k≥2 Lemma 3.3. Suppose that Assumption H holds. If 3.4 holds, then βkMPRP has property (∗ ), that is, 1 there exists a constant b > 1, such that |βkMPRP | ≤ b, 2 there exists a constant λ > 0, such that ||xk − xk−1 || ≤ λ ⇒ |βkMPRP | ≤ 1/2b. Proof. From Assumption ii, we know that 3.2 holds. By 2.1, 3.2, and 3.4, one has βkMPRP gk gk−1 · gk 2γ 2 ≤ ≤ 2 b. r gk−1 2 3.15 Mathematical Problems in Engineering 9 1 0.9 0.8 0.7 0.6 Y 0.5 0.4 0.3 0.2 0.1 0 0 1 2 3 4 5 6 7 8 9 10 X MPRP method VPRP method CG-DESCENT method DL+ method Figure 3: Performance profiles with respect to the number of function evaluations. Define λ r 2 /2Lγb. If ||xk − xk−1 || ≤ λ, then from 2.1, 3.1, 3.2, and 3.4, one has βkMPRP 2 gk − g T gk−1 g T gk − gk−1 k k ≤ gk−1 2 gk−1 2 gk · gk − gk−1 γLλ 1 . ≤ ≤ 2 2b r gk−1 2 3.16 Lemma 3.4 see 19 . Suppose that Assumption H holds. Let {xk } and {dk } be generated by 1.21.3, in which αk satisfies the Wolfe line search and 2.6. If βk ≥ 0 has the property (∗ ) and 3.4 holds, then there exits λ > 0, for any Δ ∈ Z and k0 ∈ Z , for all k ≥ k0 , such that Rλk,Δ > Δ , 2 3.17 where Rλk,Δ {i ∈ Z : k ≤ i ≤ k Δ − 1, ||xi − xi−1 || ≥ λ}, |Rλk,Δ | denotes the number of the Rλk,Δ . Theorem 3.5. Suppose that Assumption H holds. Let {xk } and {dk } be generated by 1.2-1.3, in which αk satisfies the Wolfe line search and 2.6, βk βkMPRP , then one has lim infgk 0. k → ∞ 3.18 10 Mathematical Problems in Engineering Table 1: The numerical results of the modified PRP method. Problem ROSE FROTH BADSCP BADSCB BEALE HELIX BRAD GAUSS MEYER GULF BOX SING WOOD KOWOSB BD OSB1 BIGGS OSB2 JENSAM VARDIM WATSON PEN2 Dim 2 2 2 2 2 3 3 3 3 3 3 4 4 4 4 5 6 11 6 7 8 9 10 11 3 5 6 8 9 10 12 15 5 6 7 8 10 12 15 20 5 10 15 20 30 40 50 60 NI 24 11 26 11 21 25 20 3 1 1 1 67 33 57 26 1 121 341 12 13 11 12 26 NaN 4 6 5 7 7 7 7 8 59 387 1768 3934 4319 1892 1527 3001 111 185 154 178 123 147 152 163 NF 109 80 227 89 75 76 73 8 1 2 1 263 150 222 127 1 449 900 49 56 53 65 133 NaN 40 57 65 72 78 81 90 92 200 1281 5834 13373 15102 6762 5552 11308 439 845 774 989 610 700 744 813 NG 90 61 210 79 59 61 61 6 1 2 1 228 117 195 96 1 396 811 32 35 30 38 94 NaN 26 38 43 47 50 52 58 60 167 1134 5191 11920 13451 6007 4933 10107 393 752 679 889 534 617 651 720 CPU 0.3651 0.0594 0.2000 0.1085 0.1449 0.1754 0.1380 0.0164 0.0063 0.0173 0.0574 0.5000 0.2421 0.4000 0.1995 0.0157 1.0000 1.3000 0.0900 0.1872 0.1678 0.1160 0.2604 NaN 0.0135 0.0296 0.0270 0.0327 0.0647 0.0646 0.0647 0.0948 0.2000 1.4000 6.0000 14.0000 17.0000 9.0000 7.0000 19.0000 0.4000 1.5000 0.5000 0.6000 0.4000 0.5000 1.2000 0.7000 Mathematical Problems in Engineering 11 Table 1: Continued. Problem PEN1 TRIG ROSEX SINGX BV IE Dim 5 10 20 30 50 100 200 300 10 20 50 100 200 300 400 500 100 200 300 400 500 1000 1500 2000 100 200 300 400 500 1000 1500 2000 200 300 400 500 600 1000 1500 2000 200 300 400 500 600 1000 1500 2000 NI 30 88 32 73 72 29 28 27 41 56 49 61 56 52 57 53 26 26 26 26 26 26 26 26 78 79 73 89 91 93 82 80 1813 636 226 188 86 21 11 2 6 6 6 6 6 6 6 6 NF 151 415 155 350 346 189 198 201 92 136 106 137 116 106 116 109 123 123 123 123 123 123 123 123 320 335 308 367 374 385 347 341 4326 1501 516 420 190 40 20 6 13 13 13 13 13 13 13 13 NG 125 357 124 290 285 147 152 150 82 127 103 127 114 101 114 108 103 103 103 103 103 103 103 103 283 293 269 324 330 342 306 299 4063 1418 487 398 184 37 19 5 7 7 7 7 7 7 7 7 CPU 0.2742 0.9000 0.3349 0.7000 0.3000 0.2458 0.4759 1.0464 0.3817 0.5634 0.1949 0.3857 2.0205 11.6394 44.9734 89.8125 0.2323 0.2583 0.3078 0.4697 0.6781 2.4474 5.3979 9.9364 0.8000 0.8000 0.8000 1.6000 2.2000 8.0000 15.8000 28.4000 9.0000 5.4000 2.7000 3.2000 1.9000 0.9963 1.0900 0.5456 0.3063 0.6698 1.1916 1.8511 2.6615 7.3635 16.6397 29.4927 12 Mathematical Problems in Engineering Table 2: The numerical results of the VPRP method. Problem TRID ROSE FROTH BADSCP BADSCB BEALE HELIX BRAD GAUSS MEYER GULF BOX SING WOOD KOWOSB BD OSB1 BIGGS OSB2 JENSAM VARDIM WATSON Dim 200 300 400 500 600 1000 1500 2000 2 2 2 2 2 3 3 3 3 3 3 4 4 4 4 5 6 11 6 7 8 9 10 11 3 5 6 8 9 10 12 15 5 6 7 8 10 12 15 20 NI 35 36 37 35 36 35 36 37 24 12 99 13 12 74 30 3 1 1 1 101 174 71 42 1 113 264 9 11 10 17 17 6 4 6 5 7 7 7 7 8 193 342 1451 6720 NaN 3507 5271 NaN NF 81 83 83 78 80 79 84 85 111 78 394 30 48 203 100 7 1 2 1 341 482 234 161 1 375 667 33 39 42 90 124 76 40 57 65 72 78 81 90 92 535 1002 4157 19530 NaN 10432 15817 NaN NG 74 75 75 73 76 75 79 79 85 59 336 19 35 175 80 4 1 2 1 289 417 203 125 1 330 603 17 17 19 57 84 46 26 38 43 47 50 52 58 60 473 890 3678 17300 NaN 9234 14006 NaN CPU 0.3327 0.3587 0.3731 0.4935 0.6862 1.7180 4.0501 7.5866 0.1585 0.0698 0.4000 0.0448 0.0402 0.5000 0.2027 0.0085 0.0058 0.0052 0.0561 0.7000 0.6000 0.3000 0.2451 0.0063 0.8000 1.5000 0.0696 0.1137 0.0883 0.1850 0.1435 0.1111 0.0290 0.0203 0.0273 0.036 0.0715 0.0458 0.0672 0.0622 1.4000 2.2000 5.0000 20.0000 NaN 13.0000 24.0000 NaN Mathematical Problems in Engineering 13 Table 2: Continued. Problem PEN2 PEN1 TRIG ROSEX SINGX BV Dim 5 10 15 20 30 40 50 60 5 10 20 30 50 100 200 300 10 20 50 100 200 300 400 500 100 200 300 400 500 1000 1500 2000 100 200 300 400 500 1000 1500 2000 200 300 400 500 600 1000 1500 2000 NI 129 90 434 941 771 248 952 927 32 90 28 76 64 23 21 28 36 56 45 58 64 52 60 59 24 24 24 24 24 24 24 24 169 199 365 627 129 229 100 128 NaN 7278 3837 1842 898 133 19 2 NF 499 379 1467 2959 2531 978 2788 2822 151 383 160 334 344 160 174 192 82 125 93 120 135 102 132 127 111 111 111 111 111 111 111 111 562 676 1181 2025 431 788 329 428 NaN 12990 6707 3236 1562 232 36 6 NG 434 328 1298 2612 2193 860 2511 2435 124 324 121 279 280 122 128 143 71 114 85 113 128 99 125 116 85 85 85 85 85 85 85 85 483 576 1031 1756 367 675 280 363 NaN 12989 6706 3235 1561 231 35 5 CPU 1.0000 0.3000 2.2000 3.0000 3.0000 1.6000 3.0000 4.0000 0.3752 0.8000 0.1119 0.3000 0.5000 0.2212 0.4081 1.0207 0.4048 0.6413 0.3426 0.4573 2.0248 12.4258 52.5691 101.6207 0.2798 0.2808 0.3136 0.4271 0.6181 2.1821 4.7644 8.8878 0.6000 1.4000 3.7000 9.0000 2.5000 16.9000 16.0000 36.0000 NaN 55.0000 42.0000 26.0000 17.6000 6.2000 2.0484 0.5156 14 Mathematical Problems in Engineering Table 2: Continued. Problem IE TRID Dim 200 300 400 500 600 1000 1500 2000 200 300 400 500 600 1000 1500 2000 NI 7 7 7 7 7 7 7 7 33 35 35 35 37 34 37 37 NF 15 15 15 15 15 15 15 15 75 79 78 78 82 76 86 86 NG 8 8 8 8 8 8 8 8 71 75 74 74 78 72 81 80 CPU 0.3596 0.7986 1.4202 2.2147 3.1811 8.8316 19.7838 34.7553 0.3758 0.3654 0.3912 0.5188 0.7388 1.6832 4.1950 7.5255 1 0.9 0.8 0.7 0.6 Y 0.5 0.4 0.3 0.2 0.1 0 0 1 2 3 4 5 6 7 8 9 10 X MPRP method VPRP method CG-DESCENT method DL+ method Figure 4: Performance profiles with respect to the number of gradient evaluations. Proof. To obtain this result, we proceed by contradiction. Suppose that 3.18 does not hold, which means that there exists r > 0 such that gk ≥ r, so, we know that Lemmas 3.2 and 3.4 hold. for k ≥ 1, 3.19 Mathematical Problems in Engineering 15 Table 3: The numerical results of the CG-DESCENT method. Problem ROSE FROTH BADSCP BADSCB BEALE HELIX BRAD GAUSS MEYER GULF BOX SING WOOD KOWOSB BD OSB1 BIGGS OSB2 JENSAM VARDIM WATSON PEN2 Dim 2 2 2 2 2 3 3 3 3 3 3 4 4 4 4 5 6 11 6 7 8 9 10 11 3 5 6 8 9 10 12 15 5 6 7 8 10 12 15 20 5 10 15 20 30 40 50 60 NI 36 12 40 16 11 66 47 3 1 1 1 62 103 77 53 1 128 379 NaN 12 11 NaN 5 21 4 6 5 7 7 7 7 8 135 421 1822 2607 NaN 3370 5749 5902 128 147 663 734 810 1488 744 755 NF 132 64 213 101 45 179 143 10 1 2 1 198 298 222 204 1 395 915 NaN 51 50 NaN 59 148 40 57 65 72 78 81 90 92 391 1186 5278 7589 NaN 10111 17442 18524 485 571 2262 2452 2438 4502 2342 2457 NG 107 48 189 88 33 152 122 8 1 2 1 164 247 192 162 1 341 827 NaN 32 26 NaN 34 105 26 38 43 47 50 52 58 60 336 1043 4655 6716 NaN 8930 15368 16282 421 486 1996 2181 2264 3960 2056 2188 CPU 0.2371 0.0462 0.2000 0.1212 0.0405 0.4397 0.2292 0.0093 0.0085 0.0068 0.0600 0.3000 0.5000 0.4000 0.3000 0.0084 0.7000 1.2000 NaN 0.1114 0.0844 NaN 0.0421 0.1749 0.0246 0.0174 0.0266 0.0226 0.0323 0.0495 0.0610 0.0583 0.5000 1.4000 6.0000 9.0000 NaN 12.0000 27.0000 36.0000 1.1000 0.4000 2.0000 3.0000 3.0000 5.0000 2.0000 3.0000 16 Mathematical Problems in Engineering Table 3: Continued. Problem PEN1 TRIG ROSEX SINGX BV IE Dim 5 10 20 30 50 100 200 300 10 20 50 100 200 300 400 500 100 200 300 400 500 1000 1500 2000 100 200 300 400 500 1000 1500 2000 200 300 400 500 600 1000 1500 2000 200 300 400 500 600 1000 1500 2000 NI 39 94 35 76 81 31 25 26 33 60 43 53 59 51 58 51 36 34 35 31 31 37 34 32 77 54 99 79 69 101 66 69 NaN NaN 4509 1635 925 247 18 2 7 7 7 7 7 7 7 7 NF 174 404 162 347 375 184 171 186 74 145 95 116 123 109 121 110 124 125 133 121 129 142 132 128 227 172 301 245 215 322 210 210 NaN NaN 8044 2926 1605 418 38 6 15 15 15 15 15 15 15 15 NG 141 345 127 286 313 141 126 139 65 132 90 108 118 105 114 105 99 102 107 100 106 116 106 103 187 141 248 207 180 271 174 173 NaN NaN 8043 2925 1604 417 37 5 8 8 8 8 8 8 8 8 CPU 0.3298 0.8000 0.1220 0.3000 0.4000 0.2434 0.4106 1.0160 0.2781 0.5098 0.3620 0.4474 1.7555 12.6411 45.0506 89.7073 0.1230 0.1539 0.3939 0.5789 0.8736 3.5602 7.4845 12.9238 0.7000 0.2376 1.0000 1.3000 1.6000 8.6000 12.4000 23.1000 NaN NaN 63.0000 34.0000 24.5000 16.9000 3.0564 0.6883 0.3546 0.7927 1.4039 2.2022 3.1297 8.6892 19.5607 34.7423 Mathematical Problems in Engineering 17 Table 3: Continued. Problem TRID Dim 200 300 400 500 600 1000 1500 2000 NI 31 33 32 34 35 36 36 35 NF 70 73 72 76 78 81 84 80 NG 58 65 61 71 74 77 79 73 CPU 0.2691 0.2857 0.4264 0.6853 0.9484 2.6056 5.8397 9.9491 We also define uk dk /||dk ||, then for all l, k ∈ Z l ≥ k, one has xl − xk−1 l xi − xi−1 · ui−1 ik l l si−1 · uk−1 si−1 ui−1 − uk−1 , ik 3.20 ik where si−1 xi − xi−1 , that is, l si−1 · uk−1 xl − xk−1 − ik l si−1 ui−1 − uk−1 . 3.21 ik From Assumption H, we know that there exists a constant ξ > 0 such that x ≤ ξ, for x ∈ V. 3.22 From 3.21 and the above inequality, one has l l si−1 ≤ 2ξ si−1 · ui−1 − uk−1 . ik 3.23 ik Let Δ be a positive integer and Δ ∈ 8ξ/λ, 8ξ/λ 1 where λ has been defined in Lemma 3.4. From Lemma 3.2, we know that there exists k0 such that i≥k0 ui1 − ui 2 ≤ 1 . 4Δ 3.24 18 Mathematical Problems in Engineering Table 4: The numerical results of the DL method. P ROSE FROTH BADSCP BADSCB BEALE HELIX BARD GAUSS MEYER GULF BOX SING WOOD KOWOSB BD OSB1 BIGGS OSB2 JENSAM VARDIM WASTON PEN2 Dim 2 2 2 2 2 3 3 3 3 3 3 4 4 4 4 5 6 11 6 7 8 9 10 11 3 5 6 8 9 10 12 15 5 6 7 8 10 12 15 20 5 10 15 20 30 40 50 60 It 25 9 19 11 11 50 20 2 1 1 1 91 148 76 67 1 59 279 9 9 NaN 15 17 5 4 6 5 7 7 7 7 8 62 346 1247 2034 5973 3200 1865 6420 83 128 277 290 767 675 617 NaN f iter 110 65 146 56 50 144 76 5 1 2 1 295 402 236 208 1 211 699 35 35 NaN 91 136 80 40 57 65 72 78 81 90 92 208 1055 3443 6104 17601 9291 5402 18320 329 519 1040 1113 2231 1929 2223 NaN Grade iter 87 50 133 47 39 122 64 3 1 2 1 250 349 204 177 1 189 614 19 17 NaN 57 96 50 26 38 43 47 50 52 58 60 173 927 3085 5408 15699 8259 4790 16360 283 449 922 981 2040 1706 1962 NaN CPU 0.1192 0.0283 0.1003 0.0434 0.0355 0.2616 0.0824 0.0079 0.0087 0.0062 0.0601 0.5000 0.8000 0.4000 0.3000 0.0083 0.3000 1.4000 0.0570 0.1276 NaN 0.1148 0.2084 0.0882 0.0153 0.0192 0.0184 0.0219 0.0438 0.0242 0.0442 0.0535 0.2000 1.8000 4.0000 7.0000 21.0000 12.0000 8.0000 30.0000 0.7000 1.0000 0.9000 0.9000 3.0000 3.0000 2.0000 NaN Mathematical Problems in Engineering 19 Table 4: Continued. P PEN1 TRIG ROSEX SINGX BV IE Dim 5 10 20 30 50 100 200 300 10 20 50 100 200 300 400 500 100 200 300 400 500 1000 1500 2000 100 200 300 400 500 1000 1500 2000 200 300 400 500 600 1000 1500 2000 200 300 400 500 600 1000 1500 2000 It 33 78 24 73 66 22 27 29 34 59 46 55 56 52 56 54 25 25 25 25 25 25 25 25 113 117 215 113 110 137 152 110 NaN NaN 5624 3314 1618 258 18 2 6 6 6 6 6 6 6 6 f iter 165 353 119 367 356 154 178 195 82 140 94 120 116 110 115 121 110 110 110 110 110 110 110 110 384 397 735 384 367 452 494 367 NaN NaN 8458 4854 2291 312 30 6 13 13 13 13 13 13 13 13 Grade iter 139 298 94 305 293 117 136 147 71 129 88 111 111 105 113 116 87 87 87 87 87 87 87 87 329 341 634 329 313 388 420 313 NaN NaN 8457 4853 2290 311 29 5 7 7 7 7 7 7 7 7 CPU 0.2615 0.7000 0.0842 0.3000 0.6000 0.1957 0.4346 1.0408 0.2910 0.5161 0.3865 0.4625 1.6476 12.3062 44.8831 97.7043 0.0859 0.2019 0.2697 0.4169 0.5990 2.2166 4.8426 9.1263 1.0000 0.5000 2.1000 1.6000 2.2000 9.2000 22.1000 30.2000 NaN NaN 49.0000 38.0000 25.0000 8.7000 1.7251 0.5129 0.3075 0.6731 1.1939 1.8608 2.6785 7.3685 16.5794 29.5854 20 Mathematical Problems in Engineering Table 4: Continued. P TRID Dim 200 300 400 500 600 1000 1500 2000 It 33 35 36 35 37 34 37 37 f iter 75 79 80 78 82 76 86 86 Grade iter 71 75 76 73 78 72 81 80 CPU 0.2606 0.2866 0.3653 0.4857 0.6950 1.6550 4.1380 7.4702 From the Cauchy-Schwartz inequality and 3.24, for all i ∈ k, k Δ − 1 , one has ui−1 − uk−1 ≤ i−1 uj − uj−1 jk ⎛ ⎞1/2 i−1 2 ≤ i − k1/2 ⎝ uj − uj−1 ⎠ 3.25 jk ≤ Δ1/2 · 1 4Δ 1/2 1 . 2 By Lemma 3.4, we know that there exists k ≥ k0 such that Rλk,Δ > Δ . 2 3.26 It follows from 3.23, 3.25, and 3.26 that λΔ λ λ 1 kΔ−1 < Rk,Δ < si−1 ≤ 2ξ. 4 2 2 ik 3.27 From 3.27, one has Δ < 8ξ/λ, which is a contradiction with the definition of Δ. Hence, lim infgk 0, k → ∞ 3.28 which completes the proof. 4. Numerical Results In this section, we compare the modified PRP conjugate gradient method, denoted the MPRP method, to VPRP method, CG-DESCENT method, and DL method under the strong Wolfe line search about problems 20 with the given initial points and dimensions. The parameters Mathematical Problems in Engineering 21 are chosen as follows: δ 0.01, σ 0.1, v 1.25, η 0.01, and t 0.1. If ||gk || ≤ 10−6 is satisfied, we will stop the program. The program will be also stopped if the number of iteration is more than ten thousands. All codes were written in Matlab 7.0 and run on a PC with 2.0 GHz CPU processor and 512 MB memory and Windows XP operation system. The numerical results of our tests with respect to the MPRP method, VPRP method, CG-DESCENT method, and DL method are reported in Tables 1, 2, 3, 4, respectively. In the tables, the column “Problem” represents the problem’s name in 20 , and “CPU,” “NI,” “NF,” and “NG” denote the CPU time in seconds, the number of iterations, function evaluations, gradient evaluations, respectively. “Dim” denotes the dimension of the tested problem. If the limit of iteration was exceeded, the run was stopped, and this is indicated by NaN. In this paper, we will adopt the performance profiles by Dolan and Moré 21 to compare the MPRP method to the VPRP method, CG-DESCENT method, and DL method in the CPU time, the number of iterations, function evaluations, and gradient evaluations performance, respectively see Figures 1, 2, 3, 4. In figures, X τ −→ 1 size p ∈ P : log2 rp,s ≤ τ , np Y P rp,s ≤ τ : 1 ≤ s ≤ ns . 4.1 Figures 1–4 show the performance of the four methods relative to CPU time, the number of iterations, the number of function evaluations, and the number of gradient evaluations, respectively. For example, the performance profiles with respect to CPU time means that for each method, we plot the fraction P of problems for which the method is within a factor τ of the best time. The left side of the figure gives the percentage of the test problems for which a method is the fastest; the right side gives the percentage of the test problems that are successfully solved by each of the methods. The top curve is the method that solved of the most problems in a time that was within a factor τ of the best time. Obviously, Figure 1 shows that MPRP method outperforms VPRP method, CGDESCENT method, and DL method for the given test problems in the CPU time. Figures 2–4 show that the MPRP method also has the best performance with respect to the number of iterations and function and gradient evaluations since it corresponds to the top curve. So, the MPRP method is computationally efficient. 5. Conclusions We have proposed a modified PRP method on the basic of the PRP method, which can generate sufficient descent directions with inexact line search. Moreover, we proved that the proposed modified method converge globally for general nonconvex functions. The performance profiles showed that the proposed method is also very efficient. Acknowledgments The authors wish to express their heartfelt thanks to the referees and Professor Piermarco Cannarsa for their detailed and helpful suggestions for revising the paper. This work was supported by The Nature Science Foundation of Chongqing Education Committee KJ091104 and Chongqing Three Gorge University 09ZZ-060. 22 Mathematical Problems in Engineering References 1 E. Polak and G. Ribire, “Note sur la xonvergence de directions conjugees,” Rev Francaise Informat Recherche Operatinelle 3e Annee, vol. 16, pp. 35–43, 1969. 2 B. T. Polak, “The conjugate gradient method in extreme problems,” USSR Computational Mathematics and Mathematical Physics, vol. 9, pp. 94–112, 1969. 3 M. Al-Baali, “Descent property and global convergence of the Fletcher-Reeves method with inexact line search,” IMA Journal of Numerical Analysis, vol. 5, no. 1, pp. 121–124, 1985. 4 W. W. Hager and H. Zhang, “A new conjugate gradient method with guaranteed descent and an efficient line search,” SIAM Journal on Optimization, vol. 16, no. 1, pp. 170–192, 2005. 5 Y. H. Dai and L. Z. Liao, “New conjugacy conditions and related nonlinear conjugate gradient methods,” Applied Mathematics and Optimization, vol. 43, no. 1, pp. 87–101, 2001. 6 W. W. Hager and H. Zhang, “A survey of nonlinear conjugate gradient methods,” Pacific Journal of Optimization, vol. 2, no. 1, pp. 35–58, 2006. 7 M. J. D. Powell, “Nonconvex minimization calculations and the conjugate gradient method,” in Numerical Analysis (Dundee, 1983), vol. 1066 of Lecture Notes in Mathematics, pp. 122–141, Springer, Berlin, Germany, 1984. 8 M. J. D. Powell, “Convergence properties of algorithms for nonlinear optimization,” SIAM Review, vol. 28, no. 4, pp. 487–500, 1986. 9 J. C. Gilbert and J. Nocedal, “Global convergence properties of conjugate gradient methods for optimization,” SIAM Journal on Optimization, vol. 2, no. 1, pp. 21–42, 1992. 10 L. Zhang, W. Zhou, and D. Li, “A descent modified Polak-Ribière-Polyak conjugate gradient method and its global convergence,” IMA Journal of Numerical Analysis, vol. 26, no. 4, pp. 629–640, 2006. 11 Z. Wei, G. Y. Li, and L. Qi, “Global convergence of the Polak-Ribière-Polyak conjugate gradient method with an Armijo-type inexact line search for nonconvex unconstrained optimization problems,” Mathematics of Computation, vol. 77, no. 264, pp. 2173–2193, 2008. 12 L. Grippo and S. Lucidi, “A globally convergent version of the Polak-Ribière conjugate gradient method,” Mathematical Programming, vol. 78, no. 3, pp. 375–391, 1997. 13 G. Yu, Y. Zhao, and Z. Wei, “A descent nonlinear conjugate gradient method for large-scale unconstrained optimization,” Applied Mathematics and Computation, vol. 187, no. 2, pp. 636–643, 2007. 14 J. Zhang, Y. Xiao, and Z. Wei, “Nonlinear conjugate gradient methods with sufficient descent condition for large-scale unconstrained optimization,” Mathematical Problems in Engineering, vol. 2009, Article ID 243290, 16 pages, 2009. 15 Y. Xiao, Q. Wang, and D. Wang, “Notes on the Dai-Yuan-Yuan modified spectral gradient method,” Journal of Computational and Applied Mathematics, vol. 234, no. 10, pp. 2986–2992, 2010. 16 Z. Wan, Z. Yang, and Y. Wang, “New spectral PRP conjugate gradient method for unconstrained optimization,” Applied Mathematics Letters, vol. 24, no. 1, pp. 16–22, 2011. 17 L. Zhang, W. Zhou, and D. Li, “Global convergence of a modified Fletcher-Reeves conjugate gradient method with Armijo-type line search,” Numerische Mathematik, vol. 104, no. 4, pp. 561–572, 2006. 18 G. Zoutendijk, “Nonlinear programming, computational methods,” in Integer and Nonlinear Programming, J. A. badie, Ed., pp. 37–86, North-Holland, Amsterdam, Netherlands, 1970. 19 Y. H. Dai and Y. Yuan, Nonlinear Conjugate Gradient Method, vol. 10, Shanghai Scientific & Technical, Shanghai, China, 2000. 20 J. J. Moré, B. S. Garbow, and K. E. Hillstrom, “Testing unconstrained optimization software,” Association for Computing Machinery Transactions on Mathematical Software, vol. 7, no. 1, pp. 17–41, 1981. 21 E. D. Dolan and J. J. Moré, “Benchmarking optimization software with performance profiles,” Mathematical Programming, vol. 91, no. 2, pp. 201–213, 2002. Advances in Operations Research Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Advances in Decision Sciences Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Mathematical Problems in Engineering Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Algebra Hindawi Publishing Corporation http://www.hindawi.com Probability and Statistics Volume 2014 The Scientific World Journal Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Differential Equations Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Volume 2014 Submit your manuscripts at http://www.hindawi.com International Journal of Advances in Combinatorics Hindawi Publishing Corporation http://www.hindawi.com Mathematical Physics Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Journal of Complex Analysis Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 International Journal of Mathematics and Mathematical Sciences Journal of Hindawi Publishing Corporation http://www.hindawi.com Stochastic Analysis Abstract and Applied Analysis Hindawi Publishing Corporation http://www.hindawi.com Hindawi Publishing Corporation http://www.hindawi.com International Journal of Mathematics Volume 2014 Volume 2014 Discrete Dynamics in Nature and Society Volume 2014 Volume 2014 Journal of Journal of Discrete Mathematics Journal of Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Applied Mathematics Journal of Function Spaces Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Optimization Hindawi Publishing Corporation http://www.hindawi.com Volume 2014 Hindawi Publishing Corporation http://www.hindawi.com Volume 2014