Literature DB >> 28529434

A modified three-term PRP conjugate gradient algorithm for optimization models.

Yanlin Wu1.   

Abstract

The nonlinear conjugate gradient (CG) algorithm is a very effective method for optimization, especially for large-scale problems, because of its low memory requirement and simplicity. Zhang et al. (IMA J. Numer. Anal. 26:629-649, 2006) firstly propose a three-term CG algorithm based on the well known Polak-Ribière-Polyak (PRP) formula for unconstrained optimization, where their method has the sufficient descent property without any line search technique. They proved the global convergence of the Armijo line search but this fails for the Wolfe line search technique. Inspired by their method, we will make a further study and give a modified three-term PRP CG algorithm. The presented method possesses the following features: (1) The sufficient descent property also holds without any line search technique; (2) the trust region property of the search direction is automatically satisfied; (3) the steplengh is bounded from below; (4) the global convergence will be established under the Wolfe line search. Numerical results show that the new algorithm is more effective than that of the normal method.

Entities:  

Keywords:  conjugate gradient; sufficient descent; trust region

Year:  2017        PMID: 28529434      PMCID: PMC5415590          DOI: 10.1186/s13660-017-1373-4

Source DB:  PubMed          Journal:  J Inequal Appl        ISSN: 1025-5834            Impact factor:   2.491


Introduction

We consider the optimization models defined by where the function is continuously differentiable. There exist many similar professional fields of science that can revert to the above optimization models (see, e.g., [2-21]). The CG method has the following iterative formula for (1.1): where is the kth iterate point, the steplength is , and the search direction is designed by where is the gradient and is a scalar. At present, there are many well-known CG formulas (see [22-46]) and their applications (see, e.g., [47-50]), where one of the most efficient formulas is the PRP [34, 51] defined by where is the gradient, , and is the Euclidian norm. The PRP method is very efficient as regards numerical performance, but it fails as regards the global convergence for the general functions under Wolfe line search technique and this is a still open problem; many scholars want to solve it. It is worth noting that a recent work of Yuan et al. [52] proved the global convergence of PRP method under a modified Wolfe line search technique for general functions. Al-Baali [53], Gilbert and Nocedal [54], Toouati-Ahmed and Storey [55], and Hu and Storey [56] hinted that the sufficient descent property may be crucial for the global convergence of the conjugate gradient methods including the PRP method. Considering the above suggestions, Zhang, Zhou, and Li [1] firstly gave a three-term PRP formula where . It is not difficult to deduce that holds for all k, which implies that the sufficient descent property is satisfied. Zhang et al. proved that the three-term PRP method has global convergence under Armijo line search technique for general functions but this fails for the Wolfe line search. The reason may be the trust region feature of the search direction that cannot be satisfied for this method. In order to overcome this drawback, we will propose a modified three-term PRP formula that will have not only the sufficient descent property but also the trust region feature. In the next section, a modified three-term PRP formula is given and the new algorithm is stated. The sufficient descent property, the trust region feature, and the global convergence of the new method are established in Section 3. Numerical results are reported in the last section.

The modified PRP formula and algorithm

Motivated by the above observation, the modified three-term PRP formula is where , , and are constants. It is easy to see that the difference between (1.5) and (2.1) is the denominator of the second and the third terms. This is a little change that will guarantee another good property for (2.1) and impel the global convergence for Wolfe conditions.

Algorithm 1

New three-term PRP CG algorithm (NTT-PRP-CG-A) Initial given parameters: , , , , , . Let and . , stop. Get stepsize by the following Wolfe line search rules: and Let . If the condition holds, stop the program. Calculate the search direction by (2.1). Set and go to Step 2.

The sufficient descent property, the trust region feature, and the global convergence

It has been proved that, even for the function ( is a constant) and the strong Wolfe conditions, the PRP conjugate gradient method may not yield a descent direction for an unsuitable choice (see [24] for details). An interesting feature of the new three-term CG method is that the given search direction is sufficiently descent.

Lemma 3.1

The search direction is defined by (2.1) and it satisfies and for all , where is a constant.

Proof

For , it is easy to get and , so (3.1) is true and (3.2) holds with . If , by (2.1), we have Then (3.1) is satisfied. By (2.1) again, we obtain where the last inequality follows from . Thus (3.2) holds for all with . The proof is complete. □

Remark

(1) Equation (3.1) is the sufficient descent property and the inequality (3.2) is the trust region feature. And these two relations are verifiable without any other conditions. (2) Equations (3.1) and (2.2) imply that the sequence generated by Algorithm 1 is descent, namely holds for all k. To establish the global convergence of Algorithm 1, the normal conditions are needed.

Assumption A

The defined level set is bounded with given point . The function f has a lower bound and it is differentiable. The gradient g is Lipschitz continuous where a constant.

Lemma 3.2

Suppose that Assumption A holds and NTT-PRP-CG-A generates the sequence . Then there exists a constant such that Using (3.5) and (2.3) generate where the last equality follows from (3.1). By (3.2), we get Setting completes the proof. □ The above lemma shows that the steplengh has a lower bound, which is helpful for the global convergence of Algorithm 1.

Theorem 3.1

Let the conditions of Lemma 3.2 hold and be generated by NTT-PRP-CG-A. Thus we get By (2.2), (3.1), and (3.6), we have Summing the above inequality from to ∞, we have which means that The proof is complete. □

Numerical results and discussion

This section will report the numerical experiment of the NTT-PRP-CG-A and the algorithm of Zhang et al. [1] (called Norm-PRP-A), where the Norm-PRP-A is the Step 4 of Algorithm 1 that is replaced by: Calculate the search direction by (1.5). Since the method is based on the search direction (1.5), we only compare the numerical results between the new algorithm and the Norm-PRP-A. The codes of these two algorithms are written by Matlab and the problems are chosen from [57, 58] with given initial points. The tested problems are listed in Table 1. The parameters are , , , , . The program uses the Himmelblau rule: Set if , otherwise set . The program stops if or hold, where we choose and . For the choice of the stepsize in (2.2) and (2.3), the largest cycle number is 10 and the current stepsize is accepted. The dimensions of the test problems accord to large-scale variables with 3,000, 12,000, and 30,000. The runtime environment is MATLAB R2010b and run on a PC with CPU Intel Pentium(R) Dual-Core CPU at 3.20 GHz, 2.00 GB of RAM, and the Windows 7 operating system.
Table 1

Test problems

No. Problems \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\boldsymbol{x_{0}}$\end{document}x0
1Extended Freudenstein and Roth function[0.5,−2,…,0.5,−2]
2Extended trigonometric function[0.2,0.2,…,0.2]
3Extended Rosenbrock function[−1.2,1,−1.2,1,…,−1.2,1]
4Extended White and Holst function[−1.2,1,−1.2,1,…,−1.2,1]
5Extended Beale function[1,0.8,…,1,0.8]
6Extended penalty function[1,2,3,…,n]
7Perturbed quadratic function[0.5,0.5,…,0.5]
8Raydan 1 function[1,1,…,1]
9Raydan 2 function[1,1,…,1]
10Diagonal 1 function[1/n,1/n,…,1/n]
11Diagonal 2 function[1/1,1/2,…,1/n]
12Diagonal 3 function[1,1,…,1]
13Hager function[1,1,…,1]
14Generalized tridiagonal 1 function[2,2,…,2]
15Extended tridiagonal 1 function[2,2,…,2]
16Extended three exponential terms function[0.1,0.1,…,0.1]
17Generalized tridiagonal 2 function[−1,−1,…,−1,−1]
18Diagonal 4 function[1,1,…,1,1]
19Diagonal 5 function[1.1,1.1,…,1.1]
20Extended Himmelblau function[1,1,…,1]
21Generalized PSC1 function[3,0.1,…,3,0.1]
22Extended PSC1 function[3,0.1,…,3,0.1]
23Extended Powell function[3,−1,0,1,…]
24Extended block diagonal BD1 function[0.1,0.1,…,0.1]
25Extended Maratos function[1.1,0.1,…,1.1,0.1]
26Extended Cliff function[0,−1,…,0,−1]
27Quadratic diagonal perturbed function[0.5,0.5,…,0.5]
28Extended Wood function[−3,−1,−3,−1,…,−3,−1]
29Extended Hiebert function[0,0,…,0]
30Quadratic QF1 function[1,1,…,1]
31Extended quadratic penalty QP1 function[1,1,…,1]
32Extended quadratic penalty QP2 function[1,1,…,1]
33Quadratic QF2 function[0.5,0.5,…,0.5]
34Extended EP1 function[1.5.,1.5.,…,1.5]
35Extended tridiagonal-2 function[1,1,…,1]
36BDQRTIC function (CUTE)[1,1,…,1]
37TRIDIA function (CUTE)[1,1,…,1]
38ARWHEAD function (CUTE)[1,1,…,1]
39NONDIA (Shanno-78) function (CUTE)[−1,−1,…,−1]
40NONDQUAR function (CUTE)[1,−1,1,−1,…,1,−1]
41DQDRTIC function (CUTEr)[3,3,3...,3]
42EG2 function (CUTE)[1,1,1...,1]
43DIXMAANA function (CUTE)[2,2,2,…,2]
44DIXMAANB function (CUTE)[2,2,2,…,2]
45DIXMAANC function (CUTE)[2,2,2,…,2]
46DIXMAANE function (CUTE)[2,2,2,…,2]
47Partial perturbed quadratic function[0.5,0.5,…,0.5]
48Broyden tridiagonal function[−1,−1,…,−1]
49Almost perturbed quadratic function[0.5,0.5,…,0.5]
50Tridiagonal perturbed quadratic function[0.5,0.5,…,0.5]
51EDENSCH function (CUTE)[0,0,…,0]
52VARDIM function (CUTE)[1 − 1/n,1 − 2/n,…,1 − n/n]
53STAIRCASE S1 function[1,1,…,1]
54LIARWHD function (CUTEr)[4,4,…,4]
55DIAGONAL 6 function[1,1,…,1]
56DIXON3DQ function (CUTE)[−1,−1,…,−1]
57DIXMAANF function (CUTE)[2,2,2,…,2]
58DIXMAANG function (CUTE)[2,2,2,…,2]
59DIXMAANH function (CUTE)[2,2,2,…,2]
60DIXMAANI function (CUTE)[2,2,2,…,2]
61DIXMAANJ function (CUTE)[2,2,2,…,2]
62DIXMAANK function (CUTE)[2,2,2,…,2]
63DIXMAANL function (CUTE)[2,2,2,…,2]
64DIXMAAND function (CUTE)[2,2,2,…,2]
65ENGVAL1 function (CUTE)[2,2,2,…,2]
66FLETCHCR function (CUTE)[0,0,…,0]
67COSINE function (CUTE)[1,1,…,1]
68Extended DENSCHNB function (CUTE)[1,1,…,1]
69DENSCHNF function (CUTEr)[2,0,2,0,…,2,0]
70SINQUAD function (CUTE)[0.1,0.1,…,0.1]
71BIGGSB1 function (CUTE)[0,0,…,0]
72Partial perturbed quadratic PPQ2 function[0.5,0.5,…,0.5]
73Scaled quadratic SQ1 function[1,2,…,n]
74Scaled quadratic SQ2 function[1,2,…,n]
Test problems Table 2 report the test numerical results of the NTT-PRP-CG-A and the Norm-PRP-A, and we notate:
Table 2

Numerical results

No. Dimension NTT-PRP-CG-A Norm-PRP-A
Ni Nfg CPU time Ni Nfg CPU time
13,00015430.46800331920.546004
12,00015430.842405561581.778411
30,00015431.482009361132.730018
23,000571310.374402551260.374402
12,000631441.138807621420.920406
30,000661523.08882661522.511616
33,000541860.1248011173750.202801
12,000672330.2340011444790.514803
30,000732380.5304031595221.62241
43,000591980.2964022075950.936006
12,000341390.7332052648014.305628
30,000742564.1184262286188.907657
53,00023680.093601391060.124801
12,00023690.265202391090.390003
30,00021640.826805471351.279208
63,000801850.124801801850.093601
12,0001032320.4056031032320.343202
30,0001022351.2168081022350.998406
73,0001,0002,0021.0452073579430.421203
12,0001,0002,0023.166828352,2572.808018
30,0001,0002,0029.7812631,0002,7799.734462
83,00021470.046819460.0312
12,00020440.09360119460.093601
30,00020440.29640219460.265202
93,00012260.031212260.0312
12,00012260.046812260.0624
30,00012260.20280112260.156001
103,0002130.03122130.0312
12,0002130.1248012130.093601
30,0002130.3120022130.280802
113,000811940.171601241010.0624
12,000912470.76440515590.202801
30,00011350.43680313500.280802
123,00017360.046814330.0624
12,00019400.17160114330.124801
30,00019400.49920314330.343202
133,00023860.09360122840.078
12,000421110.452403421110.468003
30,0002130.3588022130.327602
143,0006150.7176056150.733205
12,0006157.0044455135.709637
30,0003814.2584913813.587687
153,00038851.794011661763.04202
12,0004110217.9245156016928.09578
30,0004411475.39528368194120.245571
163,00020420.062420420
12,00024500.17160124500.156001
30,00024500.48360324500.436803
173,00024550.15600131710.218401
12,00033730.76440529740.717605
30,000481033.04201930811.996813
183,0003100.015613430.0312
12,0003100.031213430.0156
30,0003100.031214470.124801
193,000390390
12,000390.0468390.0312
30,000390.124801390.124801
203,00033820.031226740.0312
12,00011610.06245350.0312
30,0005350.09360120670.218401
213,00025590.09360127630.0624
12,00027630.24960226600.187201
30,00025580.53040327630.530403
223,0006310.03127420
12,0006310.06245210.0624
30,0006310.2184015210.124801
233,0001343830.6708043349861.52881
12,0001474162.6520174521,3097.73765
30,0001143305.30403429185412.776482
243,00028900.0624501260.109201
12,000311080.249602601460.405603
30,00028970.686404671601.170007
253,00028560.031228560.0312
12,0007160.01562317740.748805
30,0007160.03122137742.028013
263,000651520.124801651520.124801
12,000721660.514803721660.468003
30,000791801.51321791801.341609
273,00031940.06241043270.156001
12,000431370.1872012026550.639604
30,0001043291.1544073841,2314.024826
283,000401240.046831760.0312
12,00031910.12480138950.124801
30,000401070.54600332780.265202
293,0004190.03121002870.124801
12,0004190.0156842400.312002
30,0004190.093601932640.842405
303,0001,0002,0020.8424054461,2050.436803
12,0001,0002,0022.6364177542,0102.074813
30,0001,0002,0028.3304531,0002,7218.065252
313,00029660.046829660.0624
12,00034780.15600134780.156001
30,00034780.42120334780.452403
323,000481000.093601481000.093601
12,00037800.28080237800.234001
30,00036800.78000536800.670804
333,000370370
12,000250250.0312
30,000250.0312250
343,000480.0312480.0312
12,0007140.06247140.0312
30,00010200.15600110200.124801
353,00012240.031212240
12,00021420.09360121420.093601
30,0004100.0936014100.0312
363,00014481.138807451483.244821
12,0008286.91084412036995.831414
30,000175555.427155162483488.922734
373,0007761,5590.7332051,0002,6881.107607
12,0001,0002,0063.3228211,0002,7333.556823
30,0001,0002,0119.8280635061,3784.960832
383,0009300.031227810.0312
12,00010320.046821600.140401
30,00011340.14040124690.312002
393,00026520.062426520
12,00029580.09360129580.093601
30,00023460.18720123460.171601
403,0005541,3325.8812381,0002,85611.013671
12,0001,0002,22839.7334551,0002,89243.352678
30,0001,0002,247100.7454461,0002,866108.186694
413,00027680.078491330.0312
12,00028690.093601501360.124801
30,00037910.390002391010.374402
423,0006240.03126240
12,0006240.06246240.0624
30,0006240.1872016240.156001
433,00028600.20280128600.218401
12,00030640.93600630640.858005
30,00032682.52721631662.230814
443,00046960.35880246960.296402
12,000491021.51321491021.404009
30,000521084.024826521083.728424
453,00019440.20280119440.124801
12,00020460.60840420460.577204
30,00020461.5444120461.48201
463,0001172440.9204061082960.967206
12,0001653405.1168331203264.009226
30,00019540015.67810112634110.576868
473,00027668.2992534410212.963683
12,000318793.74100149141150.150963
30,000691821,163.84546852561,490.683156
483,00032741.76281127631.310408
12,0005010330.154993297419.546925
30,00042100112.726323378794.209004
493,0001,0002,0020.8580055751,5930.577204
12,0001,0002,0022.7924188852,3772.527216
30,0001,0002,0029.4848611,0002,7388.143252
503,0001,0002,00257.23676737099823.727752
12,0001,0002,002617.732769202,495676.233135
30,0001,0002,0022,467.967021,0002,7202,856.471911
513,00023480.12480123480.140401
12,00023480.81120523480.452403
30,00023481.15440723481.216808
523,0001212760.4368031212760.374402
12,0001383162.0904131383161.684811
30,0001503444.664431503444.61763
533,0001,0002,0090.9984061,0002,7061.170008
12,0001,0002,0093.7596241,0002,6613.369622
30,0001,0002,0098.5020541,0002,7819.594061
543,00032870.06242035770.343202
12,00013410.1092012016071.029607
30,00042990.4836033621,1124.836031
553,00021440.54600421440.530403
12,00023487.48804823487.441248
30,000245030.654197245029.983392
563,0004308860.3588025071,3970.608404
12,0004308861.4508096131,6672.043613
30,0004308863.5412234911,3374.492829
573,0001452961.154407551320.468003
12,0002074207.75325691792.246414
30,00026553619.500125771966.27124
583,0001072230.873606812020.670804
12,0001242573.931225912433.07322
30,00014229310.514467982618.205653
593,000771660.639604521370.405603
12,0001072264.508429601521.934412
30,000942037.082445721815.803237
603,0004889833.9780261113030.967206
12,0001753605.5224351062933.650423
30,00019439814.47689314037711.856076
613,0001452961.185608561420.468003
12,0002064186.692443701792.277615
30,00026453419.390924922477.75325
623,0001533141.232408631630.717605
12,0002394867.332047862142.761218
30,00031363423.166148962618.127652
633,0002094301.9344121383781.388409
12,0001,0002,00937.7990421644486.489642
30,0001,0002,00987.22015919152118.532919
643,00029640.26520229640.218401
12,00031681.04520731680.936006
30,00032702.34001532702.324415
653,00022511.90321219451.59121
12,000173814.586094173814.258491
30,000173861.167992173859.420781
663,0001,0002,00357.9855727332,29350.684725
12,0001,0002,003618.637566214671171.757101
30,00041110.37406758157163.879051
673,0006370.03129590.0312
12,00010630.499203482310.577204
30,0005270.12480110540.296402
683,00035720.031235720.0312
12,00038780.12480138780.109201
30,00039800.34320239800.374402
693,00027580.031230640.0624
12,00028600.14040132680.187201
30,00029620.42120333700.468003
703,00025821.9500131293868.876457
12,0005218446.8471143479119.621567
30,000136252.790738193598597.967433
713,0001,0002,0040.8892064491,2470.468003
12,0001,0002,0044.1964276611,7792.106014
30,0001,0002,0047.2384466061,6455.506835
723,0007062,011228.8378671,0002,845323.405673
12,0005691,5891,742.468777852,2342,412.040662
30,0002296543,931.3812011,0002,81317,084.27791
733,0001,0002,0020.9360064901,3070.421203
12,0001,0002,0023.2916219002,4602.605217
30,0001,0002,0027.5660481,0002,7357.940451
743,0001,0002,0020.8736063981,0610.374402
12,0001,0002,0024.3992287952,1202.262015
30,0001,0002,0027.5192481,0002,6827.86245
Numerical results No. the test problems number. Dimension: the variables number. Ni: the iteration number. Nfg: the function and the gradient value number. CPU time: the CPU time of operating system in seconds. A new tool was given by Dolan and Moré [59] to analyze the performance of the algorithms. Figures 1-3 show that the efficiency of the NTT-PRP-CG-A and the Norm-PRP-A relate to Ni, Nfg, and CPU time, respectively. It is easy to see that these two algorithms are effective for those problems and the given three-term PRP conjugate gradient method is more effective than that of the normal three-term PRP conjugate gradient method. Moreover, the NTT-PRP-CG-A has good robustness. Overall, the presented algorithm has some potential property both in theory and numerical experiment, which is noticeable.
Figure 1

Performance profiles of the algorithms for the test problems (Ni).

Figure 3

Performance profiles of the algorithms for the test problems (CPU time).

Performance profiles of the algorithms for the test problems (Ni). Performance profiles of the algorithms for the test problems (Nfg). Performance profiles of the algorithms for the test problems (CPU time).

Conclusions

In this paper, based on the PRP formula for unconstrained optimization, a modified three-term PRP CG algorithm was presented. The proposed method possesses sufficient descent property also holds without any line search technique, and we have automatically the trust region property of the search direction. Under the Wolfe line search, the global convergence was proven. Numerical results showed that the new algorithm is more effective compared with the normal method.
  3 in total

1.  Incremental Support Vector Learning for Ordinal Regression.

Authors:  Bin Gu; Victor S Sheng; Keng Yeow Tay; Walter Romano; Shuo Li
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2014-08-12       Impact factor: 10.451

2.  Structural Minimax Probability Machine.

Authors:  Bin Gu; Xingming Sun; Victor S Sheng
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2016-04-14       Impact factor: 10.451

3.  Two New PRP Conjugate Gradient Algorithms for Minimization Optimization Models.

Authors:  Gonglin Yuan; Xiabin Duan; Wenjie Liu; Xiaoliang Wang; Zengru Cui; Zhou Sheng
Journal:  PLoS One       Date:  2015-10-26       Impact factor: 3.240

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.