Issue |
Wuhan Univ. J. Nat. Sci.
Volume 29, Number 5, October 2024
|
|
---|---|---|
Page(s) | 403 - 411 | |
DOI | https://doi.org/10.1051/wujns/2024295403 | |
Published online | 20 November 2024 |
Mathematics
CLC number: O221.2
A New Two-Parameter Family of Nonlinear Conjugate Gradient Method Without Line Search for Unconstrained Optimization Problem
无约束优化问题的一种新的无线搜索的两参数族非线性共轭梯度法
School of Statistics and Mathematics, Inner Mongolia University of Finance and Economics, Hohhot 010070, Inner Mongolia Autonomous Region, China
Received:
28
November
2023
This paper puts forward a two-parameter family of nonlinear conjugate gradient (CG) method without line search for solving unconstrained optimization problem. The main feature of this method is that it does not rely on any line search and only requires a simple step size formula to always generate a sufficient descent direction. Under certain assumptions, the proposed method is proved to possess global convergence. Finally, our method is compared with other potential methods. A large number of numerical experiments show that our method is more competitive and effective.
摘要
针对无约束优化问题,提出了一种无需线搜索的两参数族非线性共轭梯度法。该方法的主要特点是不依赖于任何线搜索,仅需要一个简单的步长公式总能产生充分下降的方向。在一定的假设条件下,证明了该方法具有全局收敛性。最后,我们的方法与其他数值效果较好的方法进行了比较。大量的数值实验表明,该方法更具有竞争力和有效性。
Key words: unconstrained optimization / conjugate gradient method without line search / global convergence
关键字 : 无约束优化 / 无线搜索共轭梯度法 / 全局收敛性
Cite this article: ZHU Tiefeng. A New Two-Parameter Family of Nonlinear Conjugate Gradient Method Without Line Search for Unconstrained Optimization Problem[J]. Wuhan Univ J of Nat Sci, 2024, 29(5): 403-411.
Biography: ZHU Tiefeng, male, Associate professor, research direction: optimization theory and methods, reliability statistics. E-mail: tfzhu2016@163.com
Fundation item: Supported by 2023 Inner Mongolia University of Finance and Economics, General Scientific Research for Universities directly under Inner Mongolia, China (NCYWT23026), and 2024 High-quality Research Achievements Cultivation Fund Project of Inner Mongolia University of Finance and Economics, China (GZCG2479)
© Wuhan University 2024
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
0 Introduction
Unconstrained optimization methods are widely used in the fields of nonlinear
dynamic systems and engineering computation to obtain the numerical solution of the optimal control problem[1,2]. In this paper, we consider the following unconstrained optimization problem:
where is sufficiently smooth high-dimensional function. The nonlinear conjugate gradient (CG) methods are highly useful for solving this kind of problem because they can avoid storing and remembering matrices[3,4]. In general, the iterative formula of the CG method is obtained by
where is step size along the search direction , and defined by
where denotes the gradient . is scalar and it is chosen differently so that there are different CG methods corresponding to the different . Some well-known formulas for are given by
(Hestenes-Stiefel (HS) method[5]),
(Fletcher-Reeves (FR) method[6]),
(Polak-Ribiere-Polyak (PRP) method[7]),
(Conjugate Descent (CD) method[8]),
(Liu -Storey (LS) method[9]),
and
(Dai and Yuan (DY) method[10]),
where is Euclidean norm and "T" stands for the transpose,. In general, DY and FR methods based on inexact line search have good convergence performance, but the computational efficiency is not as good as that of PRP method. In order to establish methods with both good numerical performance and convergence properties, many scholars have proposed improved conjugate gradient methods in recent years[11-14]. They showed that the method is globally convergent if the following strong Wolfe line search conditions for are satisfied,
where .
The line search in the conjugate gradient method is usually chosen by a strong Wolfe line search, however, the line search skill usually brings computational burden, especially in solving large-scale nonlinear unconstrained optimization problem. To overcome this problem, Sun and Zhang[3] introduced five CG methods respectively without line search in which the line search step is replaced by a simple step size formula :
where , is chosen such that , is Lipchitz constant defined in Assumption 1 below, and {Qk} is a sequence of positive definite matrices satisfying for positive constants and such that , .
Chen and Sun[4] extended this technique to two-parameter family CG method, which can be deemed as generalization of the HS and LS without line search. Yu[15] and Narushima[16] applied this method to memory gradient method and obtained their global convergence, respectively. Yin and Chen[17] showed that three-term CG methods without line search are globally convergence. Other modified CG methods without line search reader may see Refs.[18-20].
As supplementary of these results, in this paper, we continue their research and would show that the proposed CG method without line search in which the line search step is replaced by fixed step-length formula (5), is global convergence for following new :
where , . This motivation mainly comes from Refs.[3, 4]. Our main aim is to develop a new method and obtain better property of the new method while keeping its simple structure. It should be noted that the formula (6) includes some important special sub-classes. For example, when , the proposed CG method can be deemed as DY method without line search; when , the proposed CG method can be deemed as FR method without line search. Hence, the proposed CG method (6) without line search can be deemed as generalization of the DY and FR.
The rest of this paper is organized as follows. In Section 1, we first give some preliminary results on the CG method without line search and discuss our method sufficient descent property. Furthermore, we prove the global convergence of the proposed method without line search and present algorithm frame. A large amount of numerical experiments are given for illustrative purposes in Section 2. Conclusions appear in Section 3.
1 Analysis of Global Convergence and Algorithm Frame
1.1 Analysis of Global Convergence
We adopt the following assumption on function of which is commonly used in the literature.
Assumption 1[3] The objective function is in a neighborhood of the level set and is bounded. Here, by we mean that the gradient is Lipschitz continuous with modulus i.e., there exist such that
for any .
Assumption 2[3] The function is and strongly convex on . In other words, there exists such that
Lemma 1 Suppose that is given by (2), (3) and (5). Then
holds for all k, where
and
Proof See Lemma 1 in Ref. [3].
Lemma 2 Suppose that Assumption 1 holds and that is given by (2), (3) and (5). Then
Proof See Lemma 5 in Ref. [3].
Lemma 3 Suppose that Assumption 2 holds and is given by (6). Then .
Proof We prove Lemma 3 by induction. For , it is easy to obtain that . Suppose holds, now we prove that holds.
By (3), (6) and Lemma 1, we have
From Corollary 3 in Ref.[3], we know . From the induction hypothesis, we have . Furthermore, one has
which is due to , and . From above analysis, we have
where and are not equal to zero at the same time. Thus, we obtain . This finishes our proof.
If there exists a constant such that (), then we say the search direction of the method satisfies the sufficient descent condition. Lemma 3 means that the search direction is the sufficient descent direction, which is often required in the convergence analysis of CG method.
Theorem 1 Suppose that Assumptions 1, 2 hold and that is given by (6). Then
where is a positive constant.
Proof By (3), we have
squaring both sides of it, we obtain
By Lemma 1 and substituting the expression (6) into above formula, we have
which is due to that is deduced in Lemma 3. By in Lemma 3 , we have
Now by that is deduced in Ref. [3], and the two inequalities above, we obtain
Let , we obtain
The proof is completed. There holds for all k if Assumption 1 is valid.
Theorem 2 Suppose that Assumptions 1, 2 hold and that sequence is given by (2), (3), (5) and (6). Then .
Proof Suppose, by contradiction, that the stated conclusion is not true. Then, in view of , there exists a constant , such that for all k, and by Theorem 1 we have .
Thus, we have a recursive equation which leads to
Thus, equation (11) means that
where , and , are constants. From equation (12) we have , which is contradictory to Lemma 2. Hence we obtain .
Theorem 2 denotes that our proposed CG method without line search possesses global convergence.
1.2 Algorithm Frame
Based on the discussion above, now we can describe the algorithm frame for solving the unconstrained optimization problem (1) as follows:
Step 0 Given an initial point , constants , , , , and let .
Step 1 If a stopping criterion is satisfied, then stop; otherwise, go to Step 2.
Step 2 Compute a step size by formula (6).
Step 3 Let , compute and by (3) and (6).
Step 4 Let and go to Step 1.
From above algorithm, we note that the step-length depends on the Lipschitz constant, but in general, the Lipschitz constant is not known in previously. It should be noted that if is a twice continuous differentiable strictly convex function, the Lipschitz constant can be replaced by a positive constant[15]. So, we choose .
2 Numerical Experiments
In this section, in order to show the performance of the given method, we test our proposed method with fixed step-length (5) (denoted by OMFSL), method (6) with the strong Wolfe line search (4) (denoted by SSWLS), and some other considered algorithms without line search such as DY and FR method [3] and Chen and Sun's method[4] via 24 test problems from Ref.[21]. We use the same convergence criteria in Ref.[21].
The parameters are chosen as , , and . All codes are written in MATLAB 7.5 and ran on Lenovo with 1.90 GHz CPU processor, 2.43 GB RAM memory, and Windows XP operating system. In our numerical experiments the no-line-search method with Qk=I (the unit matrix) show bad convergence behavior. Thus we consider BFGS updates for Qk in formula (5).
The test results are shown in Tables 1 and 2. The results for each problem by each method are shown in the form of , where denote the number of the iteration at the terminate iteration, the total number of calculations of the gradient value of the objective function, the total number of computations of the objective function value, the value of function at the terminate iteration, the CPU time in seconds, respectively. indicates the variable dimension of the test problems.
From Tables 1 and 2, it is easy to see that OMFSL performs better than other considering methods. Meanwhile, we presented the Dolan and Moré[22] performance profiles for OMFSL and other considered methods. Note that the performance ratio is the probability for a solver for the tested problems with the factor of the smallest cost. In Figs.1-5, the performance of OMFSL with fixed step-length (5) and other considered methods are presented.
Fig. 1 Performance profile on the absolute errors of versus |
Fig. 2 Performance profile on the number of iteration |
Fig. 3 Performance profile on the NG |
Fig. 4 Performance profile on the NF |
Fig. 5 Performance profile on the Tcpu |
As we can see from Fig.1, OMFSL is superior to all other considering methods for the absolute errors of versus (the function value at the optimal solution). Figures 2-5 shows that OMFSL is better than all other considered methods for the number of iteration, the number of gradient evaluations (NG), the number of function value evaluations (NF) and CPU time (Tcpu). In conclusion, we can see that OMFSL without line search is very competitive for the test problems, and OMFSL is alternative for solving nonlinear unconstrained optimization problems.
Numerical results for the tested problems
Results of and Tcpu for the tested problems
3 Conclusion
In this paper, we have combined a two-parameter conjugate gradient method defined by formula (6) with Sun-Zhang's step-length formula defined by formula (5) and have proved its global convergence property under the appropriate assumptions. This step-length formula (5) might be practical in case that the line search is expensive or hard. We allow certain flexibility in selecting the sequence {Qk} in practical computation, and we also can consider other updates for Qk. Our proofs require that the function is at least and the level set is bounded. Reducing the requirement of optimization function and finding completely constant step-length might be topic of further research.
References
- Zhang L M, Gao H T, Chen Z Q, et al. Multi-objective global optimal parafoil homing trajectory optimization via Gauss pseudo spectral method[J]. Nonlinear Dynamics, 2013, 72(1): 1-8. [NASA ADS] [CrossRef] [MathSciNet] [Google Scholar]
- Jiang X Z, Jian J B. A sufficient descent Dai-Yuan type nonlinear conjugate gradient method for unconstrained optimization problems[J]. Nonlinear Dynamics, 2013, 72(1): 101-112. [NASA ADS] [CrossRef] [MathSciNet] [Google Scholar]
- Sun J, Zhang J P. Global convergence of conjugate gradient methods without line search[J]. Annals of Operations Research, 2001, 103(1): 161-173. [CrossRef] [MathSciNet] [Google Scholar]
- Chen X D, Sun J. Global convergence of a two-parameter family of conjugate gradient methods without line search[J]. Journal of Computational and Applied Mathematics, 2002, 146(1): 37-45. [NASA ADS] [CrossRef] [MathSciNet] [Google Scholar]
- Hestenes M R, Stiefel E. Methods of conjugate gradients for solving linear systems[J]. Journal of Research of the National Bureau of Standards, 1952, 49(6): 409-436. [CrossRef] [Google Scholar]
- Fletcher R, Reeves C M. Function minimization by conjugate gradients[J]. The Computer Journal, 1964, 7(2): 149-154. [CrossRef] [MathSciNet] [Google Scholar]
- Polyak B T. The conjugate gradient method in extremal problems[J]. USSR Computational Mathematics & Mathematical Physics, 1969, 9(4): 94-112. [CrossRef] [Google Scholar]
- Flecher R. Practical Methods of Optimization, Vol1: Unconstrained Optimization[M]. New York: John Wiley & Sons, 1987. [Google Scholar]
- Liu Y, Storey C. Efficient generalized conjugate gradient algorithms[J]. Journal of Optimization Theory and Application, 1991, 69(1): 129-137. [CrossRef] [MathSciNet] [Google Scholar]
- Dai Y H, Yuan Y X. A nonlinear conjugate gradient with a strong global convergence property[J]. SIAM Journal on Optimization, 1999, 10(1): 177-182. [CrossRef] [MathSciNet] [Google Scholar]
- Deepho J, Abubakar A B, Malik M, et al. Solving unconstrained optimization problems via hybrid CD-DY conjugate gradient methods with applications[J]. Journal of Computational and Applied Mathematics, 2022, 405: 113823. [CrossRef] [MathSciNet] [Google Scholar]
- Abubakar A B, Kumam P, Malik M, et al. A hybrid conjugate gradient based approach for solving unconstrained optimization and motion control problems[J]. Mathematics and Computers in Simulation, 2022, 201: 640-657. [CrossRef] [MathSciNet] [Google Scholar]
- Goncalves M L N, Lima F S, Prudente L F. A study of Liu-Storey conjugate gradient methods for vector optimization[J]. Applied Mathematics and Computation, 2022, 425: 127099. [CrossRef] [Google Scholar]
- Chen C L, Luo L L, Han C H, et al. Global convergence of an extended descent algorithm without line search for unconstrained optimization[J]. Journal of Applied Mathematics and Physics, 2018, 6(1): 130-137. [CrossRef] [Google Scholar]
- Yu Z S. Global convergence of a memory gradient method without line search[J]. Journal of Applied Mathematics & Computing, 2008, 26(1): 545-553. [CrossRef] [MathSciNet] [Google Scholar]
- Narushima Y S. A memory gradient method without line search for unconstrained optimization[J]. SUT Journal of Mathematics, 2006, 42(2): 191-206. [CrossRef] [MathSciNet] [Google Scholar]
- Yin L, Chen X D. Global convergence of two kinds of three-term conjugate gradient methods without line search[J]. Asia Pacific Journal of Operational Research, 2013, 30(1):1-10. [Google Scholar]
- Li X, Chen X D. Global convergence of shortest-residual family of conjugate gradient methods without line search[J]. Asia Pacific Journal of Operational Research, 2005, 22(4): 529-538. [CrossRef] [MathSciNet] [Google Scholar]
- Du S Q, Chen Y Y. Global convergence of a modified spectral FR conjugate gradient method[J]. Applied Mathematics and Computation, 2008, 202(2): 766-770. [CrossRef] [MathSciNet] [Google Scholar]
- Zhu H,Chen X D. Global convergence of a special case of the Dai-Yuan family without line search[J]. Asia-Pacific Journal of Operational Research, 2008, 25(3): 411-420. [CrossRef] [MathSciNet] [Google Scholar]
- Zhu T F, Yan Z Z, Peng X Y. A modified nonlinear conjugate gradient method for engineering computation[J]. Mathematical Problems in Engineering, 2017, 2017(1): 1425857. [CrossRef] [PubMed] [Google Scholar]
- Dolan E D, Moré J J. Benchmarking optimization software with performance profiles[J]. Mathematical Programming, 2002, 91(2): 201-213. [CrossRef] [MathSciNet] [Google Scholar]
All Tables
All Figures
Fig. 1 Performance profile on the absolute errors of versus | |
In the text |
Fig. 2 Performance profile on the number of iteration | |
In the text |
Fig. 3 Performance profile on the NG | |
In the text |
Fig. 4 Performance profile on the NF | |
In the text |
Fig. 5 Performance profile on the Tcpu | |
In the text |
Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.
Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.
Initial download of the metrics may take a while.