Abstract

In this paper, based on the HS method and a modified version of the PRP method, a hybrid conjugate gradient (CG) method is proposed for solving large-scale unconstrained optimization problems. The CG parameter generated by the method is always nonnegative. Moreover, the search direction possesses the sufficient descent property independent of line search. Utilizing the standard Wolfe–Powell line search rule to yield the stepsize, the global convergence of the proposed method is shown under the common assumptions. Finally, numerical results show that the proposed method is promising compared with two existing methods.

1. Introduction

Consider the problem of minimizing over :where is continuously differentiable. Throughout, the gradient of at is denoted by , i.e., . We know that conjugate gradient (CG) methods are very popular and effective for solving unconstrained optimization problems (1), especially for large-scale case by means of their simplicity and low memory requirements. These preferred features greatly promote their applications in various areas such as image deblurring and denoising, neural network, compressed sensing, and others. We refer the interested readers to some recent works [13] and references therein for more details. The numerical results reported in [1] reveal that the CG method has great potential in solving image restoration problems.

Generally, the iterative formula of the CG method for solving problem (1) can be read aswhere is called the stepsize computed by some line search. Here, is commonly known as the search direction, which is defined as follows:where is the so-called CG parameter and is the abbreviation of , i.e., . The two key factors that affect the numerical performance of the CG method are the stepsize and the CG parameter. First, we outline several well-known line search criteria in the literature.(a)The exact line search rule: calculate a stepsize satisfying(b)The standard (weak) Wolfe–Powell (WWP) line search rule: calculate a stepsize satisfyingandwhere .(c)The strong Wolfe–Powell (SWP) line search rule: calculate a stepsize satisfying (5) and

On the other hand, different CG methods are determined by different CG parameters. The well-known CG methods include the Fletcher–Reeves (FR) [4], Polak–Ribière–Polyak (PRP) [5, 6], Hestenes–Stiefel (HS) [7], Liu–Storey (LS) [8], Fletcher (CD) [9], and Dai–Yuan (DY) [10] methods, and their CG parameters are, respectively, given bywhere and stands for the Euclidean norm. The methods yielded by the above CG parameters are called the classical CG methods, and their convergence analysis and numerical performance have been extensively studied (see, e.g., [412]). It has been shown that the above formulas for the CG parameters are equivalent when is convex quadratic and the stepsize is obtained by carrying out the exact line search rule (4). However, their numerical performance strongly depends on the CG parameter . The FR, CD, and DY methods possess good convergence, but the numerical performance for these methods is somewhat unsatisfactory for solving general unconstrained nonlinear optimization problems [1214]. On the contrary, it has been shown that the convergence properties of PRP, HS, and LS methods are not so well, but they often possess better computational performance [1214]. Therefore, in the past few decades, based on the above formulas, plenty of formulas for are designed for CG methods that possess both good global convergence properties and promising numerical performance (see [1216] and references therein).

To our knowledge, the first hybrid CG method in the literature was proposed by Touati-Ahmed and Storey [17] (TS method), where is computed as

Apparently, the TS method has some good properties of FR and PRP methods since is a hybrid of and . Combined with HS and DY methods, Dai and Yuan [18] proposed another hybrid CG method (hHD method), in which the hybrid CG parameter is obtained by

When the WWP line search rule is used to compute the stepsize, the resulting search direction in [18] is a descent one and the global convergence for the hHD method is proved. Moreover, the numerical experiments reported in [18] illustrated that the hHD method is competitive and practicable. For other closely related works, we refer the readers to [18, 19] and the references therein. It is worth noting that the CG parameters defined in [1719] are restricted to positive values. As explicated in [19], this restriction in turn results in global convergence of the algorithm. In recent years, many hybrid CG methods were proposed on the basis of the methodology of discrete combinations of several CG parameters (see, e.g., [1, 13, 2023]). The combination parameter is computed by some secant equations [13, 20], the conjugacy condition [21, 22], or by minimizing the least-squares problem consisting of the unknown search direction and an existing one (see [23] and the references therein).

In 2016, Wei et al. [24] introduced a modified PRP method, usually called the WYL method, where the corresponding parameter is yielded by

Under the assumption that generated by Wei et al. [24] satisfies the so-called sufficient descent condition the WYL method is globally convergent under the WWP line search rule and possesses superior numerical performance. Subsequently, Dai and Wen [25] proposed two improved CG methods with sufficient descent property. The CG parameters in [25] are defined aswhere . Clearly, the search direction yielded by satisfies the sufficient descent condition without depending on any line search. However, the sufficient descent property associated with relies on the WWP line search rule.

Based on the above observations, it is interesting to design a hybrid CG method such that the CG parameter is nonnegative and the resulting search direction possesses the sufficient descent property independent of line search technique. Motivated by the methods in [24, 25] and considering that the HS method performs best among the classical CG methods, a new formula for the CG parameter is given bywhere . It is not difficult to see that is a hybrid of , , and . Interestingly, the above parameter is always nonnegative. To see this, let be the angle between and . Thus, we know from (14) thatwhich further implies

Moreover, plugging the CG parameter into (3), we can show that the resulting search direction possesses the sufficient descent property independent of line search technique (see Lemma 1 below).

The structure of this paper is organized as follows. In Section 2, our algorithm framework is presented, and the sufficient descent property with respect to the resulting search direction is discussed in detail. Section 3 is devoted to establishing the convergence of the proposed method with the WWP line search rule. In the last section, some preliminary numerical results are reported to verify the efficiency of the presented method.

2. The Algorithm

In this section, we first propose the algorithm framework for solving problem (1), in which we do not specify which line search rule generates the stepsize. Subsequently, we analyze the sufficient descent property for the search direction. By inserting the WWP line search rule into the algorithm framework, our hybrid CG method is proposed.

The following lemma shows that the direction sequence generated by Algorithm 1 possesses the sufficient descent property independent of any line search.

Step 0: (initialization) given the initial point , . Set and .
Step 1: stop if .
Step 2: compute the stepsize by an appropriate line search rule.
Step 3: generate the new iteration point by and compute according to (14).
Step 4: compute the next direction by (3), set , and go to Step 1.

Lemma 1. Let be a sequence generated by Algorithm 1. Then, for some constant , it holds that

Proof. When , it follows from the definition of in (3) that . So, the relation in (17) holds when . Now, consider the case . If , it follows from (3) thatSuppose that for all . It then follows from (3), (15), and (16) thatwhich completes the proof.

For convenience, in the following statements, we call the method generated by Algorithm 1 with the WWP line search rule as the hHPR CG method.

3. Convergence

In this section, we analyze the convergence for the hHPR CG method. For this goal, the following common assumptions are necessary.

Assumption 1. (i)The level set is bounded. Here, is the given initial point.(ii)In some neighborhood of the level set , the objective function is continuously differentiable, and its gradient is Lipschitz continuous, i.e., there exists a constant such that

The following lemma provides the convergence for the PRP-type CG method, which was originally introduced in [19].

Lemma 2. Consider the general CG method (2) and (3) with the following three properties:(i)The CG parameter is always nonnegative, i.e., for all .(ii)The line search satisfies (5) and (6) and the sufficient descent condition.(iii)Property holds. Then,

Property 1. Consider a method of forms (2) and (3). Suppose thatWe say that the method has property , if for all , there exist constants and such that , and if where , then we have .
From (16) and Lemmas 1 and 2, to obtain the global convergence of the hHPR CG method, we only prove that our method owns property .

Lemma 3. Consider the method of forms (2) and (3) in which . If Assumption 1 holds, then satisfies property .

Proof. Considering the method of forms (2) and (3) and using the constants and in (22), we have from (16) thatLet and . If , we obtain from Assumption 1(ii) and (15) thatTherefore, the proof is completed.

With (16) and Lemmas 13 at hand, one can establish the global convergence of the hHPR CG method.

Theorem 1. Let be a sequence generated by the hHPR CG method. If Assumption 1 holds, then .

4. Numerical Experiments

In this section, we verify the efficiency and robustness of the hHPR CG method (hHPR for short) by solving some classical tested problems and compare it with two well-known CG methods: DHS and DPRP in [25].

For the tested problems, some of them are from the well-known CUTE library in [26] and the others come from [27]. Moreover, their dimensions range from 2 to 1000000. All codes were written in MATLAB R2016a, and the numerical experiments were conducted on a Dell PC with Intel Core CPU 3.00 GHz and 16.00 GB RAM. For the aforementioned methods, we reset the search direction by taking once an ascent direction occurs. For the sake of fairness, all the stepsizes are yielded by the WWP line search rule following a bisection algorithm proposed in [28], and the corresponding parameters are set to and . Moreover, we adopt the strategy described in [29] to compute the initial stepsize.

Let for hHPR, and let for DHS and DPRP. Denote the iteration numbers, the CPU time in seconds, and the final value of by , and , respectively. If or , we stop the program. If the latter requirement holds, i.e., , we use “-” to denote Itr, Tcpu, and .

The numerical results are listed in Tables 1 and 2, where “TP” denotes the tested problems used in numerical experiments and “Dim” stands for the dimension of the tested problems.

As we all know, the performance profile introduced in [30] is very useful in measuring the performance of numerical algorithms. Figures 1 and 2 plot the performance profiles of hHPR, DHS, and DPRP in terms of Itr and Tcpu, respectively. Based on the left side of Figures 1 and 2, the proposed method is clearly above the other two curves, and this in turn shows that compared with DHS and DPRP, our proposed method is efficient and encouraging. On the other hand, based on the right side of Figures 1 and 2, our proposed method can successfully solve about 90% of the tested problems and clearly outperforms the other two methods.

Data Availability

All the datasets used in this paper are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The corresponding author acknowledges the Natural Science Foundation of Guangxi Province (grant no. 2021GXNSFAA075001).