Abstract

We develop a sufficient descent method for solving large-scale unconstrained optimization problems. At each iteration, the search direction is a linear combination of the gradient at the current and the previous steps. An attractive property of this method is that the generated directions are always descent. Under some appropriate conditions, we show that the proposed method converges globally. Numerical experiments on some unconstrained minimization problems from CUTEr library are reported, which illustrate that the proposed method is promising.

1. Introduction

In this paper, we consider the unconstrained optimization problem where is a continuously differentiable function, and its gradient at point is denoted by , or for the sake of simplicity. is the number of variables, which is automatically assumed to be large. Large-scale optimization is one of the important research areas both in optimization theory and algorithm design. There exist some kinds of effective methods available for solving (1.1), as for instance, inexact Newton, limited memory quasi-Newton, conjugate gradient, spectral gradient, and subspace methods [14].

The iterative formula the method is given by where is a steplength, and is a search direction. Generally, we say that is a descent direction of at if . Furthermore, if there exists a positive constant such that , then is often called a sufficient descent direction. The descent property is very important for the iterative method to be global convergent, especially for the conjugate gradient method [5]. Unfortunately, for the earliest spectral gradient method [6] and the PRP conjugated gradient method, their generated directions are not always descent. Therefore, in the last years, much effort has been done to smooth away the drawback and develop new methods, for example, [7].

As is wellknown, if Armijo line search is used, the standard PRP method can cycle infinitely without approaching any optimal point when solving nonconvex minimization problems. To overcome this drawback, Hager and Zhang [8] made a slight modification on the PRP formula. Consequently, the resulting method own the remarkable property: for any line search, the generated direction always satisfies the sufficient descent condition . Moreover, Zhang et al. [9] further studied the PRP formula and proposed a three-term modified PRP method, in which the generated directions satisfy and this property is independent of any line search. Additionally, for improving the numerical performance of the standard FR method, Zhang et al. [10] developed a modified FR method where the direction also satisfy (1.3).

Although much progress has been made in designing a sufficient descent direction in a conjugate gradient method, it seemingly receive little attention in other methods. Very recently, An et al. [5] proposed a robust technique to construct a sufficient descent direction for unconstrained optimization. The descent direction is a linear combination of the steepest descent direction and the projection of the original direction, that is, where is a scalar and is an original direction. Based on the definition of , it is easy to deduce that the sufficient descent condition (1.3) holds. This technique has been applied to the PSB updated formula and showed that the resulting method converges globally and superlinearly for an uniformly convex function. Moreover, the direction in [11] is truly a special case of (1.4).

We note that (1.4) can be considered as a general form of a nonlinear conjugate gradient method, in which where is a scalar. Comparing with (1.4), we see that the scalar takes the place of the coefficient matrix of . In this paper, we further study the sufficient descent technique (1.4), and propose a simple sufficient descent method for solving unconstrained optimization problems. Our motivation is simple, we choose as the gradient of the previous step in (1.4). The search direction in this proposed method always satisfies a sufficient descent condition. Under some conditions, we show that the algorithm converges globally by using a special line search. The performance of the method on some CUTEr test problems indicates that it is encouraging.

We organize this paper as follows. In the next section, we construct the sufficient descent direction. In Section 3, we state the steps of our new algorithm with a special line search. We also report some experiments on some large-scale unconstrained optimization problems. Finally, we conclude this paper with some remarks in the last section. Throughout this paper, the symbol denotes the Euclidean norm of a vector.

2. New Search Direction

This section aims to state the new direction formula and investigate its properties. Now, if we take as and set in (1.4), we get Obviously, the direction is a linear combination of the gradient at the current and the previous steps. Additionally, to obtain the global convergence of PRP method, Cheng [11] introduced a descent direction defined as We note that (2.2) is only a special choice of (1.4), in which is replaced by and is chosen as the scalar . If we denote then it is easy to see that for any , we have which indicates that is a symmetric and positive semidefinite matrix.

3. Sufficient Descent Method

As we have stated in the previous section, the directions in (2.1) satisfy a sufficient descent condition. In this section, we list our algorithm and establish its global convergence.

Firstly, we state the steps of the simple sufficient descent method. In order to achieve the global convergence of our method, we consider the backtracking line search of Grippo and Lucidi (GL) [12]. That is, for given , , , find the first () such that satisfies

Now, we are ready to state the steps of the Simple Sufficient Descent (SSD) method.

Algorithm 3.1 (SSD). We have the following steps.Step 1. Given . Let , and . Set . Step 2. If , then stop.Step 3. Compute by (2.1).Step 4. Find the steplength satisfying (3.1). Let .Step 5. Set , go to Step 2.

The remainder of this section is devoted to investigate the global convergence of Algorithm 3.1. We first state some assumptions.

Assumption. Function is continuously differentiable and the level set is bounded.

Assumption. The gradient of is Lipschitz continuous, that is, there exists a constant such that

We first prove the following lemma.

Lemma 3.4. Let the sequence be generated by Algorithm 3.1. Suppose that Assumptions 3.2 and 3.3 hold. Then,

Proof. We have the following cases.Case 1. Consider firstly the GL line search (3.1). If , then may not satisfy (3.1). That is, By the mean value theorem and Assumption 3.3, there exists such that where is the Lipschitz constant of . Substituting the last inequality into (3.4), we get This implies that there is a constant such that Case 2. If , then we have which shows that (3.7) holds with . Moreover, (3.7) indicates the lemma claims.

Theorem 3.5. Let be generated by Algorithm 3.1 with GL backtracking line search (3.1). Then

Proof. From Assumption 3.3, we know that there exists a positive constant such that By the definition of in (2.1), we have which shows that is bounded. We get from (3.1) that Consequently, we have For the sake of contradiction, we suppose that (3.9) does not hold. Then there is a constant such that for all .
If , we get from (3.13) that , which shows . This contradicts the assumption (3.9).
If , that is, there is an infinite index set such that It follows from the line search step in Algorithm 3.1 that when is sufficiently large, does not satisfy (3.1). Then from the first part of the proof of Lemma 3.4 we see that (3.6) holds. Since is bounded and , (3.6) implies . This also yields a contradiction. The proof is complete.

4. Numerical Results

In this section, we test the feasibility and effectiveness of the Algorithm SSD. The algorithm is implemented in Fortran77 code using double precision arithmetic. All runs are performed on a PC ( Intel Pentium Dual E2140 1.6 GHz, 256 MB SDRAM) with Linux operations system. The algorithm stops if the infinity-norm of the final gradient is below 10−5, that is The iteration is also stopped if the number of iterations exceeds 10000 or the number of function evaluations reaches 20000. Our experiments are performed on the subset of the nonlinear unconstrained problems from CUTEr [13] collection. The second-order derivatives of all the selected problems are available. Since we are interested in large problems, we refined this selection by considering only problems where the number of variables is at least 50. The parameter in line search (3.1) is taken as: , , and . The numerical results of the algorithms SSD are listed in Table 1. The columns have the following meanings: Problem: name of the test problem,Dim: dimension of the test problem,Iter: number of iterations,Nf: number of function evaluations,Time: CPU time in seconds,Fv: final function value,Norm-2: -norm of the final gradient, and Norm-0: -norm of the final gradient.

In addition, we also present extensive numerical results of the state-of-the-art algorithm CG_DESCENT [8, 14]. CG_DESCENT is a conjugate gradient descent method for solving large-scale unconstrained optimization problems. A new nonmonotone line search were used in CG_DESCENT which makes this algorithm very efficient. The Fortran code can be obtained from http://www.math.ufl.edu/~hager/. When running of CG_DESCENT, default values are used for all parameters. The results are summarized in Table 2.

Observing the tables, we see that SSD works well, as it reached a stationary point based on the stopping criterion (4.1) for almost all these test problems. Although SSD requires large number of iterations or function evaluations, it needs less time consuming comparatively. However, the numerical comparisons tell us that the state-of-the-art algorithm CG_DESCENT performs a little better than SSD, but for some specific problems, the enhancement of SSD is still noticeable. The numerical experiments also show that SSD can potentially used to solve unconstrained optimization problems with higher dimensions.

5. Concluding Remarks

In this paper, motivated by a descent technique, we have proposed a simple sufficient descent method for solving large-scale unconstrained optimization problems. At each iteration, the generated directions are only related to the gradient information of two successive points. Under some mild conditions, we have shown that this method is global convergent. The numerical results indicate that it works well on some selected problems from CUTEr library. We think there are several other issues that could be lead to improvements. A first point that should be considered is probably the choice of a nonmonotone line search technique. We have adopted the monotone scheme of Grippo and Lucidi (3.1), but this is not the only possible choice, and the framework developed in [15, 16] may allow one to achieve a better performance. Similar to [5], another important point that should be further investigated is a modification of the search direction without violating the nice sufficient descent property. To this end, it is interesting to note that this method is capable of solving nonlinear systems of equations without Jacobian information and cheaply computing.

Acknowledgments

The authors thank Professor Zengxin Wei of Guangxi University for his careful reading. The authors are also grateful two anonymous referees for their useful comments and suggestions to improve the paper. The work is supported by the Chinese Postdoctoral Foundation Grant no. 20090461094, the Natural Science Foundation of China Grant no. 11001075, the Natural Science Foundation of Henan Province Eduction Department Grant no. 2010B110004.