Research Article  Open Access
Smoothing Nonmonotone BarzilaiBorwein Gradient Method and Its Application to Stochastic Linear Complementarity Problems
Abstract
A new algorithm for nonsmooth boxconstrained minimization is introduced. The method is a smoothing nonmonotone BarzilaiBorwein (BB) gradient method. All iterates generated by this method are feasible. We apply this method to stochastic linear complementarity problems. Numerical results show that our method is promising.
1. Introduction
In this paper, we consider a problemwhere , and . If is differentiable on the feasible set, there are many methods for (1), such as trust region method [1], projected gradient method [2], projected BarzilaiBorwein method [3], Newton method [4], and activeset projected trust region [5]. If is semismooth, there are few methods for (1). In this paper, we only consider that is locally Lipschitzian, but not necessarily differentiable.
In [6], a smoothing projected gradient method (SPG) was introduced for nonsmooth optimization problems with nonempty closed convex set. This algorithm is easy to implement. At each iteration, authors approximate the objective function by a smooth function with a fixed smoothing parameter and employ the classical projected gradient method to obtain a new point. If a certain criterion is satisfied, then authors update the smoothing parameter using the new point for the next iteration.
The main motivation for the current work come from the numerical results of [6]. Through analyzing the algorithm [6], we find that it takes a lot of time to calculate projection when test problems are largescale. In order to avoid this shortcoming, we propose a smoothing nonmonotone BB gradient method. In our method, we use an active set and a nonmonotone line search strategy. The search direction consists of two parts: some of the components are simply defined; the other components are determined by the BarzilaiBorwein gradient method. We apply it to stochastic linear complementarity problems.
Throughout this paper, will be the Euclidean norm. For all , the orthogonal projection of onto a set will be denoted by . For a given matrix , will be the th row of .
The paper can be outlined as follows. In Section 2, we describe our method. In Section 3, stochastic linear complementarity problems are introduced. In Section 4, we apply our method to stochastic linear complementarity problems and numerical results are illustrated and discussed. Finally, we make some concluding remarks in Section 5.
2. Smoothing Nonmonotone BB Gradient Method
In this section, we propose a smoothing nonmonotone BB gradient method for (1), where is a general locally Lipschitz continuous function.
Definition 1. Let be a locally Lipschitz continuous function. One calls a smoothing function of , if is continuously differentiable in for any , and, for any , and is nonempty and bounded.
Let . Consider the associated set at the stationary point of function : where is a fixed parameter. For , we define index set , , and as follows:where . The set is an estimate of the active set at point . For simplicity, we abbreviate , , and defined by (5) as , , and , respectively. We determine by the following process:It is easy to show .
Now, we state algorithm as follows.
Algorithm 2. Choose positive constants , , a tolerance parameter , integer , constants , and parameters ; choose an initial point , . Set .Stop if . Otherwise, set If , choose , and set . Let , and go to Step 1.Determine , , and according to (5).Compute direction according to (6).Compute the step size by the Armijo line search, satisfying Set , , and .Compute , , and . If , ; otherwise, . If , ; otherwise, . Set whereLet and .Let Go to Step 2.
Remark 3. In Step 2 of Algorithm 2, we use to update the smoothing parameter.
Remark 4. Algorithm 2 employs a twoloop approach: the outer loop updates smoothing parameter , and the inner loop computes a stationary point of . In Step 4, the alternate BB step is used to compute the direction .
Remark 5. Note that , so there exists an infinite subsequence such that This implies that every limit point generated by Algorithm 2 is a stationary point of (1).
The convergence of Algorithm 2 is similar to literature [6, 7], so we omit the process here.
3. Stochastic Linear Complementarity Problems
Let be a probability space with being a subset of . Suppose that the probability distribution is known. The stochastic linear complementarity problem (SLCP) [8] is to find such thatwhere and for are random matrices and vectors. Throughout this paper, we assume that and are measurable functions of and satisfy where stands for the expectation. If only contains a single realization, then (12) reduces to the standard LCP. The LCP has been studied by many researchers [9–14]. Since, in many practical problems, some elements may involve uncertain data, problem (12) has been receiving much attention in recent literature [8, 15–20].
In general, there is no satisfying (12) for almost . A deterministic formulation for the SLCP provides a decision vector which is optimal in a certain sense. Different deterministic formulations may yield different solutions that are optimal in different senses. There are two reformulations of (12) that have been proposed: the expected value (EV) formulation [15] and the expected residual minimization (ERM) formulation [8]. In this paper, we concentrate on ERM. ERM is to find a vector that minimizes the expected residual of the SLCP ; that is,where is defined by denotes the th component of the vector . Here, is an NCP function which has the property In this paper, we choose .
Let ; by [6], we know that is semismooth and is a smoothing function of , where Hence, is a smoothing function of , where and is defined by .
In [6], authors use sample average approximation (SAA) [21], which replaces (14) by its approximationHere, the sample is generated by Monte Carlo sampling method, following the same probability distribution as . So, a smoothing function of is In the next section, we apply Algorithm 2 to (20).
4. Numerical Results
The test problems are randomly generated. The procedure of generating the tested problems is employed from [16, 20]. We omit the procedure here. All these problems are tested in Matlab (version 7.5).
Several parameters are needed to generate the problem: , , and . A vector is randomly generated. When the parameter , then is the unique global solution of the test problems and . If , the global solution is unknown.
Algorithm 2 contains several parameters; we use the values in all numerical experiments. We terminate the iteration if one of the following conditions is satisfied:
We start from the same randomly generated initial point and compare our method with SPG. The numerical results can be seen in Tables 12. Let if is differentiable at . In Tables 12, denote the number of variables; and present at the finial iterates and , respectively; and denote the value of at the final iterates and , respectively.


The results reported in Tables 12 show that smoothing nonmonotone BB gradient method is quite promising. From results shown in Tables 12, we observe that, firstly, Algorithm 2 has less time for all test problems. Secondly, the function values of Algorithm 2 and SPG are close for most test problems. Through analysis, we find that Algorithm 2 drops faster than SPG in each step. This is mainly attributed to the active set and BB step.
5. Concluding Remarks
In this paper, we present a smoothing nonmonotone BarzilaiBorwein gradient method for nonsmooth boxconstrained minimization. The main idea of our method is to use a parametric smoothing approximation function in nonmonotone BB gradient method. In order to have a high speed of convergence, we calculate twostep size. We use BB method and inexact line search, which make sure that the algorithm has high efficiency. We apply it to stochastic linear complementarity problems. Numerical results show that our method is promising.
Conflict of Interests
The author declares that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This project was supported by the National Natural Science Foundation of China (Grants nos. 61362021 and 11361018), Guangxi Fund for Distinguished Young Scholars (2012GXSFFA060003), Guangxi Natural Science Foundation (no. 2014GXNSFFA118001), the Scientific Research Foundation of the Higher Education Institutions of Guangxi, China (Grant no. ZD2014050), Guangxi Natural Science Foundation (no. 2014GXNSFAA118003), and Guangxi Key Laboratory of Automatic Detecting Technology and Instruments (no. YQ15112).
References
 S. Bellavia, M. Macconi, and B. Morini, “An affine scaling trustregion approach to boundconstrained nonlinear systems,” Applied Numerical Mathematics, vol. 44, no. 3, pp. 257–280, 2003. View at: Publisher Site  Google Scholar  MathSciNet
 P. H. Calamai and J. J. Mor, “Projected gradient methods for linearly constrained problems,” Mathematical Programming, vol. 39, no. 1, pp. 93–116, 1987. View at: Publisher Site  Google Scholar  MathSciNet
 Y.H. Dai and R. Fletcher, “Projected barzilaiborwein methods for largescale boxconstrained quadratic programming,” Numerische Mathematik, vol. 100, no. 1, pp. 21–47, 2005. View at: Publisher Site  Google Scholar  MathSciNet
 F. Facchinei, J. Júdice, and J. Soares, “An active set Newton algorithm for largescale nonlinear programs with box constraints,” SIAM Journal on Optimization, vol. 8, no. 1, pp. 158–186, 1998. View at: Publisher Site  Google Scholar  MathSciNet
 L. Qi, X. J. Tong, and D. H. Li, “Activeset projected trustregion algorithm for boxconstrained nonsmooth equations,” Journal of Optimization Theory and Applications, vol. 120, no. 3, pp. 601–625, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 C. Zhang and X. Chen, “Smoothing projected gradient method and its application to stochastic linear complementarity problems,” SIAM Journal on Optimization, vol. 20, no. 2, pp. 627–649, 2009. View at: Publisher Site  Google Scholar  MathSciNet
 Y. Xiao and Q. Hu, “Subspace BarzilaiBorwein gradient method for largescale bound constrained optimization,” Applied Mathematics and Optimization, vol. 58, no. 2, pp. 275–290, 2008. View at: Publisher Site  Google Scholar  MathSciNet
 X. Chen and M. Fukushima, “Expected residual minimization method for stochastic linear complementarity problems,” Mathematics of Operations Research, vol. 30, no. 4, pp. 1022–1038, 2005. View at: Publisher Site  Google Scholar  MathSciNet
 B. Chen, X. Chen, and C. Kanzow, “A penalized fischerburmeister NCPfunction,” Mathematical Programming, vol. 88, no. 1, pp. 211–216, 2000. View at: Publisher Site  Google Scholar  MathSciNet
 F. Facchinei and J.S. Pang, FiniteDimensional Variational Inequalities and Complementarity Problems, Springer, New York, NY, USA, 2003.
 A. Fischer, “Solution of monotone complementarity problems with locally lipschitzian functions,” Mathematical Programming, vol. 76, pp. 513–532, 1997. View at: Publisher Site  Google Scholar  MathSciNet
 F. Facchinei and C. Kanzow, “A nonsmooth inexact newton method for the solution of largescale nonlinear complementarity problems,” Mathematical Programming, vol. 76, no. 3, pp. 493–512, 1997. View at: Publisher Site  Google Scholar  MathSciNet
 D. Sun, “A regularization Newton method for solving nonlinear complementarity problems,” Applied Mathematics and Optimization, vol. 40, no. 3, pp. 315–339, 1999. View at: Publisher Site  Google Scholar  MathSciNet
 D. Sun, R. S. Womersley, and H. Qi, “A feasible semismooth asymptotically Newton method for mixed complementarity problems,” Mathematical Programming, vol. 94, pp. 167–187, 2002. View at: Publisher Site  Google Scholar  MathSciNet
 G. Grkan, A. Y. Ozge, and S. M. Robinson, “Samplepath solution of stochastic variational inequalities,” Mathematical Programming, vol. 84, no. 2, pp. 313–333, 1999. View at: Publisher Site  Google Scholar  MathSciNet
 G. L. Zhou and L. Caccetta, “Feasible semismooth newton method for a class of stochastic linear complementarity problems,” Journal of Optimization Theory and Applications, vol. 139, no. 2, pp. 379–392, 2008. View at: Publisher Site  Google Scholar  MathSciNet
 H. Fang, X. J. Cheng, and M. Fukushima, “Stochastic ${r}_{0}$ matrix linear complementarity problems,” SIAM Journal on Optimization, vol. 18, no. 2, pp. 482–506, 2007. View at: Publisher Site  Google Scholar
 G.H. Lin and M. Fukushima, “New reformulations for stochastic nonlinear complementarity problems,” Optimization Methods and Software, vol. 21, no. 4, pp. 551–564, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 G.H. Lin, X. Chen, and M. Fukushima, “New restricted NCP functions and their applications to stochastic NCP and stochastic MPEC,” Optimization, vol. 56, no. 56, pp. 641–953, 2007. View at: Publisher Site  Google Scholar  MathSciNet
 X. Chen, C. Zhang, and M. Fukushima, “Robust solution of monotone stochastic linear complementarity problems,” Mathematical Programming, vol. 117, no. 12, pp. 51–80, 2009. View at: Publisher Site  Google Scholar  MathSciNet
 A. J. Kleywegt, A. Shapiro, and T. HomemdeMello, “The sample average approximation method for stochastic discrete optimization,” SIAM Journal on Optimization, vol. 12, no. 2, pp. 479–502, 2002. View at: Publisher Site  Google Scholar  MathSciNet
Copyright
Copyright © 2015 Xiangli Li. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.