Research Article  Open Access
A Modified Spectral PRP Conjugate Gradient Projection Method for Solving LargeScale Monotone Equations and Its Application in Compressed Sensing
Abstract
In this paper, we develop an algorithm to solve nonlinear system of monotone equations, which is a combination of a modified spectral PRP (PolakRibièrePolyak) conjugate gradient method and a projection method. The search direction in this algorithm is proved to be sufficiently descent for any line search rule. A line search strategy in the literature is modified such that a better step length is more easily obtained without the difficulty of choosing an appropriate weight in the original one. Global convergence of the algorithm is proved under mild assumptions. Numerical tests and preliminary application in recovering sparse signals indicate that the developed algorithm outperforms the stateoftheart similar algorithms available in the literature, especially for solving largescale problems and singular ones.
1. Introduction
In many fields of sciences and engineering, solution of a nonlinear system of equations is a fundamental problem. For example, in [1, 2], both a Nash economic equilibrium problem and a signal processing problem were formulated into a nonlinear system of equations. Owing to complexity, in the past five decades, numerous algorithms and some software packages in virtue of those powerful algorithms have been developed for solving the nonlinear system of equations. See, for example, [1, 3–16] and the references therein. Nevertheless, in practice, no any algorithm can efficiently solve all the systems of equations arising from sciences and engineering. It is significant to develop a specific algorithm to solve the problems with different analytic and structural features [17, 18].
In this paper, we consider the following nonlinear system of monotone equations:where is a continuous and monotone function; that is to say, satisfiesIt has been shown that the solution set of problem (1) is convex if it is nonempty [3]. In addition, throughout the paper, the space is equipped with the Euclidean norm and the inner product , for .
Aiming at solution of problem (1), many efficient methods were developed recently. Only by incomplete enumeration, we here mention the trust region method [19], the Newton and the quasiNewton methods [4–6, 19], the GaussNewton methods [7, 8], the LevenbergMarquardt methods [20–22], the derivativefree methods and its modified versions [9–16, 23–27], the derivativefree conjugate gradient projection method [14], the modified PRP (PolakRibièrePolyak) conjugate gradient method [11], the TPRP method [10], the PRPtype method [28], the projection method [23], the FRtype method [9], and the modified spectral conjugate gradient projection method [13]. Summarily, the spectral gradient methods and the conjugate gradient methods are more popular in solving a largescale nonlinear system of equations than the Newton and the quasiNewton methods. One of the former’s advantages lies in that there is no requirement of computing and storing the Jacobian matrix or its approximation.
Specifically, Li introduced a class of methods for largescale nonlinear monotone equations in [10], which include the SGliked method, the MPRP method, and the TPRP method. Chen [28] proposed a PRP method for largescale nonlinear monotone equations. A descent modified PRP method and FRtype methods were presented in [9, 11], respectively. Liu and Li proposed a projection method for convex constrained monotone nonlinear equations in [23]. Two derivativefree conjugate gradient projection methods were presented for such a system in [14]. Three extensions of conjugate gradient algorithms were developed in [24–26], respectively. Based on the projection technique in [29], Wan and Liu proposed a modified spectral conjugate gradient projection method (MSCGP) to solve a nonlinear monotone system of symmetric equations in [13]. Then, in [2], MSCGP was successfully applied into recovering sparse signals and restoring blurred images.
It is noted that the main idea of [13, 23] is to construct a search direction by projection technique such that it is sufficiently descent. In virtue of derivativefree and low storage properties, numerical experiments indicated that the developed algorithm in [13] is more efficient to solve largescale nonlinear monotone systems of equations than the similar ones available in the literature.
In [30], Yang et al. proposed a modified spectral PRP conjugate gradient method for solving unconstrained optimization problem. It was proved that the search direction at each iteration is a descent direction of objective function and global convergence was established under mild conditions. Our research interest in this paper is to study how to extend this method to solution of problem (1). Specifically, we should address the following issues:
(1) Without need of derivative information of the function , how to determine the spectral and conjugate parameters to get a sufficiently descent search direction at each iteration?
(2) To ensure global convergence of algorithm, how to choose an appropriate step length for the given search direction? Particularly, monotonicity of should be utilized to design a new iterate scheme?
(3) What about the numerical performance of new iteration scheme? Especially, whether it is more efficient or not than the similar algorithms available in the literature.
Note that a new line search rule was proposed in [31] for solving nonlinear monotone equations with convex constraints, and it was shown that, in virtue of this line search, the developed algorithm has good numerical performance. However, the presented line search is involved with choice of a weight. Since it may be difficult to choose an appropriate weight in the practical implementation of algorithm, we attempt to overcome this difficulty in this paper.
Summarily, we intend to propose a modified spectral PRP conjugate gradient derivativefree projection method for solving largescale nonlinear equations. Global convergence of this method will be proved, and numerical tests will be conducted by implementing the developed algorithm to solve benchmark largescale test problems and to reconstruct sparse signals in compressive sensing.
The rest of this paper is organized as follows. In Section 2, we first state the idea to propose a new spectral PRP conjugate gradient method. Then, a new algorithm is developed. Global convergence is established in Section 3. Section 4 is devoted to numerical experiments. Preliminary application of the algorithm is presented in Section 5. Some conclusions are drawn in Section 6.
2. Development of New Algorithm
In this section, we will state how to develop a new algorithm in detail.
2.1. Projection Method
Generally, to solve (1), we need to construct an iterative format as follows:where is called a step length and is a search direction. Let . If satisfiesthen a projection method can be obtained for solving Problem (1). Actually, by monotonicity of , it holds thatfor any solution of (1), . With such a , we define a hyperplane:From (4) and (5), it is clear that strictly separates the iterate point from the solution . Thus, the projection of onto is closer to than . Consequently, the iterative formatis referred to as the projection method proposed in [29]. Both analytic properties and numerical results have shown efficiency and robustness of the projectionbased algorithms for monotone system of equations [10, 13, 14]. In this paper, we intend to propose a new spectral conjugate gradient method also in virtue of the above projection technique.
2.2. A Modified Spectral PRP Conjugate Gradient Method
In the projection method (7), it is noted that must satisfy (4). That is to say, should be a search direction satisfyingVery recently, Wan et al. [13] proposed a modified spectral conjugate gradient projection method for solving nonlinear monotone symmetric equations, where was chosen byand and are computed byrespectively, , and . It was proved in [13] that given by (9) and (10) is sufficiently descent and satisfies .
Note that a modified spectral PRP conjugate gradient method was proposed for solving unconstrained optimization problems in [30]. Similar to the idea in [13], we can extend the method in [30] to solution of problem (1). Specifically, we compute and in (9) byrespectively. Although (11) gives different choices of and from (10), we can also prove the following result.
Proposition 1. Let be given by (9) and (11). Then, for any , the following equality holds:
Proof. For , we haveFor , we haveWe now prove that ifholds for (), then (12) also holds for .
Actually,where the forth equality follows condition (15). Consequently, by mathematical induction, (12) holds for any .
Proposition 1 ensures that the idea of projection method can be incorporated into design of iteration schemes to solve (1) as the search direction is determined by (9) and (11).
2.3. Modified Line Search Rule
Since it is critical to choose an appropriate step length to improve the performance of the iterate scheme (3), as well as determination of search directions, we now present an inexact line search rule to determine in (3).
Very recently, Ou and Li [31] presented a line search rule as follows: find a nonnegative step length such that the following inequality is as follows:where is a fixed search direction, is a given initial step length, and are two given constants, and is specified byIn [31], it was required that in (18) satisfies . Clearly, and are the weights for the values 1 and , respectively. In the practical implementation, it is may be difficult to choose an appropriate . To overcome this difficulty, we choose in (18) byIt is sure that, for the new method as a combination of (9), (11), (17), and (19), we need to establish its convergence theory and to further test its numerical performance.
Remark 2. In [31], to ensure that welldefined the line search (17) is welldefined, it is assumed that satisfieswhere is a given constant. By Proposition 1, it is clear that chosen by (9) and (11) satisfies (20) as .
2.4. Development of New ProjectionBased Algorithm
With the above preparation, we are in a position to develop an algorithm to solve problem (1) by combining the projection technique and the new methods to determine a search direction and a step length.
We now present the computer procedure of Algorithm 1.
Remark 3. Since Algorithm 1 does not involve computing the Jacobian matrix of or its approximation, both information storage and computational cost of the algorithm are lower. In virtue of this advantage, Algorithm 1 is helpful to solution of largescale problems. In next section, we will prove that Algorithm 1 is applicable even if is nonsmooth. Our numerical tests in Section 4 will further show that Algorithm 1 can find a singular solution of problem (1) (see problem 5 and Table 5).
Remark 4. Compared with the algorithm developed in [13], problem (1) is not assumed to be a symmetric system of equations.
3. Convergence
In this section, we are going to study global convergence of Algorithm 1.
Apart from different choices of search direction and step length, Algorithm 1 can be treated as a variant of the projection algorithm in [29]. So, similar to some critical points of establishing global convergence in [29], we attempt to prove that Algorithm 1 is globally convergent. Very recently, locally linear convergence was proved in [32] for some PRPtype projection methods.
We first state the following mild assumptions.
Assumption 5. The function is monotone on .
Assumption 6. The solution set of problem (1) is nonempty.
Assumption 7. The function is Lipschitz continuous on ; namely, there exists a positive constant such that for all ,
Under these assumptions, we can prove that Algorithm 1 has the following nice properties.
Lemma 8. Let be a sequence generated by Algorithm 1. If Assumptions 5, 6, and 7 hold, then
(1) for any , such that ,(2) The sequence is bounded.
(3) If is a finite sequence, then the last iterate point is a solution of problem (1); otherwise,and(4) The sequence is bounded. Hence, there exists a constant such that .
Proof. (1) Let be any point such that . Then, by monotonicity of , we haveFrom (7), it is also easy to verify that is the projection of onto the halfspace:Thus, it follows from (25) that belongs to this halfspace. From the basic properties of projection operator [33], we know thatConsequently,The desired result (22) is directly obtained from (28).
(2) From (28), it is clear that the sequence is nonnegative and decreasing. Thus, is a convergent sequence. It is concluded that is bounded.
(3) From (28), we knowThus,Since the sequence is bounded, the series is convergent. Consequently,The third result has been proved.
(4) For any , by Lipschitz continuity, we haveSince is convergent, we conclude that is bounded. Consequently, for all , there exists a constant such that .
Lemma 8 indicates that, for the sequence generated by Algorithm 1, the sequence is decreasing and convergent, and the sequence is bounded, where is any solution of problem (1).
Lemma 9. Suppose that Assumptions 5, 6, and 7 hold. Let be a sequence generated by Algorithm 1. If there exists a constant such that, for any positive integer ,Then, the sequence of directions is bounded; i.e., there exists a constant such that, for any positive integer ,
Proof. From (9), (11), (12), and the results of Lemma 8, it follows thatFrom (24), we know that there exist a positive integer and a positive number () such that, for all , Hence,Let . TakeThen, holds for any positive integer .
Lemma 10. Suppose that Assumptions 5, 6, and 7 hold. Let and be two sequences generated by Algorithm 1. Then, the line search rule (17) of Step 2 in Algorithm 1 is welldefined.
Proof. Our aim is to show that the line search rule (17) terminates finitely with a positive step length . By contradiction, suppose that, for some iterate indexes such as , condition (17) does not hold. As a result, for all ,From (18) and the termination condition of Algorithm 1, it follows that, for all ,By taking the limit as in both sides of (39), we haveEquations (41) contradicts the fact that for all . That is to say, the line search rule terminates finitely with a positive step length ; i.e., the line search step of Algorithm 1 is welldefined.
With the above preparation, we are now state the convergence result of Algorithm 1.
Theorem 11. Suppose that Assumptions 5, 6, and 7 hold. Let be a sequence generated by Algorithm 1. Then,
Proof. For the sake of contradiction, we suppose that the conclusion is not true. Then, by the definition of inferior limit, there exists a constant such that, for any ,Consequently, fromit follows that for any .
From (7), (17), and (40), we getCombining (24) and (45), we obtainSince for any , we haveClearly, does not satisfy in (17). It says thatBy Lemmas 8 and 9, we know that the two sequences and are bounded. Without loss of generality, we choose a subset such thatTaking the limit in the two sides of (48) as (), it holds thatOn the other hand, from (43), we know thatBy taking the limit in the two sides of (51) for , we getIt contradicts (50). Thus, the proof of Theorem 11 has been completed.
Remark 12 (only with being generated by (19)). As is determined by (18), the proofs are similar.
Since the proof of Theorem 11 does not involve differentiability of , let alone nonsingularity of its Jacobian matrix, we know that the following result holds.
Corollary 13. For any nonsmooth or singular function , let be a sequence generated as Algorithm 1 is used to solve problem (1). Under Assumptions 5, 6, and 7, it holds that
It should be pointed out that the global convergence of Algorithm 1 depends on assumption on monotonicity of . For nonmonotonic function , Algorithm 1 may be not applicable.
4. Numerical Experiments
In this section, by numerical experiments, we are going to study the effectiveness and robustness of Algorithm 1 for solving largescale system of equations.
We first collect some benchmark test problems available in the literature.
Problem 14 (see [5]). The elements of are given by
Problem 15 (see [5]). The elements of are given by
Problem 16 (see [15]). The elements of are given by
Problem 17 (see [28]). The elements of are given byClearly, problem 4 is a linear system of equations, which is used to test Algorithm 1 in this special case.
Problem 18 (see [15]). The elements of are given by
Problem 19 (see [15]). The elements of are given by
Clearly, apart from problem 4, all the others are nonlinear system of equations. Problem 15 is nonsmooth at point . The size of all the test problems is variable. If this size is larger than 1000, the problem can be regarded to be largescale. We solve all the test problems with different sizes by Algorithm 1, especially in comparison with the existing seven similar algorithms, such as those developed very recently in [10, 13–15, 28].
All the algorithms are coded in MATLAB R2010b and run on a personal computer with a 2.2GHZ CPU processor, 8GB memory, and Windows 10 operation system. The relevant parameters of algorithms are specified by where . The termination condition is .
Numerical performance of all the algorithms is reported in Tables 1, 2, and 3. Table 1 shows the numerical performance of all the eight algorithms with the fixed initial points: for problems 1, 5, and 6; for problems 2 and 4; for problem 3. Table 2 demonstrates the numerical performance of all the eight algorithms with initial points randomly generated by Matlab’s Code “rand(n,1)”. Table 3 shows the numerical performance of our algorithm with dimension of 1000000.



For simplification of statement, we use the following notations:
Dim: the dimension of test problems.
Ni: the number of iterations.
Nf: the number of function evaluations.
MPPRPM: the developed algorithm with determined by (19) in this paper.
MPPRPW: the developed algorithm with generated by (18) in this paper.
MSDFPB: the modified spectral derivativefree projectionbased algorithm in [13].
PRP: the PRP conjugate gradient derivativefree projectionbased methods in [28].
MPRP: the modified PRP conjugate gradient derivativefree projectionbased methods in [10].
TPRP: the twoterm PRP conjugate gradient derivativefree projectionbased methods in [10].
DFPB1: the steepest descent derivativefree projectionbased methods in [14] with search direction as follows:where , , , and .
DFPB2: the steepest descent derivativefree projectionbased methods in [14] with in (62) being replaced by
MHS: the MHSPRP conjugate gradient derivativefree projectionbased methods in [15].
From the results in Tables 1, 2, and 3, it follows that our algorithm (MPPRP) outperforms the other seven algorithms, no matter how to choose the initial points (see the italicized results). Especially, it seems to more efficiently solve largescale test problems. Actually, Table 3 shows that MPPRPM can solve the first five Problems with dimension of 1000000 in less time, compared with the other algorithms.
In order to further measure the efficiency difference of all the eight algorithms, we calculate the average number of iteration, the average consumed CPU time, and their standard deviations, respectively. In Table 4, ANi and StdNi stand for the average number of iteration and its standard deviation, respectively. ACT and StdCT represent the average consumed CPU time and its standard deviation, respectively. The average number of function evaluation and its standard deviation are denoted by ANf and StdNf, respectively. Clearly, StdNi, StdNf, and StdCT can show robustness of all the algorithms.


The results in Table 4 demonstrate that both of MPPRPM and MPPRPW outperform the other seven algorithms.
In the end of this section, we use our algorithm to solve a singular system of equations. The next test problem is a modified version of problem 1.
Problem 20. The elements of are given by The initial point is fixed as .
Note that problem 7 is singular since zero is its solution and the Jacobian matrix is singular. We implement Algorithm 1 to solve problem 7 with different dimensions to test whether it can find the singular solution or not.
The results in Table 5 indicate that Algorithm 1 is also efficient to solve the singular system of equations.
5. Preliminary Application in Compressed Sensing
In this section, we will apply our algorithm to solve an engineering problem originated from compressed sensing of sparse signals.
Let be a linear operator, let be a sparse or a nearly sparse original signal, and let be an observed value which satisfies the following linear equations:The original signal is desirable to be reconstructed from the linear equations . Unfortunately, it is often that this linear equation is underdetermined or illconditioned in practice, and has infinitely many solutions. From the fundamental principles of compressed sensing, it is a reasonable approach to seek the sparsest one among all the solutions, which contains the least nonzero components.
In [2], it was shown that the compressed sensing of sparse signals can be formulated the following nonlinear system of equations:where where , , and , for all with , and is an dimensional vector with all elements being one. Clearly, (66) is nonsmooth.
Implement Algorithm 1 to solve the compressed sensing problem of sparse signals with different compressed ratios. In this experiment, we consider a typical compressive sensing scenario, where the goal is to reconstruct an length sparse signal from observations. We test the three algorithms (two popular algorithms: CGD in [34] and SGCS in [35], and MPPRPM) under three CS ratios (CSR): , 0.25, 0.5 corresponding to different numbers of measurements with , respectively. The original sparse signal contains 32(64) randomly nonzero elements. The measurement is disturbed by noise, i.e., , where is the Gaussian noise distributed as with . is the Gaussian matrix generated by commend randn(, ) in Matlab. The quality of restoration is measured by the mean of squared error (MSE) to the original signal ; that is,where is the restored signal. The iterative process starts at the measurement image, i.e., , and terminates when the relative change between successive iterates falls below . It says thatwhere and is chosen as suggested in [36]. The other parameters in this experiment are same as those in the numerical experiments conducted in Section 4. Numerical efficiency is shown in Table 6.

Clearly, from the results in Table 6, it follows that, for any CS ratio, MPPRPM can recover the sparse signals more efficiently without reduction of recovery quality (see the italicized results in Table 6). If the sparsity level 32 is replaced by 64 in the 2048length original signal, the corresponding results are also presented in Table 6, denoted by .
6. Conclusions
In this paper, we have presented a modified spectral PRP conjugate gradient derivativefree projectionbased method for solving the largescale nonlinear monotonic equations, where the search direction is proved to be sufficiently descent for any line search rule, and the step length is chosen by a line search which can overcome the difficulty of choosing an appropriate weight.
Under mild assumptions, global convergence theory has been established for the developed algorithm. Since our algorithm does not involve computing the Jacobian matrix or its approximation, both information storage and computational cost of the algorithm are lower. That is to say, our algorithm is helpful to solution of largescale problems. In addition, it has been shown that our algorithm is also applicable to solve a nonsmooth system of equations or a singular one.
Numerical tests have demonstrated that our algorithm outperforms the others by costing less number of function evaluations, less number of iterations, or less CPU time to find a solution with the same tolerance, especially in comparison with some similar algorithms available in the literature. Efficiency of our algorithm has also been shown by reconstructing sparse signals in compressed sensing.
In future research, due to its satisfactory numerical efficiency, the proposed method in this paper can be extended into solving more largescale nonlinear system of equations from many fields of sciences and engineering.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
We declare that all the authors have no any conflicts of interest about submission and publication of this paper.
Authors’ Contributions
Zhong Wan conceived and designed the research plan and wrote the paper. Jie Guo performed the mathematical modelling and numerical analysis and wrote the paper.
Acknowledgments
This research is supported by the National Natural Science Foundation of China (Grant no. 71671190) and Opening Foundation from State Key Laboratory of Developmental Biology of Freshwater Fish, Hunan Normal University.
References
 S. Huang and Z. Wan, “A new nonmonotone spectral residual method for nonsmooth nonlinear equations,” Journal of Computational and Applied Mathematics, vol. 313, pp. 82–101, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 Z. Wan, J. Guo, J. Liu, and W. Liu, “A modified spectral conjugate gradient projection method for signal recovery,” Signal, Image and Video Processing, vol. 12, no. 8, pp. 1455–1462, 2018. View at: Publisher Site  Google Scholar
 J. M. Ortega and W. C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables, Siam, New York, NY, USA, 1970. View at: MathSciNet
 J. E. Dennis and B. R. Schnabel, “Numerical methods for nonlinear equations and unconstrained optimization,” Classics in Applied Mathematics, p. 16, 1983. View at: Google Scholar
 W. Zhou and D. Li, “Limited memory {BFGS} method for nonlinear monotone equations,” Journal of Computational Mathematics, vol. 25, no. 1, pp. 89–96, 2007. View at: Google Scholar  MathSciNet
 J. M. Martinez, “A family of quasiNewton methods for nonlinear equations with direct secant updates of matrix factorizations,” SIAM Journal on Numerical Analysis, vol. 27, no. 4, pp. 1034–1049, 1990. View at: Publisher Site  Google Scholar  MathSciNet
 G. Fasano, F. Lampariello, and M. Sciandrone, “A truncated nonmonotone GaussNewton method for largescale nonlinear leastsquares problems,” Computational Optimization and Applications, vol. 34, no. 3, pp. 343–358, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 D. Li and M. Fukushima, “A globally and superlinearly convergent GaussNewtonbased BFGS method for symmetric nonlinear equations,” SIAM Journal on Numerical Analysis, vol. 37, no. 1, pp. 152–172, 1999. View at: Publisher Site  Google Scholar  MathSciNet
 Z. Papp and S. Rapajić, “FR type methods for systems of largescale nonlinear monotone equations,” Applied Mathematics and Computation, vol. 269, pp. 816–823, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 Q. Li and D.H. Li, “A class of derivativefree methods for largescale nonlinear monotone equations,” IMA Journal of Numerical Analysis (IMAJNA), vol. 31, no. 4, pp. 1625–1635, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 L. Zhang, W. Zhou, and D. H. Li, “A descent modified PolakRibierePolyak conjugate gradient method and its global convergence,” IMA Journal of Numerical Analysis (IMAJNA), vol. 26, no. 4, pp. 629–640, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 W. La Cruz, J. M. Martnez, and M. Raydan, “Spectral residual method without gradient information for solving largescale nonlinear systems of equations,” Mathematics of Computation, vol. 75, no. 255, pp. 1429–1448, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 Z. Wan, W. Liu, and C. Wang, “A modified spectral conjugate gradient projection method for solving nonlinear monotone symmetric equations,” Pacific Journal of Optimization. An International Journal, vol. 12, no. 3, pp. 603–622, 2016. View at: Google Scholar  MathSciNet
 M. Ahookhosh, K. Amini, and S. Bahrami, “Two derivativefree projection approaches for systems of largescale nonlinear monotone equations,” Numerical Algorithms, vol. 64, no. 1, pp. 21–42, 2013. View at: Publisher Site  Google Scholar  MathSciNet
 Q.R. Yan, X.Z. Peng, and D.H. Li, “A globally convergent derivativefree method for solving largescale nonlinear monotone equations,” Journal of Computational and Applied Mathematics, vol. 234, no. 3, pp. 649–657, 2010. View at: Publisher Site  Google Scholar  MathSciNet
 K. Amini, A. Kamandi, and S. Bahrami, “A doubleprojectionbased algorithm for largescale nonlinear systems of monotone equations,” Numerical Algorithms, vol. 68, no. 2, pp. 213–228, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 Y. Li, Z. Wan, and J. Liu, “Bilevel programming approach to optimal strategy for vendormanaged inventory problems under random demand,” The ANZIAM Journal, vol. 59, no. 2, pp. 247–270, 2017. View at: Publisher Site  Google Scholar  MathSciNet
 T. Li and Z. Wan, “New adaptive BarzilarBorwein step size and its application in solving large scale optimization problems,” The ANZIAM Journal, vol. 61, no. 1, pp. 76–98, 2019. View at: Publisher Site  Google Scholar  MathSciNet
 X. J. Tong and L. Qi, “On the convergence of a trustregion method for solving constrained nonlinear equations with degenerate solutions,” Journal of Optimization Theory and Applications, vol. 123, no. 1, pp. 187–211, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 K. Levenberg, “A method for the solution of certain nonlinear problems in least squares,” Quarterly of Applied Mathematics, vol. 2, pp. 164–168, 1944. View at: Publisher Site  Google Scholar  MathSciNet
 D. Marquardt, “An algorithm for leastsquares estimation of nonlinear parameters,” SIAM Journal on Applied Mathematics, vol. 11, no. 2, pp. 431–441, 1963. View at: Publisher Site  Google Scholar  MathSciNet
 C. Kanzow, N. Yamashita, and M. Fukushima, “LevenbergMarquardt methods with strong local convergence properties for solving nonlinear equations with convex constraints,” Journal of Computational and Applied Mathematics, vol. 173, no. 2, pp. 321–343, 2005. View at: Publisher Site  Google Scholar  MathSciNet
 J. K. Liu and S. J. Li, “A projection method for convex constrained monotone nonlinear equations with applications,” Computers & Mathematics with Applications, vol. 70, no. 10, pp. 2442–2453, 2015. View at: Publisher Site  Google Scholar
 G. Yuan and M. Zhang, “A threeterms PolakRibièrePolyak conjugate gradient algorithm for largescale nonlinear equations,” Journal of Computational and Applied Mathematics, vol. 286, pp. 186–195, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 G. Yuan, Z. Meng, and Y. Li, “A modified Hestenes and Stiefel conjugate gradient algorithm for largescale nonsmooth minimizations and nonlinear equations,” Journal of Optimization Theory and Applications, vol. 168, no. 1, pp. 129–152, 2016. View at: Publisher Site  Google Scholar  MathSciNet
 G. Yuan and W. Hu, “A conjugate gradient algorithm for largescale unconstrained optimization problems and nonlinear equations,” Journal of Inequalities and Applications, vol. 2018, no. 1, article 113, 2018. View at: Publisher Site  Google Scholar  MathSciNet
 L. Zhang and W. Zhou, “Spectral gradient projection method for solving nonlinear monotone equations,” Journal of Computational and Applied Mathematics, vol. 196, no. 2, pp. 478–484, 2006. View at: Publisher Site  Google Scholar  MathSciNet
 W. Cheng, “A PRP type method for systems of monotone equations,” Mathematical and Computer Modelling, vol. 50, no. 12, pp. 15–20, 2009. View at: Publisher Site  Google Scholar  MathSciNet
 M. V. Solodov and B. F. Svaiter, “A globally convergent inexact Newton method for systems of monotone equations,” in Reformulation: Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods, pp. 355–369, Springer, Boston, MA, USA, 1998. View at: Publisher Site  Google Scholar  MathSciNet
 Z. Wan, Z. Yang, and Y. Wang, “New spectral PRP conjugate gradient method for unconstrained optimization,” Applied Mathematics Letters, vol. 24, no. 1, pp. 16–22, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 Y. Ou and J. Li, “A new derivativefree SCGtype projection method for nonlinear monotone equations with convex constraints,” Applied Mathematics and Computation, vol. 56, no. 12, pp. 195–216, 2018. View at: Publisher Site  Google Scholar  MathSciNet
 W. Zhou and D. Li, “On the Qlinear convergence rate of a class of methods for monotone nonlinear equations,” Pacific Journal of Optimization, vol. 14, pp. 723–737, 2018. View at: Google Scholar
 B. T. Polyak, Introduction to Optimization, Optimization Software, New York, NY, USA, 1987. View at: MathSciNet
 Y. Xiao and H. Zhu, “A conjugate gradient method to solve convex constrained monotone equations with applications in compressive sensing,” Journal of Mathematical Analysis and Applications, vol. 405, no. 1, pp. 310–319, 2013. View at: Publisher Site  Google Scholar  MathSciNet
 Y. Xiao, Q. Wang, and Q. Hu, “Nonsmooth equations based method for l_{1} problems with applications to compressed sensing,” Nonlinear Analysis. Theory, Methods and Applications, vol. 74, no. 11, pp. 3570–3577, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 S. Kim, K. Koh, M. Lustig et al., “A method for largescale 1regularized least squares problems with applications in signal processing and statistics,” Technical Report, Dept. of Electrical Engineering, Stanford University, Stanford, Calif, USA, 2007. View at: Google Scholar
Copyright
Copyright © 2019 Jie Guo and Zhong Wan. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.