Research Article  Open Access
M. Y. Waziri, Z. A. Majid, "An Enhanced MatrixFree Secant Method via PredictorCorrector Modified Line Search Strategies for Solving Systems of Nonlinear Equations", International Journal of Mathematics and Mathematical Sciences, vol. 2013, Article ID 814587, 6 pages, 2013. https://doi.org/10.1155/2013/814587
An Enhanced MatrixFree Secant Method via PredictorCorrector Modified Line Search Strategies for Solving Systems of Nonlinear Equations
Abstract
Diagonal updating scheme is among the cheapest Newtonlike methods for solving system of nonlinear equations. Nevertheless, the method has some shortcomings. In this paper, we proposed an improved matrixfree secant updating scheme via line search strategies, by using the steps of backtracking in the Armijotype line search as a step length predictor and WolfeLike condition as corrector. Our approach aims at improving the overall performance of diagonal secant updating scheme. Under mild assumptions, the global convergence results have been presented. Numerical experiments verify that the proposed approach is very promising.
1. Introduction
Consider the problem with .
The mapping is assumed to satisfy the following standard assumptions:(1)is continuously differentiable in an open convex set ;(2)there exists a solution of (1) in such that and ;(3)the Jacobian is local Lipschitz continuous at .
The famous scheme for finding the solution to (1) is the Newton method. The method is simple to implement, and it produces an iterative sequence from any given initial guess in the neighborhood of via where .
The attractive features of this method are that it is easy to implement and converges rapidly [1]. However, the major difficulty of Newtontype method is matrix storage requirements especially when handling large systems of nonlinear equations [1]. To overcome such difficulty, the simple modification on the Newton method is the fixed Newton method. Fixed Newton method for the determination of solution is given by This method avoids computation and storing the Jacobian in each iterations (except at ). However, it still requires to solve systems of linear equations and consumes more CPU time as the system’s dimension increases [2, 3].
QuasiNewton method is another variant of Newtontype methods. It replaces the Jacobian or its inverse with an approximation which can be updated at each iteration and is given as where the matrix is the approximation of the Jacobian at . The main idea behind quasiNewton’s method is to eliminate the evaluation cost of the Jacobian matrix [4, 5]. However, the most critical part of such quasiNewton method is storing the full matrix of the approximate Jacobian, which can be a very expensive task as the dimension of systems increases [6]. In contrast, this paper presents an improved variant of quasiNewton update via the steps of backtracking in the Armijotype line search and Wolfe–like condition and incorporating restarting strategy whenever the updating matrix is singular or nearly singular. The anticipation has been to improve the overall performance of the diagonal quasiNewton method.
We organized the rest of this paper as follows. In the next section, we present the proposed method. Section 3 presents convergence results. Numerical experiments are reported in Section 4, and finally, conclusion is given in Section 5.
2. The Improved Secant Diagonal Updating
It is well known that it is not always feasible to compute the full elements of the Jacobian matrix of the given nonlinear function or it may be very expensive. We often have to approximate the Jacobian matrix by some other approach, and the famous method of doing so is quasiNewton's method [4]. The basic idea underlining this approach has been to reduce the evaluation cost of the Jacobian matrix. This new approach generates a sequence of points via where is a step length, and is a diagonal approximation of the inverse Jacobian matrix which can be updated in each iterations. To achieve this, we incorporate some new line search strategies, via the steps of backtracking in the Armijotype line search as predictor and then improved via Wolfelike condition as a corrector. The Armijo rule is among the inexact line search methods which guarantees a sufficient degree of accuracy to ensure the algorithm convergence. Nevertheless, the scheme needs high floating points operations and function call. In this paper, we present a simple line search strategy, which is less computational cost, floating points operations and CPU time consumptions respectively compared to classical Armijor line search. The approach is as follows.
Given and , the proposed approach finds the appropriate , such that In addition, the new strategy is implemented in an iterative way using a fixed initial value of as follows.
Algorithm 1 (ArmijoLike).
Step 0. Set , , and .
Step 1. . Choose as the step size; stop. Otherwise , .
Waziri et al. [7] have set the step length , for all ; this approach is mostly used in many Newtonlike methods. Here, we continue in the spirit of diagonal updating, using a new line search strategy to obtain a good step length in every iteration, anticipating to produce a more accurate approximation of the Jacobian inverse matrix and then employing restating strategy whenever the updating matrix is undefined. To this end, would be obtained almost similar to the diagonal updating scheme presented in [7] in which instead of . Now, let the deviation between and denoted as be minimized under some norms; the optimal solution is given as where , , , , and is the trace operation, respectively. The updated formula for the proposed matrix is given as [7] To safeguard on the possibilities of generating undefined , we let , whenever this situation happens
Now, we can describe the algorithm for our proposed approach as follows.
Algorithm 2 (EMFM).
Step 1. Choose an initial guess , , , and let .
Step 2. Compute , and if , stop.
Step 3. Compute .
Step 4. If , retain and go to 5. Otherwise, set and repeat 4.
Step 5. If , retain and go to 6. Otherwise, set and repeat 5.
Step 6. Let .
Step 7. If , stop. Otherwise go to Step 8.
Step 8. If , where , compute , and if not, .
Step 9. Set and go to .
3. Convergence Analysis
We present the convergence result of EMFM method by proving the existence of the step length . We will make the following assumptions on nonlinear system .
Assumption 3. (i) is differentiable in an open convex set in .(ii)There exists such that , and is continuous for all . (iii) satisfies Lipschitz condition of order one; that is, there exists a positive constant such that for all .(iv) There exist constants such that for all and .
To this end, we proceed by given the following result for the step length generated by the proposed strategies.
Theorem 4. Assume that is a strictly convex function. Suppose that the new strategies are employed with and positive exist, for all sufficiently large . Then, the iterates generated by the line search algorithm have the property that
Proof. From condition (iii) of Assumption 3 and the fact that , we have
Recall that
Equation (13) gives
Since and , hence, it follows that
After little simplifications, we obtain
Hence, exist and are positive.
We continue to show the convergence of the iterates by recalling that
Then, we have
for a finite , and (18) yields
Since the algorithm terminates at , it follows from (19) that
We have
Therefore, due to , it implies that
hence
To show the convergence results of EMFM method, we require to show that the updating matrix is bounded above and below by some positive constants. Hence, we can state the following result on the boundedness of by assuming that, without loss of generality, the updating matrix (8) is always used; then, we have the following.
Theorem 5. Let satisfy Assumption 3, and let for all finite . Let be the sequence generated by (8). If the given nonsingular satisfies for some constants and , then the sequence is bounded for all finite .
Proof. Since , it follows that
For and assuming , we have
where is the largest element among .
After multiplying (26) by and substituting , we have
Since , then (27) turns into
From Assumption 3 and , (28) becomes
where .
Since for , it follows that
Hence, we obtain
Suppose that ; then,
From the fact that , it follows that
where .
4. Numerical Results
In this section, we consider some benchmark problems to illustrate the performance of the method proposed in this paper for solving largescale systems of nonlinear equations when compared to some Newtonlike methods. The computations are performed in MATLAB 7.0 using double precision computer, and the stopping rule used is The identity matrix has been chosen as an initial approximate Jacobian inverse. We further design the codes to terminate whenever one of the following happens:(i)the number of iteration is at least 250, but no point of that satisfies (34) is obtained; (ii)CPU time in seconds reaches 250; (iii)There is insufficient memory to initiate the run.
The performances of these methods are be compared in terms of number of iterations and CPU time in seconds. In the following, some details on the benchmarks test problems are presented.
Problem 1. System of nonlinear equations is as follows:
Problem 2. Extended Trigonometric function of Spedicator [8] is as follows:
Problem 3. System of nonlinear equations is as follows:
(A1) Nonadiabatic Stirred Tank Reactors. The first application is the model of [9]. The model deals with two continuous nonadiabatic stirred tank reactors. The reactors are in a series, at steady state having a recycle component, and with an exothermic firstorder irreversible reaction. By eliminating certain variables, the model results into two nonlinear systems of equations as follows:
The dimensionless temperatures of the two reactors are represented by and . The parameters , , , , and are given as 1, 1000, 2, 2, and 22, respectively.
(A2) Navigation by Range Measurements. Consider two beacons determining position by measuring the distances and , where is the position of beacon 1, is the position of beacon 2, and is an unknown point (see Figure 1). This can be modeled into a two nonlinear systems of equations as follows:
The parameters , , , and are given as , , 14, and 16, respectively.
Table 1 shows the number of iterations and CPU time for these five methods, respectively. In Table 1, the value “Dim” denotes the dimension of the systems, “NI” denotes number of iterations, while “CPU” is the CPU time in seconds respectively. We analyze the performance of each method via execution time, floating points operations, and storage locations. One can observe that EMFM has the smallest number of iterations compared to the classical diagonal updating (IVDN) proposed by Waziri et al. [7]. This shows that the line search strategies presented in this paper have increased the convergence speed of the classical diagonal updating method.

If we compare the performance of all methods, in terms of CPU time, it is clear that EMFM method consumes less CPU time than the others and still keeping memory requirement and CPU time in seconds to only . All five methods are able to obtain the solution of A2, but EMFM method consumes less CPU time in second (0.001) compared to the other 4 methods. Moreover, for A1, still proposed method has shown a promising performance with less storage locations (2 locations) whereas NM, FN, and BM, respectively, required 4 locations for each.
5. Conclusion
From the fact that there is a rapid development in research on solving nonlinear systems, nevertheless, the dimension of the nonlinear system is most of the times so large that it requires a very costly arithmetic operations when using some other Newtonlike methods; so we usually employ cheap iterative approach, and the good candidate is classical diagonal updating. Notwithstanding, the classical updating scheme has some lapses. One is that it usually needs high number of iterations. The other is that the method generally has slow convergence behavior. This paper proposes an enhanced diagonal secant updating scheme based on the steps of backtracking in the Armijotype line search and then improved via Wolfelike condition. Our approach aims at improving the overall performance of diagonal secant updating scheme. The algorithm only requires to store a row vector while ignoring all the off and low diagonal elements and therefore largely reduces memory locations. In addition, as it uses two line search strategies (predictor and corrector) to obtain a new iterates point, the spectral properties of the diagonal updating scheme is improved, and rapid convergence property is gained.
Computational experiment suggests that it is very vital for diagonal updating scheme to use line search strategy. EMFM method has very good solving speed and the best performance among the Newtonlike methods. Finally, it can be concluded that this approach would certainly be quite useful for solving largescale systems of nonlinear equations.
References
 J. E. Dennis, Jr. and R. B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear Equations, Prentice Hall, Englewood Cliffs, NJ, USA, 1983. View at: MathSciNet
 M. Y. Waziri, W. J. Leong, M. A. Hassan, and M. Monsi, “A New Newton method with diagonal Jacobian approximation for systems of NonLinear equations,” Journal of Mathematics and Statistics, vol. 6, no. 3, pp. 246–252, 2010. View at: Publisher Site  Google Scholar
 M. Y. Waziri, W. J. Leong, and M. Mamat, “A twostep matrixfree secant method for solving largescale systems of nonlinear equations,” Journal of Applied Mathematics, vol. 2012, Article ID 348654, 9 pages, 2012. View at: Publisher Site  Google Scholar  MathSciNet
 C. G. Broyden, “A class of methods for solving nonlinear simultaneous equations,” Mathematics of Computation, vol. 19, pp. 577–593, 1965. View at: Google Scholar  MathSciNet
 K. Natasa and L. Zorna, “Newtonlike method with modification of the righthand vector,” Mathematics of Computation, vol. 71, pp. 237–250, 2001. View at: Google Scholar
 C. T. Kelley, Iterative Methods for Linear and Nonlinear Equations, vol. 16, SIAM, Philadelphia, Pa, USA, 1995. View at: Publisher Site  MathSciNet
 M. Y. Waziri, W. J. Leong, and M. A. Hassan, “Diagonal Broydenlike method for largescale systems of nonlinear equations,” Malaysian Journal of Mathematical Sciences, vol. 6, no. 1, pp. 59–73, 2012. View at: Google Scholar  MathSciNet
 E. Spedicator, “Computational experience with quasNewton algorithms for minimization problems of moderately large size,” Tech. Rep. CISEN175, Segrate, Milano, Italy, 1975. View at: Google Scholar
 J. Sinkule, “Multiplicity and stability in a sequence of two nonadiabatic nonisothermal CSTR,” Chemical Engineering Sceinces, vol. 35, pp. 987–996, 1980. View at: Google Scholar
Copyright
Copyright © 2013 M. Y. Waziri and Z. A. Majid. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.