Iterative Methods for Nonlinear Equations or Systems and Their Applications 2014View this Special Issue
Research Article | Open Access
A Class of Steffensen-Type Iterative Methods for Nonlinear Systems
A class of iterative methods without restriction on the computation of Fréchet derivatives including multisteps for solving systems of nonlinear equations is presented. By considering a frozen Jacobian, we provide a class of m-step methods with order of convergence . A new method named as Steffensen-Schulz scheme is also contributed. Numerical tests and comparisons with the existing methods are included.
In this paper, we take into account the following system of nonlinear equations: wherein each function maps a vector of the -dimensional space into the real line . This system contains nonlinear equations with unknowns and could be expressed in a more simplified form by defining a function mapping . Hence, the nonlinear system (1) can be written in the form , where the functions , are the coordinate functions of ; see for more details  and also for application issues the work .
We assume that is a smooth function of in the open convex set . There is a strong incentive to use derivative information as well as function values in order to solve traditionally the system (1). The most famous solver for such a problem is Newton's iteration , which is defined as
In this iterative method, we use the Jacobian matrix, that is, , with entries . To be more precise, there are plenty of solvers to tackle the problem (1) or its scalar case, such as those in [3–6]. Among such methods, the third-order iterative methods like the Halley and Chebyshev methods  are considered less practically from a computational point of view because they need to compute the expensive second-order Fréchet derivatives in contrast to the quadratically convergent method (2), in which the first-order Fréchet derivative needs to be calculated. As a matter of fact, for the considered problem (1), the first Fréchet derivative is a matrix with entries, while the second-order Fréchet derivative has entries (without considering the symmetry). On this account, it needs a large amount of operations in order to evaluate the second derivative for each iteration.
To avoid using the Jacobian whose computation is time-consuming for large scale systems, in Chapter 7 of , authors presented the definition of divided difference in -dimensional space as an matrix with elements to prevent computing the first-order Fréchet derivative. Note that (3) is a component-to-component definition. Traub in the pioneer book  introduced another tool named as to estimate the Jacobian matrix and to derive Steffensen's method for nonlinear systems as follows: wherein with .
The primary aim of the present study is to achieve high rate of convergence in order to solve (1) using (5). Hence, we first propose a new iterative method with fourth-order convergence to find both real and complex solutions. The new method does not even need the evaluation of one first-order Fréchet derivative, let alone the higher-order ones. Next, by considering a frozen Jacobian matrix, we suggest a general -step class of iterative methods with arbitrary order of convergence and higher computational efficiency index.
The rest of this paper is prepared as follows. In Section 2, the construction of a scheme is offered. It also includes the analysis of convergence and shows that the suggested method has fourth order. In Section 3, we will extend the new method to present a multistep class of iterations. Section 4 contains a discussion about the implementation and the efficiency of the iterative methods. Section 5 contains a contribution of this study by introducing Steffensen-Schulz iterative method for the first time. This is followed by Section 6 where some numerical tests will be furnished to illustrate the accuracy and efficiency of the proposed approach. Section 7 ends the paper where short conclusions of the study by pointing out future research aspects are given.
In , authors presented the following iterative method: wherein and it possesses the fourth-order convergence to approximate the simple solution of the system (1). As could be seen, it includes the evaluations and three divided difference operators of order one based on (5), to possess the fourth-order convergence. This process is costly for challenging systems of nonlinear equations.
Now, in order to reach the fourth order of convergence without imposing the computation of even first-order Fréchet derivative or three divided difference operators, we consider a three-step structure with the same correcting factor as follows: wherein we could use the component-by-component approximation (3) : or the estimation introduced by Traub (5) as follows:
Note that (8) and (9) are not equal. Per computing step, the method (6) requires computing at three different points without the computation of the Fréchet derivatives which are costly for large scale problems.
The following theorem will be demonstrated by means of the -dimensional Taylor expansion of the functions using the estimation (8) in (7). We here include some of the basic notions which are important in the proof. Let be sufficiently Fréchet differentiable in . By using the notation introduced in , the th derivative of at , , is the -linear function such that . Thus, we have the following:(1),(2), for all permutation of .
Hence, we take into account , and also . It is well known that, for lying in a neighborhood of a solution of the nonlinear system , Taylor's expansion might be written as follows (assuming that the Jacobian matrix is nonsingular): , where , . We observe that since and .
In addition, we can express as , wherein is the identity matrix of the same order to the Jacobian matrix. Therefore, . We also in the sequel denote as the error in the th iteration. The equation where is a -linear function , is called the error equation and is the order of convergence.
Remark 1. which is the error in the th iteration is a vector and would be a matrix.
Theorem 2. Let be sufficiently Fréchet differentiable at each point of an open convex neighborhood of , that is, a simple solution of the system . Let one suppose that is continuous and nonsingular in . Then the sequence obtained using the iterative method (7) using (8) converges to with convergence rate , and the error equation reads
Proof. Similar notation and terminology as in , yields to , and, by taking into consideration , we write where , . Then, and subsequently the expression for would be The Taylor expansion in the second step using (14) yields Therefore, Next, from an analogous reasoning as in (15) and (16), we obtain the error equation (11). Consequently, taking into account (11), it can be concluded that the order of convergence of the proposed method is four.
Note that it could be shown, in a similar way to the previous theorem, that (17) possesses fourth order of convergence. The implementation of (17) depends on the involved linear algebra problems. An interesting point in the new method (17) is that the LU decomposition of needs to be done only once, and it could effectively be used three times per computing step to increase the rate of convergence without imposing much computational burden.
3. An -Step Class
This section presents a general class of multistep iteration methods. In fact, the new scheme (17) can simply be improved by considering the Jacobian matrix to be frozen. In such a way, we are able to propose a general -step multipoint class of iterative methods in the following structure: wherein is used in the linear system , . We remark that, in this structure, the LU factorization of the Jacobian matrix would be computed only once. This reduces the computational load of the linear algebra problems in implementing (18).
In the iterative process (18) each added step will impose one more -dimensional function, whose cost is while the convergence order will be improved to , wherein is the order of the previous substeps. Considering the well-known mathematical induction, it would be easy to deduce the following Theorem for (18).
Theorem 3. Using the same conditions as in Theorem 2, the -step iterative process (18) has the local convergence order using evaluations of the function and one first-order divided difference operator per full iterations.
Proof. The proof of this theorem is based on mathematical induction and is straightforward.
As an example, the five-step sixth-order method from the new class has the following structure:
In the iterative method (17), one may solve one linear system of equations per computing step, with three right-hand side vectors , , and , and a similar procedure must be applied for (19). In such a case, one could compute a factorization of the matrix and use it repeatedly. It is known that the cost (number of products/quotients) of solving the associated linear system by LU decomposition is (including the LU factorization and two triangular systems), where is the size of the system. Moreover, if one has systems with the same matrix, then the final cost is .
The computational cost for (17) is as follows: evaluations of scalar functions for , evaluations of scalar functions , and evaluations of scalar functions for and (since we have computed above and before) for the estimation (5).
We provide a comparison of efficiency indices for the methods (2) denoted by NM, (4) denoted by SM, (6) denoted by ZM, and the new methods (17) denoted by PM4 and (19) denoted by PM6 based on the computational efficiency index which is also known as operational-efficiency index in the literature .
The log plot of the efficiencies according to the definition of this index for an iterative method, which is given by , where is the order of convergence and stands for the total computational cost per iteration in terms of the number of functional evaluations and the number of products/quotients in finding the LU decomposition and two triangular systems, is given in Figure 1. It is clearly obvious that the new method (19) for higher has dominance with respect to the other well-known methods.
Note that , and for the proposed methods and .
Although the new scheme does not have the drawbacks of NM or ZM in terms of computing Fréchet derivatives or three different divided difference operators, a significant focus should be allocated for large scale nonlinear systems. In fact, for large scale sparse nonlinear systems of equations, we have two main shortcomings, first putting numeric values into the Jacobian matrix for Newton-type methods and second the LU decomposition, which is time-consuming. The suggested method in this work does not need the computations of the Fréchet derivatives and, due to this, it resolves the first problem. For solving the second problem, the new scheme should be coded in the inexact form (see, e.g., [12, 13]) using very efficient solvers for the solution of linear systems of equations such as GMRES.
5. A Multiplication-Rich Scheme
A contribution of this study lies in a simple trick that could be used in this section to produce derivative-free schemes which are inversion-free as well. As a matter of fact, the general class (18) is free from Fréchet derivative per computing step and this makes it nice but we could use a simple trick to make the process inversion-free. That is to say, the computation of the inverse of in (18) is tough since it could be a dense matrix. Although we use the LU decomposition to proceed per cycle, we could apply an iterative inverse finder to avoid the computation of inverse (or solving a system) by imposing further matrix-matrix multiplications.
Such a procedure could be done using the well-known Schulz-type methods (see, e.g., ). Here, we apply the Schulz inverse finder for this purpose, for one step of the matrix iterative method (18) as follows: wherein and , are the largest and smallest singular values of . Unfortunately, convergence of this iterative scheme happens for initial matrices so close to the zeros and it might not be quadratic. In fact, the Steffensen method (SM) and its variants obtained by the class (18), for instance, the approach (19), should become inversion-free easily and efficiently.
We now illustrate the basins of attraction for SM2 and PM4 in the complex square of , when the stopping criterion is . Hopefully, the convergence radius for tackling nonlinear problems can become broadened by introducing the free nonzero parameter , which would be a constant array in the multivariate case . Introducing yields the following form of Steffensen's method:
Figure 2(a) shows the SM without imposing small values for , that is, with , and Figure 2(b) is the basins of attraction of PM4. Figure 3 shows the basins of attraction of SM and PM4, but with small values of .
(a) in SM
(b) in PM4
(a) in SM
(b) in PM4
It is clear that choosing small value for results in larger basins of attraction . Hence, we revise this idea and propose an inversion-free method for solving nonlinear systems of equation in what follows: wherein .
We name this multiplication-rich method as the Steffensen-Schulz iteration since it is derivative-free and inversion-free at the same time. It only requires one matrix inversion in the whole process of computing .
6. Numerical Testing
We employ here the second-order method of Newton (2), the second-order scheme of Steffensen (4), the fourth-order scheme of Zheng et al. (6), and the proposed fourth-order derivative-free method (17) to compare the numerical results obtained from these methods in solving test nonlinear systems. We also put (22) into test for Test Problems 1 and 4.
Test 1. As the first problem, we take into account the following hard system of 10 nonlinear equations with 10 unknowns:
In this test problem, the approximate solution up to 5 decimal places is the following vector: , .
Test 2. We consider the following nonlinear system: where its solution is the vector for odd .
Test 3. We consider the following large scale nonlinear system: where its solution is the vector .
We report the numerical results for solving Tests 1–3 in Tables 1–3 based on the initial values. Note that an important aspect in implementing iterative method for solving nonlinear systems is to find a robust initial guess to guarantee the convergence. Some discussions regarding this are given in [16, 17].
The residual norm along with the number of iterations and computational time using the command Timing in Mathematica 8  is reported in Tables 1–3. The computer specifications are Microsoft Windows XP Intel(R), Pentium(R) 4 CPU, and 3.20 GHz with 4 GB of RAM.
An efficient way to observe the behavior of the order of convergence is to use the local computational order of convergence (COC) that can be defined by , in the -dimensional case. We have used this index in the numerical comparisons listed in Tables 1–3 for each different method to illustrate their numerical orders.
In numerical comparisons, we have chosen the fixed point arithmetic to be 200 using the command SetAccuarcy[expr,200] in the written codes. Note that, for some iterative methods, their residual norm at the specified iteration will exceed the bound of ; thus, we consider such approximations as the exact solution and denoted the residual norm by 0 in some cells of the tables.
Results in Table 1 show that the new scheme can be considered for complex solutions of hard nonlinear systems. In this test, due to the fact that the dimension of the nonlinear system is low, we have used the LU decomposition to prevent the computation of 3 linear systems. Computational time reported in Table 1 verifies the fact that derivative-free methods with few number of divided difference operators are reliable. Note that Figure 4(a) reveals the residual fall of (22) for solving Test 1. It also reveals a quadratic convergence in the obtained residuals per full cycle.
(a) Convergence history of (22) for Test 1
(b) Convergence history of (22) for Test 4
In order to tackle large scale nonlinear systems, we have included Tests 2 and 3 in this work. As could be seen from Tables 2 and 3, the cases for and are considered, respectively. It is obvious that for large scale nonlinear systems we have two difficulties in implementation; first the LU decomposition takes time to be attained and second finding the numeric values of the Jacobian matrix in the Newton-type methods is costly. However, to have a fair comparison, we have not defined the pattern of the Jacobian in the computations and used LU decomposition in solving the considered linear systems.
Test 4. Consider the mixed Hammerstein integral equation :
where , and the kernel is given by
In order to solve this nonlinear integral equation, we transform the above equation into a finite-dimensional problem by using Gauss-Legendre quadrature formula given as , where the abscissas and the weights are determined for by Gauss-Legendre quadrature formula. Denoting the approximation of by (), we obtain the system of nonlinear equations
where, for , we have
wherein the abscissas and the weights are known.
Using the initial approximation , we apply the proposed method (22) which is multiplication-rich to find the final solution vector of the nonlinear integral equation (28). Figure 4(b) puts on show the residual fall for solving the nonlinear integral equation (26) using (22) when is the size of the nonlinear system of equations.
From numerical results in this section, it is clear that the accuracy in successive iterations increases, showing stable nature of the methods. Also, like the existing methods, the presented method shows consistent convergence behavior. From the calculation of computational order of convergence, it is also verified that order of convergence is preserved.
7. Concluding Summary
We have presented a class of iterative methods for finding the solution of nonlinear systems. The construction of the suggested scheme let us achieve high convergence order by avoiding the computation of the Jacobian matrix, whose evaluation takes time for large scale nonlinear systems.
Some different numerical tests have been used to compare the consistency and stability of the proposed iterations in contrast to the existing methods. The numerical results obtained in Section 6 reverified the theoretical aspects of the paper. We have also revealed that the methods can efficiently be used for complex zeros as well.
Further modifications could be done to make the method hybrid so as to have a trust region. In sum, we can conclude that the novel iterative methods have an acceptable performance in solving systems of nonlinear equations.
Conflict of Interests
The authors declare that they do not have any conflict of interests in their submitted paper.
The authors wish to thank the anonymous referees for their valuable suggestions on the first version of this paper.
- J. F. Traub, Iterative Methods for the Solution of Equations, Prentice-Hall Series in Automatic Computation, Prentice-Hall, New York, NY, USA, 1964.
- E. S. Alaidarous, M. Z. Ullah, F. Ahmad, and A. S. Al-Fhaid, “An efficient higher-order quasilinearization method for solving nonlinear BVPs,” Journal of Applied Mathematics, vol. 2013, Article ID 259371, 11 pages, 2013.
- Q. Zheng, F. Huang, X. Guo, and X. Feng, “Doubly-accelerated Steffensen's methods with memory and their applications on solving nonlinear ODEs,” Journal of Computational Analysis and Applications, vol. 15, no. 5, pp. 886–891, 2013.
- F. Soleymani and F. Soleimani, “Novel computational derivative-free methods for simple roots,” Fixed Point Theory, vol. 13, no. 1, pp. 247–258, 2012.
- F. Soleymani, “On a novel optimal quartically class of methods,” Far East Journal of Mathematical Sciences, vol. 58, no. 2, pp. 199–206, 2011.
- W.-X. Wang, Y.-L. Shang, W.-G. Sun, and Y. Zhang, “Finding the roots of system of nonlinear equations by a novel filled function method,” Abstract and Applied Analysis, vol. 2011, Article ID 209083, 9 pages, 2011.
- M. T. Darvishi and B.-C. Shin, “High-order Newton-Krylov methods to solve systems of nonlinear equations,” Journal of the Korean Society for Industrial and Applied Mathematics, vol. 15, no. 1, pp. 19–30, 2011.
- J. M. Ortega and W. C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables, Academic Press, New York, NY, USA, 1970.
- Q. Zheng, P. Zhao, and F. Huang, “A family of fourth-order Steffensen-type methods with the applications on solving nonlinear ODEs,” Applied Mathematics and Computation, vol. 217, no. 21, pp. 8196–8203, 2011.
- M. Grau-Sánchez, M. Noguera, and S. Amat, “On the approximation of derivatives using divided difference operators preserving the local convergence order of iterative methods,” Journal of Computational and Applied Mathematics, vol. 237, no. 1, pp. 363–372, 2013.
- A. Cordero, J. L. Hueso, E. Martínez, and J. R. Torregrosa, “A modified Newton-Jarratt's composition,” Numerical Algorithms, vol. 55, no. 1, pp. 87–99, 2010.
- Z.-Z. Bai and H.-B. An, “A globally convergent Newton-GMRES method for large sparse systems of nonlinear equations,” Applied Numerical Mathematics, vol. 57, no. 3, pp. 235–252, 2007.
- S. C. Eisenstat and H. F. Walker, “Globally convergent inexact Newton methods,” SIAM Journal on Optimization, vol. 4, no. 2, pp. 393–422, 1994.
- F. Soleymani and P. S. Stanimirovic, “A higher order iterative method for computing the Drazin inverse,” The Scientific World Journal, vol. 2013, Article ID 708647, 11 pages, 2013.
- A. Cordero, F. Soleymani, J. R. Torregrosa, and S. Shateyi, “Basins of attraction for various Steffensen-type methods,” Journal of Applied Mathematics, vol. 2014, Article ID 539707, 17 pages, 2014.
- F. Toutounian, J. Saberi-Nadjafi, and S. H. Taheri, “A hybrid of the Newton-GMRES and electromagnetic meta-heuristic methods for solving systems of nonlinear equations,” Journal of Mathematical Modelling and Algorithms, vol. 8, no. 4, pp. 425–443, 2009.
- S. Wagon, Mathematica in Action, Springer, New York, NY, USA, 3rd edition, 2010.
- M. Trott, The Mathematica Guide-Book for Numerics, Springer, New York, NY, USA, 2006.
Copyright © 2014 F. Soleymani et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.