Research Article | Open Access
Janak Raj Sharma, Puneet Gupta, "An Efficient Family of Traub-Steffensen-Type Methods for Solving Systems of Nonlinear Equations", Advances in Numerical Analysis, vol. 2014, Article ID 152187, 11 pages, 2014. https://doi.org/10.1155/2014/152187
An Efficient Family of Traub-Steffensen-Type Methods for Solving Systems of Nonlinear Equations
Based on Traub-Steffensen method, we present a derivative free three-step family of sixth-order methods for solving systems of nonlinear equations. The local convergence order of the family is determined using first-order divided difference operator for functions of several variables and direct computation by Taylor's expansion. Computational efficiency is discussed, and a comparison between the efficiencies of the proposed techniques with the existing ones is made. Numerical tests are performed to compare the methods of the proposed family with the existing methods and to confirm the theoretical results. It is shown that the new family is especially efficient in solving large systems.
The problem of finding solution of the system of nonlinear equations , where , is an open convex domain in , by iterative methods is an important and challenging task in numerical analysis and many applied scientific branches. One of the basic procedures for solving nonlinear equations is the quadratically convergent Newton method (see [1, 2]): where is the inverse of the first Fréchet derivative of the function .
In many practical situations, it is preferable to avoid the calculation of derivative of the function . In such situations, it is preferable to use only the computed values of and to approximate by employing the values of at suitable points. For example, a basic derivative free iterative method is the Traub-Steffensen method , which also converges quadratically and follows the scheme where is the inverse of the first-order divided difference of and ; is an arbitrary nonzero constant. Throughout this paper, is used to denote the th iteration function of convergence order . For , the scheme (2) reduces to the well-known Steffensen method .
In recent years, many derivative free higher-order methods of great efficiency are developed for solving scalar equation ; see [5–14] and the references therein. For systems of nonlinear equations, however, the construction of efficient higher-order derivative free methods is a difficult task and therefore not many such methods can be found in the literature. Recently, based on Steffensen's scheme, that is, when in (2), a family of seventh-order methods has been proposed in . Some important members of this family, as shown in , are given as follows: Per iteration, both methods use four functions, five first-order divided differences, and three matrix inversions. The notable feature of these algorithms is their simple design which makes them easily implemented to solve systems of nonlinear equations. Here, the fourth-order method is the generalization of the method proposed by Ren et al. in  and is the generalization of the method by Liu et al. .
In this paper, our aim is to develop derivative free iterative methods that may satisfy the basic requirements of generating quality numerical algorithms, that is, the algorithms with (i) high convergence speed, (ii) minimum computational cost, and (iii) simple design. In this way, we here propose a derivative free family of sixth-order methods. The scheme is composed of three steps of which the first two steps consist of any derivative free fourth-order method with the base as the Traub-Steffensen iteration (2) whereas the third step is weighted Traub-Steffensen iteration. The algorithm of the present contribution is as simple as the methods (3) and (4) but with an additional advantage that it possesses high computational efficiency, especially when applied for solving large systems of equations.
The rest of the paper is summarized as follows. The sixth-order scheme with its convergence analysis is presented in Section 2. In Section 3, the computational efficiency of new methods is discussed and is compared with the methods which lie in the same category. Various numerical examples are considered in Section 4 to show the consistent convergence behavior of the methods and to verify the theoretical results. Section 5 contains the concluding remarks.
2. The Method and Its Convergence
Based on the above considerations of a quality numerical algorithm, we begin with the following three-step iteration scheme: where denotes any derivative free fourth-order scheme and , , are some parameters to be determined.
In order to find the convergence order of scheme (5), we first define divided difference operator for multivariable function (see ). The divided difference operator of is a mapping defined by Expanding in Taylor series at the point and then integrating, we have where .
Let . Assuming that exists and then developing and its first three derivatives in a neighborhood of , we have where and . are symmetric operators that are used later on.
We can now analyze behavior of the scheme (5) through the following theorem.
Theorem 1. Let the function be sufficiently differentiable in an open neighborhood of its zero and is a fourth-order iteration function which satisfies where and . If an initial approximation is sufficiently close to , then the local order of convergence of method (5) is at least 6 provided , , and .
Proof. Let . Then, using (8), it follows that
Employing (7) for , , and and then using (9)–(11), we write
Expanding in the formal power developments, the inverse of preceding operator can be written as
Using (8) and (15) to the required terms in the first step of (5), we find that
Equation (7) for , , and yields
Similarly, substituting , , and in (7), we obtain
Then, using (15), (17), and (18) to the required terms, we find that
Postmultiplying (19) by and simplifying
Taylor series of about yields
Then, using (20) and (21) in the third step of (5), we obtain the error equation as
In order to find , , and , it will be sufficient to equate the factors , , and to and then solving the resulting system of equations
we obtain , , and .
Thus, for this set of values, the above error equation reduces to which shows the sixth order of convergence and hence the result follows.
Finally, the sixth-order family of methods is expressed by wherein
Thus, the scheme (25) defines a new three-step family of derivative free sixth-order methods with the first two steps as any fourth-order scheme whose base is the Traub-Steffensen method (3). Some simple members of this family are as follows.
Method-I. The first method, which is denoted by , is given by where is the fourth-order method as given in the formula (3). It is clear that this formula uses four functions, three first-order divided differences, and two matrix inversions per iteration.
Method-II. The second method, that we denote by , is given as where is the fourth-order method shown in (4). This method also requires the same evaluations as in the above method.
3. Computational Efficiency
Here, we estimate the computational efficiency of the proposed methods and compare it with the existing methods. To do this, we will make use of efficiency index, according to which the efficiency of an iterative method is given by , where is the order of convergence and is the computational cost per iteration. For a system of nonlinear equations in unknowns, the computational cost per iteration is given by (see ) Here, denotes the number of evaluations of scalar functions used in the evaluation of and , and denotes the number of products needed per iteration. The divided difference of is an matrix with elements given as (see [17, 18]) In order to express the value of in terms of products, a ratio between products and evaluations of scalar functions and a ratio between products and quotients are required.
To compute in any iterative function, we evaluate scalar functions , and if we compute a divided difference , then we evaluate scalar functions, where and are computed separately. We must add quotients from any divided difference, products for multiplication of a matrix with a vector or of a matrix by a scalar, and products for multiplication of a vector by a scalar. In order to compute an inverse linear operator, we solve a linear system, where we have products and quotients in the LU decomposition and products and quotients in the resolution of two triangular linear systems.
The computational efficiency of the present sixth-order methods and is compared with the existing fourth-order methods and and with the seventh-order methods and . In addition, we also compare the present methods with each other. Let us denote efficiency indices of by and computational cost by . Then, taking into account the above and previous considerations, we have
3.1. Comparison between Efficiencies
To compare the computational efficiencies of the iterative methods, say against , we consider the ratio It is clear that if , the iterative method is more efficient than .
versus Case. For this case, the ratio (37) is given by which shows that for , , and . Thus, we have for , , and .
versus Case. In this case, the ratio (37) takes the following form In this case, it is easy to prove that for , , and , which implies that .
versus Case. The ratio (37) yields which shows that for , , and . Thus, we conclude that for , , and .
versus Case. In this case, the ratio (37) is given by With the same range of , as in the previous case and , the ratio , which implies that .
versus Case. In this case, the ratio (37) is given by It is easy to show that for , , and , which implies that for this range of values of the parameters .
versus Case. The ratio (37) is given by With the same range of , as in the previous cases and , we have , which implies that .
versus Case. The ratio (37) yields which shows that for , , and , so it follows that .
versus Case. For this case, the ratio (37) is given by In this case, also it is not difficult to prove that for , , and , which implies that .
We summarize the above results in the following theorem.
Theorem 2. For and , we have the following: (i) for ;(ii) for ;(iii) for ;(iv) for ;(v) for ;(vi) for ;(vii) for . Otherwise, the comparison depends on , , and .
4. Numerical Results
In this section, some numerical problems are considered to illustrate the convergence behavior and computational efficiency of the proposed methods. The performance is compared with the existing methods , , , and . All computations are performed using the programming package Mathematica  using multiple-precision arithmetic with 4096 digits. For every method, we analyze the number of iterations needed to converge to the solution such that . In numerical results, we also include the CPU time utilized in the execution of program which is computed by the Mathematica command “TimeUsed”. In order to verify the theoretical order of convergence, we calculate the computational order of convergence using the formula (see ) taking into consideration the last three approximations in the iterative process.
To connect the analysis of computational efficiency with numerical examples, the definition of the computational cost (29) is applied, according to which an estimation of the factor is claimed. For this, we express the cost of the evaluation of the elementary functions in terms of products, which depends on the computer, the software, and the arithmetics used (see [21, 22]). In Table 1, the elapsed CPU time (measured in milliseconds) in the computation of elementary functions and an estimation of the cost of the elementary functions in product units are displayed. The programs are performed in the processor, Intel (R) Core (TM) i5-480M CPU @ 2.67 GHz (64-bit Machine) Microsoft Windows 7 Home Basic 2009, and are compiled by Mathematica 7.0 using multiple-precision arithmetic. It can be observed from Table 1 that, for this hardware and the software, the computational cost of quotient with respect to product is .
The present methods and are tested by using the values , , and for the parameter . The following problems are chosen for numerical tests.
Problem 1. Considering the system of two equations, In this problem, are the values used in (31)–(36) for calculating computational costs and efficiency indices of the considered methods. The initial approximation chosen is and the solution is .
Problem 3. Next, consider the following boundary value problem (see ): Assume the following partitioning of the interval : Let us define . If we discretize the problem by using the numerical formula for second derivative, we obtain a system of nonlinear equations in variables: In particular, we solve this problem for so that by selecting as the initial value. The solution of this problem is and concrete values of the parameters are .
Problem 4. Consider the system of fifteen equations (see ): In this problem, the concrete values of the parameters are . The initial approximation assumed is and the solution of this problem is
Problem 5. Consider the system of fifty equations: In this problem, are the values used in (31)–(36) for calculating computational costs and efficiency indices. The initial approximation assumed is for obtaining the solution .
Problem 6. Lastly, consider the nonlinear and nondifferentiable integral equation of mixed Hammerstein type (see ): where ; and the kernel is given as follows: We transform the above equation into a finite-dimensional problem by using Gauss-Legendre quadrature formula given as where the abscissas and the weights are determined for by Gauss-Legendre quadrature formula. Denoting the approximation of by , we obtain the system of nonlinear equations: where In this problem, are the values used in (31)–(36) for calculating computational costs and efficiency indices. The initial approximation assumed is for obtaining the solution
Table 2 shows the numerical results obtained for the considered problems by various methods. Displayed in this table are the number of iterations , the computational order of convergence , the computational costs in terms of products, the computational efficiencies , and the mean elapsed CPU time (e-time). Computational cost and efficiency are calculated according to the corresponding expressions given by (31)–(36) by using the values of parameters and as calculated in each problem, while taking in each case. The mean elapsed CPU time is calculated by taking the mean of 50 performances of the program, where we use as the stopping criterion in single performance of the program.