Research Article  Open Access
On a New Iterative Scheme without Memory with Optimal Eighth Order
Abstract
The purpose of this paper is to derive and discuss a threestep iterative expression for solving nonlinear equations. In fact, we derive a derivativefree form for one of the existing optimal eighthorder methods and preserve its convergence order. Theoretical results will be upheld by numerical experiments.
1. Introduction
Assume that is sufficiently smooth and that is its simple zero; that is, . This paper concerns with numerical solution of nonlinear scalar equations by iterative expressions. Considering a known optimal eighthorder method with derivative and the conjecture of Cordero and Torregrosa [1], we construct a family of derivativefree methods without memory for solving a nonlinear equation.
To shortly review the literature, we remind readers of the following. Kung and Traub in [2] have provided a class of step derivativeinvolved methods including evaluations of the function and one of its first derivatives per full iteration to reach the convergence rate . They also have given a step derivativefree family of one parameter (consuming evaluations of the function) to again achieve the optimal convergence rate .
Remark 1 (KungTraub’s conjecture [2]). Multipoint iterative methods without memory, requiring function evaluations per iteration, have the order of convergence at most . Multipoint methods which satisfy the KungTraub conjecture (still unproved) are called optimal methods.
Some wellknown methods with eighth order of convergence can be found at [3]. As another example, Liu and Wang [4] suggested some optimal eighthorder methods using four evaluations per full cycle in what follows: where the efficiency index is 1.682. Reference [4] also suggested the following threestep approach with the same number of evaluations and efficiency index:
In what follows, in Section 2, the main derivation is provided to design a new derivativefree family with optimal eighthorder convergence for nonlinear equations. Therein, we confirm the conjecture of CorderoTorregrosa as well. Section 3 illustrates the accuracy of the new obtained threestep family of iterative methods by comparing the results for some nonlinear test functions. Finally, in Section 4, a conclusion will be drawn.
2. A New DerivativeFree Family
There are a number of papers (see, e.g., [1] and the references therein) about the idea of removing derivatives from the iteration function in order to avoid defining new functions and calculate iterates only by using the function that describes the problem and also trying to preserve the convergence order. The interest of these methods is to be applied on nonlinear equations when there are many problems for obtaining and evaluating the derivatives involved or when there is no analytical function to derive.
Hence, our focus in this work is to derive a method without the use of derivatives for nonlinear equations.
Remark 2 (Cordero and Torregrosa’s conjecture [1]). Every time that one applies the approximation of the derivative , with , on an optimal method with the order , one needs for preserving the order of convergence.
We begin by reminding the readers of the threestep iterative method without memory proposed in [5] with optimal eighth order of convergence:
The main aim is to follow Remark 2 and to present a derivativefree form of (3) with optimal eighth order of convergence. Therefore, using the approximation , we present the following formulation (): wherein
We shall see that the order of convergence for (4) reaches to the optimal case, that is, 8, with only four function evaluations per full iteration, which means that the proposed uniparametric family of derivativefree methods possesses the high efficiency index 1.682 and can be viewed as the derivativefree formulation of (3).
Theorem 3. Let be a simple zero of a sufficiently differentiable function for an open interval , which includes as an initial approximation of . Then, the family of derivativefree methods (4) is of optimal order eight.
Proof. To find the asymptotic error constant of (4) where , , we expand any terms of (4) around the simple root in the th iterate. Thus, we write where and wherein . Hence, we obtain In the same vein, we have and for the second substep, we have At this time, Taylor’s series expansion of around the root is needed. We find that and subsequently Considering these Taylor’s series expansions in the last step of (4) will result in the following final error equation: This shows that the iterative family of derivativefree methods without memory (4) is of optimal order eight. The proof is complete.
Remark 4. Theorem 3 clearly supports the conjecture of CorderoTorregrosa for providing lowcomplexity derivativefree iterative methods without memory out of optimal methods with derivative.
Note that each method of (4) reaches the efficiency index , which is greater than of optimal fourthorder techniques and of optimal Newton’s method. It has also the same computational efficiency index with (1), (2), and (3).
Remark 5. It must be remarked that, firstly, the paper [6] studied the multipoint iterative schemes using divided differences for selfacceleration of classical methods.
We here state that the free nonzero parameter in (4) gives us the ability to increase the convergence order of (4) more. Such an acceleration in order is known as with memorization (see, e.g., [7]) according the classification of Traub [8] for nonlinear solvers. To be more precise, choosing would yield an acceleration of convergence.
Anyhow, since the simple zero and subsequently are not known, one should give an approximation for (14) using an approximation polynomial in the domain . Toward this goal, if we consider to be Newton’s interpolatory polynomial of fourth degree passing through the five available nodes , , , , and at the end of each cycle, then one has the following approximation: using a suitable . Consequently, one is able to derive the following accelerated iterative method with memory:
Obviously, if fewer nodes are used for the interpolating polynomial, slower acceleration is achieved. An increase of convergence is achieved in this way without additional functional evaluations, making the proposed root solvers (16) efficient. This acceleration will be seen in Section 3.
Theorem 6. Let the function be sufficiently differentiable in a neighborhood of its simple zero . If an initial approximation is sufficiently close to , then the order of convergence of (16) is at least .
Proof. Let be a sequence of approximations generated by an iterative method with order . The error relation with the selfaccelerating parameter for (16) is in what follows: wherein is the asymptotic error constant. Using a symbolic computation and (13), we attain that Substituting the value of from (18) in (17), one may obtain Thus, it is easy to obtain wherein and are two constants and subsequently with two solutions . Clearly the value for is acceptable and would be the convergence order of the method (16) with memory. The proof is complete.
3. Numerical Testing
The objective of this section is to provide a comparison between the presented schemes and the already known methods in the literature.
For numerical reports here, we have used the optimal eighthorder threestep method (1) as (LW8) with , the optimal eighthorder threestep method (3) as (SM8), our optimal threestep eighthorder method (4) with , and the accelerated method with memory (16) denoted by (APM) with .
The results are summarized in Tables 1 and 2 after some full iterations. As they show, novel schemes are comparable with all of the methods. All numerical instances were performed by Mathematica 8 using 1000 fixed floating point arithmetic [9].


We have computed the root of each test function for the initial guess while the iterative schemes were stopped when . As can be seen, the obtained results in Tables 1 and 2 are in harmony with the analytical procedure given in Section 2.
The computational order of convergence (COC) has also been computed by
Example 7. In this test, we compare the behavior of different methods for finding the complex solution of the following nonlinear equation: using the initial approximation where . The results for this test are given in Table 1.
Example 8. We here compare the behavior of different methods for finding the solution of using the initial approximation where . The results for this test are given in Table 2.
It should be mentioned that our method (4) cannot be easily extended for nonlinear systems. The reason is that the weight functions used in (4) do not contain a finite difference operator in the denominators. Such an extension might be pursued for future studies. However, a simple extended version of (4) for the dimensional case can be written in what follows: wherein and it possesses only fifth order of convergence. Note that the extended version of Steffensen’s method has been written by wherein with . Now we apply (25) to solve a nonlinear integral equation, and keeping the rate of convergence at eight will remain as an open problem for future works.
Example 9. Consider the mixed Hammerstein integral equation [10]: where , , and the kernel is given by
In order to solve this nonlinear integral equation, we transform the above equation into a finitedimensional problem by using GaussLegendre quadrature formula given as where the abscissas and the weights are determined for by GaussLegendre quadrature formula. Denoting the approximation of by , we obtain the system of nonlinear equations where, for , we have wherein the abscissas and the weights are known.
Using the initial approximation , we apply the proposed method (25) denoted by PMS with which is multiplicationrich to find the final solution vector of the nonlinear integral equation (31). Table 3 shows the residuals in norm, when is the size of the nonlinear system of equations.

4. Concluding Remarks
Solving nonlinear equations is a classical problem which has interesting applications in various branches of science and engineering (see, e.g., [11]). In this study, we have described an iterative method without memory to find a simple root of a nonlinear equation on an open interval .
The derived scheme was developed by applying the conjecture of CorderoTorregrosa and it was proved that it converges to the simple zero of a nonlinear equation with optimal eighth order of convergence. This shows that it has the optimal efficiency index 1.682. We, furthermore, discussed how to increase the order of convergence via with memorization. Some examples have also been included to support the theoretical parts.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Authors’ Contribution
The authors have made the same contribution. All authors read and approved the final paper.
References
 A. Cordero and J. R. Torregrosa, “Lowcomplexity rootfinding iteration functions with no derivatives of any order of convergence,” Journal of Computational and Applied Mathematics, 2014. View at: Publisher Site  Google Scholar
 H. T. Kung and J. F. Traub, “Optimal order of onepoint and multipoint iteration,” Journal of the Association for Computing Machinery, vol. 21, pp. 643–651, 1974. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 A. Iliev and N. Kyurkchiev, Nontrivial Methods in Numerical Analysis: Selected Topics in Numerical Analysis, LAP LAMBERT Academic Publishing, 2010.
 L. Liu and X. Wang, “Eighthorder methods with high efficiency index for solving nonlinear equations,” Applied Mathematics and Computation, vol. 215, no. 9, pp. 3449–3454, 2010. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 F. Soleymani, S. Karimi Vanani, H. I. Siyyam, and I. A. AlSubaihi, “Numerical solution of nonlinear equations by an optimal eighthorder class of iterative methods,” Annali dell'Universita di Ferrara, vol. 59, no. 1, pp. 159–171, 2013. View at: Publisher Site  Google Scholar  MathSciNet
 V. Hristov, A. Iliev, and N. Kyurkchiev, “A note on the convergence of nonstationary finitedifference analogues,” Computational Mathematics and Mathematical Physics, vol. 45, pp. 194–201, 2005. View at: Google Scholar
 T. Lotfi, A. Cordero, J. R. Torregrosa, M. A. Abadi, and M. M. Zadeh, “On generalization based on Bi et al. iterative methods with eighthorder convergence for solving nonlinear equations,” The Scientific World Journal, vol. 2014, Article ID 272949, 8 pages, 2014. View at: Publisher Site  Google Scholar
 J. F. Traub, Iterative Methods for the Solution of Equations, PrenticeHall, New Jersey, NJ, USA, 1964. View at: MathSciNet
 R. Hazrat, Mathematica: A ProblemCentered Approach, Springer, Berlin, Germany, 2010.
 F. Soleymani, M. Sharifi, S. Shateyi, and F. K. Haghani1, “A class of Steffensentype iterative methods for nonlinear systems,” Journal of Applied Mathematics, vol. 2014, Article ID 705375, 9 pages, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 N. Kyurkchiev and A. Iliev, “On some multipoint methods arising from optimal in the sense of KungTraub algorithms,” International Journal on Mathematical Methods and Models in Biosciences, vol. 2, no. 1, Article ID 1305155, 2013. View at: Publisher Site  Google Scholar  MathSciNet
Copyright
Copyright © 2014 M. Sharifi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.