The Scientific World Journal

Volume 2014, Article ID 727490, 6 pages

http://dx.doi.org/10.1155/2014/727490

## On a New Iterative Scheme without Memory with Optimal Eighth Order

^{1}Department of Mathematics, Islamic Azad University, Shahrekord Branch, Shahrekord, Iran^{2}Department of Mathematics and Applied Mathematics, University of Venda, Thohoyandou 0950, South Africa

Received 16 July 2014; Revised 27 July 2014; Accepted 4 August 2014; Published 28 August 2014

Academic Editor: Emran Tohidi

Copyright © 2014 M. Sharifi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The purpose of this paper is to derive and discuss a three-step iterative expression for solving nonlinear equations. In fact, we derive a derivative-free form for one of the existing optimal eighth-order methods and preserve its convergence order. Theoretical results will be upheld by numerical experiments.

#### 1. Introduction

Assume that is sufficiently smooth and that is its simple zero; that is, . This paper concerns with numerical solution of nonlinear scalar equations by iterative expressions. Considering a known optimal eighth-order method with derivative and the conjecture of Cordero and Torregrosa [1], we construct a family of derivative-free methods without memory for solving a nonlinear equation.

To shortly review the literature, we remind readers of the following. Kung and Traub in [2] have provided a class of -step derivative-involved methods including evaluations of the function and one of its first derivatives per full iteration to reach the convergence rate . They also have given a -step derivative-free family of one parameter (consuming evaluations of the function) to again achieve the optimal convergence rate .

*Remark 1 (Kung-Traub’s conjecture [2]). *Multipoint iterative methods without memory, requiring function evaluations per iteration, have the order of convergence at most . Multipoint methods which satisfy the Kung-Traub conjecture (still unproved) are called optimal methods.

Some well-known methods with eighth order of convergence can be found at [3]. As another example, Liu and Wang [4] suggested some optimal eighth-order methods using four evaluations per full cycle in what follows: where the efficiency index is 1.682. Reference [4] also suggested the following three-step approach with the same number of evaluations and efficiency index:

In what follows, in Section 2, the main derivation is provided to design a new derivative-free family with optimal eighth-order convergence for nonlinear equations. Therein, we confirm the conjecture of Cordero-Torregrosa as well. Section 3 illustrates the accuracy of the new obtained three-step family of iterative methods by comparing the results for some nonlinear test functions. Finally, in Section 4, a conclusion will be drawn.

#### 2. A New Derivative-Free Family

There are a number of papers (see, e.g., [1] and the references therein) about the idea of removing derivatives from the iteration function in order to avoid defining new functions and calculate iterates only by using the function that describes the problem and also trying to preserve the convergence order. The interest of these methods is to be applied on nonlinear equations when there are many problems for obtaining and evaluating the derivatives involved or when there is no analytical function to derive.

Hence, our focus in this work is to derive a method without the use of derivatives for nonlinear equations.

*Remark 2 (Cordero and Torregrosa’s conjecture [1]). *Every time that one applies the approximation of the derivative , with , on an optimal method with the order , one needs for preserving the order of convergence.

We begin by reminding the readers of the three-step iterative method without memory proposed in [5] with optimal eighth order of convergence:

The main aim is to follow Remark 2 and to present a derivative-free form of (3) with optimal eighth order of convergence. Therefore, using the approximation , we present the following formulation (): wherein

We shall see that the order of convergence for (4) reaches to the optimal case, that is, 8, with only four function evaluations per full iteration, which means that the proposed uniparametric family of derivative-free methods possesses the high efficiency index 1.682 and can be viewed as the derivative-free formulation of (3).

Theorem 3. *Let be a simple zero of a sufficiently differentiable function for an open interval , which includes as an initial approximation of . Then, the family of derivative-free methods (4) is of optimal order eight.*

*Proof. *To find the asymptotic error constant of (4) where , , we expand any terms of (4) around the simple root in the th iterate. Thus, we write
where and
wherein . Hence, we obtain
In the same vein, we have
and for the second substep, we have
At this time, Taylor’s series expansion of around the root is needed. We find that
and subsequently
Considering these Taylor’s series expansions in the last step of (4) will result in the following final error equation:
This shows that the iterative family of derivative-free methods without memory (4) is of optimal order eight. The proof is complete.

*Remark 4. *Theorem 3 clearly supports the conjecture of Cordero-Torregrosa for providing low-complexity derivative-free iterative methods without memory out of optimal methods with derivative.

Note that each method of (4) reaches the efficiency index , which is greater than of optimal fourth-order techniques and of optimal Newton’s method. It has also the same computational efficiency index with (1), (2), and (3).

*Remark 5. *It must be remarked that, firstly, the paper [6] studied the multipoint iterative schemes using divided differences for self-acceleration of classical methods.

We here state that the free nonzero parameter in (4) gives us the ability to increase the convergence -order of (4) more. Such an acceleration in -order is known as with memorization (see, e.g., [7]) according the classification of Traub [8] for nonlinear solvers. To be more precise, choosing would yield an acceleration of convergence.

Anyhow, since the simple zero and subsequently are not known, one should give an approximation for (14) using an approximation polynomial in the domain . Toward this goal, if we consider to be Newton’s interpolatory polynomial of fourth degree passing through the five available nodes , , , , and at the end of each cycle, then one has the following approximation: using a suitable . Consequently, one is able to derive the following accelerated iterative method with memory:

Obviously, if fewer nodes are used for the interpolating polynomial, slower acceleration is achieved. An increase of convergence is achieved in this way without additional functional evaluations, making the proposed root solvers (16) efficient. This acceleration will be seen in Section 3.

Theorem 6. *Let the function be sufficiently differentiable in a neighborhood of its simple zero . If an initial approximation is sufficiently close to , then the -order of convergence of (16) is at least .*

*Proof. *Let be a sequence of approximations generated by an iterative method with order . The error relation with the self-accelerating parameter for (16) is in what follows:
wherein is the asymptotic error constant. Using a symbolic computation and (13), we attain that
Substituting the value of from (18) in (17), one may obtain
Thus, it is easy to obtain
wherein and are two constants and subsequently
with two solutions . Clearly the value for is acceptable and would be the convergence -order of the method (16) with memory. The proof is complete.

#### 3. Numerical Testing

The objective of this section is to provide a comparison between the presented schemes and the already known methods in the literature.

For numerical reports here, we have used the optimal eighth-order three-step method (1) as (LW8) with , the optimal eighth-order three-step method (3) as (SM8), our optimal three-step eighth-order method (4) with , and the accelerated method with memory (16) denoted by (APM) with .

The results are summarized in Tables 1 and 2 after some full iterations. As they show, novel schemes are comparable with all of the methods. All numerical instances were performed by Mathematica 8 using 1000 fixed floating point arithmetic [9].

We have computed the root of each test function for the initial guess while the iterative schemes were stopped when . As can be seen, the obtained results in Tables 1 and 2 are in harmony with the analytical procedure given in Section 2.

The computational order of convergence (COC) has also been computed by

*Example 7. *In this test, we compare the behavior of different methods for finding the complex solution of the following nonlinear equation:
using the initial approximation where . The results for this test are given in Table 1.

*Example 8. *We here compare the behavior of different methods for finding the solution of
using the initial approximation where . The results for this test are given in Table 2.

It should be mentioned that our method (4) cannot be easily extended for nonlinear systems. The reason is that the weight functions used in (4) do not contain a finite difference operator in the denominators. Such an extension might be pursued for future studies. However, a simple extended version of (4) for the -dimensional case can be written in what follows: wherein and it possesses only fifth order of convergence. Note that the extended version of Steffensen’s method has been written by wherein with . Now we apply (25) to solve a nonlinear integral equation, and keeping the rate of convergence at eight will remain as an open problem for future works.

*Example 9. *Consider the mixed Hammerstein integral equation [10]:
where , , and the kernel is given by

In order to solve this nonlinear integral equation, we transform the above equation into a finite-dimensional problem by using Gauss-Legendre quadrature formula given as where the abscissas and the weights are determined for by Gauss-Legendre quadrature formula. Denoting the approximation of by , we obtain the system of nonlinear equations where, for , we have wherein the abscissas and the weights are known.

Using the initial approximation , we apply the proposed method (25) denoted by PMS with which is multiplication-rich to find the final solution vector of the nonlinear integral equation (31). Table 3 shows the residuals in norm, when is the size of the nonlinear system of equations.

#### 4. Concluding Remarks

Solving nonlinear equations is a classical problem which has interesting applications in various branches of science and engineering (see, e.g., [11]). In this study, we have described an iterative method without memory to find a simple root of a nonlinear equation on an open interval .

The derived scheme was developed by applying the conjecture of Cordero-Torregrosa and it was proved that it converges to the simple zero of a nonlinear equation with optimal eighth order of convergence. This shows that it has the optimal efficiency index 1.682. We, furthermore, discussed how to increase the -order of convergence via with memorization. Some examples have also been included to support the theoretical parts.

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Authors’ Contribution

The authors have made the same contribution. All authors read and approved the final paper.

#### References

- A. Cordero and J. R. Torregrosa, “Low-complexity root-finding iteration functions with no derivatives of any order of convergence,”
*Journal of Computational and Applied Mathematics*, 2014. View at Publisher · View at Google Scholar - H. T. Kung and J. F. Traub, “Optimal order of one-point and multipoint iteration,”
*Journal of the Association for Computing Machinery*, vol. 21, pp. 643–651, 1974. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - A. Iliev and N. Kyurkchiev,
*Nontrivial Methods in Numerical Analysis: Selected Topics in Numerical Analysis*, LAP LAMBERT Academic Publishing, 2010. - L. Liu and X. Wang, “Eighth-order methods with high efficiency index for solving nonlinear equations,”
*Applied Mathematics and Computation*, vol. 215, no. 9, pp. 3449–3454, 2010. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus - F. Soleymani, S. Karimi Vanani, H. I. Siyyam, and I. A. Al-Subaihi, “Numerical solution of nonlinear equations by an optimal eighth-order class of iterative methods,”
*Annali dell'Universita di Ferrara*, vol. 59, no. 1, pp. 159–171, 2013. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus - V. Hristov, A. Iliev, and N. Kyurkchiev, “A note on the convergence of nonstationary finite-difference analogues,”
*Computational Mathematics and Mathematical Physics*, vol. 45, pp. 194–201, 2005. View at Google Scholar - T. Lotfi, A. Cordero, J. R. Torregrosa, M. A. Abadi, and M. M. Zadeh, “On generalization based on Bi et al. iterative methods with eighth-order convergence for solving nonlinear equations,”
*The Scientific World Journal*, vol. 2014, Article ID 272949, 8 pages, 2014. View at Publisher · View at Google Scholar - J. F. Traub,
*Iterative Methods for the Solution of Equations*, Prentice-Hall, New Jersey, NJ, USA, 1964. View at MathSciNet - R. Hazrat,
*Mathematica: A Problem-Centered Approach*, Springer, Berlin, Germany, 2010. - F. Soleymani, M. Sharifi, S. Shateyi, and F. K. Haghani1, “A class of Steffensen-type iterative methods for nonlinear systems,”
*Journal of Applied Mathematics*, vol. 2014, Article ID 705375, 9 pages, 2014. View at Publisher · View at Google Scholar · View at MathSciNet - N. Kyurkchiev and A. Iliev, “On some multipoint methods arising from optimal in the sense of Kung-Traub algorithms,”
*International Journal on Mathematical Methods and Models in Biosciences*, vol. 2, no. 1, Article ID 1305155, 2013. View at Publisher · View at Google Scholar · View at MathSciNet