The Scientific World Journal

Volume 2015 (2015), Article ID 934260, 12 pages

http://dx.doi.org/10.1155/2015/934260

## A General Class of Derivative Free Optimal Root Finding Methods Based on Rational Interpolation

Centre for Advanced Studies in Pure and Applied Mathematics, Bahauddin Zakariya University, Multan 60800, Pakistan

Received 22 May 2014; Accepted 18 August 2014

Academic Editor: S. A. Mohiuddine

Copyright © 2015 Fiza Zafar et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We construct a new general class of derivative free -point iterative methods of optimal order of convergence using rational interpolant. The special cases of this class are obtained. These methods do not need Newton’s iterate in the *…*first step of their iterative schemes. Numerical computations are presented to show that the new methods are efficient and can be seen as better alternates.

#### 1. Introduction

The problem of root finding has been addressed extensively in the last few decades. In 1685, the first scheme to find the roots of nonlinear equations was published by John Wallis. Its simplified description was published in 1690 by Joseph Raphson and was called Newton-Raphson method. In 1740, Thomas Simpson was the first to introduce Newton’s method as an iterative method for solving nonlinear equations. The method is quadratically convergent but it may not converge to real root if the initial guess does not lie in the vicinity of root and is zero in the neighborhood of the real root. This method is a without-memory method. Later, many derivative free methods were defined, for example, Secant’s and Steffensen’s methods. However, most of the derivative free iterative methods are with-memory methods; that is, they require old and new information to calculate the next approximation. Inspite of the drawbacks of Newton’s method, many multipoint methods for finding simple root of nonlinear equations have been developed in the recent past using Newton’s method as the first step. However, many higher order convergent derivative free iterative methods have also appeared most recently by taking Steffensen type methods at the first step. A large number of optimal higher order convergent iterative methods have been investigated recently up to order sixteen [1–3]. These methods used different interpolating techniques for approximating the first derivative.

In the era of 1960–1965, many authors used rational function approximation for finding the root of nonlinear equation, for example, Tornheim [4], Jarratt, and Nudds [5].

In 1967, Jarratt [6] effectively used rational interpolation of the form to approximate for constructing a with-memory scheme involving first derivative. The order of the scheme was 2.732 and its efficiency was 0.2182.

In 1987, Cuyt and Wuytack [7] described a with-memory iterative method involving first derivative based on rational interpolation and they also discussed two special cases of their scheme having order 1.84 with efficiency 0.1324 and order 2.41 with efficiency 0.1910.

In 1990, Field [8] used rational function to approximate the root of a nonlinear equation as follows: where , the correction in each iterate, is the root of the numerator of the Pade approximant to the Taylor series:

He proved that converges to the root with order , where and are degrees of the denominator and numerator of the Pade approximant.

In 1974, Kung and Traub [9] conjectured that a multipoint iterative scheme without memory for finding simple root of nonlinear equations requiring functional evaluations for one complete cycle can have maximum order of convergence with maximal efficiency index . Multipoint methods with this property are usually called optimal methods. Several researchers [1–3, 10] developed optimal multipoint iterative methods based on this hypothesis.

In 2011, Soleymani and Sharifi [10] developed a four-step without-memory fourteenth order convergent iterative method involving first derivative having an efficiency index of 1.6952. The first derivative at the fourth step is approximated using rational interpolation as follows: where is an optimal eighth order convergent method and the rational interpolant is given as:

In 2012, Soleymani et al. [3] developed a three-step derivative free eighth order method using rational interpolation as follows: where is any two-step fourth order convergent derivative free iterative method. They used the same rational interpolant as given in (5). The constants are determined using the interpolating conditions.

In 2012, Soleymani et al. [2] added his contribution by developing a sixteenth order four-point scheme using Pade approximation. The scheme required four evaluations of functions and one evaluation of first derivative and achieved optimal order of sixteen and and an efficiency index 1.741. The scheme was of the form where is eighth order optimal method. They used the rational interpolant of the following form:

Recently in 2013, Sharma et al. [1] developed a three-step eighth order method and its extension to four-step sixteenth order method using rational interpolation. In the scheme, the first three steps are any arbitrary eight order convergent method. The fourth step is the root of the numerator of the method, which is given as where, and rational polynomial of the following form: The efficiency index of the above sixteenth order method is 1.741. The method involves one derivative evaluation.

In this paper, we present a general class of derivative free -point iterative method which satisfies Kung and Tarub’s Hypothesis [9]. Proposed schemes require functional evaluations to acquire the convergence order and efficiency index can have . The contents of the paper are summarized as follows. In Section 2, we present a general class of -point iterative scheme and its special cases with second, fourth, eighth, and sixteenth order convergence. Section 3 consists of the convergence analysis of the iterative methods discussed in Section 2. In the last section of the paper, we give concluding remarks and some numerical results to show the effectiveness of the proposed methods.

#### 2. Higher Order Derivative Free Optimal Methods

In this section, we give a general class of -point iterative method involving functional evaluations having order of convergence . Thus, the scheme is optimal in the sense of the conjecture of Kung and Traub [9].

Consider a rational polynomial of degree as follows: where,

We approximate by rational function given by (12) to construct a general class of -point iterative scheme. Then, the root of nonlinear equation is the root of the numerator of the rational interpolant of degree for the -point method. The unknowns are determined by the following interpolating conditions:

Then, the general -point iterative method is given by Now, we are going to derive its special cases. For in (14)-(15), We find and such that So, the two-point iterative scheme becomes The iterative scheme (18) is the same as given by Steffensen [11] for ; thus, is a particular case of our scheme given by (14)-(15).

For , we have

We find , , and using the following conditions: By using conditions (20), we have Now, using (21), we have the following three-point iterative scheme: For , we have the following rational interpolant: such that The conditions (24) are used to determine the unknowns , , , and . Thus, we attain a four-point iterative method as follows: where, For , we have the following five-point iterative scheme: where, such that The interpolating conditions (29) yield where, and , , are given as in (26). Hence, we obtain the following iterative method: We, now, give the convergence analysis of the proposed iterative methods (18), (22), (25), and (32).

#### 3. Convergence Analysis

Theorem 1. *
Let us consider as the simple root of sufficiently differentiable function in the neighborhood of the root for interval . If is sufficiently close to , then, for every , the iterative methods defined by (18) and (22) are second and fourth order convergent, respectively, with the error equations given by
**
respectively, where,
*

*Proof. *Let , where is the root of and is the error at step. Now, using Taylor expansion of about the root , we have
where is defined by (34). Taylor’s expansions for and are

Using (35), (36), and (37), we have
which shows that the method (18) is quadratically convergent for all . Again using Taylor expansion of , we have
Now, using (35)–(39), we see that the order of convergence of the method (22) is four and the error equation is given by

*Theorem 2. Let us consider as the simple root of sufficiently differentiable function in the neighborhood of the root for interval . If is sufficiently close to , then for all , the iterative methods defined by (25) and (32) are eighth and sixteenth order convergent, respectively, with the error equations given by
where,
*

*Proof. *Let , where is the root of and is the error in the approximation at iteration. We will use (35), (37), (39), and (40) up to in this result and set
Again, using Taylor expansion of , we have
Now, using (35)–(39), (43), and (44), we see that the method (25) has eighth order convergence with the error equation
To find the error equation of (32), we use (45) up to and set
Taylor expansion of is
Hence, using (35)–(39), (43), (44), (46), and (47), we see that iterative method (32) is sixteenth order convergent with the error equation given by

*Remark 3. *From Theorems 1 and 2, it can be seen that the iterative schemes (18), (22), (25), and (32) are second, fourth, eighth, and sixteenth order convergent requiring two, three, four, and five functional evaluations, respectively. Hence, the proposed iterative schemes (18), (22), (25), and (32) are optimal in the sense of the hypothesis of Kung and Traub [9] with the efficiency indices , , , . Also, it is clear that (14)-(15) is a general -point scheme with optimal order of convergence . The efficiency index of this scheme is .

*4. Numerical Results*

*4. Numerical Results*

*In this section, we present some test functions to demonstrate the performance of the newly developed sixteenth order scheme (32) (FNMS-16). For the sake of comparison, we consider the existing higher order convergent methods based on rational interpolation. We consider the fourteenth order method of Soleymani and Sharifi (4) (SS-14), the sixteenth order method of Soleymani et al.(7) (SSS-16), and the sixteenth order method of Sharma et al.(9) (SGG-16). All the computations for the above-mentioned methods are performed using Maple 16 with 4000 decimal digits precision. The test functions given in Table 1 are taken from [1, 2, 10]. We used almost all types of nonlinear functions, polynomials, and transcendental functions to test the new methods. Table 2 shows that the newly developed sixteenth order methods are comparable with the existing methods of this domain in terms of significant digits and number of function evaluations per iteration. In many examples, the newly developed methods perform better than the existing methods. It can also be seen from the tables that, for the choice of initial guess, near to the exact root or far from the exact root, the performance of the new methods is better.*