/ / Article
Special Issue

## Iterative Methods for Nonlinear Equations or Systems and Their Applications 2014

View this Special Issue

Research Article | Open Access

Volume 2014 |Article ID 737305 | https://doi.org/10.1155/2014/737305

Young Ik Kim, Young Hee Geum, "A New Biparametric Family of Two-Point Optimal Fourth-Order Multiple-Root Finders", Journal of Applied Mathematics, vol. 2014, Article ID 737305, 7 pages, 2014. https://doi.org/10.1155/2014/737305

# A New Biparametric Family of Two-Point Optimal Fourth-Order Multiple-Root Finders

Revised17 Jun 2014
Accepted20 Jun 2014
Published14 Sep 2014

#### Abstract

We construct a biparametric family of fourth-order iterative methods to compute multiple roots of nonlinear equations. This method is verified to be optimally convergent. Various nonlinear equations confirm our proposed method with order of convergence of four and show that the computed asymptotic error constant agrees with the theoretical one.

#### 1. Introduction

It is not surprising that modified Newton’s method  in the simple form is most widely used to find the approximate multiple root of known multiplicity for a given nonlinear equation . Recall that numerical scheme (1) is a one-point optimal method with quadratic convergence. In order to find numerical solution for multiple roots of nonlinear equations more accurately, many researchers have made enormous efforts in developing higher-order methods with improved convergence.

In this paper, we extend modified Newton’s method and propose two-point optimal fourth-order multiple-root finders by evaluating two derivatives and one function per iteration. The optimality will be pursued based on Kung-Traub’s conjecture  in which the convergence order of any multipoint method  without memory can reach at most for evaluations of functions or derivatives.

The contents of this paper consist of what follows. Described in Section 2 are previous studies on multiple-root finders. Section 3 proposes a new biparametric family of two-point optimal fourth-order multiple-root finders. It fully treats method development and convergence analysis. Derivation of the error equations for the proposed schemes is an important task for ensuring convergence behavior. In Section 4, a variety of numerical examples are presented for a wide selection of test functions. It is important to compare the convergence behavior of the proposed schemes with that of existing methods. We confirm that the proposed methods well show the convergence behavior predicted by the developed theory.

#### 2. Preliminary Review of Previous Studies

A number of interesting fourth-order multiple-root finders can be found in papers . Among these, we especially introduce five studies as follows. Shengguo et al.  introduced the following fourth-order method which needs evaluations of one function and two derivatives per iteration for chosen in a neighborhood of the sought zero of with known multiplicity : where , , , and .

J. R. Sharma and R. Sharma  constructed the following fourth-order scheme with , , , and : Li et al.  presented the fourth-order method with : Zhou et al.  proposed the following fourth-order iterative scheme with : Kanwar et al.  developed the fourth-order optimal multipoint iterative method for multiple zeros: where , , and .

#### 3. Method Development and Convergence Analysis

We first suppose that a function has a multiple root with integer multiplicity and is analytic in a small neighborhood of . Then a new iteration method free of second derivatives is proposed below to find an approximate root of multiplicity , given an initial guess sufficiently close to : where with , , , , , and are parameters to be chosen for maximal order of convergence [17, 18]. We establish a main theorem describing the convergence analysis regarding proposed scheme (7) and find out how to select parameters , , and for optimal fourth-order convergence.

Definition 1 (error equation, asymptotic error constant, and order of convergence). Let be a sequence converging to and let be the th iterate error. If there exist real numbers and such that the following error equation holds then or is called the asymptotic error constant and is called the order of convergence [17, 18].

Theorem 2. Let have a zero with integer multiplicity and be analytic in a small neighborhood of . Let and for . Let be an initial guess chosen in a sufficiently small neighborhood of . Let , be two free constant parameters. Let , , , and . Then iterative methods (7) are optimal and of order four and possess the following error equation: where .

Proof. The optimality on convergence order of proposed scheme (7) is clear in the sense of Kung-Traub due to three functional evaluations. Hence, it suffices to determine the constant parameters for fourth-order convergence.
Applying the Taylor’s series expansion about , we get the following relations: where , , and for .
Dividing (11) by (12), we have where , , .
For algebraic convenience, we introduce a parameter defined by ; that is, to obtain Evaluating from (11) with being replaced by in (14), we find Substituting (11)–(15) into (7), we obtain the error equation: where , , and coefficients , depend on the parameters , , , , , and and the function .
Solving , for and , respectively, we get We substitute , into and put . Solving independently of and , that is, solving for and , we obtain Substituting into (17) and (18) with , we get the following relations: By the aid of symbolic computation of Mathematica , we arrive at the relation below: where . As a result, the proof is completed.

Remark 3. We observe that error equation (20) contains only one free parameter , being independent of . Table 1 shows typically chosen parameters and and defines various methods , .

 Method Y1 , Y2 Y3 Y4

#### 4. Numerical Examples and Conclusion

We have performed a variety of numerical experiments with Mathematica Version 5  to confirm the theory developed in Section 3. In these experiments, we assign , via Mathematica command \$MinPrecision = 300, as the minimum number of precision digits to achieve the specified sufficient accuracy. It is crucial to compute with high accuracy for desired numerical results. When zero is not exactly known, it is replaced by a highly accurate value which has larger number of significant digits than the assigned minimum number of precision digits. To deal with numerical results more effectively, we first define To properly display numerical results, we need to define the th computational error for . We need further terminologies as defined below.

Definition 4 (computational asymptotic error constant and computational convergence order). Assume that theoretical asymptotic error constant and convergence order are known (usually via main theorem). Define as the computational asymptotic error constant and as the computational convergence order. Then we find that is equal or close to , while is equal or close to .
If has the same accuracy of \$MinPrecision as that of , then would be nearly zero and hence computing would unfavorably break down. Computed values of are accurate up to 300 significant digits. For current experiments is found to be accurate enough about up to 400 significant digits. To supply such , a set of following Mathematica commands are used: Although the number of significant digits of and is and , respectively, the limited paper space allows us to list both of them only up to 15 significant digits. We set the error bound to for .

As a first example, we select a function having a multiple zero with . We choose as an initial guess. We take another function with a root . We select as an initial value. The order of convergence and the asymptotic error constant are clearly shown in Tables 2 and 3 revealing a good agreement with the theory in Section 3. Taking another function with a root with multiplicity , we select as an initial value. In this example, we also find that the order of convergence is four and the computational asymptotic error constant well approaches the theoretical value . The computational convergence order and the computational asymptotic error constant are certainly shown in Tables 24 reaching a good agreement with the theory. It is certain that these methods need one evaluation of the function and two evaluations of the first derivative .

 0 0.00667483 0.0642487 0.2482221894 1 0.1363900716 4.21814 2 0.2482218636 4.00000 3 0.2482221894 4.00000 4 0.
 0 5.77466 0.113932 0.2202805193 1 0.1796711117 4.09332 2 0.2202681897 4.00001 3 0.2202805193 4.00000 4
 0 0.00250952 0.0979858 2.458883957 1 1.91755 2.024494357 4.08382 2 2.458901785 4.00000 3 2.458883957 4.00000 4 2.458883957 4.00000 5