Iterative Methods for Nonlinear Equations or Systems and Their Applications 2014
View this Special IssueResearch Article  Open Access
Young Ik Kim, Young Hee Geum, "A New Biparametric Family of TwoPoint Optimal FourthOrder MultipleRoot Finders", Journal of Applied Mathematics, vol. 2014, Article ID 737305, 7 pages, 2014. https://doi.org/10.1155/2014/737305
A New Biparametric Family of TwoPoint Optimal FourthOrder MultipleRoot Finders
Abstract
We construct a biparametric family of fourthorder iterative methods to compute multiple roots of nonlinear equations. This method is verified to be optimally convergent. Various nonlinear equations confirm our proposed method with order of convergence of four and show that the computed asymptotic error constant agrees with the theoretical one.
1. Introduction
It is not surprising that modified Newton’s method [1] in the simple form is most widely used to find the approximate multiple root of known multiplicity for a given nonlinear equation . Recall that numerical scheme (1) is a onepoint optimal method with quadratic convergence. In order to find numerical solution for multiple roots of nonlinear equations more accurately, many researchers have made enormous efforts in developing higherorder methods with improved convergence.
In this paper, we extend modified Newton’s method and propose twopoint optimal fourthorder multipleroot finders by evaluating two derivatives and one function per iteration. The optimality will be pursued based on KungTraub’s conjecture [2] in which the convergence order of any multipoint method [3] without memory can reach at most for evaluations of functions or derivatives.
The contents of this paper consist of what follows. Described in Section 2 are previous studies on multipleroot finders. Section 3 proposes a new biparametric family of twopoint optimal fourthorder multipleroot finders. It fully treats method development and convergence analysis. Derivation of the error equations for the proposed schemes is an important task for ensuring convergence behavior. In Section 4, a variety of numerical examples are presented for a wide selection of test functions. It is important to compare the convergence behavior of the proposed schemes with that of existing methods. We confirm that the proposed methods well show the convergence behavior predicted by the developed theory.
2. Preliminary Review of Previous Studies
A number of interesting fourthorder multipleroot finders can be found in papers [4–16]. Among these, we especially introduce five studies as follows. Shengguo et al. [14] introduced the following fourthorder method which needs evaluations of one function and two derivatives per iteration for chosen in a neighborhood of the sought zero of with known multiplicity : where , , , and .
J. R. Sharma and R. Sharma [12] constructed the following fourthorder scheme with , , , and : Li et al. [15] presented the fourthorder method with : Zhou et al. [16] proposed the following fourthorder iterative scheme with : Kanwar et al. [8] developed the fourthorder optimal multipoint iterative method for multiple zeros: where , , and .
3. Method Development and Convergence Analysis
We first suppose that a function has a multiple root with integer multiplicity and is analytic in a small neighborhood of . Then a new iteration method free of second derivatives is proposed below to find an approximate root of multiplicity , given an initial guess sufficiently close to : where with , , , , , and are parameters to be chosen for maximal order of convergence [17, 18]. We establish a main theorem describing the convergence analysis regarding proposed scheme (7) and find out how to select parameters , , and for optimal fourthorder convergence.
Definition 1 (error equation, asymptotic error constant, and order of convergence). Let be a sequence converging to and let be the th iterate error. If there exist real numbers and such that the following error equation holds then or is called the asymptotic error constant and is called the order of convergence [17, 18].
Theorem 2. Let have a zero with integer multiplicity and be analytic in a small neighborhood of . Let and for . Let be an initial guess chosen in a sufficiently small neighborhood of . Let , be two free constant parameters. Let , , , and . Then iterative methods (7) are optimal and of order four and possess the following error equation: where .
Proof. The optimality on convergence order of proposed scheme (7) is clear in the sense of KungTraub due to three functional evaluations. Hence, it suffices to determine the constant parameters for fourthorder convergence.
Applying the Taylor’s series expansion about , we get the following relations:
where , , and for .
Dividing (11) by (12), we have
where , , .
For algebraic convenience, we introduce a parameter defined by ; that is, to obtain
Evaluating from (11) with being replaced by in (14), we find
Substituting (11)–(15) into (7), we obtain the error equation:
where , , and coefficients , depend on the parameters , , , , , and and the function .
Solving , for and , respectively, we get
We substitute , into and put . Solving independently of and , that is, solving for and , we obtain
Substituting into (17) and (18) with , we get the following relations:
By the aid of symbolic computation of Mathematica [19], we arrive at the relation below:
where . As a result, the proof is completed.
Remark 3. We observe that error equation (20) contains only one free parameter , being independent of . Table 1 shows typically chosen parameters and and defines various methods , .

4. Numerical Examples and Conclusion
We have performed a variety of numerical experiments with Mathematica Version 5 [19] to confirm the theory developed in Section 3. In these experiments, we assign , via Mathematica command $MinPrecision = 300, as the minimum number of precision digits to achieve the specified sufficient accuracy. It is crucial to compute with high accuracy for desired numerical results. When zero is not exactly known, it is replaced by a highly accurate value which has larger number of significant digits than the assigned minimum number of precision digits. To deal with numerical results more effectively, we first define To properly display numerical results, we need to define the th computational error for . We need further terminologies as defined below.
Definition 4 (computational asymptotic error constant and computational convergence order). Assume that theoretical asymptotic error constant and convergence order are known (usually via main theorem). Define as the computational asymptotic error constant and as the computational convergence order. Then we find that is equal or close to , while is equal or close to .
If has the same accuracy of $MinPrecision as that of , then would be nearly zero and hence computing would unfavorably break down. Computed values of are accurate up to 300 significant digits. For current experiments is found to be accurate enough about up to 400 significant digits. To supply such , a set of following Mathematica commands are used:
Although the number of significant digits of and is and , respectively, the limited paper space allows us to list both of them only up to 15 significant digits. We set the error bound to for .
As a first example, we select a function having a multiple zero with . We choose as an initial guess. We take another function with a root . We select as an initial value. The order of convergence and the asymptotic error constant are clearly shown in Tables 2 and 3 revealing a good agreement with the theory in Section 3. Taking another function with a root with multiplicity , we select as an initial value. In this example, we also find that the order of convergence is four and the computational asymptotic error constant well approaches the theoretical value . The computational convergence order and the computational asymptotic error constant are certainly shown in Tables 2–4 reaching a good agreement with the theory. It is certain that these methods need one evaluation of the function and two evaluations of the first derivative .


