- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Journal of Applied Mathematics
Volume 2013 (2013), Article ID 369067, 7 pages
A Two-Parameter Family of Fourth-Order Iterative Methods with Optimal Convergence for Multiple Zeros
Department of Applied Mathematics, Dankook University, Cheonan 330-714, Republic of Korea
Received 25 October 2012; Revised 2 December 2012; Accepted 19 December 2012
Academic Editor: Fazlollah Soleymani
Copyright © 2013 Young Ik Kim and Young Hee Geum. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
We develop a family of fourth-order iterative methods using the weighted harmonic mean of two derivative functions to compute approximate multiple roots of nonlinear equations. They are proved to be optimally convergent in the sense of Kung-Traub’s optimal order. Numerical experiments for various test equations confirm well the validity of convergence and asymptotic error constants for the developed methods.
A development of new iterative methods locating multiple roots for a given nonlinear equation deserves special attention on both theoretical and numerical interest, although prior knowledge about the multiplicity of the sought zero is required . Traub  discussed the theoretical importance of multiple-root finders, although the multiplicity is not known a priori by stating: “since the multiplicity of a zero is often not known a priori, the results are of limited value as far as practical problems are concerned. The study is, however, of considerable theoretical interest and leads to some surprising results.” This motivates our analysis for multiple-root finders to be shown in this paper. In case the multiplicity is not known, interested readers should refer to the methods suggested by Wu and Fu  and Yun [4, 5].
Various iterative schemes finding multiple roots of a nonlinear equation with the known multiplicity have been proposed and investigated by many researchers [6–12]. Neta and Johnson  presented a fourth-order method extending Jarratt's method. Neta  also developed a fourth-order method requiring one-function and three-derivative evaluations per iteration grounded on a Murakami’s method . Shengguo et al.  proposed the following fourth-order method which needs evaluations of one function and two derivatives per iteration for chosen in a neighborhood of the sought zero of with known multiplicity as follows: where , , , and with the following error equation: where − + , , and for .
Based on Jarratt  scheme for simple roots, J. R. Sharma and R. Sharma  developed the following fourth order of convergent scheme: where , , , and and derived the error equation below: where for .
The above error equation can be expressed in terms of as follows: where − + .
We now proceed to develop a new iterative method finding an approximate root of a nonlinear equation , assuming the multiplicity of is known. To do so, we first suppose that a function has a multiple root with integer multiplicity and is analytic in a small neighborhood of . Then we propose a new iterative method free of second derivatives below with an initial guess sufficiently close to as follows: where with , and as parameters to be chosen for maximal order of convergence [2, 19]. One should note that is obtained from Taylor expansion of about up to the first-order terms with weighted harmonic mean  of and .
Theorem 1 shows that proposed method (6) possesses 2 free parameters and . A variety of free parameters and give us an advantage that iterative scheme (6) can develop various numerical methods. One can often have a freedom to select best suited parameters and for a sought zero . Several interesting choices of and further motivate our current analysis. As seen in Table 1, we consider five kinds of methods Y1, Y2, Y3, Y4, and Y5 list selected parameters , and the corresponding values , respectively.
In this paper, we investigate the optimal convergence of the fourth-order methods for multiple-root finders with known multiplicity in the sense of optimal order claimed by Kung-Traub  and derive the error equation. We find that our proposed schemes require one evaluation of the function and two evaluations of first derivative and satisfy the optimal order. In addition, through a variety of numerical experiments we wish to confirm that the proposed methods show well the convergence behavior predicted by the developed theory.
2. Convergence Analysis
In this section, we describe a choice of parameters , and in terms of and to get fourth-order convergence for our proposed scheme (6).
Theorem 1. Let have a zero with integer multiplicity and be analytic in a small neighborhood of . Let , . Let be an initial guess chosen in a sufficiently small neighborhood of . Let , , , and . Let be two free constant parameters. Then iterative method (6) is of order four and defines a two-parameter family of iterative methods with the following error equation: where and
Proof. Using Taylor's series expansion about , we thave the following relations:
where , , and for .
Dividing (10) by (11), we obtain where , , and .
Expressing in terms of a new parameter for algebraic simplicity, we get Since can be expressed from in (11) with substituted by from (13), we get
With the aid of symbolic computation of Mathematica , we substitute (10)–(14) into proposed method (6) to obtain the error equation as where , and the coefficient may depend on parameters , and .
Solving and for and , respectively, we get after simplifications Putting , we have where and .
Observe that is satisfied with . Solving for , we get Substituting into (16) and (19) with , we can rearrange these expressions to obtain Calculating by the aid of symbolic computation of Mathematica , we arrive at the error equation below: where + with .
It is interesting to observe that error equation (23) has only one free parameter , being independent of . Table 1 shows typically chosen parameters and and defines various methods derived from (6). Method Y5 results in the iterative scheme (1) that Shengguo et al.  suggested.
3. Numerical Examples and Conclusion
In this section, we have performed numerical experiments using Mathematica Version 5 program to convince that the optimal order of convergence is four and the computed asymptotic error constant agrees well with the theoretical value . To achieve the specified sufficient accuracy and to handle small number divisions appearing in asymptotic error constants, we have assigned as the minimum number of digits of precision by the command and set the error bound to for . We have chosen the initial values close to the sought zero to get fourth-order convergence. Although computed values of are truncated to be accurate up to 250 significant digits and the inexact value of is approximated to be accurate enough about up to 400 significant digits (with the command , , ), we list them up to 15 significant digits because of the limited space.
As a first example with a double zero and an initial guess , we select a test function . As a second experiment, we take another test function with a root of multiplicity and with an initial value .
Taking another test function with a root of multiplicity , we select as an initial value.
Throughout these examples, we confirm that the order of convergence is four and the computed asymptotic error constant approaches well the theoretical value . The order of convergence and the asymptotic error constant are clearly shown in Tables 2, 3, and 4 reaching a good agreement with the theory developed in Section 2.
The additional test functions listed below further confirm the convergence behavior of our proposed method (6).
Table 5 shows the convergence behavior of among methods S, J, Y1, Y2, Y3, and Y4, where S denotes the method proposed by Shengguo et al. , J the method proposed by J. R. Sharma and R. Sharma , the methods Y1 to Y4 are described in Table 1. It is certain that proposed method (6) needs one evaluation of the function and two evaluations of the first derivative per iteration. Consequently, the corresponding efficiency index  is found to be , which is optimally consistent with the conjecture of Kung-Traub . For the particularly chosen test functions in these numerical experiments, methods Y1 to Y4 have shown better accuracy than methods S and J.
Nevertheless, the favorable performance of proposed scheme (6) is not always expected since no iterative method always shows best accuracy for all the test functions. If we look at the asymptotic error equation closely, we should note that the computational accuracy is sensitively dependent on the structures of the iterative methods, the sought zeros, convergence orders, the test functions, and good initial values.
It is important to properly choose enough number of precision digits. If is small, gets much smaller, as increases. If the number of precision is small and error bound is not small enough, the term causes a great loss of significant digits due to magnified round-off errors. This hinders us from verifying and more accurately.
Bold-face numbers in Table 5 refer to the least error until the prescribed error bound is met. This paper has confirmed optimal fourth-order convergence proved the correct error equation for proposed iterative methods (6), using the weighted harmonic mean of two derivatives to find approximate multiple zeros of nonlinear equations. We remark that the error equation of (6) contains only one free parameter , being independent of .
We still further need to discuss some aspects of root finding for ill-conditioned problems as well as sensitive dependence of zeros on initial values for iterative methods. As is well known, a high-degree (say, degree higher than 20, taking the multiplicity of a zero into account) polynomial is very likely to be ill conditioned. In this case, small changes in the coefficients can greatly alter the zeros. The small changes can occur as a result of rounding process in computing coefficients. Minimal round-off errors may improve the root finding of ill-conditioned problems. Certainly multi-precision arithmetic should be used in conjunction with optimized algorithms reducing round-off errors. High-order methods with the asymptotic error constant of small magnitude are preferred for locating zeros with relatively good accuracy. Locating zeros for ill-conditioned problems is generally believed to be a difficult task.
It is also important to properly choose close initial values near the root for guaranteed convergence of the proposed method. Indeed, initial values are chaotic  to the convergence of the root . The following statement quoted from  is not too much to emphasize the importance of selected initial values: “A point that belongs to the non-convergent region for a particular value of the parameter can be in the convergent region for another parameter value, even though the former might have a higher order of convergence than the second. This then indicates that showing whether a method is better than the other should not be done through solving a function from a randomly chosen initial point and comparing the number of iterations needed to converge to a root.”
Since our current analysis aims on the convergence of the proposed method, initial values [25–27] are selected in a small neighborhood of for guaranteed convergence. Thus the chaotic behavior of on the convergence should be separately treated under the different subject in future analysis. On the one hand, future research may be more strengthened with the graphical analysis on the convergence including chaotic fractal basins of attractions. On the other hand, rational approximations  provide rich resources of future research on developing new high-order optimal methods for multiple zeros.
The first author (Y. I. Kim) was supported by the Research Fund of Dankook University in 2011; the corresponding author (Y. H. Geum) was also supported by the National Research Foundation of Korea funded by the Ministry of Education, Science, and Technology (Project No. 2011-0014638). In addition, the authors would like to give special thanks to anonymous referees for their valuable suggestions and comments on this paper.
- A. Iliev and N. Kyurkchiev, Nontrivial Methods in Numerical Analysis (Selected Topics in Numerical Analysis), Lambert Academic, Saarbrucken, Germany, 2010.
- J. F. Traub, Iterative Methods for the Solution of Equations, Prentice-Hall Series in Automatic Computation, Prentice-Hall, Englewood Cliffs, NJ, USA, 1964.
- X. Wu and D. Fu, “New high-order convergence iteration methods without employing derivatives for solving nonlinear equations,” Computers & Mathematics with Applications, vol. 41, no. 3-4, pp. 489–495, 2001.
- B. I. Yun, “Iterative methods for solving nonlinear equations with finitely many roots in an interval,” Journal of Computational and Applied Mathematics, vol. 236, no. 13, pp. 3308–3318, 2012.
- B. I. Yun, “A derivative free iterative method for finding multiple roots of nonlinear equations,” Applied Mathematics Letters, vol. 22, no. 12, pp. 1859–1863, 2009.
- L. Atanassova, N. Kjurkchiev, and A. Andreev, “Two-sided multipointmethods of high order for solution of nonlinear equations, numerical methods and applications,” in Proceedings of the International Conference on Numerical Methods and Applications, pp. 33–37, Sofia, Bulgaria, August 1989.
- C. Dong, “A family of multipoint iterative functions for finding multiple roots of equations,” International Journal of Computer Mathematics, vol. 21, pp. 363–367, 1987.
- S. G. Li, L. Z. Cheng, and B. Neta, “Some fourth-order nonlinear solvers with closed formulae for multiple roots,” Computers & Mathematics with Applications, vol. 59, no. 1, pp. 126–135, 2010.
- X. Li, C. Mu, J. Ma, and L. Hou, “Fifth-order iterative method for finding multiple roots of nonlinear equations,” Numerical Algorithms, vol. 57, no. 3, pp. 389–398, 2011.
- M. S. Petković, L. D. Petković, and J. Džunić, “Accelerating generators of iterative methods for finding multiple roots of nonlinear equations,” Computers & Mathematics with Applications, vol. 59, no. 8, pp. 2784–2793, 2010.
- J. R. Sharma and R. Sharma, “Modified Chebyshev-Halley type method and its variants for computing multiple roots,” Numerical Algorithms, vol. 61, no. 4, pp. 567–578, 2012.
- X. Zhou, X. Chen, and Y. Song, “Constructing higher-order methods for obtaining the multiple roots of nonlinear equations,” Journal of Computational and Applied Mathematics, vol. 235, no. 14, pp. 4199–4206, 2011.
- B. Neta and A. N. Johnson, “High-order nonlinear solver for multiple roots,” Computers & Mathematics with Applications, vol. 55, no. 9, pp. 2012–2017, 2008.
- B. Neta, “Extension of Murakami's high-order non-linear solver to multiple roots,” International Journal of Computer Mathematics, vol. 87, no. 5, pp. 1023–1031, 2010.
- T. Murakami, “Some fifth order multipoint iterative formulae for solving equations,” Journal of Information Processing, vol. 1, pp. 138–139, 1978.
- L. Shengguo, L. Xiangke, and C. Lizhi, “A new fourth-order iterative method for finding multiple roots of nonlinear equations,” Applied Mathematics and Computation, vol. 215, no. 3, pp. 1288–1292, 2009.
- P. Jarratt, “Some efficient fourth order multipoint methods for solving equations,” BIT Numerical Mathematics, vol. 9, pp. 119–124, 1969.
- J. R. Sharma and R. Sharma, “Modified Jarratt method for computing multiple roots,” Applied Mathematics and Computation, vol. 217, no. 2, pp. 878–881, 2010.
- B. Neta, Numerical Methods for the Solution of Equations, Net-A-Sof, Calif, USA, 1983.
- R. Sharma, “Some more inequalities for arithmetic mean, harmonic mean and variance,” Journal of Mathematical Inequalities, vol. 2, no. 1, pp. 109–114, 2008.
- H. T. Kung and J. F. Traub, “Optimal order of one-point and multipoint iteration,” Journal of the Association for Computing Machinery, vol. 21, pp. 643–651, 1974.
- S. Wolfram, The Mathematica Book, Wolfram Media, Champaign, Ill, USA, 4th edition, 1999.
- D. Gulick, Encounters with Chaos, Prentice-Hall, New York, NY, USA, 1992.
- H. Susanto and N. Karjanto, “Newton's method's basins of attraction revisited,” Applied Mathematics and Computation, vol. 215, no. 3, pp. 1084–1090, 2009.
- F. Soleymani and S. Shateyi, “Two optimal eighth-order derivative-free classes of iterative methods,” Abstract and Applied Analysis, vol. 2012, Article ID 318165, 14 pages, 2012.
- F. Soleymani, S. Shateyi, and H. Salmani, “Computing simple roots by an optimal sixteenth-order class,” Journal of Applied Mathematics, vol. 2012, Article ID 958020, 13 pages, 2012.
- F. Soleymani, D. K. R. Babajee, S. Shateyi, and S. S. Motsa, “Construction of optimal derivative-free techniques without memory,” Journal of Applied Mathematics, vol. 2012, Article ID 497023, 24 pages, 2012.