/ / Article

Research Article | Open Access

Volume 2013 |Article ID 687382 | 7 pages | https://doi.org/10.1155/2013/687382

# A Two-Point Newton Method Suitable for Nonconvergent Cases and with Super-Quadratic Convergence

Revised25 Dec 2012
Accepted17 Feb 2013
Published19 Mar 2013

#### Abstract

An iterative formula based on Newton’s method alone is presented for the iterative solutions of equations that ensures convergence in cases where the traditional Newton Method may fail to converge to the desired root. In addition, the method has super-quadratic convergence of order 2.414 (i.e., ). Newton method is said to fail in certain cases leading to oscillation, divergence to increasingly large number, or offshooting away to another root further from the desired domain or offshooting to an invalid domain where the function may not be defined. In addition when the derivative at the iteration point is zero, Newton method stalls. In most of these cases, hybrids of several methods such as Newton, bisection, and secant methods are suggested as substitute methods and Newton method is essentially blended with other methods or altogether abandoned. This paper argues that a solution is still possible in most of these cases by the application of Newton method alone without resorting to other methods and with the same computational effort (two functional evaluations per iteration) like the traditional Newton method. In addition, the proposed modified formula based on Newton method has better convergence characteristics than the traditional Newton method.

#### 1. Introduction

Iterative procedures for solutions of equations are routinely employed in many science and engineering problems. Starting with the classical Newton methods, a number of methods for finding roots of equations have come to exist, each of which has its own advantages and limitations. The Newton method of root finding is based on the iterative formula: Newton’s method displays a faster quadratic convergence near the root while it requires evaluation of the function and its derivative at each step of the iteration. However, when the derivative evaluated is zero, Newton method stalls. For low values of the derivative, the Newton iteration offshoots away from the current point of iteration and may possibly converge to a root far away from the intended domain. For certain forms of equations, Newton method diverges or oscillates and fails to converge to the desired root. In addition, the convergence of Newton method can be slow near roots of multiplicity although modifications can be made to increase the rate of convergence .

Modifications of the Newton method with higher order convergence have been proposed that require also evaluation of a function and its derivatives. An example of such methods is a third order convergence method by Weerakoon and Fernando  that requires evaluation of one function and two first derivatives. A fourth order iterative method, according to Traub  also requires evaluation of one function and two derivative evaluations. Grau-Sánchez and Díaz-Barrero  gave a compositing of function evaluation at a point and its derivative to improve the convergence of Newton’s method from 2 to 4. Recently other methods of fifth, sixth, seventh, and higher order convergence have been proposed . While higher order Newton methods ensure faster convergence, their stability for certain equation forms may have the same problem similar to that of the traditional Newton method.

The secant method does not require evaluation of derivatives. However, the rate of convergence is about 1.618, the convergence may be a problem for some forms of equations, and the secant method may fail to converge in those cases. Muller’s method is an extension of the secant method to a quadratic polynomial . It requires three functional evaluations to start with but continues with one function evaluation afterwards. The method does not require derivatives and the rate of convergence near the root is superlinear, that is, about 1.84. However, Muller’s method can converge to a complex root from an initial real number . Muller’s method also requires the three points to be distinct. If two of the points coincide, the method degenerates to secant method.

Hybrids of methods are also used to provide stability and ensure convergence to a desired root. For example, Newton or secant methods can be combined with bisection to bracket the root by a small interval so that a good initial guess is available for applying Newton method with quadratic convergence. Dekker’s method  combines bisection method with that of secant method. The method starts by bracketing the root between two initial points that have functional values opposite in sign. The secant estimate of is compared with the bisection of the interval, the one estimate resulting in small interval with the point of smaller magnitude functional value is chosen, and the iteration continues. Brent’s method  is a root finding algorithm that combines root bracketing, bisection, and inverse quadratic interpolation. It is a modification of Dekker’s method to avoid slow convergence when the difference between consecutive estimates of is arbitrarily small. In such cases bisection is used for the next root estimate.

The Leap-frogging Newton method  uses the Newton method as an intermediate step for the estimation of the root followed by the application of the secant method using the initial point and the intermediate point found by Newton method. The method has cubic convergence and works in certain pathological cases where Newton method may fail. However, being dependent on Newton method for the intermediate step, the method may suffer from the same drawbacks of using the traditional Newton method mentioned above.

#### 2. Method Development

It will be shown that the iterative formula for the two-point method will take the form The method starts by selecting two points lying on a curve , namely    and  . A line connecting the two points is drawn from the point    to the point    as shown in Figure 1. A new variable    (the cotangent of the angle between this line and the vertical) is defined so that In general, for any other point lying on the curve, the variable    is defined so that: Next, Newton method of finding roots will be applied taking    as the independent variable and as the dependent variable. Denoting that is the estimate of the root for which    from the Newton method with the corresponding value of   and applying Newton method gives Substituting for and the equivalent expressions in terms of and values, namely, into (5) above gives the following expression: The derivative is evaluated from and using the formula The evaluation of the derivative in turn gives Therefore, can now be written in terms of ,   and as follows: Substituting the above expression for in the equation gives Solving for the root estimate and further rearranging results in Continuing the iteration using the above formula, for the th step of the iteration, the  th estimate of the root will take the form Denoting   by   will result in the expression

#### 3. Proof of Super-Quadratic Convergence

Recalling the iteration formula of (14), Defining the error at the th iteration to be , where    is the root of the equation desired, the errors at the th and th iteration are also defined similarly as follows: The iteration formula in (14) can now be rewritten in terms of the error terms as follows: Expanding the , and , terms about the root using Taylor series expansion, where   for  .

The Taylor series expansions of    and about are formulated similarly as follows: The algebraic expression of (18), after substituting the previous Taylor series forms, was simplified using MATLAB program. After, the fourth order error terms in and were discarded and the error terms of the denominator were also discarded compared to the dominant term of , the resulting expression will be as follows: Therefore, the error sequence has the simplified expression Defining positive real terms and so that Near the root, the above expression approaches the constant terms in  . Therefore, the power of the error term shall approach zero: The positive solution of    is the one for which the error sequence converges to zero. Therefore, the iteration formula of (14) near the root has a convergence of order 2.414 which is super-quadratic.

#### 4. Stability of the Proposed Method

The proposed method shows stability near points where application of the traditional Newton method may result in oscillation, divergence, or offshooting away from the desired root or offshooting to a possibly invalid domain where the function may not be defined. This stability characteristic may be studied by examining the iteration formula again which takes the following form: where the variable    is defined as The estimate for the root at the th iteration, , therefore, can be taken as the weighted sum of the and values at the th and th iteration, respectively. The weighing factors are () for and () for . When the method converges to the root, the value of approaches unity. This is found by examining the behavior of near the root and noting that approaches 0 near the root; that is, The iterative formula will, near the root, approach the following expression: The iteration, therefore, moves away from with a weighing factor approaching zero and gives weighing factor of 1 to .

Near points where the derivative of the function may approach zero, the    value approaches infinity and the root estimate weighs heavily in favor of rather than . This can be seen from the following limit evaluation of : The iteration, therefore, moves away from with a weighing factor approaching zero and gives weighing factor of near to 1 to .

Near points where the derivative of the function may approach zero, Newton method typically displays oscillation or offshooting behavior. The proposed two-point modified Newton procedure however shows stability without displaying oscillation and offshooting tendencies. For example, for the function , the derivative of    (i.e., ) is zero at . Starting Newton method near this point, say , will result in offshooting of the iteration to a different domain.

As shown in Figure 2, for a starting point    where the function has near zero derivative, the traditional Newton method of shoots to a root further from the nearest root of . The Newton method as such converges to . On the other hand, the proposed two-point Newton method shows stability and converges to the nearest root of .

Another example of the stability of the proposed method is when Newton method diverges instead of converging to the desired root. Figures 3 and 4 show comparison of the iterative values of    for the proposed method and the traditional Newton method for the function which is often used to illustrate the pathological condition with respect to application of Newton’s method which leads to an oscillating divergence to infinity. By contrast the proposed method displays an oscillating convergence to the desired root albeit with slower rate of convergence of order 1.0.

#### 5. Application Examples

Equations used to test efficiency of root finding methods are used here to evaluate the number of iterations required to reach to a specified level of convergence (Table 1). The stopping criterion used for the iteration process is given by The rate of convergence towards the root for each step of the iteration is evaluated using the formula Table 1 shows comparison of the proposed two-point Newton method with the Newton and secant methods for a number of equations used to test efficiency of root finding methods elsewhere. A super-quadratic convergence with which the proposed method converges to the root is mostly evident with    values being close to 2.414 during most of the iterations. It can also be seen from Table 1 that a less number of iterations are required to reach convergence for the proposed method than those required for Newton and secant methods. For example, for equations with multiple roots such as , shown in Table 1, Newton method displays linear convergence while the proposed method converges super-quadratically with an order of 2.414. Similar order of convergence is also observed for the equation .

 Function Root Starting point Secant method Newton method Two-point Newton method 1.365230013414100 0.5 10 8 6 1 9 6 5 −1.404491648215340 −1 9 7 5 −3 9 7 6 −2.0000000000000 −3 168 119 83 1.4 116 81 60 2.00000000000000 1.5 25 17 8 2.5 12 8 6 3.5 16 11 8 = sin( ) ·  + ln( ) −0.603231971557215 −0.8 8 7 5 −0.65 8 5 4 3.000000000000000 4 29 20 14 4.5 39 28 18 1.857183860207840 2 8 5 4 0.5 11 8 5
##### 5.1. Examples Where the Proposed Method Works While Newton Method Fails

The advantage of the use of the proposed two-point Newton method is shown for cases where the Newton method and in several cases also the secant method fail to converge to the root. Table 2 shows the results of the iteration for several examples of equations. In all of the examples listed, Newton method fails to converge whereas the proposed two-point Newton method converges. Secant method also fails to converge in several of the examples cited in Table 2. For example, in the case of  , Newton method starting with fails with the second iteration because the estimated value is a negative number whose logarithm is undefined. For the case    Newton method always diverges to increasingly large number for any starting    value while the proposed two-point Newton method converges super-quadratically near the root. For  , Newton method also diverges whereby each iteration gives an estimate of the root which is twice the previous value and with alternate signs leading to oscillating divergence. On the other hand application of the proposed two-point modification of Newton method leads to oscillating convergence at somehow reduced rate of convergence (of order 1 as explained above).

 Function Root Comparison of number of iterations required Starting point Secant method Newton method Two-point Newton method 1.887207676120680 1 11 Oscillates 7 0.5 23 Oscillates 6 1.000000000000000 3.0 Fails Fails 5 = arctan 0.0000000000000 3.0 Diverges Diverges 6 −3.0 Diverges Diverges 6 −1.167303978261420 2.0 Oscillates Oscillates 12 3.0 14 Oscillates 15 4.00000000000000 3.0 10 Oscillates 7 5.0 8 Oscillates 6 0.00000000000000 1.0 Oscillates Diverges 101 −1.0 Oscillates Diverges 101 1.679630610428450 3.0 Diverges Diverges 8 0.101025848315685 −1.0 Diverges Diverges 11

For the equation  , both of Newton and secant methods continue to diverge to increasingly large values while the proposed two-point Newton method shows stability of convergence. For the polynomials of different degrees cited in the Table 2, Newton and secant methods display oscillation for the starting values shown in the Table 2 while the proposed two-point Newton method displays stability and super-quadratic convergence for iteration near the root.

#### 6. Conclusion

A numerical procedure of root finding using two-point modification of Newton method has been presented. It is proved that the method has a super-quadratic convergence of order about 2.414. The method is based on application of Newton iteration formula by taking as the independent variable the cotangent of the angle between the line connecting the two successive points of iteration with the vertical and as the dependent variable the given function . The resulting iteration formula for root estimation is shown to be the weighted sum of the estimates of the two previous iterations with a weighing factor that penalizes the iteration point having undesirable characteristics such as a near zero derivative. For example, near a point where the derivative is zero, the weighing factor for that point will be near zero effectively moving the iteration away from that undesirable point.

Application examples have been given to demonstrate that the proposed method requires fewer number of iterations for convergence to a root than the traditional Newton and secant methods. The method offers a particular advantage for cases where the traditional Newton method and its variants of various order convergence may not converge. A number of examples are given where the proposed method converges to a root in a stable manner without oscillation, divergence, or offshooting from the desired domain of the root, whereas the Newton method and in several of the cases also the secant method display undesirable behaviors which prevent convergence to the desired root.

In terms of computational effort, the proposed method requires one function evaluation and one derivative evaluation at each step of the iteration except for the first step where two functional evaluations are required. The proposed method, therefore, requires essentially the same number of functional evaluations as the traditional Newton method while offering a super-quadratic convergence of order 2.414.

1. C. F. Gerald and P. O. Wheatley, Applied Numerical Analysis, 5th edition, 1994. View at: Zentralblatt MATH
2. S. Weerakoon and T. G. I. Fernando, “A variant of Newton's method with accelerated third-order convergence,” Applied Mathematics Letters, vol. 13, no. 8, pp. 87–93, 2000.
3. J. F. Traub, Iterative Methods for the Solution of Equations, Prentice-Hall, Englewood Cliffs, NJ, USA, 1964. View at: MathSciNet
4. M. Grau-Sánchez and J. L. Díaz-Barrero, “A technique to composite a modified Newton's method for solving nonlinear equations,” Annals of the University of Bucharest, vol. 2, no. 1, pp. 53–61, 2011.
5. J. R. Sharma and R. K. Guha, “A family of modified Ostrowski methods with accelerated sixth order convergence,” Applied Mathematics and Computation, vol. 190, no. 1, pp. 111–115, 2007.
6. C. Chun, “Some improvements of Jarratt's method with sixth-order convergence,” Applied Mathematics and Computation, vol. 190, no. 2, pp. 1432–1437, 2007.
7. J. Kou and X. Wang, “Sixth-order variants of Chebyshev-Halley methods for solving non-linear equations,” Applied Mathematics and Computation, vol. 190, no. 2, pp. 1839–1843, 2007.
8. J. Kou, “On Chebyshev-Halley methods with sixth-order convergence for solving non-linear equations,” Applied Mathematics and Computation, vol. 190, no. 1, pp. 126–131, 2007.
9. J. Kou and Y. Li, “An improvement of the Jarratt method,” Applied Mathematics and Computation, vol. 189, no. 2, pp. 1816–1821, 2007.
10. J. Kou, Y. Li, and X. Wang, “Some modifications of Newton's method with fifth-order convergence,” Journal of Computational and Applied Mathematics, vol. 209, no. 2, pp. 146–152, 2007.
11. S. K. Parhi and D. K. Gupta, “A sixth order method for nonlinear equations,” Applied Mathematics and Computation, vol. 203, no. 1, pp. 50–55, 2008.
12. D. E. Muller, “A method for solving algebraic equations using an automatic computer,” Mathematical Tables and Other Aids to Computation, vol. 10, pp. 208–215, 1956.
13. W. R. Mekwi, Iterative methods for roots of polynomials [M.S. thesis], University of Oxford, 2001.
14. T. J. Dekker, “Finding a zero by means of successive linear interpolation,” in Constructive Aspects of the Fundamental Theorem of Algebra, B. Dejon and P. Henrici, Eds., Wiley-Interscience, London, UK, 1969. View at: Google Scholar | Zentralblatt MATH
15. R. P. Brent, Algorithms for Minimization without Derivatives, chapter 4, Prentice-Hall, Englewood Cliffs, NJ, USA, 1973. View at: MathSciNet
16. A. B. Kasturiarachi, “Leap-frogging Newton's method,” International Journal of Mathematical Education in Science and Technology, vol. 33, no. 4, pp. 521–527, 2002. View at: Publisher Site | Google Scholar | MathSciNet