Introduction

Many problems in science and engineering require solving a nonlinear scalar equation f(x)=0 [116]. As a result, solving nonlinear equations is an important part of scientific computing. There exist various iterative methods for solving nonlinear scalar equations. We are interested in sixth-order iterative methods, and their dynamics, to find a simple zero, that is f(γ)=0 and f(γ)≠0, of a nonlinear equation f(x)=0. There exist many sixth-order iterative methods (see, e.g., [57, 12, 13, 15]) for solving nonlinear scalar equations. The paper develops a scheme for constructing sixth-order iterative methods or family of methods. The scheme unifies existing sixth-order iterative methods. It is shown that various existing sixth-order iterative methods can be generated by the scheme through a proper choice of the independent parameters.

Let us first explore the existing sixth-order convergent iterative methods and family of methods. Lately, Sharma and Guha [13], through a modification of the well-known Ostrowski’s method [11], developed the following sixth-order convergent method (SG):

y n = x n f ( x n ) f ( x n ) , z n = y n f ( x n ) f ( x n ) 2 f ( y n ) f ( y n ) f ( x n ) , x n + 1 = z n f ( x n ) + af ( y n ) f ( x n ) + ( a 2 ) f ( y n ) f ( z n ) f ( x n ) ,
(1)

where aR. From here onwards, the preceding method is referred through the initials of the authors, i.e., SG. Neta [7] proposed a family, consisting of three steps and one parameter, of sixth-order convergent iterative methods (NETA):

y n = x n f ( x n ) f ( x n ) , z n = y n f ( x n ) + α f ( y n ) f ( x n ) + ( α 2 ) f ( y n ) f ( y n ) f ( x n ) , x n + 1 = z n f ( x n ) f ( y n ) f ( x n ) 3 f ( y n ) f ( z n ) f ( x n ) .
(2)

From here onwards, the above method is referred to as NETA. We observe that for α=0, the second step of the methods NETA and SG is the same, while for the choice a=−1, the third step of the methods NETA and SG is the same. In [6], Grau and Díaz-Barrero developed yet another sixth-order variant of Ostrowski’s method:

y n = x n f ( x n ) f ( x n ) , z n = y n f ( x n ) f ( x n ) 2 f ( y n ) f ( y n ) f ( x n ) , x n + 1 = z n f ( x n ) f ( x n ) 2 f ( y n ) f ( z n ) f ( x n ) .
(3)

The term f(x n )/(f(x n )−2f(y n ))is referred to as Ostrowski’s correction factor [6]. From here onwards, the above method is referred to as GD. Through a suitable modification of Ostrowski’s method, Chun and Ham [5] derived the following family of sixth-order iterative methods:

y n = x n f ( x n ) f ( x n ) , z n = y n f ( x n ) f ( x n ) 2 f ( y n ) f ( y n ) f ( x n ) , x n + 1 = z n ( u n ) f ( z n ) f ( x n ) ,
(4)

where u n =f(y n )/f(x n ) and (t) represents a real valued function satisfying (0)=1 and (0)=2. If one chooses (t)= 1 + β t 1 + ( β 2 ) t , one obtains the method SG, while if one chooses (t)= 1 1 2 t , one obtains the method GD. From here onwards, the above method is referred to as CH. It may be noticed that the methods SG, CH, and GD are formulated by modifying Ostrowski’s method; as a result, the first two steps of these methods are the same. Through a graceful modification of the well-known Kung and Traub method [17], recently Chun and Neta [16] proposed the following sixth-order iterative method:

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 1 f ( y n ) f ( x n ) 2 , x n + 1 = z n f ( z n ) f ( x n ) 1 1 f ( y n ) f ( x n ) f ( z n ) f ( x n ) 2 .
(5)

From here onwards, the above method is referred to as CN.

In this work, we propose a scheme for forming sixth-order convergent iterative methods. The methods formulated through the scheme consist of three steps. During each iteration, the developed methods require three functional evaluations and one evaluation of the derivative of the function. It is also shown that methods, such as SG, NETA, GD, and CH, that require three functional and one derivative evaluations can be formed through the proposed scheme. In consequence, the proposed scheme offers a unification of the sixth-order iterative methods, which is the main purpose of this work. The rest of the paper is organized as follows: The ‘Scheme for construct- ing sixth-order iterative methods’ section presents the scheme. In the ‘Unification of sixth-order iterative meth- ods’ section, through the scheme, we generate various well-known sixth-order iterative methods. In the ‘Numer- ical work’ section, numerical and dynamical comparisons of various methods are shown. Finally, the ‘Conclusions’ section concludes the article.

Scheme for constructing sixth-order iterative methods

Our aspiration is to develop a sixth-order iterative scheme which unifies sixth-order methods published in the literature. For this purpose, we consider the following three-step iterative scheme:

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 + j = 1 m a j f ( y n ) f ( x n ) j , x n + 1 = z n f ( z n ) f ( x n ) 1 + k = 1 l b k μ 1 f ( y n ) + μ 2 f ( z n ) f ( x n ) k .
(6)

Here, a j , b k , μ1, μ2, m, and l are independent parameters. The parameters a j , b k , μ1, μ 2 R, while the parameters m and l are positive integers. The parameters are established through the following convergence theorem:

Theorem 1. Let γ be a simple zero of a sufficiently differentiable function f:DRR in an open interval D. If the initialization x0 is sufficiently close to γ, then the scheme (6) defines sixth-order iterative methods iff a1=2and b1=2/μ1, and the error equation for the family of methods is given as

e n + 1 = c 2 c 3 c 1 5 c 2 2 + a 2 c 2 2 6 c 2 2 + c 2 2 b 2 μ 1 2 + c 3 c 1 c 1 5 × e n 6 + O e n 7 ,
(7)

where e n =x n γ and c m =fm(γ)/m! with m≥1.

Proof. The Taylor expansion of the function f (x) around the solution γ is given as

f( x n )= c 1 e n + c 2 e n 2 + c 3 e n 3 + c 4 e n 4 +O e n 5 .
(8)

Furthermore, from the preceding equation, we have

f ( x n )= c 1 +2 c 2 e n +3 c 3 e n 2 +4 c 4 e n 3 +O e n 4 .
(9)

Dividing Equations 8 and 9,

f ( x ) f ( x ) = e n c 2 c 1 e n 2 2 c 3 c 1 c 2 2 c 1 2 e n 3 3 c 4 c 1 2 7 c 2 c 3 c 1 + 4 c 2 3 c 1 3 e n 4 + O e n 5 .
(10)

From the first step of our scheme, we write

y n γ= e n f ( x n ) f ( x n ) ;
(11)

substituting Equation 10 into the preceding equation yields

y n γ = c 2 c 1 e n 2 + 2 c 3 c 1 c 2 2 c 1 2 e n 3 + 3 c 4 c 1 2 7 c 2 c 3 c 1 + 4 c 2 3 c 1 3 e n 4 + O e n 5 .
(12)

Expanding f(y n ), around the solution γ, through the Taylor series and using f(γ)=0,

f ( y n ) = j = 1 c j ( y n γ ) ;

substituting Equation 12 into the above equation, we obtain

f ( y n ) = c 2 e n 2 + 2 c 3 c 1 c 2 2 c 1 e n 3 + 3 c 4 c 1 2 7 c 2 c 3 c 1 + 5 c 2 3 c 1 2 e n 4 + O e n 5 .
(13)

Dividing Equations 8 and 13 gives

f ( y n ) f ( x n ) = c 2 c 1 e n + 2 c 3 c 1 3 c 2 2 c 1 2 e n 2 + 3 c 4 c 1 2 10 c 2 c 3 c 1 + 8 c 2 3 c 1 3 e n 3 + O e n 4 .
(14)

From the second step of our scheme (6), we may write

z n = x n f ( x n ) f ( x n ) 1 + f ( y n ) f ( x n ) 1 + a 1 f ( y n ) f ( x n ) + a 2 f ( y n ) f ( x n ) 2 + ;
(15)

substituting f(x n )/f(x n ), from Equation 10, and f(y n )/f(x n ), from Equation 14, into the above equation yields

z n γ = γ c 2 2 a 1 2 c 1 2 e n 3 c 2 4 a 1 c 3 c 1 7 a 1 c 2 2 + a 2 c 2 2 7 c 3 c 1 + 9 c 2 2 c 1 3 × e n 4 + O e n 5 .
(16)

The Taylor expansion of f(z n )around the solution γ is given as

f( z n )= k = 1 c k ( z n γ ) k ;
(17)

substituting z n γ, from Equation 16, into the above equation returns

f ( z n ) = c 2 2 a 1 2 c 1 e n 3 c 2 4 a 1 c 3 c 1 7 a 1 c 2 2 + a 2 c 2 2 7 c 3 c 1 + 9 c 2 2 c 1 2 × e n 4 + O e n 5 .
(18)

From the third step of our scheme (6), we have

x n + 1 = z n f ( z n ) f ( x n ) 1 + b 1 μ 1 f ( y n ) f ( x n ) + μ 2 f ( z n ) f ( x n ) + b 2 μ 1 f ( y n ) f ( x n ) + μ 2 f ( z n ) f ( x n ) 2 + ;

substituting Equations 9, 14, and 18 into the above equation furnishes the following error relation:

x n + 1 = γ + c 2 3 ( 2 + b 1 μ 1 ) a 1 2 e n 4 c 1 3 + c 2 2 c 1 4 11 a 1 11 b 1 μ 1 + 6 b 1 a 1 μ 1 + 20 c 1 c 3 + 2 + a 1 μ 1 2 b 2 + 12 a 1 + a 2 + 19 μ 1 + 4 + 4 a 1 a 1 2 μ 2 b 1 2 a 2 + 18 a 1 26 c 2 2 e n 5 + c 2 c 1 5 16 a 1 16 b 1 μ 1 + 28 + 9 a 1 b 1 μ 1 c 1 2 c 2 c 4 + 12 a 1 b 1 μ 1 20 b 1 μ 1 + 33 20 a 1 c 1 2 c 3 2 + 131 b 1 μ 1 89 a 1 b 1 μ 1 28 b 1 μ 2 15 b 2 μ 1 2 + 8 a 1 b 2 μ 1 2 + 8 b 1 μ 1 a 2 + 30 b 1 μ 2 a 1 8 b 1 μ 2 a 1 2 15 a 2 + 125 a 1 167 c 1 c 2 2 c 3 + 86 a 1 b 1 μ 1 15 b 1 μ 1 a 2 58 b 1 μ 2 a 1 + 17 b 1 μ 2 a 1 2 15 a 1 b 2 μ 1 2 + 4 a 2 b 1 μ 2 + a 2 b 2 μ 1 2 8 b 2 μ 1 μ 2 + a 1 b 3 μ 1 3 + b 1 μ 1 a 3 2 a 2 b 1 μ 2 a 1 + 8 b 2 μ 1 μ 2 a 1 2 b 2 μ 1 μ 2 a 1 2 a 1 2 98 a 1 + 24 a 2 2 a 3 + 108 + 48 b 1 μ 2 111 b 1 μ 1 + 25 b 2 μ 1 2 2 b 3 μ 1 3 c 2 4 e n 6 + O ( e n 7 ) .
(19)

From the above error relation, we may deduce that the three-step scheme (6) will define sixth-order methods, or family of methods, if the following three equations are satisfied simultaneously:

( a 1 2 ) ( b 1 μ 1 2 ) = 0 , 11 a 1 11 b 1 μ 1 + 6 b 1 a 1 μ 1 + 20 = 0 , 2 + a 1 μ 1 2 b 2 + 12 a 1 + a 2 + 19 μ 1 + 4 + 4 a 1 a 1 2 μ 2 b 1 2 a 2 + 18 a 1 26 = 0 .

From a simple calculation, we see that a solution is

a 1 = 2 and b 1 = 2 μ 1 .

Substituting a1=2and b1=2/μ1 in Equation 19 produces the required error equation (7). This proves our theorem. □

Consequently, this work contributes the following three-step sixth-order convergent iterative scheme for solving nonlinear equations:

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + j = 2 m a j f ( y n ) f ( x n ) j , x n + 1 = z n f ( z n ) f ( x n ) 1 + 2 μ 1 μ 1 f ( y n ) + μ 2 f ( z n ) f ( x n ) + k = 2 l b k μ 1 f ( y n ) + μ 2 f ( z n ) f ( x n ) k .
(20)

In the preceeding scheme, the parameters a j , b k (with j≥2 and k≥2), and μ m (with m=1,2) are free to choose. From here onwards, the above scheme is referred to as USS for short. Accordingly, USS presents opportunities to form various sixth-order methods. The next section explores few interesting choices of these parameters to formulate methods and family of methods from the published literature.

Unification of sixth-order iterative methods

Let us now derive methods from the published literature. Let us first construct the family of methods developed by Sharma and Guha [13]. For this purpose, we consider

a j = 2 j for j 2 , m = , μ 1 = 1 , b k = 2 ( 2 β ) k 1 for k 2 , l = , μ 2 = 0 .

Here, βR. Substituting these values in the second and third steps of the proposed scheme USS, we get

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + j = 2 2 j f ( y n ) f ( x n ) j , x n + 1 = z n f ( z n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + k = 2 2 ( 2 β ) k 1 f ( y n ) f ( x n ) k .

Using 1 + r + r2 + r3 + ⋯=1/(1−r)for |r|<1 in the second and third steps of the preceeding equation, we obtain the method, SG, developed by Sharma and Guha [13]. Now, to formulate the family of methods developed by Chun and Ham [5], we consider

a j = 2 j for j 2 , m = , μ 1 = 1 , b k = ω k for k 2 , l = , μ 2 = 0 .

Here, aR. Substituting the preceding choices in the second and third steps of the scheme USS, we get

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + j = 2 2 j f ( y n ) f ( x n ) j , x n + 1 = z n f ( z n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + k = 2 ω k f ( y n ) f ( x n ) k .
(21)

Using 1 + r + r2 + r3 + ⋯=1/(1−r)for |r|<1 in the second step of the preceding equation, we obtain the second step of the method, CH, developed by Chun and Ham [5]. The third step of the method CH is given as x n + 1 = z n f( z n )/ f ( x n )( u n ) (see Equation 4). Here, u n =f(y n )/f(x n ) and (t) is a real valued function satifying (0)=1 and (0)=2. Through a simple calculation, the function ( u n ) may be expressed as

f ( y n ) f ( x n ) = 1 + 2 f ( y n ) f ( x n ) + k = 2 ω k f ( y n ) f ( x n ) k .

As a consequence, the formulated method (21), through the scheme USS, is the method developed by Chun and Ham [5]. The method of Grau and Díaz-Barrero [6] can be derived by considering the following:

a j = 2 j for j 2 , m = , μ 1 = 1 , b k = 2 k for k 2 , l = , μ 2 = 0 ,

in the second and third steps of the developed scheme USS. To derive the family of methods developed by Neta et al. (2), the choices are

a j = 2 ( ( a 2 ) ) j 1 for j 2 , m = , μ 1 = 1 , b k = 2 ( ( 3 ) ) k 1 for k 2 , l = , μ 2 = 0 .

Here, aR. To formulate the recently developed sixth-order method CN by Chun and Neta [16], the parameters, in the scheme USS, are chosen as follows:

a j = ( j + 1 ) for j 2 , m = , μ 1 = 1 , b k = ( k + 1 ) for k 2 , l = , μ 2 = 1 .

Substituting the above values in the scheme USS, we get

y n = x n f ( x n ) f ( x n ) , z n = y n f ( y n ) f ( x n ) 1 + 2 f ( y n ) f ( x n ) + j = 2 ( j + 1 ) f ( y n ) f ( x n ) j , x n + 1 = z n f ( z n ) f ( x n ) 1 + 2 f ( y n ) + f ( z n ) f ( x n ) + k = 2 ( k + 1 ) f ( y n ) + f ( z n ) f ( x n ) k .

Using 1 + 2r + 3r2 + 4r3 + ⋯=1/(1−r)2for |r|<1 in the second and third steps of the preceding method yields the method of Chun and Neta [16].

Numerical work

Let { x n } n = 0 be a sequence, generated by an iterative method, converging to γ and e n =x n γ. If there exist a real number ξ∈ [1,) and a nonzero constant C such that

lim n | e n + 1 | / | e n | ξ = C ,

then ξ is called the convergence order of the sequence and the constant C is called the asymptotic error constant. From the preceding relation, the computational order of convergence (COC) is approximated as follows:

ρ log | ( x n + 1 γ ) / ( x n γ ) | log | ( x n γ ) / ( x n 1 γ ) | .

All the computations are performed in the programming language C+ + . For numerical precision, the C+ + library ARPREC [1] is being used. For the convergence of the method, it is required that the distance of two consecutive iterates (|xn + 1x n |) and the absolute value of the function (|f(x n )|), also referred to as residual, be less than 10−300. The maximum allowed iteration is 200.

Solving nonlinear equations

The methods are tested for the following functions:

f 1 ( x ) = x 3 + 4 x 2 10 , γ 1 . 365 . f 2 ( x ) = x 2 exp ( x ) 3 x + 2 , γ 0 . 257 . f 3 ( x ) = x exp ( x 2 ) sin 2 ( x ) + 3 cos ( x ) + 5 , γ 1 . 207 . f 4 ( x ) = x 4 + sin Π x 2 5 , γ = 2 . f 5 ( x ) = e x sin x + log ( 1 + x 2 ) , γ = 0 . f 6 ( x ) = 2 + x 2 sin Π x 2 + 1 x 4 + 1 17 3 + 1 17 , γ = 2 .

Various free parameters are chosen as in the method SG: a=2, in the method NETA: α=5, and in the method CH: β=3. The outcome of the numerical experimentation is presented in Table 1. Table 1 reports the number of function evaluations and COC during the second-to-the-last iterative step. COC is rounded to the nearest significant digits. In Table 1, we can see that the methods GD and CN show better results.

Table 1 The (number of function evaluations, COC) for various sixth-order iterative methods

Dynamic behavior of various sixth-order methods

Let f (x) be a complex polynomial and γbe one of its zeros. Furthermore, let x 0 C be a starting point for an iterative method. Then, the sequence { x n } n = 0 , generated by the iterative methods, may converge or may not converge to the zero γ. If the sequence converges to the zero γ, then the starting point x 0 is attracted to γ. The basin of attraction, corresponding to a zero γof the complex polynomial f (x), is the set of all the starting points x 0 which are attracted to γ.

To make the figures, first we take a rectangle D, of size [−1. 5,1 . 5]×[−1. 5,1,5], and then divide the rectangle into 1,000×1,000grids [18–20]. Further, we apply the iterative method starting at each grid point. The iterative methods converge if the residual, in a maximum of 10 iterations, is less than 10−15. Given that the iterative method does not generate a residual less than 10−10 in the maximum allowed iterations, we say that the initial point does not converge to any root. Let us denote the number of mean iterations by μ ¯ and the percentage of diverging points by ω[14].

We consider the polynomial f(x)= x n 1,xC for finding the n th roots of unity. The n th roots of unity are given by [18–20]

α j = cos 2 Π ( j 1 ) n + i sin 2 Π ( j 1 ) n , j = 1 , 2 , , n.

The outcome of our numerical experimentation is reported in Figures 1 and 2. From these figures, we notice that the methods GD and CN which require the least number of iterations μ ¯ while diverging at the least number of points show better results.

Figure 1
figure 1

The attraction basins for the polynomial f(z)=z3−1 . (a) SG, μ ¯ =4.46 and ω =10 . 37, a =2; (b) CN, μ ¯ =3.98 and ω =1 . 98; (c) GD, μ ¯ =3.94 and ω =0 . 56; (d) NETA, μ ¯ =4.25 and ω =10 . 95, α =5.

Figure 2
figure 2

The attraction basins for the polynomial f(z)=z7−1 . (a) SG, μ ¯ =5.86 and ω =30 . 33, a =2; (b) CN, μ ¯ =5.74 and ω =23 . 20; (c) GD, μ ¯ =5.19 and ω =11 . 48; (d) NETA, μ ¯ =6.49 and ω=45. 33, α=5.

Conclusions

This work has developed a scheme to unify various sixth-order iterative methods. Comparison among iterative methods, by using the basins of attraction and also through numerical computations, is also presented. Ideas presented in this work can be further developed and extended to include iterative methods of higher orders such as seven or eight.