1 Introduction

In this paper, we are concerned with the problem of approximating a locally unique solution x of the nonlinear equation

F(x)=0,
(1.1)

where F is twice Fréchet-differentiable operator defined on a nonempty open and convex subset of a Banach space X with values in a Banach space Y.

Many problems from computational sciences and other disciplines can be brought in a form similar to equation (1.1) using mathematical modeling [13]. The solutions of these equations can rarely be found in a closed form. That is why most solution methods for these equations are iterative. The study about convergence matter of iterative procedures is usually based on two types: semi-local and local convergence analysis. The semi-local convergence matter is, based on the information around an initial point, to give conditions ensuring the convergence of the iterative procedure; while the local one is, based on the information around a solution, to find estimates of the radii of convergence balls.

In this paper, we provide a semi-local convergence analysis for Halley’s method defined by [46]

x n + 1 = x n Γ F ( x n ) F ( x n ) 1 F( x n )
(1.2)

for each n0, where

Γ F (x)= ( I L F ( x ) ) 1 , L F (x)= 1 2 F ( x ) 1 F (x) F ( x ) 1 F(x).

The convergence of Halley’s method has a long history and has been studied by many authors (cf. [1, 2, 47] and the references therein). The most popular conditions for the semi-local convergence of Halley’s method are given as follows.

(C1) There exists x 0 D such that F ( x 0 ) 1 L(Y,X), the space of bounded linear operator from Y into X;

(C2) F ( x 0 ) 1 F( x 0 )η for any η>0;

(C3) F ( x 0 ) 1 F (x)M for each x in D and M>0;

(C4) F ( x 0 ) 1 [ F (x) F (y)]Kxy for each x,yD and K>0.

The corresponding sufficient convergence condition [1, 7] is given by

η 4 K + M 2 M M 2 + 2 K 3 K ( M + M 2 + 2 K ) .
(1.3)

There are simple examples to show that (C4) is not satisfied. As an example, let X=Y=R, D=[0,+) and define F(x) on D by

F(x)= 4 15 x 5 2 + x 2 +x+1.

Then we have

| F ( x ) F ( y ) | =| x y |= | x y | x + y .

Therefore, there is no constant K satisfying (C4). Other examples where (C4) is not satisfied can be found in [2].

Using the recurrent relations, Equerro and Hernández [7] expanded the applicability of Halley’s method by dropping condition (C4) and replacing (1.3) by

Mη< 4 6 5 .
(1.4)

In the present study, we show how to expand even further the applicability of Halley’s method using (C1), (C2), (C3) and the center-Lipschitz condition:

(C5) F ( x 0 ) 1 [ F (x) F ( x 0 )]Lx x 0 for each xD and L>0.

We have that

LM
(1.5)

holds in general and M L can be arbitrarily large [2, 4]. The semi-local convergence analysis of Halley’s method requires obtaining upper bounds on the norms F ( x ) 1 F ( x 0 ). In the literature, (C3) is used to obtain

F ( x ) 1 F ( x 0 ) 1 1 M x x 0
(1.6)

for each xD. However, if we use (C5) less expensive and tighter if L<M (see (1.5)), the estimate

F ( x ) 1 F ( x 0 ) 1 1 L x x 0
(1.7)

for each xD is obtained. This modification can lead to a tighter convergence analysis and weaker sufficient convergence conditions or error bounds or the location of the solution for Halley’s method (see numerical examples). The new approach has already led to such advantages in the case of Newton-type methods [1, 2, 46].

We use a different approach than recurrent relations in our semi-local convergence analysis. The paper is organized as follows. Section 2 contains the semi-local convergence of Halley’s method, whereas the numerical examples are presented in the concluding section.

2 Semi-local convergence

We present the semi-local convergence analysis of Halley’s method. We shall use an additional condition.

Define

a= M η 2 , η 0 = η 1 a .

(C6) Suppose that

( L + M 2 ) η<1,
(2.1)

and there exists

b(0,1L η 0 ),
(2.2)

which is the minimal positive zero of a function g on (0,1L η 0 ) given by

g(t)= a 1 L η 0 1 t + a 1 a ( 1 L η 0 1 t ) ( 1 a ( 1 a ) ( 1 L η 0 1 t ) 2 ( a 1 L η 0 1 t + a 1 a ) ) t.
(2.3)

Then we have

b= a 1 L η 0 1 b + a 1 a ( 1 L η 0 1 b ) ( 1 a ( 1 a ) ( 1 L η 0 1 b ) 2 ( a 1 L η 0 1 b + a 1 a ) ) .
(2.4)

Define

R 0 = η 0 1 b ,α=L R 0 .

Then we have

a ( 0 , 1 ) , α ( 0 , 1 ) , a ( 1 a ) ( 1 α ) 2 ( a 1 α + a 1 a ) = M 2 η 0 4 ( 1 α ) 2 ( η 1 α + η 0 ) < 1 .
(2.5)

We can rewrite (2.4) as

b= M 2 ( η 1 α + η 0 ) ( 1 α ) ( 1 M 2 η 0 4 ( 1 α ) 2 ( η 1 α + η 0 ) ) .

Let us introduce another condition needed in our main theorem.

(C7) Suppose that

d= M 2 ( 1 α ) η + M 2 η 0 ( 1 α ) ( 1 M 2 ( 1 α ) 2 η ) <1.

We refer to (C1), (C2), (C3), (C5), (C6) and (C7) as the (C) conditions. Let U(x,R), U ¯ (x,R) stand, respectively, for the open and closed balls in X with center x and radius R>0. Then we can show the following semi-local convergence result for Halley’s method.

Theorem 2.1 Let F:DXY be continuously twice Fréchet differentiable, where X, Y are Banach spaces and D is open and convex. Suppose the (C) conditions and U ¯ ( x 0 , R 0 )D. Then the Halley sequence { x n } generated by (1.2) is well defined, remains in U( x 0 , R 0 ) for all n0 and converges to a unique solution x U ¯ ( x 0 , R 0 ) of the equation F(x)=0. Moreover, the following error estimate holds: for each n0,

x n + 2 x n + 1 b x n + 1 x n .
(2.6)

Furthermore, if there exists R R 0 such that

U ¯ ( x 0 , R ) D
(2.7)

and

L 2 ( R 0 + R ) <1,
(2.8)

then x is the only solution of equation F(x) in U ¯ ( x 0 , R ).

Proof We have, using (C1)-(C3), that

I ( I L F ( x 0 ) ) = L F ( x 0 ) = 1 2 F ( x 0 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 0 ) 1 2 F ( x 0 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 0 ) 1 2 M η = a < 1 .
(2.9)

It follows from (2.9) and the Banach lemma on invertible operators [2, 3] that ( I L F ( x 0 ) ) 1 exists and

( I L F ( x 0 ) ) 1 1 1 L F ( x 0 ) 1 1 a .

Then, by (C2), (1.2) and the above estimate, we get

x 1 x 0 = ( I L F ( x 0 ) ) 1 F ( x 0 ) 1 F ( x 0 ) ( I L F ( x 0 ) ) 1 F ( x 0 ) 1 F ( x 0 ) η 1 a = η 0 < R 0 .

We have, by (C5),

I F ( x 0 ) 1 F ( x 1 ) L x 1 x 0 L η 0 <L R 0 =α<1.

Hence F ( x 1 ) 1 exists and

F ( x 1 ) 1 F ( x 0 ) < 1 1 α .

In view of Halley’s iteration, we can write

[ I L F ( x 0 ) ] ( x 1 x 0 )+ F ( x 0 ) 1 F( x 0 )=0

or

F( x 0 )+ F ( x 0 )( x 1 x 0 )= 1 2 F ( x 0 ) F ( x 0 ) 1 F( x 0 )( x 1 x 0 )

or

F ( x 0 ) 1 F ( x 1 ) = F ( x 0 ) 1 [ F ( x 1 ) F ( x 0 ) F ( x 0 ) ( x 1 x 0 ) + F ( x 0 ) + F ( x 0 ) ( x 1 x 0 ) ] = 1 2 F ( x 0 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 0 ) ( x 1 x 0 ) + 0 1 F ( x 0 ) 1 F ( x 0 + θ ( x 1 x 0 ) ) ( 1 θ ) d θ ( x 1 x 0 ) 2 .

Hence we get

F ( x 0 ) 1 F ( x 1 ) 1 2 M η x 1 x 0 + 1 2 M x 1 x 0 2 M 2 ( 1 α ) η x 1 x 0 + 1 2 M x 1 x 0 2 M 2 ( η 1 α + η 0 ) x 1 x 0 .

Moreover, we have

L F ( x 1 ) = 1 2 F ( x 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 1 ) F ( x 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 1 ) 1 2 F ( x 1 ) 1 F ( x 0 ) 2 F ( x 0 ) 1 F ( x 1 ) F ( x 0 ) 1 F ( x 1 ) M 2 ( 1 α ) 2 ( M 2 ( 1 α ) η x 1 x 0 + 1 2 M x 1 x 0 2 ) M 2 ( η 1 α + η 0 ) x 1 x 0 4 ( 1 α ) 2 M 2 η 0 4 ( 1 α ) 2 ( η 1 α + η 0 ) < 1 .

Then ( I L F ( x 1 ) ) 1 exists and

( I L F ( x 1 ) ) 1 1 1 L F ( x 1 ) 1 1 M 2 ( 1 α ) 2 ( M 2 ( 1 α ) η x 1 x 0 + 1 2 M x 1 x 0 2 ) 1 1 M 2 η 0 4 ( 1 α ) 2 ( η 1 α + η 0 ) .

So, x 2 is well defined and, using (1.2), we get

x 2 x 1 F ( x 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x 1 ) 1 L F ( x 1 ) M 2 ( 1 α ) η x 1 x 0 + 1 2 M x 1 x 0 2 ( 1 α ) ( 1 M 2 ( 1 α ) 2 [ M 2 ( 1 α ) η x 1 x 0 + 1 2 M x 1 x 0 2 ] ) M 2 ( η 1 α + η 0 ) ( 1 α ) ( 1 M 2 η 0 4 ( 1 α ) 2 ( η 1 α + η 0 ) ) x 1 x 0 = b x 1 x 0 .

Therefore, we have

x 2 x 0 x 2 x 1 + x 1 x 0 b x 1 x 0 + x 1 x 0 = ( 1 + b ) x 1 x 0 = 1 b 2 1 b x 1 x 0 < x 1 x 0 1 b η 0 1 b = R 0 .

Hence we have x 2 U( x 0 , R 0 ).

The above shows the following items are true for n=0:

  1. (a)

    F ( x n + 1 ) 1 exists and

    F ( x n + 1 ) 1 F ( x 0 ) < 1 1 α ;
  2. (b)
    F ( x 0 ) 1 F ( x n + 1 ) d n + 1 ;
  3. (c)
    d n + 1 d d n d n + 1 d 0 ;
  4. (d)

    ( I L F ( x n + 1 ) ) 1 exists and

    ( I L F ( x n + 1 ) ) 1 1 1 M 2 ( 1 α ) 2 d n + 1 ;
  5. (e)

    x n + 2 is well defined and

    x n + 2 x n + 1 d n + 1 ( 1 α ) ( 1 M 2 ( 1 α ) 2 d n + 1 ) b x n + 1 x n b n + 1 x 1 x 0 ;
  6. (f)

    x n + 2 x 0 < R 0 .

Here, the sequence { d n } is defined by

d 0 = η , d n + 1 = M 2 ( 1 α ) d n x n + 1 x n + M 2 x n + 1 x n 2

for each n0. The rest will be shown by induction.

Assume that (a)-(f) are true for all natural integers nk, where k0 is a fixed integer. Then F ( x k + 2 ) exists since x k + 2 U( x 0 , R 0 ) and

I F ( x 0 ) 1 F ( x k + 2 ) L x k + 2 x 0 <L R 0 =α<1.

Hence F ( x k + 2 ) 1 exists and

F ( x k + 2 ) 1 F ( x 0 ) 1 1 L x k + 2 x 0 < 1 1 α .

Next, we estimate F ( x 0 ) 1 F( x k + 2 ). It follows from Halley’s method that

F ( x k + 2 ) = F ( x k + 2 ) F ( x k + 1 ) F ( x k + 1 ) ( x k + 2 x k + 1 ) + 1 2 F ( x k + 1 ) F ( x k + 1 ) 1 F ( x k + 1 ) ( x k + 2 x k + 1 )

or

F ( x 0 ) 1 F ( x k + 2 ) = 1 2 F ( x 0 ) 1 F ( x k + 1 ) F ( x k + 1 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x k + 1 ) ( x k + 2 x k + 1 ) + 0 1 F ( x 0 ) 1 F ( x k + 1 + θ ( x k + 2 x k + 1 ) ) ( 1 θ ) d θ ( x k + 2 x k + 1 ) 2 .

Hence we get

F ( x 0 ) 1 F ( x k + 2 ) M 2 ( 1 α ) d k + 1 x k + 2 x k + 1 + M 2 x k + 2 x k + 1 2 = d k + 2 F ( x 0 ) 1 F ( x k + 2 ) = M 2 ( d k + 1 1 α + x k + 2 x k + 1 ) x k + 2 x k + 1 F ( x 0 ) 1 F ( x k + 2 ) M 2 ( d k + 1 η 1 α + b k + 1 x 1 x 0 ) x k + 2 x k + 1 F ( x 0 ) 1 F ( x k + 2 ) M 2 ( d η 1 α + b η 0 ) x k + 2 x k + 1 F ( x 0 ) 1 F ( x k + 2 ) M 2 ( d η 1 α + b η 0 ) b k + 1 x 1 x 0 F ( x 0 ) 1 F ( x k + 2 ) M 2 ( d η 1 α + b η 0 ) b η 0 , d k + 2 M 2 ( d η 1 α + b η 0 ) x k + 2 x k + 1 d k + 2 M 2 ( d η 1 α + b η 0 ) d k + 1 ( 1 α ) ( 1 M 2 ( 1 α ) 2 d k + 1 η ) d k + 2 M 2 ( d η 1 α + b η 0 ) d k + 1 ( 1 α ) ( 1 M 2 ( 1 α ) 2 d η ) d k + 2 M 2 ( η 1 α + η 0 ) d k + 1 ( 1 α ) ( 1 M 2 ( 1 α ) 2 η ) d k + 2 = d d k + 1 d k + 2 d 0
(2.10)

and

L F ( x k + 2 ) 1 2 F ( x k + 2 ) 1 F ( x 0 ) 2 F ( x 0 ) 1 F ( x k + 2 ) F ( x 0 ) 1 F ( x k + 2 ) M 2 ( 1 α ) 2 d k + 2 M 2 4 ( 1 α ) 2 ( η 1 α + η 0 ) η 0 < 1 .

Thus ( I L F ( x k + 2 ) ) 1 exists and

( I L F ( x k + 2 ) ) 1 1 1 M 2 ( 1 α ) 2 d k + 2 1 1 M 2 4 ( 1 α ) 2 ( η 1 α + η 0 ) η 0 .

Therefore, x k + 3 is well defined. Moreover, we obtain

x k + 3 x k + 2 ( I L F ( x k + 2 ) ) 1 F ( x k + 2 ) 1 F ( x 0 ) F ( x 0 ) 1 F ( x k + 2 ) d k + 2 ( 1 α ) ( 1 M 2 ( 1 α ) 2 d k + 2 ) M 2 ( η 1 α + η 0 ) x k + 2 x k + 1 ( 1 α ) ( 1 M 2 4 ( 1 α ) 2 ( η 1 α + η 0 ) η 0 ) = b x k + 2 x k + 1 b k + 2 x 1 x 0 .

Furthermore, we have

x k + 3 x 0 x k + 3 x k + 2 + x k + 2 x k + 1 + + x 1 x 0 ( b k + 2 + b k + 1 + + 1 ) x 1 x 0 = 1 b k + 3 1 b x 1 x 0 < η 0 1 b = R 0 .

Hence we deduce that x k + 3 U( x 0 , R 0 ), which completes the induction for (a)-(f).

Let m be a natural integer. Then we have

x k + m x k x k + m x k + m 1 + x k + m 1 x k + m 2 + + x k + 1 x k ( b m 1 + + b + 1 ) x k + 1 x k 1 b m 1 b b k x 1 x 0 .

It follows that { x k } is a complete sequence in a Banach space X and as such it converges to some x U ¯ ( x 0 , R 0 ) (since U ¯ ( x 0 , R 0 ) is a closed set). By letting k in (2.10), we obtain F( x )=0. We also have

x x k b k 1 b x 1 x 0 .

To show the uniqueness part, let y be a solution of the equation F(x)=0 in U ¯ ( x 0 , R ). Let T= 0 1 F ( x 0 ) 1 F ( x +θ( y x ))dθ. Using (C5), we have in turn that

I T = 0 1 F ( x 0 ) 1 [ F ( x + θ ( y x ) ) F ( x 0 ) ] d θ L 0 1 [ ( 1 θ ) x x 0 + θ y x 0 ] d θ L 2 ( R 0 + R ) < 1 .

It follows from the Banach lemma on invertible operators [24] that T 1 L(Y,X). Using the identity

0= F ( x 0 ) 1 ( F ( y ) F ( x ) ) = F ( x 0 ) 1 T ( y x ) ,

we deduce y = x . This completes the proof. □

3 Numerical examples

In this section, we give some examples to show the application of our theorem.

Example 3.1 Let us define a scalar function F(x)= x 3 0.64 on D=(0,1) with initial point x 0 =0.85. Then we have that

F (x)=3 x 2 , F (x)=6x.
(3.1)

So, F( x 0 )=0.025875, F ( x 0 )=2.1675 and η=0.011937716. Moreover, we have, for any xD,

| F ( x 0 ) 1 F ( x ) | = 2 | x | x 0 2 2 x 0 2 =2.76816609
(3.2)

and

| F ( x 0 ) 1 [ F ( x ) F ( x 0 ) ] | = | x + x 0 | x 0 2 |x x 0 |2.560553633|x x 0 |.
(3.3)

That is, we can choose M=2.76816609 and L=2.560553633 in conditions (C3) and (C5), respectively. Hence, we obtain a=0.016522791, η 0 =0.012138274 and 1L η 0 =0.968919297. Furthermore, it is easy to get b=0.035021656 by using iterative methods such as the secant method. Then we have R 0 =0.012578805, α=0.032208705 and d=0.035628901. So, conditions (C6) and (C7) are satisfied. It is clear that

U ¯ ( x 0 , R 0 )=[0.837421195,0.862578805]D.

Now, all the conditions in Theorem 2.1 are true and Theorem 2.1 applies. We can compare our results to the ones in [7]. Using (1.4), we get

α 0 =Mη=0.03304558< 4 6 5 .

So, the conditions of [7] are also satisfied. The uniqueness ball is U ¯ ( x 0 , r 0 η), where

r 0 = 2 3 α 0 2 ( 1 3 α 0 + α 0 2 ) =1.053745860.

Then we get R 1 = r 0 η=0.012579319> R 0 . That is, we provide better information on the location of the solution. Moreover, by (2.7) and (2.8), we can set R =0.15, which extends the uniqueness ball from U ¯ ( x 0 , R 0 ) to U ¯ ( x 0 , R ).

Example 3.2 In this example, we provide an application of our results to a special nonlinear Hammerstein integral equation of the second kind. Consider the integral equation

x(s)=1+ 4 5 0 1 G(s,t)x ( t ) 3 dt
(3.4)

for any s[0,1], where G is the Green kernel on [0,1]×[0,1] defined by

G(s,t)={ t ( 1 s ) , t s ; s ( 1 t ) , s t .
(3.5)

Let X=Y=C[0,1] and D be a suitable open convex subset of X 1 :={xX:x(s)>0,s[0,1]}, which is given below. Define a mapping F:DY by

[ F ( x ) ] (s)=x(s)1 4 5 0 1 G(s,t)x ( t ) 3 dt
(3.6)

for any s[0,1]. The first and second derivatives of F are given by

[ F ( x ) y ] (s)=y(s) 12 5 0 1 G(s,t)x ( t ) 2 y(t)dt
(3.7)

for any s[0,1] and

[ F ( x ) y z ] (s)= 24 5 0 1 G(s,t)x(t)y(t)z(t)dt
(3.8)

for any s[0,1], respectively.

We use the max-norm. Let x 0 (s)=1 for all s[0,1]. Then, for any yD, we have

[ ( I F ( x 0 ) ) y ] (s)= 12 5 0 1 G(s,t)y(t)dt
(3.9)

for any s[0,1], which means

I F ( x 0 ) 12 5 max s [ 0 , 1 ] 0 1 G(s,t)dt= 12 5 × 8 = 3 10 <1.
(3.10)

It follows from the Banach theorem that F ( x 0 ) 1 exists and

F ( x 0 ) 1 1 1 3 10 = 10 7 .
(3.11)

On the other hand, it follows from (3.6) that F( x 0 )= 4 5 max s [ 0 , 1 ] 0 1 G(s,t)dt= 1 10 . Then we get η= 1 7 .

Note that F (x) is not bounded in X or its subset X 1 . Take into account that a solution x of equation (1.1) with F given by (3.6) must satisfy

x 1 1 10 x 3 0,
(3.12)

i.e., x ρ 1 =1.153467305 and x ρ 2 =2.423622140, where ρ 1 and ρ 2 are the positive roots of the real equation z1 1 10 z 3 =0. Consequently, if we look for a solution such that x < ρ 1 X 1 , we can consider D={x:x X 1  and x<r}, with r( ρ 1 , ρ 2 ), as a nonempty open convex subset of X. For example, choose r=1.7.

Using (3.7) and (3.8), we have, for any x,y,zD,

| [ ( F ( x ) F ( x 0 ) ) y ] ( s ) | = 12 5 | 0 1 G ( s , t ) ( x ( t ) 2 x 0 ( t ) 2 ) y ( t ) d t | 12 5 0 1 G ( s , t ) | x ( t ) + x 0 ( t ) | | x ( t ) x 0 ( t ) | y ( t ) d t 12 5 0 1 G ( s , t ) ( r + 1 ) | x ( t ) x 0 ( t ) | y ( t ) d t , s [ 0 , 1 ]
(3.13)

and

| [ F ( x ) y z ] ( s ) | = 24 5 0 1 G(s,t)x(t)y(t)z(t)dt,s[0,1].
(3.14)

Then we get

F ( x ) F ( x 0 ) 12 5 × 1 8 (r+1)x x 0 = 81 100 x x 0
(3.15)

and

F ( x ) 24 5 × r 8 = 51 50 .
(3.16)

Now, we can choose constants in Theorem 2.1 as follows:

M = 51 35 , L = 81 70 , a = 51 490 , η 0 = 70 439 , ( L + M 2 ) η = 66 245 , 1 L η 0 = 358 439 , b = 0.4266517573 , R 0 = 0.2781089940 , α = 0.3218118359 , d = 0.5138822165 .
(3.17)

We also have

U ¯ ( x 0 , R 0 )D.
(3.18)

So, all the conditions in Theorem 2.1 are satisfied and Theorem 2.1 applies. Consequently, the sequence { x n } generated by Halley’s method (1.2) with initial point x 0 converges to the unique solution x of equation (1.1) on U ¯ ( x 0 , R 0 ). Moreover, by (2.7) and (2.8), we can set R =0.7, which extends the uniqueness ball from U ¯ ( x 0 , R 0 ) to U ¯ ( x 0 , R ).