1 Introduction

In 1994, Censor and Elfving [1] first introduced the split feasibility problem (SFP) in finite-dimensional Hilbert spaces for modeling inverse problems which arise from phase retrievals and in medical image reconstruction [2]. It was found that the SFP can also be used to model intensity-modulated radiation therapy (IMRT) (see [36]). Very recently, Xu [7] considered the SFP in the framework of infinite-dimensional Hilbert spaces. In this setting, the SFP is formulated as the problem of finding a point x with the property

x CandA x Q,
(1.1)

where C and Q are the nonempty closed convex subsets of the infinite-dimensional real Hilbert spaces H 1 and H 2 , respectively. Let AB( H 1 , H 2 ), where B( H 1 , H 2 ) denotes the family of all bounded linear operators from H 1 to H 2 .

We use Γ to denote the solution set of the SFP, i.e.,

Γ={xC:AxQ}.

Assume that the SFP is consistent (i.e., (1.1) has a solution) so that Γ is closed, convex and nonempty. A special case of the SFP is the following convex constrained linear inverse problem:

find xCsuch that Ax=b,
(1.2)

which has extensively been investigated by using the Landweber iterative method [8]:

let  x 0  be arbitrary for  n = 0 , 1 , ,  let x n + 1 = x n + γ A T ( b A x n ) .

Comparatively, the SFP has received much less attention so far due to the complexity resulting from the set Q. Therefore, whether various versions of the projected Landweber iterative method [8] can be extended to solve the SFP remains an interesting open topic.

The original algorithm given in [1] involves the computation of the inverse A 1 (assuming the existence of the inverse of A):

x k + 1 = A 1 P Q ( P A ( C ) ( A x k ) ) ,k0,

where C,Q R n are closed convex sets, A is a full rank n×n matrix and A(C)={y R n |y=Ax,xC}, and thus has not become popular.

A more popular algorithm that solves the SFP seems to be the CQ algorithm of Byrne [2, 9] which is found to be a gradient-projection method (GPM) in convex minimization. It is also a special case of the proximal forward-backward splitting method [10]. The CQ algorithm only involves the computations of the projections P C and P Q onto the sets C and Q, respectively, and is therefore implementable in the case where P C and P Q have closed-form expressions (for example, C and Q are closed balls or half-spaces). It remains, however, a challenge on the CQ algorithm in the case where the projection P C and/or P Q fail to have closed-form expressions though theoretically we can prove the (weak) convergence of the algorithm.

Recently, Xu [7] gave a continuation of the study on the CQ algorithm and its convergence. He applied Mann’s algorithm to the SFP and proposed an averaged CQ algorithm, which was proved to be weakly convergent to a solution of the SFP. He derived a weak convergence result, which shows that for suitable choices of iterative parameters (including the regularization), the sequence of iterative solutions can converge weakly to an exact solution of the SFP. He also established the strong convergence result, which shows that the minimum-norm solution can be obtained. Later, Deepho and Kumam [11] extended the results of Xu [7] by introducing and studying the modified Halpern iterative scheme for solving the split feasibility problem (SFP) in the setting of infinite-dimensional Hilbert spaces.

Throughout this paper, we always assume that the SFP is consistent, that is, the solution set Γ of the SFP is nonempty. Let f: H 1 R be a continuous differentiable function. The minimization problem

min x C f(x):= 1 2 A x P Q A x 2
(1.3)

is ill-posed. Therefore (see [7]), consider the following Tikhonov regularized problem:

min x C f α (x):= 1 2 A x P Q A x 2 + 1 2 α x 2 ,
(1.4)

where α>0 is the regularization parameter.

We observe that the gradient

f α (x)=f(x)+αI= A (I P Q )A+αI
(1.5)

is (α+ A 2 )-Lipschitz continuous and α-strongly monotone.

Define the Picard iterates

x n + 1 α = P C ( I γ ( A ( I P Q ) A + α I ) ) x n α .
(1.6)

Xu [7] showed that if SFP (1.1) is consistent, then as n, x n α x α and consequently the strong lim α 0 x α exists and is the minimum-norm solution of the SFP. Note that (1.6) is double-step iteration. Xu [7] further suggested the following single step regularized method:

x n + 1 = P C (Iγ f α n ) x n = P C ( ( 1 α n γ n ) x n γ n A ( I P Q ) A x n ) .
(1.7)

He proved that the sequence { x n } generated by (1.7) converges in norm to the minimum-norm solution of the SFP provided the parameters { α n } and { γ n } satisfy the following conditions:

  1. (i)

    α n 0 and 0< γ n < α n A 2 + α n ;

  2. (ii)

    n α n γ n =;

  3. (iii)

    | γ n + 1 γ n | + γ n | α n + 1 α n | ( α n + 1 γ n + 1 ) 2 0.

Motivated by the idea of the relaxed extragradient method and Xu’s regularization, Ceng, Ansari and Yao [12] presented the following relaxed extragradient method with regularization for finding a common element of the solution set of the split feasibility problem and the set Fix(S) of fixed points of a nonexpansive mapping S:

{ x 0 = x C chosen arbitrarily , y n = ( 1 β n ) + β n P C ( x n λ f α n ( x n ) ) , x n + 1 = γ n x n + ( 1 γ n ) S P C ( y n λ f α n ( y n ) ) , n 0 .
(1.8)

They only obtained the weak convergence of iterative algorithm (1.8).

The purpose of this paper to study and analyze an relaxed extragradient method with regularization for finding a common element of the solution set Γ of the SFP and the set solutions of fixed points for asymptotically quasi-nonexpansive mappings and a Lipschitz continuous mapping in a real Hilbert space. We prove that the sequence generated by the proposed method converges weakly to an element x ˆ in Fix(T)Γ.

2 Preliminaries

We first recall some definitions, notations, and conclusions which will be needed in proving our main results. Let H be a real Hilbert space with the inner product , and and let C be a closed and convex subset of H. Let E be a Banach space. A mapping T:EE is said to be demi-closed at origin if for any sequences { x n }E with x n x and (IT) x n 0, x =T x . A Banach space E is said to have the Opial property if for any sequence { x n } with x n x ,

lim inf n x n x < lim inf n x n y,yE with y x .

Remark 2.1 It is well known that each Hilbert space possesses the Opial property.

Definition 2.2 Let H be a real Hilbert space, let C be a nonempty and closed convex subset. We denote by Fix(T) the set of fixed points of T, that is, Fix(T)={xC:x=Tx}. Then T is said to be

  1. (i)

    nonexpansive ifTxTyxy for all x,yC;

  2. (ii)

    quasi-nonexpansive ifTxpxp for all xC and pF(T);

  3. (iii)

    asymptotically nonexpansive if there exist a sequence k n 1 and lim n k n =1 such that

    T n x T n y k n xy

for all x,yC and n1;

  1. (iv)

    asymptotically quasi-nonexpansive if there exist a sequence k n 1 and lim n k n =1 such that

    T n x p k n xp

for all xC, pF(T) and n1;

  1. (v)

    uniformlyL-Lipschitzian if there exists a constant L>0 such that

    T n x T n y Lxy

for all x,yC and n1.

Remark 2.3 By the above definitions, it is clear that:

  1. (i)

    a nonexpansive mapping is an asymptotically quasi-nonexpansive mapping;

  2. (ii)

    a quasi-nonexpansive mapping is an asymptotically-quasi nonexpansive mapping;

  3. (iii)

    an asymptotically nonexpansive mapping is an asymptotically quasi-nonexpansive mapping.

Proposition 2.4 (see [9])

We have the following assertions.

  1. (i)

    Tis nonexpansive if and only if the complementITis 1 2 -ism.

  2. (ii)

    IfTisν-ism andγ>0, thenγTis ν γ -ism.

  3. (iii)

    Tis averaged if and only if the complementITisν-ism for someν> 1 2 .

Indeed, forα(0,1), Tisα-averaged if and only ifITis 1 2 α -ism.

Proposition 2.5 (see [9, 13])

We have the following assertions.

  1. (i)

    IfT=(1α)S+αVfor someα(0,1), Sis averaged andVis nonexpansive, thenTis averaged.

  2. (ii)

    Tis firmly nonexpansive if and only if the complementITis firmly nonexpansive.

  3. (iii)

    IfT=(1α)S+αVfor someα(0,1), Sis firmly nonexpansive andVis nonexpansive, thenTis averaged.

  4. (iv)

    The composite of finite many averaged mappings is averaged. That is, if each of the mappings { T i } i = 1 n is averaged, then so is the composite T 1 T 2 T N . In particular, if T 1 is α 1 -averaged and T 2 is α 2 -averaged, where α 1 , α 2 (0,1), then the composite T 1 T 2 isα-averaged, whereα= α 1 + α 2 α 1 α 2 .

  5. (v)

    If the mappings { T i } i = 1 n are averaged and have a common fixed point, then

    i = 1 n Fix( T i )=Fix( T 1 T N ).

Lemma 2.6 (see [14], demiclosedness principle)

LetCbe a nonempty closed and convex subset of a real Hilbert spaceHand letS:CCbe a nonexpansive mapping withFix(S). If the sequence{ x n }Cconverges weakly toxand the sequence{(IS) x n }converges strongly toy, then(IS)x=y; in particular, ify=0, thenxFix(S).

Lemma 2.7 (see [15])

Let the sequences { a n } and { u n } of real numbers satisfy

a n + 1 (1+ u n ) a n ,n1,

where a n 0, u n 0and n = 1 u n <. Then

  1. (1)

    lim n a n exists;

  2. (2)

    if lim inf n a n =0, then lim n a n =0.

The following lemma gives some characterizations and useful properties of the metric projection P C in a Hilbert space.

For every point xH, there exists a unique nearest point in C, denoted by P C x, such that

x P C xxy,yC,
(2.1)

where P C is called the metric projection ofH onto C. We know that P C is a nonexpansive mapping of H onto C.

Proposition 2.8For givenxHandzC:

  1. (i)

    z= P C xif and only ifxz,yz0for allyC.

  2. (ii)

    z= P C xif and only if x z 2 x y 2 y z 2 for allyC.

  3. (iii)

    For allyH, P C x P C y,xy P C x P C y 2 .

Lemma 2.9 (see [16])

LetHbe a real Hilbert space. Then the following equations hold:

  1. (i)

    x y 2 = x 2 y 2 2xy,yfor allx,yH;

  2. (ii)

    t x + ( 1 t ) y 2 =t x 2 +(1t) y 2 t(1t) x y 2 for allt[0,1]andx,yH.

Let K be a nonempty closed convex subset of a real Hilbert space H and let F:KH be a monotone mapping. The variational inequality problem (VIP) is to find xK such that

Fx,yx0,yK.

The solution set of the VIP is denoted by VIP(K,F). It is well known that

xVI(K,F)x= P K (xλFx),λ>0.

A set-valued mapping T:H 2 H is called monotone if for all x,yH, fTx and gTy imply

xy,fg0.

A monotone mapping T:H 2 H is called maximal if its graph G(T) is not properly contained in the graph of any other monotone mapping. It is known that a monotone mapping T is maximal if and only if for (x,f)H×H, xy,fg0 for every (y,g)G(T) implies fTx. Let F:KH be a monotone and k-Lipschitz continuous mapping and let N K v be the normal cone to K at vK, that is,

N K v= { w H : v u , w 0 , u K } .

Define

Tv={ F v + N K v if  v K , if  v K .

Then T is maximal monotone and 0Tv if and only if vVI(K,F); see [15] for more details.

We can use fixed point algorithms to solve the SFP on the basis of the following observation.

Let λ>0 and assume that x Γ. Then A x Q, which implies that (I P Q )A x =0, and thus λ A (I P Q )A x =0. Hence, we have the fixed point equation (Iλ A (I P Q )A) x = x . Requiring that x C, we consider the fixed point equation

P C (Iλf) x = P C ( I λ A ( I P Q ) A ) x = x .
(2.2)

It is proved in [[7], Proposition 3.2] that the solutions of fixed point equation (2.2) are exactly the solutions of the SFP; namely, for given x H 1 , x solves the SFP if and only if x solves fixed point equation (2.2).

Proposition 2.10 (see [12])

Given x H 1 , the following statements are equivalent.

  1. (i)

    x solves the SFP;

  2. (ii)

    x solves fixed point equation (2.2);

  3. (iii)

    x solves the variational inequality problem (VIP) of finding x Csuch that

    f ( x ) , x x 0,xC,
    (2.3)

wheref= A (I P Q )Aand A is the adjoint ofA.

Proof (i) ⇔ (ii). See the proof in [[7], Proposition 3.2].

  1. (ii)

    ⇔ (iii). Observe that

    P C ( I λ A ( I P Q ) A ) x = x ( I λ A ( I P Q ) A ) x x , x x 0 , x C λ A ( I P Q ) A x , x x 0 , x C f ( x ) , x x 0 , x C ,

where f= A (I P Q )A. □

Remark 2.11 It is clear from Proposition 2.10 that

Γ:=Fix ( P C ( I λ f ) ) =VI(C,f)

for any λ>0, where Fix( P C (Iλf)) and VI(C,f) denote the set of fixed points of P C (Iλf) and the solution set of VIP.

3 Main result

Theorem 3.1LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceHand letT:CCbe a uniformlyL-Lipschitzian and asymptotically quasi-nonexpansive mappings withFix(T)Γand{ k n }[1,)for allnNsuch that n = 1 ( k n 1)<. Let{ x n }and{ y n }be the sequences inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , y n = P C ( I λ n f α n ) x n , x n + 1 = β n x n + ( 1 β n ) T n y n , n 0 ,
(3.1)

where f α n =f+ α n I= A (I P Q )A+ α n I, and three sequences{ α n }, { λ n }, and{ β n }satisfy the conditions:

  1. (i)

    n = 1 α n <,

  2. (ii)

    { λ n }[a,b]for somea,b(0, 1 A 2 )and n = 1 | λ n + 1 λ n |<,

  3. (iii)

    { β n }[c,d]for somec,d(0,1).

Then the sequences{ x n }and{ y n }converge weakly to an element x ˆ Fix(T)Γ.

Proof We first show that P C (Iλ f α ) is ζ-averaged for each λ n (0, 2 α + A 2 ), where

ζ= 2 + λ ( α + A 2 ) 4 .

Indeed, it is easy to see that f= A (I P Q )A is 1 A 2 -ism, that is,

f ( x ) f ( y ) , x y 1 A 2 f ( x ) f ( y ) 2 .

Observe that

( α + A 2 ) f α ( x ) f α ( y ) , x y = ( α + A 2 ) [ α x y 2 + f ( x ) f ( y ) , x y ] = α 2 x y 2 + α f ( x ) f ( y ) , x y + α A 2 x y 2 + A 2 f ( x ) f ( y ) , x y α 2 x y 2 + 2 α f ( x ) f ( y ) , x y + f ( x ) f ( y ) 2 = α ( x y ) + f ( x ) f ( y ) 2 = f ( x ) f ( y ) 2 .

Hence, it follows that f α =αI+ A (I P Q )A is 1 α + A 2 -ism. Thus, λ f α is 1 λ ( α + A 2 ) -ism. By Proposition 2.4(iii) the composite (Iλ f α ) is λ ( α + A 2 ) 2 -averaged. Therefore, noting that P C is 1 2 -averaged and utilizing Proposition 2.5(iv), we know that for each λ(0, 2 α + A 2 ), P C (Iλ f α ) is ζ-averaged with

ζ= 1 2 + λ ( α + A 2 ) 2 1 2 λ ( α + A 2 ) 2 = 2 + λ ( α + A 2 ) 4 (0,1).

This shows that P C (Iλ f α ) is nonexpansive. Furthermore, for { λ n }[a,b] with a,b(0, 1 A 2 ), utilizing the fact that lim n 1 α n + A 2 = 1 A 2 , we may assume that

0<a λ n b< 1 A 2 ,n0.

Consequently, it follows that for each integer n0, P C (I λ n f α n ) is ζ n -averaged with

ζ n = 1 2 + λ n ( α n + A 2 ) 2 1 2 λ n ( α n + A 2 ) 2 = 2 + λ n ( α n + A 2 ) 4 (0,1).

This immediately implies that P C (I λ n f α n ) is nonexpansive for all n0.

We divide the remainder of the proof into several steps.

Step 1. We prove that { x n } is bounded. Indeed, we take a fixed pFix(T)Γ arbitrarily. Then we get P C (I λ n f)p=p for λ n (0, 2 A 2 ). Since P C and (I λ n f α n ) are nonexpansive mappings, then we have

y n p = P C ( I λ n f α n ) x n P C ( I λ n f ) p P C ( I λ n f α n ) x n P C ( I λ n f α n ) p + P C ( I λ n f α n ) p P C ( I λ n f ) p x n p + ( I λ n f α n ) p ( I λ n f ) p = x n p + λ n f p λ n f α n p = x n p + λ n f p f α n p = x n p + λ n f p f p α n p x n p + α n λ n p .
(3.2)

Observe that

x n + 1 p = β n x n + ( 1 β n ) T n y n p β n x n p + ( 1 β n ) T n y n p β n x n p + ( 1 β n ) k n y n p β n x n p + ( 1 β n ) k n ( x n p + λ n α n p ) = β n x n p + ( 1 β n ) k n x n p + ( 1 β n ) k n α n λ n p = ( 1 + ( k n 1 ) ( 1 β n ) ) x n p + ( 1 β n ) k n α n λ n p .
(3.3)

Since n = 1 ( k n 1)<, according to Lemma 2.7 and (i), (ii) and (3.3), we obtain that

lim n x n p exists for each pFix(T)Γ.
(3.4)

This implies that { x n } is bounded and { y n } is also bounded.

It follows that

T n x n p k n x n p.

Hence { T n x n p} is bounded.

Step 2. We prove that

lim n y n T y n =0.

In fact, it follows from (3.2) that

y n p 2 = ( x n p + α n λ n p ) 2 x n p 2 + 2 α n λ n p x n p + α n 2 λ n 2 p 2 = x n p 2 + α n ( 2 λ n p x n p + α n λ n 2 p 2 ) = x n p 2 + α n M ,

where M= sup n 0 {2 λ n p x n p+ α n λ n 2 p 2 }<.

It follows that

T n y n p 2 ( k n y n p ) 2 = k n 2 y n p 2 = k n 2 x n p 2 + α n k n 2 M .

Also, observe that

x n + 1 p 2 = β n x n + ( 1 β n ) T n y n p 2 β n x n p 2 + ( 1 β n ) T n y n p 2 β n ( 1 β n ) T n y n x n 2 β n x n p 2 + ( 1 β n ) ( k n 2 x n p 2 + α n k n 2 M ) β n ( 1 β n ) T n y n x n 2 = β n x n p 2 + ( 1 β n ) k n 2 x n p 2 + ( 1 β n ) k n 2 α n M β n ( 1 β n ) T n y n x n 2 = ( k n 2 β n ( k n 2 1 ) ) x n p 2 + ( 1 β n ) k n 2 α n M β n ( 1 β n ) T n y n x n 2 .

Hence, we have

β n ( 1 β n ) T n y n x n 2 ( k n 2 β n ( k n 2 1 ) ) x n p 2 x n + 1 p 2 + ( 1 β n ) k n 2 α n M .
(3.5)

By the conditions (i), (iii) and lim n k n =1, we can conclude that

lim n T n y n x n =0.
(3.6)

Consider that since y n = P C ( x n λ n f α n x n ) and by Proposition 2.8(ii), we have

y n p 2 x n λ n f α n ( x n ) p 2 x n λ n f α n ( x n ) y n 2 = x n p 2 x n y n 2 + 2 λ n f α n ( x n ) , p y n = x n p 2 x n y n 2 + 2 λ n ( f α n ( x n ) f α n ( p ) , p x n + f α n ( p ) , p x n + f α n ( x n ) , x n y n ) x n p 2 x n y n 2 + 2 λ n ( f α n ( p ) , p x n + f α n ( x n ) , x n y n ) = x n p 2 x n y n 2 + 2 λ n [ ( α n I + f ) p , p x n + f α n ( x n ) , x n y n ] = x n p 2 x n y n 2 + 2 λ n [ α n p , p x n + f α n ( x n ) , x n y n ] = x n p 2 x n y n 2 + 2 λ n α n p , p x n + 2 λ n f α n ( x n ) , x n y n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n x n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n p + p x n = x n p 2 x n y n 2 + 2 λ n α n p , p x n 2 λ n f α n ( x n ) , y n p 2 λ n f α n ( x n ) , p x n x n p 2 x n y n 2 + 2 λ n α n p p x n 2 λ n f α n ( x n ) y n p 2 λ n f α n ( x n ) p x n .
(3.7)

Consequently, utilizing Lemma 2.9(ii) and (3.7), we conclude that

x n + 1 p 2 = β n x n + ( 1 β n ) T n y n p 2 = β n x n + ( 1 β n ) T n y n ( β n + ( 1 β n ) ) p 2 = β n x n + ( 1 β n ) T n y n β n p ( 1 β n ) p 2 = β n ( x n p ) + ( 1 β n ) ( T n y n p ) 2 = β n x n p 2 + ( 1 β n ) T n y n p 2 β n ( 1 β n ) x n T n y n 2 β n x n p 2 + ( 1 β n ) k n 2 y n p 2 β n ( 1 β n ) x n T n y n 2 = β n x n p 2 + ( 1 β n ) k n 2 [ x n p 2 x n y n 2 + 2 λ n α n p p x n 2 λ n f α n ( x n ) y n p 2 λ n f α n ( x n ) p x n ] β n ( 1 β n ) x n T n y n 2 = ( β n + ( 1 β n ) k n 2 ) x n p 2 ( 1 β n ) k n 2 x n y n 2 + 2 ( 1 β n ) k n 2 λ n α n p p x n 2 ( 1 β n ) k n 2 λ n f α n ( x n ) y n p 2 ( 1 β n ) k n 2 λ n f α n ( x n ) p x n β n ( 1 β n ) x n T n y n 2 = ( k n 2 β n ( k n 2 1 ) ) x n p 2 ( 1 β n ) k n 2 x n y n 2 + 2 ( 1 β n ) k n 2 λ n α n p p x n 2 ( 1 β n ) k n 2 λ n f α n ( x n ) y n p 2 ( 1 β n ) k n 2 λ n f α n ( x n ) p x n β n ( 1 β n ) x n T n y n 2 .

It follows that we get

( 1 β n ) k n 2 x n y n 2 2 ( 1 β n ) k n 2 λ n α n p p x n + 2 ( 1 β n ) k n 2 λ n f α n ( x n ) ( y n p + p x n ) + β n ( 1 β n ) x n T n y n 2 ( k n 2 β n ( k n 2 1 ) ) x n p 2 x n + 1 p 2 .
(3.8)

So, taking n, since lim n 0 k n =1, (i)-(iii), (3.6) and (3.8), we can conclude that

lim n 0 y n x n =0.
(3.9)

Consider

x n + 1 x n = β n x n x n + ( 1 β n ) T n y n = ( 1 β n ) x n + ( 1 β n ) T n y n ( 1 β n ) T n y n x n .
(3.10)

From (3.6) we obtain

x n + 1 x n (1 β n ) T n y n x n 0(as n).
(3.11)

Observe that

T n y n y n = T n y n x n + x n y n T n y n x n + x n y n .

So, from (3.6) and (3.9), we get

lim n T n y n y n =0.
(3.12)

We compute that

y n + 1 y n = P C ( x n + 1 λ n + 1 f α n + 1 x n + 1 ) P C ( x n λ n f α n x n ) = P C ( I λ n + 1 f α n + 1 ) x n + 1 P C ( I λ n f α n ) x n P C ( I λ n + 1 f α n + 1 ) x n + 1 P C ( I λ n + 1 f α n + 1 ) x n + P C ( I λ n + 1 f α n + 1 ) x n P C ( I λ n f α n ) x n x n + 1 x n + ( I λ n + 1 f α n + 1 ) x n ( I λ n f α n ) x n = x n + 1 x n + x n λ n + 1 f α n + 1 x n ( x n λ n f α n x n ) = x n + 1 x n + λ n f α n x n λ n + 1 f α n + 1 x n = x n + 1 x n + λ n ( f + α n ) x n λ n + 1 ( f + α n + 1 ) x n = x n + 1 x n + λ n f x n + λ n α n x n ( λ n + 1 f x n + λ n + 1 α n + 1 x n ) = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n α n x n λ n + 1 α n + 1 x n = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n α n x n λ n α n + 1 x n + λ n α n + 1 x n λ n + 1 α n + 1 x n = x n + 1 x n + ( λ n λ n + 1 ) f x n + λ n ( α n α n + 1 ) x n + ( λ n λ n + 1 ) α n + 1 x n x n + 1 x n + | λ n λ n + 1 | f x n + λ n | α n α n + 1 | x n + α n + 1 | λ n λ n + 1 | x n .

From the conditions (i), (ii) and (3.11), we obtain that

y n + 1 y n 0(as n).
(3.13)

Since T is uniformly L-Lipschitzian continuous, then

y n T y n y n y n + 1 + y n + 1 T n + 1 y n + 1 + T n + 1 y n + 1 T n + 1 y n + T n + 1 y n T y n y n y n + 1 + y n + 1 T n + 1 y n + 1 + L y n y n + 1 + L T n y n y n .

Since lim n y n + 1 y n =0 and lim n y n T n y n =0, it follows that

lim n y n T y n =0.
(3.14)

Step 3. We show that x ˆ Fix(T)Γ.

Since f= A (I P Q )A is Lipschitz continuous, from (3.9), we have

lim n f ( x n ) f ( y n ) =0.

Since { x n } is bounded, there is a subsequence { x n i } of { x n } that converges weakly to some  x ˆ .

First, we show that x ˆ Γ. Since x n y n 0, it is known that y n i x ˆ .

Put

S w 1 ={ f w 1 + N C w 1 if  w 1 C , if  w 1 C ,

where N C w 1 ={z H 1 : w 1 u,z0,uC}. Then S is maximal monotone and 0S w 1 if and only if w 1 VI(C,f); (see [17]) for more details. Let ( w 1 ,z)G(S), we have

zS w 1 =f w 1 + N C w 1 ,

and hence

zf w 1 N C w 1 .

So, we have

w 1 u,zf w 1 0,uC.

On the other hand, from

y n = P C (I λ n f α n ) x n and w 1 C,

we have

x n λ n f α n x n y n , y n w 1 0,

and

w 1 y n , y n x n λ n + f α n x n 0.

Therefore, from zf w 1 N C w 1 and y n i C, it follows that

w 1 y n i , z w 1 y n i , f w 1 w 1 y n i , f w 1 w 1 y n i , y n i x n i λ n i + f α n i x n i = w 1 y n i , f w 1 w 1 y n i , y n i x n i λ n i + f x n i α n i w 1 y n i , x n i = w 1 y n i , f w 1 f y n i + w 1 y n i , f y n i f x n i w 1 y n i , y n i x n i λ n i α n i w 1 y n i , x n i w 1 y n i , f y n i f x n i w 1 y n i , y n i x n i λ n i α n i w 1 y n i , x n i .

Hence, we obtain

w 1 x ˆ ,z0as i.

Since S is maximal monotone, we have x ˆ S 1 0, and hence x ˆ VI(C,f). Thus, it is clear that x ˆ Γ.

Next, we show that x ˆ Fix(T). Indeed, since y n i x ˆ and y n i T y n i 0 by (3.14) and Lemma 2.6, we get x ˆ Fix(T). Therefore, we have x ˆ Fix(T)Γ.

Let { x n j } be another subsequence of { x n } such that { x n j } x ¯ . Then x ¯ Fix(T)Γ. Let us show that x ˆ = x ¯ . Assume that x ˆ x ¯ . From the Opial condition [18], we have

lim n x n x ˆ = lim n i inf x n i x ˆ < lim n i inf x n i x ¯ = lim n x n x ¯ = lim n j inf x n j x ¯ < lim n j inf x n j x ˆ = lim n x n x ˆ .

This is a contradiction. Thus, we have x ˆ = x ¯ . This implies

x n x ˆ Fix(T)Γ.

Further, from x n y n 0, it follows that y n x ˆ . This shows that both sequences { x n } and { y n } converge weakly to x ˆ Fix(T)Γ. This completes the proof. □

Utilizing Theorem 3.1, we have the following new results in the setting of real Hilbert spaces.

Take T n I(identitymappings) in Theorem 3.1. Therefore the conclusion follows.

Corollary 3.2LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceH. Suppose thatΓ. Let{ x n }be a sequence inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , x n + 1 = β n x n + ( 1 β n ) P C ( I λ n f α n ) x n , n 0 ,
(3.15)

where f α n =f+ α n I= A (I P Q )A+ α n I, and the sequences{ α n }, { λ n }, and{ β n }satisfy the conditions:

  1. (i)

    n = 1 α n <,

  2. (ii)

    { λ n }[a,b]for somea,b(0, 1 A 2 )and n = 1 | λ n + 1 λ n |<,

  3. (iii)

    { β n }[c,d]for somec,d(0,1).

Then the sequence{ x n }converges weakly to an element x ˆ Γ.

Take P C I(identitymappings) in Theorem 3.1. Therefore the conclusion follows.

Corollary 3.3LetCbe a nonempty, closed, and convex subset of a real Hilbert spaceHand letT:CCbe a uniformlyL-Lipschitzian and quasi-nonexpansive mapping withFix(T)and{ k n }[1,)for allnNsuch that n = 1 ( k n 1)<. Let{ x n }be the sequence inCgenerated by the following algorithm:

{ x 0 = x C chosen arbitrarily , x n + 1 = β n x n + ( 1 β n ) T n x n , n 0 ,
(3.16)

and let the sequence{ β n }satisfy the condition{ β n }[c,d]for somec,d(0,1). Then the sequence{ x n }converges weakly to an element x ˆ Fix(T).

Remark 3.4 Theorem 3.1 improves and extends [[7], Theorem 5.7] in the following aspects:

  1. (a)

    The iterative algorithm [[7], Theorem 5.7] is extended for developing our relaxed extragradient algorithm with regularization in Theorem 3.1.

  2. (b)

    The technique of proving weak convergence in Theorem 3.1 is different from that in [[7], Theorem 5.7] because of our technique to use asymptotically quasi-nonexpansive mappings and the property of maximal monotone mappings.

  3. (c)

    The problem of finding a common element of Fix(T)Γ for asymptotically quasi-nonexpansive mappings which is more general than that for nonexpansive mappings and the problem of finding a solution of the SFP in [[7], Theorem 5.7].