1 Introduction

The variational inequality (VI) problem has been studied by many researchers because of its wide applications in optimization problems, complementary problems, fixed point problems and many more (Facchinei and Pang 2003). To solve a VI, numerous algorithms have been suggested, especially projection-type algorithms like projection, extragradient-type algorithms and its variants.

Let us briefly recall some fundamental methods for solving (pseudo)-monotone variational inequality problems which will be (re)-considered in this paper. One of the most well-known algorithms is the extragradient method (Korpelevich 1976) and its variant algorithms proposed by Censor et al. (2011a, b, 2012). The extragradient was proposed by Korpelevich for solving monotone variational inequalities and saddle point problem in finite dimensional spaces (Korpelevich 1976) and then extended to infinite dimensional Hilbert spaces in Khanh (2016) for solving monotone VIs and in Vuong (2018) for solving pseudo-monotone VIs. One of its important variants is the subgradient extragradient considered by Censor et al. (2012), which reduced the number of projections onto the feasible set. The Forward–Backward–Forward (FBF) method was proposed originally by Tseng (2000) for solving monotone inclusions, a more general model than VIs. The applicability of FBF method for solving pseudo-monotone VIs was studied recently in Boţ et al. (2020). Last, we take the Popov’s method (see Popov (1980)) and its modified version proposed by Malitsky and Semenov (2014) into account due to merits within every single iterations. One of them is that we just need compute one value of operator instead of two as in extragradient-type method.

To gain deeper insight for VIs, many researchers considered the weak sharp condition of solution set of VIs. Weak sharp solutions and its geometry condition were firstly introduced by Burke and Ferris for mathematical programming solving an optimization problem (Burke and Ferris 1993). Later, Marcotte and Zhu (1998) modified this geometry condition and introduced weak sharp solutions for VIs, simultaneously presented the finite convergence of algorithms for solving VIs. They also proved the equivalence between the weak sharpness of solution set with the dual gap function. Liu and Wu (2016a, 2016b) further studied the weak sharp of solution set of VIs with respect to primal gap function. Recently, Al-Homidan et al. (2016) used weak sharp solutions for the VIs without considering the primal or dual gap function to studied the finite termination property of sequences generated by iterative methods, such as the proximal point method, inexact proximal point method and gradient projection method. These results were also extended to non-smooth VIs as well as VIs on Hadamard manifolds and equilibrium problems (Al-Homidan et al. 2017; Kolobov et al. 2022, 2021; Nguyen et al. 2020, 2021).

In this paper, we discuss the finite convergence of projection-type methods for solving VIs problem such as extragradient method, Forward-Backward-Forward method, Popov method and their variants. For each method, we also provide an estimation for the number of iterations to guarantee the sequence converges to a solution of the VI problem. Moreover, we prove that this estimations is tight. The rest of this article is organized as follows. Section 2 introduces some preliminaries. Section 3 presents the finite convergence of the extragradient method to a weakly sharp solution set. Section 4 contains the finite convergence results of the Forward-Backward-Forward and the subgradient-extragradient methods. The finite convergence results of Popov’s method and its variants are established in Sect. 5. Finally, we provide numerical examples in the last Section.

2 Preliminaries

Let H be real Hilbert space with the inner product \(\left<\cdot ,\cdot \right>\) and a generated norm \(\Vert \cdot \Vert \). Let C be a nonempty closed convex subset of H and let F be a mapping from H to H. We consider the variational inequality problem, denoted by VI (CF), which is to find \(x^*\in C\) such that

$$\begin{aligned} \left<F(x^*), x-x^*\right>\ge 0 \quad \forall x\in C. \end{aligned}$$
(1)

We denote the solution set of VI (CF) is \(C^*\). In this article, we assume that the \(C^*\) is nonempty and we will recall some definitions about Lipschitz continuity and monotonicity of F (Karamardian and Schaible 1990) as follows

  • F is Lipschitz continuous on H if there exists \(L>0\) such that \(\Vert F(x)-F(y)\Vert \le L\Vert x-y\Vert \) for all \(x, y\in H\);

  • F is pseudo-monotone on C if there is any \(x,y\in C\) such that \(\langle F(y),x-y\rangle \ge 0\) then \(\langle F(x),x-y\rangle \ge 0\);

  • F is monotone on C if for any \(x,y\in C\) we have \(\langle F(x)-F(y),x-y\rangle \ge 0\). Obviously, if F is monotone on C, then F is pseudo-monotone on C.

The metric projection of element x in real Hilbert space H on closed convex subset C, denoted by \(P_C(x)\), is a unique element of C such that \(\Vert x-P_C(x)\Vert \le \Vert x-y\Vert \) for all \(y\in C\). We also denote dist\((x,C)=\Vert x-P_C(x)\Vert \). The metric projection has three important properties as follows (Goebel and Reich 1984).

Theorem 2.1

For any \(x, z\in H\) and \(y\in C\) we have

  1. (a)

    \(\Vert P_C(x)-P_C(z)\Vert \le \Vert x-z\Vert \) (nonexpansivity of \(P_C(.)\));

  2. (b)

    \(\langle x-P_C(x),y-P_C(x)\rangle \le 0\);

  3. (c)

    \(\Vert P_C(x)-y\Vert ^2\le \Vert x-y\Vert ^2-\Vert x-P_C(x)\Vert ^2\).

We recall the definition of weak sharp solution with respect to geometry condition and equivalent conditions as in Marcotte and Zhu (1998). Firstly, we denote by \({\mathbb {B}}\) the unit ball in H. For a given set X in H, we denote by intX the interior of X and by clX the closure of X. The polar \(X^o\) is defined by

$$\begin{aligned} X^o:=\{y\in H, \langle y,x\rangle \le 0\quad \text {for all } x\in X\}. \end{aligned}$$

Let C be a nonempty, closed, convex subset of H. The tangent cone to C at a point \(x\in C\) is defined by

$$\begin{aligned} T_C(x):=\text {cl}\left( \bigcup _{\gamma >0}\frac{X-x}{\gamma }\right) . \end{aligned}$$

The normal cone to \(x\in C\) is defined by

$$\begin{aligned} N_C(x):=\{u\in H: \langle u,y-x \rangle \le 0 \text { for all } y\in C\}. \end{aligned}$$

The solution set \(C^*\) of VI(CF) is weakly sharp if we have, for any \(x^*\in C^*\),

$$\begin{aligned} -F(x^*)\in \text {int} \left( \displaystyle \bigcap _{x\in C^*}[T_C(x) \cap N_{C^*}(x)]^o\right) . \end{aligned}$$
(2)

From (2) we have that if \(C^*\) is weakly sharp, then there exists a constant \(\alpha >0\) such that

$$\begin{aligned} \alpha {\mathbb {B}}\subset F(x^*)+[T_C(x^*)\cap N_{C^*}(x^*)]^o,\quad \text {for all } x^*\in C^*. \end{aligned}$$
(3)

It is equivalent to say that (see (Marcotte and Zhu 1998, Theorem 4.1)) for each \(x^*\in C^*\),

$$\begin{aligned} \langle F(x^*),v\rangle \ge \alpha \Vert v\Vert , \quad \text {for all } v\in T_C(x)\cap N_{C^*}(x). \end{aligned}$$
(4)

We will need the following important theorem (see (Al-Homidan et al. 2016, Theorem 2)) in the proof of finite convergence.

Theorem 2.2

Let C be a nonempty, closed, convex subset of Hilbert space H and \(F:C\rightarrow H\) be a mapping. Assume that the solution \(C^*\) of VI(CF) is nonempty, closed and convex.

  1. (a)

    If \(C^*\) is weakly sharp and F is monotone, then there exists a positive constant \(\alpha >0\) such that

    $$\begin{aligned} \langle F(x), x-P_{C^*}(x)\rangle \ge \alpha \text {dist(}x,C^*\text {), for all } x\in C. \end{aligned}$$
    (5)
  2. (b)

    If F is constant on \(C^*\) and continuous on C and (5) holds for some \(\alpha >0\), then \(C^*\) is weakly sharp.

In the rest of the paper, we will prove the finite convergence of sequences generated by three fundamental algorithms: The (Subgradient) Extragradient, the Forward-Backward-Forward and the Popov Algorithms under the monotonicity of F and weak sharp condition of the solution set \(C^*\) of VI(CF).

3 Extragradient method

In this part, we consider the extragradient algorithm as follows (Korpelevich 1976)

$$\begin{aligned} \lambda> & {} 0, x_0\in C,\\ y_n= & {} P_C(x_n-\lambda F(x_n)),\\ x_{n+1}= & {} P_C(x_n-\lambda F(y_n)). \end{aligned}$$

Firstly, we recall the important inequality relating the distances from the points generated by the extragradient algorithm to the point \(x^*\) of the solution set \(C^*\). The proof presented here is shorter than the one in Khanh (2016).

Lemma 3.1

Let \(F:C\rightarrow H\) be pseudo-monotone and Lipschitz continuous with constant L and \(x^*\) be a point in solution set \(C^*\). Let \(\{x_n\}\) and \(\{y_n\}\) be sequences generated by extragradient algorithm. Then the following inequality holds

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda L)(\Vert x_{n+1}-y_n\Vert ^2+\Vert x_n-y_n\Vert ^2). \end{aligned}$$
(6)

Proof

Since \(x^*\in C^*\subset C,\, x_{n+1}=P_C(x_n-\lambda F(y_n)\), we have from Theorem 2.1 (c) that

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2&\le \Vert x_n-\lambda F(y_n)-x^*\Vert ^2-\Vert x_n-\lambda F(y_n)-x_{n+1}\Vert ^2\nonumber \\&=\Vert x_n-x^*\Vert ^2-\Vert x_n-x_{n+1}\Vert ^2-2\lambda \langle F(y_n),x_{n+1}-x^*\rangle \nonumber \\&=\Vert x_n-x^*\Vert ^2-\Vert x_n-y_n+y_n-x_{n+1}\Vert ^2-2\lambda \langle F(y_n),y_n-x^*\rangle \nonumber \\ {}&\quad -2\lambda \langle F(y_n),x_{n+1}-y_n\rangle \nonumber \\&= \Vert x_n-x^*\Vert ^2-\Vert x_n-y_n\Vert ^2-\Vert x_{n+1}-y_n\Vert ^2-2\langle x_n-y_n,y_n-x_{n+1}\rangle \nonumber \\ {}&\quad -2\lambda \langle F(y_n),y_n-x^*\rangle \nonumber \\&\quad +2\lambda \langle F(x_n)-F(y_n),x_{n+1}-y_n\rangle -2\lambda \langle F(x_n), x_{n+1}-y_n\rangle . \end{aligned}$$
(7)

We have from \(x^*\in C^*\), \(y_n \in C\) and (1) that \(\langle F(x^*),y_n-x^*\rangle \ge 0\). Due to pseudo-monotonicity of F, we can infer that

$$\begin{aligned} \langle F(y_n),y_n-x^*\rangle \ge 0. \end{aligned}$$
(8)

Using Theorem 2.1 (b), since \(y_n=P_C(x_n-\lambda F(y_n))\), we obtain

$$\begin{aligned} \langle x_n-\lambda F(x_n)-y_n,x_{n+1}-y_n\rangle \le 0. \end{aligned}$$

This is equivalent to

$$\begin{aligned} -2\lambda \langle F(x_n),x_{n+1}-y_n\rangle \le 2\langle x_n-y_n,y_n-x_{n+1}\rangle . \end{aligned}$$
(9)

Since F is Lipschitz continuous on H with constant L, we have

$$\begin{aligned} 2\lambda \langle F(x_n)-F(y_n),x_{n+1}-y_n\rangle\le & {} 2\lambda \Vert F(x_n)-F(y_n)\Vert \Vert x_{n+1}-y_n\Vert \nonumber \\ {}\le & {} 2\lambda L\Vert x_n-y_n\Vert \Vert x_{n+1}-y_n\Vert . \end{aligned}$$
(10)

Combining (8), (9), (10) with (7), we get

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-\Vert x_n-y_n\Vert ^2-\Vert x_{n+1}-y_n\Vert ^2+2\lambda L \Vert x_n-y_n\Vert \Vert x_{n+1}-y_n\Vert . \end{aligned}$$
(11)

By using the Cauchy-Schwarz inequality \(2 \Vert x_n-y_n\Vert \Vert x_{n+1}-y_n\Vert \le \Vert x_n-y_n\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2\) in right hand side of above inequality, we obtain

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda L)(\Vert x_{n+1}-y_n\Vert ^2+\Vert x_n-y_n\Vert ^2). \end{aligned}$$

\(\square \)

Under the weak sharp condition of solution set \(C^*\), we will show the finite convergence of the sequence \(\{x_n\}\) generated by extragradient algorithm.

Theorem 3.1

Let \(F:C \rightarrow H\) be monotone, Lipschitz continuous with constant L and assume that the solution set \(C^*\) be weakly sharp with modulus \(\alpha >0\). Let \(\{x_n\} \) be the sequence generated by extragradient algorithm with \(0<\lambda < 1/L\). Then, \(\{x_n\}\) converges strongly to a point in \(C^*\) in atmost \((k+1)\) iterations with

$$\begin{aligned} k\le \frac{6\text {dist} \, (x_0,C^*)^2}{\alpha ^2\lambda ^2(1-\lambda L)}. \end{aligned}$$
(12)

Moreover, the estimation in (12) is tight.

Proof

Since

$$\begin{aligned} x_{n+1}=P_C(x_n-\lambda F(y_n))=P_C(x_n-\lambda F(x_{n+1}) +\lambda (F(x_{n+1})-F(y_n))), \end{aligned}$$

for all \(u\in C\), we have

$$\begin{aligned} \langle x_n-\lambda F(x_{n+1})+ \lambda ( F(x_{n+1}) -F(y_n))-x_{n+1},u-x_{n+1}\rangle \le 0. \end{aligned}$$

Then, we get

$$\begin{aligned} \langle F(x_{n+1}),x_{n+1}-u\rangle&\le \frac{1}{\lambda }\langle x_n-x_{n+1},x_{n+1}-u\rangle +\langle F(x_{n+1}) -F(y_n),x_{n+1}-u\rangle \nonumber \\&\le \frac{1}{\lambda }\Vert x_n-x_{n+1}\Vert \Vert x_{n+1} -u\Vert +\Vert F(x_{n+1})-F(y_n)\Vert \Vert x_{n+1}-u\Vert . \end{aligned}$$
(13)

On the other hand, from Theorem 3.1, let \(x^*\) be a point of solution set \(C^*\), we have

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le & {} \Vert x_n-x^*\Vert ^2-(1-\lambda L)\Vert x_n-y_n\Vert ^2 -(1-\lambda L)\Vert x_{n+1}-y_n\Vert ^2\\ {}\le & {} \Vert x_n-x^*\Vert ^2 \quad \text {for all}\ n. \end{aligned}$$

This implies that \(\{\Vert x_n-x^*\Vert \}\) is non-increasing sequence , therefore \(\lim _{n\rightarrow \infty }\Vert x_n-x^*\Vert \) exists. Moreover, we also have

$$\begin{aligned} \Vert x_n-x^*\Vert ^2-\Vert x_{n+1}-x^*\Vert ^2\ge (1-\lambda L)(\Vert x_n-y_n\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2) , \quad \text {for all}\ n. \end{aligned}$$
(14)

Noticing that \(1-\lambda L>0\), we infer

$$\begin{aligned}&\Vert x_n-x^*\Vert ^2-\Vert x_{n+1}-x^*\Vert ^2\nonumber \\ {}&\quad \ge \frac{1-\lambda L}{3}\Vert x_{n+1}-y_n\Vert ^2+\frac{2(1-\lambda L)}{3}(\Vert x_n-y_n\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2)\nonumber \\&\quad \ge \frac{1-\lambda L}{3}\Vert x_{n+1}-y_n\Vert ^2+ \frac{1-\lambda L}{3} (\Vert x_n-y_n\Vert +\Vert x_{n+1}-y_n\Vert )^2\nonumber \\&\quad \ge \frac{1-\lambda L}{3}(\Vert x_{n+1}-y_n\Vert ^2+\Vert x_{n+1}-x_n\Vert ^2). \end{aligned}$$
(15)

Letting \(n\rightarrow \infty \) and taking the limits in the both sides of (15), we deduce that \(\lim _{n\rightarrow \infty }\Vert x_{n+1}-y_n\Vert =0\) and \(\lim _{n\rightarrow \infty }\Vert x_{n+1}-x_n\Vert =0\).

For \(0<N\in {\mathbb {N}}\), we have from (15) that

$$\begin{aligned} \frac{1-\lambda L}{3} \sum _{i=0}^N(\Vert x_{i+1}-y_i\Vert ^2+\Vert x_{i+1}-x_i\Vert ^2)&\le \sum _{i=0}^N(\Vert x_{i}-x^*\Vert ^2-\Vert x_{i+1}-x^*\Vert ^2)\\&=\Vert x_0-x^*\Vert ^2-\Vert x_{N+1}-x^*\Vert ^2\\&\le \Vert x_0-x^*\Vert ^2. \end{aligned}$$

Since above inequality holds with any \(x^*\in C^*\), we obtain

$$\begin{aligned} \frac{1-\lambda L}{3} \sum _{i=0}^N(\Vert x_{i+1}-y_i\Vert ^2+\Vert x_{i+1}-x_i\Vert ^2)\le \text {dist}(x_0,C^*)^2. \end{aligned}$$
(16)

Let k be the smallest integer such that

$$\begin{aligned} \alpha \lambda >\Vert x_{k+1}-x_k\Vert +\Vert x_{k+1}-y_k\Vert . \end{aligned}$$
(17)

Since \(\frac{1}{\lambda }>L>0\), we can infer

$$\begin{aligned} \alpha&> \dfrac{1}{\lambda }(\Vert x_{k+1}-x_k\Vert +\Vert x_{k+1}-y_k\Vert )\nonumber \\&> \dfrac{1}{\lambda }\Vert x_{k+1}-x_k\Vert +L\Vert x_{k+1}-y_k\Vert . \end{aligned}$$
(18)

We assume that \(x_{k+1} \notin C^*\) and set \(t_{k+1}=P_{C^*}(x_{k+1})\in C\). Then, by the weak sharpness of the solution set \(C^*\) and Theorem 2.2 (a), the Lipschitz continuity and monotonicity of F and inequality (13), we have

$$\begin{aligned} \alpha \text {dist}(x_{k+1},C^*)&=\alpha \Vert x_{k+1}-t_{k+1}\Vert \\&\le \langle F(x_{k+1}),x_{k+1}-t_{k+1}\rangle \\&\le \frac{1}{\lambda }\Vert x_k-x_{k+1}\Vert \Vert x_{k+1}-t_{k+1}\Vert +\Vert F(x_{k+1})-F(y_k)\Vert \Vert x_{k+1}-t_{k+1}\Vert \\&\le \frac{1}{\lambda }\Vert x_k-x_{k+1}\Vert \Vert x_{k+1}-t_{k+1}\Vert +L\Vert x_{k+1}-y_k\Vert \Vert x_{k+1}-t_{k+1}\Vert \\&= \Vert x_{k+1}-t_{k+1}\Vert \left( \frac{1}{\lambda }\Vert x_{k+1} -x_k\Vert +L\Vert x_{k+1}-y_k\Vert \right) . \end{aligned}$$

This implies that \(\frac{1}{\lambda }\Vert x_{k+1}-x_k\Vert +L\Vert x_{k+1}-y_k\Vert \ge \alpha \), which contradicts (18). Hence, \(x_{k+1}\in C^*\). It follows from (16) that

$$\begin{aligned} \text {dist}(x_0,C^*)^2&\ge \frac{1-\lambda L}{3}\sum _{i=0}^{k-1}(\Vert x_{i+1}-x_i\Vert ^2+\Vert x_{i+1}-y_i\Vert ^2)\\ {}&\ge \frac{1-\lambda L}{6}\sum _{i=0}^{k-1}(\Vert x_{i+1}-x_i\Vert +\Vert x_{i+1}-y_i\Vert )^2\\&\ge \frac{1-\lambda L}{6}k\lambda ^2\alpha ^2, \end{aligned}$$

where the last inequality is deduced by (17). So, we obtain

$$\begin{aligned} k\le \frac{\text {6dist}(x_0,C^*)^2}{\alpha ^2\lambda ^2(1-\lambda L)}. \end{aligned}$$
(19)

To show that the above estimation is tight, let us consider a simple counter example. Let \(H={\mathbb {R}}\), \(C=[0, +\infty )\) and \(F(x) = 1\) for all \(x\in C\). Then it is clear that F is monotone and Lipschitz continuous on C with any modulus \(L>0\). The problem VI(FC) has a unique solution \(x^* = 0\), i.e. \(C^* = \{0\}\). Then (5) holds with \(\alpha =1\) and it follows from Theorem 2.2 that \(C^*\) is weakly sharp, hence (19) holds whenever \(\lambda <1/L\). Since F is Lipschitz continuous with all \(L > 0\), we deduce that (19) holds for all \(\lambda >0\). Let \(\lambda L =\frac{1}{2}\), from (19) we have that \(k\le \frac{\text {12dist}(x_0,C^*)^2}{\lambda ^2}\). Taking \(\lambda \) large enough, we conclude from (19) that \(k=0\), i.e. the algorithm converges to the solution in one step. Indeed, taking \(x_0 = a \in C\) and \(\lambda = a\), we obtain

$$\begin{aligned} y_0&= P_C (x_0 - \lambda F(x_0)) = P_C (a - a) = 0\\ x_1&= P_C (x_0 - \lambda F(y_0)) = P_C (a - a) = 0 = x^*. \end{aligned}$$

\(\square \)

4 Forward–Backward–Forward method

We consider the Forward–Backward–Forward algorithm proposed by Tseng (2000) as follows

$$\begin{aligned} \lambda> & {} 0,\quad x_0\in C,\\ y_n= & {} P_C(x_n-\lambda F(x_n)),\\ x_{n+1}= & {} y_n-\lambda (F(y_n)-F(x_n)). \end{aligned}$$

Like previous part, we recall and prove the main inequality relating the distances from the points generated by Forward-Backward-Forward algorithm to the point \(x^*\) in the solution set \(C^*\). The proof for monotone VIs was proposed in Tseng (2000).

Lemma 4.1

Let \(F:H\rightarrow H\) be pseudo-monotone and Lipschitz continuous with constant L. Let C be a nonempty closed convex subset of H and \(x^*\) be a point in solution set \(C^*\) of VI(CF). Let \(\{x_n\}\) and \(\{y_n\}\) be sequences generated by Forward–Backward–Forward algorithm. Then the following inequality holds

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda ^2 L^2)\Vert x_n-y_n\Vert ^2. \end{aligned}$$
(20)

Proof

From the equation

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2+\Vert x_{n+1}-x_n\Vert ^2-\Vert x_n-x^*\Vert ^2=2\langle x_{n+1}-x^*,x_{n+1}-x_n\rangle , \end{aligned}$$

and noticing that \(x_n-x_{n+1}=x_n-\lambda F(x_n)-y_n+\lambda F(y_n)\), we have

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2&=\Vert x_n-x^*\Vert ^2-\Vert x_{n+1}-x_n\Vert ^2+2\langle x_n-x_{n+1},x^*-x_{n+1}\rangle \nonumber \\&=|x_n-x^*\Vert ^2-\Vert x_{n+1}-x_n\Vert ^2+2\langle x_n-x_{n+1}, x^*-y_n\rangle \nonumber \\ {}&\quad +2\langle x_n-x_{n+1},y_n-x_{n+1}\rangle \nonumber \\&=|x_n-x^*\Vert ^2-\Vert x_{n+1}-x_n\Vert ^2+2\langle x_n-\lambda F(x_n)-y_n,x^*-y_n\rangle \nonumber \\ {}&\quad +2\lambda \langle F(y_n),x^*-y_n\rangle +2\langle x_n-x_{n+1},y_n-x_{n+1}\rangle . \end{aligned}$$
(21)

Since \(y_n=P_C(x_n-\lambda F(x_n))\) and \(x^*\in C^*\subset C\), we have from Theorem 2.1 (b) that

$$\begin{aligned} \langle x_n-\lambda F(x_n)-y_n,x^*-y_n\rangle \le 0. \end{aligned}$$
(22)

On the other hand, since \(x^*\) is a point of solution set \(C^*\), \(y_n\in C\) and (1)

$$\begin{aligned} \langle F(x^*),y_n-x^*\rangle \ge 0. \end{aligned}$$

In addition, F is pseudo-monotone, hence

$$\begin{aligned} \langle F(y_n),y_n-x^*\rangle \ge 0, \end{aligned}$$

or equivalently

$$\begin{aligned} \langle F(y_n),x^*-y_n\rangle \le 0. \end{aligned}$$
(23)

Next, we estimation the term \(2\langle x_n-x_{n+1},y_n-x_{n+1}\rangle \) by the Lipschitz continuity of F as follows

$$\begin{aligned} 2\langle x_n-x_{n+1},y_n-x_{n+1}\rangle&=2\Vert x_n-x_{n+1}\Vert ^2+2\langle x_n-x_{n+1},y_n-x_n\rangle \nonumber \\&=\Vert x_n-x_{n+1}\Vert ^2+\Vert x_n-y_n+\lambda (F(y_n)-F(x_n))\Vert ^2\nonumber \\&\quad +2\langle x_n-y_n+\lambda (F(y_n)-F(x_n)),y_n-x_n\rangle \nonumber \\&=\Vert x_n-x_{n+1}\Vert ^2+\Vert x_n-y_n\Vert ^2+\lambda ^2\Vert F(y_n)-F(x_n)\Vert ^2 \nonumber \\ {}&\quad +2\lambda \langle F(y_n)-F(x_n),x_n-y_n\rangle \nonumber \\&\quad -2\Vert x_n-y_n\Vert ^2+2\lambda \langle F(y_n)-F(x_n),y_n-x_n\rangle \nonumber \\&=\Vert x_{n+1}-x_n\Vert ^2-\Vert x_n-y_n\Vert ^2+\lambda ^2\Vert F(x_n)-F(y_n)\Vert ^2\nonumber \\&\le \Vert x_{n+1}-x_n\Vert ^2-\Vert x_n-y_n\Vert ^2+\lambda ^2L^2\Vert x_n-y_n\Vert ^2. \end{aligned}$$
(24)

Combining (22), (23), (24) with (21), we get

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda ^2 L^2)\Vert x_n-y_n\Vert ^2 . \end{aligned}$$

\(\square \)

Under the weak sharp condition of solution set \(C^*\), we also show the finite convergence of the feasible sequence \(\{y_n\}\) to solution set \(C^*\) of this algorithm.

Theorem 4.1

Let \(F:H\rightarrow H\) be monotone, Lipschitz continuous with constant L and assume that the solution set \(C^*\) of VI(CF) be weakly sharp with modulus \(\alpha >0\). Let \(\{x_n\},\{y_n\} \) be the sequences generated by Forward-Backward-Forward algorithm with \(0<\lambda < 1/L\). Then, \(\{y_n\}\) converges strongly to a point in \(C^*\) in at most k iterations with

$$\begin{aligned} k\le \frac{(\lambda L+1)\text {dist}(x_0,C^*)^2}{(1-\lambda L)\alpha ^2\lambda ^2}. \end{aligned}$$
(25)

Moreover, the estimation (25) is tight.

Proof

Since

$$\begin{aligned} y_{n}=P_C(x_{n}-\lambda F(x_{n})), \end{aligned}$$

for all \(w\in C\), we get

$$\begin{aligned} \langle x_n-\lambda F(x_{n})-y_n, w-y_n\rangle \le 0. \end{aligned}$$

Therefore,

$$\begin{aligned} \langle F(y_{n}),y_{n}-w\rangle&\le \frac{1}{\lambda }\langle x_n-y_{n},y_{n}-w\rangle +\langle F(y_n)-F(x_n),y_n-w\rangle \nonumber \\&\le \frac{1}{\lambda }\Vert x_n-y_n\Vert \Vert y_n-w\Vert +\Vert F(y_n)-F(x_n)\Vert \Vert y_n-w\Vert . \end{aligned}$$
(26)

On the other hand, since F is pseudo-monotone, let \(x^*\) be a point of solution set \(C^*\), from Theorem 4.1 we have

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda ^2L^2)\Vert x_n-y_n\Vert ^2\quad \text {for all}\ n. \end{aligned}$$
(27)

Since \(0<\lambda <\frac{1}{L}\), (27) implies that \(\{\Vert x_n-x^*\Vert ^2\}\) is non-increasing sequence, therefore \(\lim _{n\rightarrow \infty }\Vert x_n-x^*\Vert \) exists. Moreover, we also have

$$\begin{aligned} \Vert x_n-x^*\Vert ^2-\Vert x_{n+1}-x^*\Vert ^2\ge (1-\lambda ^2 L^2)\Vert x_n-y_n\Vert ^2 , \quad \text {for all}\ n. \end{aligned}$$
(28)

Letting \(n\rightarrow \infty \) and taking the limits in the both sides of (28), we deduce that \(\lim _{n\rightarrow \infty }\Vert x_{n}-y_n\Vert =0\).

For \(0<N\in {\mathbb {N}}\), we have from (28) that

$$\begin{aligned} (1-\lambda ^2 L^2) \sum _{i=0}^N\Vert x_{i}-y_i\Vert ^2&\le \sum _{i=0}^N(\Vert x_i-x^*\Vert ^2-\Vert x_{i+1}-x^*\Vert ^2)\\&=\Vert x_0-x^*\Vert ^2-\Vert x_{N+1}-x^*\Vert ^2\le \Vert x_0-x^*\Vert ^2. \end{aligned}$$

Since above inequality holds with any \(x^*\in C^*\), we get

$$\begin{aligned} (1-\lambda ^2 L^2) \sum _{i=0}^N\Vert x_{i}-y_i\Vert ^2\le \text {dist}(x_0,C^*)^2. \end{aligned}$$
(29)

Since \(\lim _{n\rightarrow \infty }\Vert x_{n}-y_n\Vert =0\), we choose k be the smallest integer such that

$$\begin{aligned} \frac{\alpha \lambda }{\lambda L +1}>\Vert x_{k}-y_k\Vert . \end{aligned}$$
(30)

We assume that \(y_{k} \notin C^*\) and set \(t_{k}=P_{C^*}(y_{k})\in C\). Then, by the weak sharpness of the solution set \(C^*\), the Lipschitz continuity and monotone property of F and inequality (26), we have

$$\begin{aligned} \alpha \text {dist}(y_k,C^*)&=\alpha \Vert y_k-t_k\Vert \\&\le \langle F(y_k),y_k-t_k\rangle \\&\le \frac{1}{\lambda }\Vert y_k-t_k\Vert \Vert x_k-y_k\Vert +\Vert F(x_{k})-F(y_k)\Vert \Vert y_k-t_{k}\Vert \\&\le \frac{1}{\lambda }\Vert y_k-t_k\Vert \Vert x_k-y_k\Vert +L\Vert x_k-y_k\Vert \Vert y_k-t_k\Vert \\&= \Vert y_k-t_k\Vert \left( \frac{1}{\lambda }+L\right) \Vert x_{k}-y_k\Vert . \end{aligned}$$

This implies that \(\Vert x_{k}-y_k\Vert \ge \frac{\alpha \lambda }{\lambda L+1}\), which contradicts (30). Hence, \(y_{k}\in C^*\). It follows from (29) that

$$\begin{aligned} \text {dist}(x_0,C^*)^2\ge (1-\lambda ^2L^2) \sum _{i=0}^{k-1}\Vert x_{i}-y_i\Vert ^2 \ge (1-\lambda ^2L^2) \frac{\alpha ^2\lambda ^2}{(\lambda L+1)^2} k =\frac{(1-\lambda L)\alpha ^2\lambda ^2}{1+\lambda L}k, \end{aligned}$$

where the last inequality is deduced by (30). Therefore, we get

$$\begin{aligned} k\le \frac{(\lambda L+1)\text {dist}(x_0,C^*)^2}{(1-\lambda L)\alpha ^2\lambda ^2}. \end{aligned}$$

\(\square \)

To show that the above estimation is tight, let us consider again the simple counter example as in Theorem 3.1. Let \(\lambda L =\frac{1}{2} < 1\) we deduce

$$\begin{aligned} k \le \frac{ 3 \text {dist}(x_0,C^*)^2}{ \lambda ^2 }. \end{aligned}$$

Hence, choosing \(\lambda \) large enough, we get \(k=0\), i.e. \(y_0 \in C^*\). Indeed, taking \(x_0 = a \in C\) and \(\lambda = a\), we obtain

$$\begin{aligned} y_0 = P_C (x_0 - \lambda F(x_0)) = P_C (a - a) = 0 =x^*. \end{aligned}$$

\(\square \)

Remark 4.1

In subgradient extragradient method (Censor et al. 2012), \(\{x_n\}, \{y_n\}\) are generated by the following algorithm

$$\begin{aligned} x_0\in & {} C, \lambda >0\\ y_n= & {} P_C(x_n-\lambda F(x_n))\\ T_n= & {} \{x\in H, \langle x_n-\lambda F(x_n)-y_n,x-y_n\rangle \le 0\}\\ x_{n+1}= & {} P_{T_n}(x_n-\lambda F(y_n)). \end{aligned}$$

The advantage of the subgradient extragradient method is that the projection onto the half-space \(T_n\) has an explicit formula. We have from the proof of Lemma 3.2 in Censor et al. (2012) that

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2\le \Vert x_n-x^*\Vert ^2-(1-\lambda ^2 L^2) \Vert x_n-y_n\Vert ^2,\quad \text {for all}\ n. \end{aligned}$$

Hence, as in Theorem 4.1, with \(0<\lambda <\frac{1}{L}\), if \(F: H\rightarrow H \) is monotone and Lipschitz continuous and \(C^*\) is weakly sharp, then \(\{y_n\}\) converges to a point in \(C^*\) in at most k iterations with

$$\begin{aligned} k\le \frac{(1+\lambda L)\text {dist} \, (x_0,C^*)^2}{\alpha ^2\lambda ^2(1-\lambda L)}. \end{aligned}$$

Moreover, the above estimation is tight. We omit the detailed proof.

5 Popov’s method

We continue employing above method for showing the finite convergence of the following Popov’s algorithm (Popov 1980):

$$\begin{aligned}{} & {} \lambda >0,\quad x_0,y_0\in C,\\{} & {} x_{n+1}=P_C(x_n-\lambda F(y_n)),\\{} & {} y_{n+1}=P_C(x_{n+1}-\lambda F(y_n)), \end{aligned}$$

under the weak sharp condition of solution set \(C^*\). Like above parts, our proof is based on the following inequality, which slightly improves the main estimation in Popov (1980). This estimation allows us to choose larger stepsize, i.e., \(\lambda \in \left( 0, \frac{1}{(1+\sqrt{2})L}\right) \) instead of \(\lambda \in \left( 0, \frac{1}{3L}\right) \) as in Popov (1980).

Lemma 5.1

Let \(F:C\rightarrow H\) be pseudo-monotone and Lipschitz continuous with constant L and C be a nonempty closed convex subset of H. Let \(x^*\) be a point in solution set \(C^*\) of VI(CF). Let \(\{x_n\}\) and \(\{y_n\}\) be sequences generated by Popov’s algorithm. Then the following inequality holds

$$\begin{aligned}&\Vert x_{n+1}-x^*\Vert ^2+\lambda L\Vert x_{n+1}-y_n\Vert ^2\le \Vert x_n-x^*\Vert ^2+\lambda L\Vert x_n-y_{n-1}\Vert ^2\nonumber \\&\quad -(1-(1+\sqrt{2})\lambda L)(\Vert x_n-y_n\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2) \quad \text {for all}\ n. \end{aligned}$$
(31)

Proof

Since \(x^*\) is a point in solution set \(C^*\), \(y_n\in C\) and F is pseudo-monotone on C

$$\begin{aligned} \langle F(y_n),y_n-x^*\rangle \ge 0. \end{aligned}$$

Therefore,

$$\begin{aligned} -\langle F(y_n),x_{n+1}-x^*\rangle \le \langle F(y_n),y_n-x_{n+1}\rangle . \end{aligned}$$
(32)

Since \(x_{n+1}=P_C(x_n-\lambda F(y_n))\), by Theorem 2.1 (c) we have

$$\begin{aligned} \Vert x_{n+1}-x_n\Vert ^2&\le \Vert x_n-\lambda F(y_n)-x^*\Vert ^2-\Vert x_n-\lambda F(y_n)-x_{n+1}\Vert ^2\\&=\Vert x_n-x^*\Vert ^2-\Vert x_n-x_{n+1}\Vert ^2-2\lambda \langle F(y_n),x_{n+1}-x^*\rangle \\&=\Vert x_n-x^*\Vert ^2-\Vert x_n-y_n\Vert ^2-\Vert y_n-x_{n+1}\Vert ^2-2\langle x_n-y_n,y_n -x_{n+1}\rangle \\ {}&\quad -2\langle F(y_n),x_{n+1}-x^*\rangle . \end{aligned}$$

Combining the above inequality with (32) we get

$$\begin{aligned} \Vert x_{n+1}-x_n\Vert ^2&\le \Vert x_n-x^*\Vert ^2-\Vert x_n-y_n\Vert ^2-\Vert y_n-x_{n+1}\Vert ^2-2\langle x_n-\lambda F(y_n)-y_n,y_n-x_{n+1}\rangle \nonumber \\&=\Vert x_n-x^*\Vert ^2-\Vert x_n-y_n\Vert ^2-\Vert y_n-x_{n+1}\Vert ^2-2\langle x_n-\lambda F(y_{n-1})-y_n,y_n-x_{n+1}\rangle \nonumber \\&\quad +2\lambda \langle F(y_n)-F(y_{n-1}), y_n-x_{n+1}\rangle . \end{aligned}$$
(33)

Using Theorem 2.1 (b), since \(y_n=P_C(x_n-\lambda F(y_{n-1}))\) and \(x_{n+1}\in C\) we have

$$\begin{aligned} \langle x_n-\lambda F(y_{n-1})-y_n,x_{n+1}-y_n\rangle \le 0. \end{aligned}$$
(34)

We estimation the term \(2\lambda \langle F(y_n)-F(y_{n-1}), y_n-x_{n+1}\rangle \) as follows:

$$\begin{aligned}&2\lambda \langle F(y_n)-F(y_{n-1}), y_n-x_{n+1}\rangle \nonumber \\&\quad \le 2\lambda L \Vert y_n-y_{n-1}\Vert \Vert y_n-x_{n+1}\Vert \nonumber \\&\quad \le 2\lambda L (\Vert y_n-x_n\Vert +\Vert x_n-y_{n-1}\Vert )\Vert x_{n+1}-y_n\Vert \nonumber \\&\quad \le 2\lambda L (\Vert x_n-y_n\Vert \Vert x_{n+1}-y_n\Vert +\Vert x_n-y_{n-1}\Vert \Vert x_{n+1}-y_n\Vert )\nonumber \\&\quad \le \lambda L [(1+\sqrt{2})\Vert x_n-y_n\Vert ^2+\frac{1}{1+\sqrt{2}}\Vert x_{n+1}-y_{n}\Vert ^2\nonumber \\&\qquad +\Vert x_n-y_{n-1}\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2]\nonumber \\&\quad =(1+\sqrt{2})\lambda L\Vert x_n-y_n\Vert ^2+\sqrt{2}\lambda L \Vert x_{n+1}-y_n\Vert ^2+\lambda L\Vert x_n-y_{n-1}\Vert ^2. \end{aligned}$$
(35)

Apply estimations (34) and (35) into the right side of inequality (33), we obtain

$$\begin{aligned} \Vert x_{n+1}-x^*\Vert ^2&\le \Vert x_n-x^*\Vert ^2+\lambda L\Vert x_n-y_{n-1}\Vert ^2-(1-(1+\sqrt{2})\lambda L)\Vert x_n-y_n\Vert ^2\\&\quad -(1-\sqrt{2}\lambda L) \Vert x_{n+1}-y_n\Vert ^2, \end{aligned}$$

which is equivalent to

$$\begin{aligned}&\Vert x_{n+1}-x^*\Vert ^2+\lambda L\Vert x_{n+1}-y_n\Vert ^2\le \Vert x_n-x^*\Vert ^2 +\lambda L\Vert x_n-y_{n-1}\Vert ^2\\&\quad -(1-(1+\sqrt{2})\lambda L)(\Vert x_n-y_n\Vert ^2+\Vert x_{n+1}-y_n\Vert ^2). \end{aligned}$$

\(\square \)

The following theorem will show the finite convergence of Popov’s method if the solution set \(C^*\) is weakly sharp.

Theorem 5.1

Let \(F:C\rightarrow H\) be monotone, Lipschitz continuous with constant L and C be a nonempty closed convex subset of H. Assume that the solution set \(C^*\) of VI(CF) is weakly sharp with modulus \(\alpha >0\). Let \(\{x_n\},\{y_n\} \) be the sequences generated by Popov’s algorithm with \(0<\lambda < 1/(1+\sqrt{2})L\). Then, \(\{x_n\}\) converges strongly to a point in \(C^*\) in at most \((k+1)\) iterations with

$$\begin{aligned} k\le \frac{\text {6(dist}(x_1,C^*)^2+\lambda L\Vert x_1-y_0\Vert ^2)}{\alpha ^2\lambda ^2(1-(1+\sqrt{2})\lambda L)}+1. \end{aligned}$$
(36)

Moreover, the estimation in (36) is tight.

Proof

Since

$$\begin{aligned} x_{n+1}=P_C(x_n-\lambda F(y_n))=P_C(x_n -\lambda F(x_{n+1})+\lambda (F(x_{n+1})-F(y_n)), \end{aligned}$$

for all \(u\in C\), we also have the similar result in (13)

$$\begin{aligned} \langle F(x_{n+1}),x_{n+1}-u\rangle \le \frac{1}{\lambda }\Vert x_n -x_{n+1}\Vert \Vert x_{n+1}-u\Vert +\Vert F(x_{n+1})-F(y_n)\Vert \Vert x_{n+1}-u\Vert . \end{aligned}$$
(37)

Since F is pseudo-monotone, let \(x^*\) be a point of solution set \(C^*\), from Theorem 5.1 we get

$$\begin{aligned}&\Vert x_{n+1}-x^*\Vert ^2+\lambda L\Vert x_{n+1}-y_n\Vert ^2 \le \Vert x_n-x^*\Vert ^2+\lambda L\Vert x_n-y_{n-1}\Vert ^2\nonumber \\&\quad -(1-(1+\sqrt{2})\lambda L)(\Vert x_n-y_n\Vert ^2+\Vert x_{n+1} -y_n\Vert ^2)\quad \text {for all}\ n. \end{aligned}$$
(38)

This implies that \(\{a_n\}=\{\Vert x_n-x^*\Vert ^2+\lambda L \Vert x_n-y_{n-1}\Vert ^2\}\) is non-increasing sequence , therefore \(\lim _{n\rightarrow \infty }a_n\) exists. Moreover, we also have

$$\begin{aligned} a_n-a_{n+1}\ge (1-(1+\sqrt{2})\lambda L)(\Vert x_n-y_n\Vert ^2 +\Vert x_{n+1}-y_n\Vert ^2) , \quad \text {for all}\ n. \end{aligned}$$
(39)

Noticing that \(0<\lambda <\frac{1}{(1+\sqrt{2})L}\), we infer

$$\begin{aligned} a_n-a_{n+1}&\ge \frac{1-(1+\sqrt{2})\lambda L}{3}\Vert x_{n+1}-y_n\Vert ^2+\frac{2(1-(1+\sqrt{2})\lambda L)}{3}(\Vert x_n-y_n\Vert ^2\nonumber \\ {}&\quad +\Vert x_{n+1}-y_n\Vert ^2)\nonumber \\&\ge \frac{1-(1+\sqrt{2})\lambda L}{3}\Vert x_{n+1}-y_n\Vert ^2+ \frac{1-(1+\sqrt{2})\lambda L}{3} (\Vert x_n-y_n\Vert \nonumber \\ {}&\quad +\Vert x_{n+1}-y_n\Vert )^2\nonumber \\&\ge \frac{1-(1+\sqrt{2})\lambda L}{3}(\Vert x_{n+1}-y_n\Vert ^2+\Vert x_{n+1}-x_n\Vert ^2). \end{aligned}$$
(40)

Letting \(n\rightarrow \infty \) and taking the limits in the both sides of (40), we deduce that \(\lim _{n\rightarrow \infty }\Vert x_{n+1}-y_n\Vert =0\) and \(\lim _{n\rightarrow \infty }\Vert x_{n+1}-x_n\Vert =0\).

For \(0<N\in {\mathbb {N}}\), we have from (40) that

$$\begin{aligned} \frac{1-(1+\sqrt{2})\lambda L}{3} \sum _{i=1}^N(\Vert x_{i+1}-y_i\Vert ^2+\Vert x_{i+1}-x_i\Vert ^2)&\le \sum _{i=1}^N(a_i-a_{i+1})\\&=a_1-a_{N+1}\\&\le a_1=\Vert x_1-x^*\Vert ^2+\lambda L\Vert x_1-y_0\Vert ^2. \end{aligned}$$

Since above inequality holds with any \(x^*\in C^*\), we obtain

$$\begin{aligned} \frac{1-(1+\sqrt{2})\lambda L}{3} \sum _{i=1}^N(\Vert x_{i+1}-y_i\Vert ^2+\Vert x_{i+1}-x_i\Vert ^2) \le \text {dist}(x_1,C^*)^2+\lambda L \Vert x_1-y_0\Vert ^2. \end{aligned}$$
(41)

Let k be the smallest integer such that

$$\begin{aligned} \alpha \lambda >\Vert x_{k+1}-x_k\Vert +\Vert x_{k+1}-y_k\Vert . \end{aligned}$$
(42)

Since \(\frac{1}{\lambda }>(1+\sqrt{2})L>L\), we can infer

$$\begin{aligned} \alpha&> \dfrac{1}{\lambda }(\Vert x_{k+1}-x_k\Vert +\Vert x_{k+1}-y_k\Vert )\nonumber \\&\ge \dfrac{1}{\lambda }\Vert x_{k+1}-x_k\Vert +L\Vert x_{k+1}-y_k\Vert . \end{aligned}$$
(43)

We assume that \(x_{k+1} \notin C^*\) and set \(t_{k+1}=P_{C^*}(x_{k+1})\in C\). Then, by the weak sharpness of the solution set \(C^*\), the Lipschitz continuity and monotonicity of F and inequality (37), we have

$$\begin{aligned} \alpha \text {dist}(x_{k+1},C^*)&=\alpha \Vert x_{k+1}-t_{k+1}\Vert \\&\le \langle F(x_{k+1}),x_{k+1}-t_{k+1}\rangle \\&\le \frac{1}{\lambda }\Vert x_k-x_{k+1}\Vert \Vert x_{k+1}-t_{k+1}\Vert +\Vert F(x_{k+1})-F(y_k)\Vert \Vert x_{k+1}-t_{k+1}\Vert \\&\le \frac{1}{\lambda }\Vert x_k-x_{k+1}\Vert \Vert x_{k+1}-t_{k+1}\Vert +L\Vert x_{k+1}-y_k\Vert \Vert x_{k+1}-t_{k+1}\Vert \\&\le \Vert x_{k+1}-t_{k+1}\Vert \left( \frac{1}{\lambda }\Vert x_{k+1} -x_k\Vert +L\Vert x_{k+1}-y_k\Vert \right) . \end{aligned}$$

This implies that \(\frac{1}{\lambda }\Vert x_{k+1}-x_k\Vert +L\Vert x_{k+1}-y_k\Vert \ge \alpha \), which contradicts (43). Hence, \(x_{k+1}\in C^*\). It follows from (41) that

$$\begin{aligned} \text {dist}(x_1,C^*)^2+\lambda L\Vert x_1-y_0\Vert ^2&\ge \frac{1-(1+\sqrt{2})\lambda L}{3} \sum _{i=1}^{k-1}(\Vert x_{i+1}-x_i\Vert ^2+\Vert x_{i+1}-y_i\Vert ^2)\\&\ge \frac{1-(1+\sqrt{2})\lambda L}{6} \sum _{i=1}^{k-1}(\Vert x_{i+1}-x_i\Vert +\Vert x_{i+1}-y_i\Vert )^2\\&\ge \frac{1-(1+\sqrt{2})\lambda L}{6}(k-1)\lambda ^2\alpha ^2, \end{aligned}$$

where the last inequality is deduced by (42). Therefore, we obtain

$$\begin{aligned} k\le \frac{\text {6(dist}(x_1,C^*)^2+\lambda L\Vert x_1-y_0\Vert ^2)}{\alpha ^2 \lambda ^2(1-(1+\sqrt{2})\lambda L)}+1. \end{aligned}$$

To show that the above estimation is tight, let us consider again the simple counter example as in Theorem 3.1. Let \(\lambda L = 1/3\) and \(\lambda \) large enough, we can deduce that \(k=1\). Taking \(x_0=a \in C\) , \(\lambda =a/2=y_0\), we obtain

$$\begin{aligned} x_1&= P_C (x_0 - \lambda F(y_0)) = P_C (a - a/2) = a/2\\ y_1&= P_C (x_1 - \lambda F(y_0)) = P_C (a/2 - a/2) = 0\\ x_2&= P_C(x_1-\lambda F(y_1)=P_C (a/2 - a/2) = 0 = x^*, \end{aligned}$$

which means that the algorithm converges to the solution in at-most two steps. \(\square \)

Remark 5.1

Malitsky and Semenov (2014) modified Popov’s algorithm by using the technique of the subgradient extragradient method:

$$\begin{aligned} \lambda> & {} 0,\quad x_0,y_0\in C,\\ x_1= & {} P_C(x_0-\lambda F(y_0)), y_1=P_C(x_1-\lambda F(y_0)).\\ H_n= & {} \{z\in H: \langle x_n-\lambda F(y_{n-1})-y_n,z-y_n\rangle \ge 0\}.\\ x_{n+1}= & {} P_{H_n}(x_n-\lambda F(y_n)),\\ y_{n+1}= & {} P_C(x_{n+1}-\lambda F(y_n)). \end{aligned}$$

Using the similar technique as in Theorem 5.1, we can prove that the sequence \(\{y_n\}\) generated by the above modified Popov’s algorithm converges to a point in the solution set if n is sufficient large.

6 Numerical illustration

In this section, we illustrate a general example in \({\mathbb {R}}^n\) space and two particular cases to show the finite convergence of sequences generated by above algorithms.

Example 6.1

Let \(H={\mathbb {R}}^n\) endowed with inner product \(\langle \cdot ,\cdot \rangle \) and corresponding norm \(\Vert \cdot \Vert =\Vert \cdot \Vert _2\). Let \(C=\{(x_1,x_2,\ldots ,x_{n}): 0< a\le x_i\le b, i= \overline{1,n}\}\) is closed, convex subset of \({\mathbb {R}}^n\) and \(F:C\rightarrow {\mathbb {R}}^{n}\), defined by

$$\begin{aligned} F(x)=(x_1,x_2,\ldots ,x_{n-1},0) \quad \text {for} \quad x=(x_1,x_2,\ldots ,x_{n-1},x_{n})\in C. \end{aligned}$$

We consider the variational inequality problem VI(CF).

Obviously, for \(x=(x_1, x_2,\ldots ,x_n), y=(y_1, y_2,\ldots ,y_n) \in {\mathbb {R}}^n\) we have

$$\begin{aligned} \Vert F(x)-F(y)\Vert =\left( \sum _{i=1}^{n-1}(x_i-y_i)^2\right) ^{\frac{1}{2}} \le \left( \sum _{i=1}^{n}(x_i-y_i)^2\right) ^{\frac{1}{2}}=\Vert x-y\Vert , \end{aligned}$$

then F is Lipschitz continuous with \(L=1\).

On the other hand, we have that the problem which is to minimize the function \(f:{\mathbb {R}}^n\rightarrow R\) defined by

$$\begin{aligned} f(x_1,x_2,\ldots ,x_n)=\sum _{i=1}^{n-1}\frac{1}{2}x_i^{2}, \end{aligned}$$

over C has same solution set with VI(CF) because f is convex, differentiable function and \(F=\nabla f\). Hence, we can see the solution set of VI(CF) is

$$\begin{aligned} C^*=\{(a,a,\ldots ,a,a, [a,b])\}. \end{aligned}$$

To check the weakly sharp property of \(C^*\), we use Theorem 2.2 (b). It is obvious that \(F(x)=(a,a,\ldots ,a,0) \ \forall \ x\in C^*\), therefore F is constant on \(C^*\). Let \(x=(x_1,x_2,\ldots ,x_n)\in C\), thus \(a\le x_1,x_2, \ldots ,x_n\le b \), we have

$$\begin{aligned} \sum _{i=1}^{n-1}x_i(x_i-a)&\ge \sum _{i=1}^{n-1}a(x_i-a)\nonumber \\&= a\sqrt{\left( \sum _{i=1}^{n-1}(x_i-a)\right) ^2}\nonumber \\&\ge a\sqrt{\sum _{i=1}^{n-1}(x_i-a)^2}. \end{aligned}$$
(44)

Using inequality 44, noticing that \(P_{C^*}(x)=(a,a,\ldots ,a,x_n)\) and \( x-P_C{^*}(x)=(x_1-a,x_2-a,\ldots ,x_{n-1}-a,0)\), we get

$$\begin{aligned} \langle F(x), x-P_{C^*}(x)\rangle&=\langle (x_1,x_2,\ldots , x_{n-1},0),(x_1-a,x_2-a,\ldots ,x_{n-1}-a,0)\rangle \nonumber \\&=\sum _{i=1}^{n-1}x_i(x_i-a)\nonumber \\&\ge a\sqrt{\sum _{i=1}^{n-1}(x_i-a)^2}= a\text {dist}(x,C^*), \end{aligned}$$
(45)

which means the inequality in Theorem 2.2 (b) is satisfied with \(\alpha =a>0\). Thus, \(C^*\) is weakly sharp.

To show our results visually, we consider two following examples in \({\mathbb {R}}^3\) and \({\mathbb {R}}^{100}\).

Let \(H={\mathbb {R}}^3\) and \(a=1, b=10\), then \(C=[1,10]\times [1,10]\times [1,10]\) is a cube in \({\mathbb {R}}^3\), \(F: C \rightarrow {\mathbb {R}}^3\) is defined by the formula \(F(x)=(x_1,x_2,0)\) where \(x=(x_1,x_2,x_3)\in C\) , which means F is the projection from a point to Oxy-plane. We consider the variational inequality problem VI(CF).

We choose starting point \(x_0=(10,10,5)\) for ExtraGradient algorithm and Forward-Backward-Forward algorithm with step size \(\lambda =0.5<1/L\). For Popov’s algorithm, we take \(x_0=(10,10,5)\), \(y_0=(10,10,1)\) and \(\lambda =0.3 < (\sqrt{2}-1)/L\). It is clear from Fig. 1 that for ExtraGradient algorithm, the sequence \(\{x_n\}\) converges to point \((1,1,5) \in C^*\) after 8 iterations. Similar results are obtained with the iterative sequences generated by Forward-Backward-Forward algorithm and Popov’s algorithm, as displayed in Figs. 2 and 3, respectively.

Fig. 1
figure 1

Performance of ExtraGradient algorithm in \({\mathbb {R}}^3\), where the solution set \(C^*\) is in red

Fig. 2
figure 2

Performance of Forward–Backward–Forward algorithm in \({\mathbb {R}}^3\), where the solution set \(C^*\) is in red

Fig. 3
figure 3

Performance of Popov’s algorithm in \({\mathbb {R}}^3\), where the solution set \(C^*\) is in red

Fig. 4
figure 4

Performance of EG, FBF and Popov algorithms in \({\mathbb {R}}^{100}\)

In the second experiment, we take \(H={\mathbb {R}}^{100}\) and \(a=1, b=100\). We choose the same random starting point \(x_0\) for ExtraGradient, Forward-Backward-Forward algorithm and Popov’s algorithm with \(\lambda =0.5<1/L\) , and one more random starting point \(y_0\) for Popov’s algorithm with \(\lambda =0.4<(\sqrt{2}-1)/L\). After applying three above algorithms, the result is demonstrated in Fig. 4. The x-axis stands for the number of steps while y-axis stands for the distance from points generated by above algorithms to solution set \(C^*\).

We can see from from Fig. 4 that, both ExtraGradient and Forward-Backward-Forward algorithms terminate after 11 steps, meanwhile the Popov’s algorithm shows a faster convergence rate and terminates after 7 steps. It is also noticed that in this particular example, the ExtraGradient and Forward-Backward-Forward algorithms are identical. The reason is that since the vector \(x_n - \lambda F(x_n) \in C\) for all n, the projection operator \(P_C\) does not contribute to the iteration process.