1 Introduction

We consider the following variational relation problem: find x ¯ X such that

  1. (1)

    x ¯ is a fixed point of S, that is, x ¯ S( x ¯ ),

  2. (2)

    R( x ¯ ,y) holds for every yT( x ¯ ),

where X and Y are nonempty sets, S is a set-valued mapping from X to itself, T is a set-valued mapping from X to Y, and R(x,y) is a relation linking xX and yY. In an abstract setting, the relation R is represented by a subset of the product space X×Y so that R(x,y) holds if and only if the point (x,y) belongs to that set. In practice, however, R is often given by equality/inequality of real-valued functions, or by inclusion/intersection of set-valued mappings on X×Y. For instance, given a real-valued function ϕ on X×Y, a variational relation can be defined by any of the following equality and inequalities: ϕ(x,y)=0, ϕ(x,y)0, ϕ(x,y)>0 or ϕ(x,y)0. When two set-valued mappings G 1 and G 2 are given on X×Y with values in a nonempty set Z, a variational relation can be defined by any of the following inclusions and intersections: G 1 (x,y) G 2 (x,y), G 1 (x,y) G 2 (x,y), G 1 (x,y) G 2 (x,y)= or G 1 (x,y) G 2 (x,y). A mixture of the above relations is also possible.

The variational relation problem was introduced in [1] and studied in a number of recent works [213]. It encompasses a large class of problems of applied mathematics including optimization problems, variational inequalities, variational inclusions, equilibrium problems etc., and offers a unifying treatment of problems that come from different areas and have a similar structure. Existence conditions of solutions to variational relation problems were analyzed in great generality, the stability of solutions of a parametric variational relation was also studied with respect to the continuity of set-valued mappings, and very recently a numerical method was developed to solve variational relation problems when the data are linear [14]. As far as we know, all conditions established for existence of solutions of variational relation problems in the above cited papers utilize intersection theorems or fixed point theorems involving the KKM property of set-valued mappings in one or another form [15]. In the present paper, we wish to give existence conditions by exploiting Mizoguchi and Takahashi’s fixed point theorem for contraction mappings and propose an algorithm to compute solutions to (VRP), which seems to be new in the theory of variational relations. Actually, we shall study a particular model of (VRP) in which R linearly depends on the decision variable x. Throughout we assume

  • X and Y are nonempty closed sets in the finite dimensional Euclidean spaces R n and R m , respectively,

  • S(x)=X for every xX,

  • R(x,y) holds if and only if Axg(y)0 with A a k×n matrix, g a vector function from Y to R k .

Thus, our problem consists of finding a vector x ¯ X such that A x ¯ g(y) for all yT( x ¯ ). This problem can also be transformed to a quasi-variational inequality problem or a quasi-equilibrium problem [1620], but its resolution, as far as we know, is not known, because a general monotonicity hypothesis often requested for quasi-variational inequality problems and quasi-equilibrium problems is not satisfied. If we define a set-valued mapping Γ:XX by

Γ(x)= { z X : A z g ( y )  for all  y T ( x ) } ,

then the (VRP) above is equivalent to the fixed point problem: find x ¯ X such that x ¯ Γ( x ¯ ). Essentially we shall exploit this equivalent formulation of (VRP) to establish existence conditions and to develop a solving method.

The paper is structured as follows. In the next section, we present two preliminary results which constitute major tools of our study: a fixed point theorem by Mizoguchi and Takahashi [21], which generalizes the Nadler fixed point principle [22], and an error bound or Hoffman’s constant for a system of linear inequalities (see [17, 19, 2328]). Section 3 is devoted to sufficient conditions for existence of solutions of (VRP). In the last section, we propose an algorithm to compute a solution of (VRP) and illustrate it by some numerical examples.

2 Preliminaries

Contraction mappings

The famous Banach contraction principle states that if (X,d) is a complete metric space and if f is a real contraction function on X, then f has a fixed point. This principle was generalized to the case of set-valued mappings by Nadler [22] and Markin [29] with the help of the Hausdorff distance. Since then a lot of investigation has been made in order to weaken the contraction hypothesis (see [21, 3034] and many references given in these). In the present paper, we are particularly interested in a theorem by Mizoguchi and Takahashi [21], which can elegantly be applied to our model. Let us recall it in details and make a discussion on its generalization.

Let D 1 and D 2 be nonempty subsets of X. The Hausdorff distance between D 1 and D 2 is defined by

h( D 1 , D 2 )=max { sup x D 1 d ( x , D 2 ) ; sup y D 2 d ( y , D 1 ) } ,

where d(x, D i ) is the distance from x to D i , i=1,2. In R n the Hausdorff distance can be given by

h( D 1 , D 2 )=inf { r > 0 : D 1 D 2 + B ( 0 , r ) , D 2 D 1 + B ( 0 , r ) } ,

where B(0,r) is the closed ball of radius r and centered at the origin of the space. Let F be a set-valued mapping from X to the space of closed subsets of X and >0. We say that F is -Lipschitz on X if for every x, x X, one has

h ( F ( x ) , F ( x ) ) x x .

When <1, an -Lipschitz mapping is called a contraction mapping. We recall the following result by Mizoguchi and Takahachi [21], which generalizes Nadler’s theorem on fixed points of a contraction mapping [22]: The set-valued mapping F, whose values are assumed nonempty closed and bounded, has a fixed point in X if the following conditions hold:

  1. (i)

    there exists a function γ:(0,)[0,1) such that lim sup r t + γ(r)<1 for each t[0,);

  2. (ii)

    for every x,yX, one has h(F(x),F(y))γ(d(x,y))d(x,y).

Further developments of this result can be found in [30, 3538]. In particular the following theorem by Ciric (Theorem 2.2 [30]) is quite general: if F has closed values and if there exists a real function ϕ:[0,)[a,1) for some a(0,1) such that

  • lim sup s t + ϕ(s)<1 for every t0;

  • for every xX, there is yF(x) satisfying

    ϕ ( d ( x , y ) ) d ( x , y ) d ( x , F ( x ) ) , d ( y , F ( y ) ) ϕ ( d ( x , y ) ) d ( x , y ) ,

then F admits a fixed point.

It is unfortunate that this theorem does not fully generalize Mizoguchi and Takahashi’s theorem because in the latter theorem it is not requested that the function ϕ takes its values bigger than a strictly positive number a. Another observation we can make is the fact that Mizoguchi and Takahashi’s theorem remains valid with the same argument even when F has closed values that are not necessarily bounded.

Error bound for a linear system

We consider a linear system Axb in R n , where A is a k×n matrix and b is a k vector. The solution set to this system is denoted P, which is a polyhedral convex set. It is clear that x is a solution to this system if and only if d(Axb, R + k )=0, where R + k is the positive octant of the space. When x is not a solution, it is important to measure the distance from x to the solution set P. This question has completely been studied starting with the famous Hoffman constant in [26] and subsequential works [23, 24, 27, 28] among many others. Here is a formula for the distance from x to P that can be obtained by elementary arguments from linear algebra (see [24]):

d(x,P)= max I I max λ R | I | { 0 } λ , A I x b I A I T λ ,

where |I| denotes the cardinal of I, ℐ the set of index subsets I{1,,k} satisfying two conditions below:

  1. (1)

    I is nonempty and the family of rows a i , iI of the matrix A is linearly independent,

  2. (2)

    there is some yP such that I is contained in the active index set I(y) at y, that is, a i y= b i , iI; A I is the submatrix of A consisting of the rows a i , iI, and A I T is its transpose.

It follows that

d(x,P)αd ( A x b , R + k ) =α ( A x b ) + ,

where ( A x b ) + is the vector obtained from Axb by substituting all negative components by zeros. The constant α>0 is called an error bound of the system or Hoffman’s constant. The best error bound, denoted α too, is obtained by maximizing the function d(x,P)/ ( A x b ) + over xP. It is given by the formula

α= max I I [ γ ( I ) ] 1 / 2 ,

where γ(I) is the smallest Pareto eigenvalue of the matrix A I A I T (see [39]), or equivalently, it is the optimal value of the problem of minimizing y T A I A I T y over the set {y R + | I | : y 2 =1}. If we denote by P the solution set to the perturbed system Ax b , then

h ( P , P ) α b b
(2.1)

whenever both P and P are nonempty.

3 Existence conditions

In this section we establish some sufficient conditions for existence of solutions to the variational problem (VRP) introduced in the beginning of Section 2. Mizoguchi and Takahashi’s fixed point theorem and the error bound (2.1) are the main tools we use for this purpose. Given a real function u(x) on X, we define a value function (called also a marginal function) of u by β(x)= inf z T ( x ) u(z). Stability and sensitivity of this function is one of the indispensable parts of the theory of optimization. We refer the interested reader to the books [17, 19, 25] for greater details. The next result will be needed in the sequel.

Lemma 3.1 Assume that for every xX, the value function β(x) is finite and that there are functions ϕ and ψ from [0,) to [0,) with ψ nondecreasing such that

h(T(x),T(y)ϕ ( d ( x , y ) ) d(x,y),
(3.1)
|u(x)u(y)|ψ ( d ( x , y ) ) d(x,y)
(3.2)

for every x,yX. Then

|β(x)β(y)|ϕ ( d ( x , y ) ) ψ ( d ( x , y ) ϕ ( d ( x , y ) ) ) d(x,y).

Proof Let x,yX be given. For each ϵ>0, choose zT(x) such that β(x)>u(z)ϵ. Choose also wT(y) such that

d(z,w)h ( T ( x ) , T ( y ) ) .
(3.3)

Then by (3.2) we have

β(y)β(x)u(w)u(z)+ϵψ ( d ( w , z ) ) d(w,z)+ϵ.

Combining this, (3.1) and (3.3), we deduce

β ( y ) β ( x ) ψ ( h ( T ( x ) , T ( y ) ) ) h ( T ( x ) , T ( y ) ) + ϵ ψ ( d ( x , y ) ϕ ( d ( x , y ) ) ) ϕ ( d ( x , y ) ) d ( x , y ) + ϵ .

Switching the roles of x and y in the above inequality and taking into account the fact that ϵ is arbitrarily chosen, we obtain |β(x)β(y)|ϕ(d(x,y))ψ(d(x,y)ϕ(d(x,y)))d(x,y) as requested. □

In the remaining part of this section, we assume that for every xX, the values b i (x)= inf z T ( x ) g i (z), i=1,,k, are finite and that the system Azb(x), zX is consistent. The vector whose components are b i (x), i=1,,k, is denoted b(x).

Theorem 3.1 Assume that there are functions ϕ and ψ i , i=1,,k, on [0,) to [0,) with ψ i nondecreasing and satisfying the following properties:

  1. (i)

    h(T(x),T(y))ϕ(d(x,y))d(x,y) for x,yX;

  2. (ii)

    | g i (x) g i (y)| ψ i (d(x,y))d(x,y) for x,yX and i=1,,k;

  3. (iii)

    lim sup s t + ϕ(s) i = 1 k ψ i ( s ϕ ( s ) ) 2 < 1 α for all t>0.

Then (VRP) admits a solution.

Proof The aim is to prove that the set-valued mapping Γ has a fixed point. It is clear that zΓ(x) if and only if a i z g i (y) for every yT(x), or equivalently, a i z b i (x), i=1,,k. By this, Γ(x) consists of the solutions to the system Azb(x). According to (2.1), we have

h ( Γ ( x ) , Γ ( y ) ) α b ( x ) b ( y ) for all x,yX.

We apply Lemma 3.1 to obtain

h ( Γ ( x ) , Γ ( y ) ) α i = 1 k | b i ( x ) b i ( y ) | 2 α ϕ ( d ( x , y ) ) i = 1 k ψ i ( d ( x , y ) ϕ ( d ( x , y ) ) ) 2 d ( x , y ) .

Consider the real function γ(t)=αϕ(t) i = 1 k ψ i ( t ϕ ( t ) ) 2 for every t0. Then the hypotheses of Mizoguchi and Takahashi’s theorem are satisfied for the set-valued mapping Γ, by which it admits a fixed point. Consequently, (VRP) has a solution. □

When the mapping T and the function g are Lipschitz, we derive the following result.

Corollary 3.1 Assume that T is κ-Lipschitz and g 1 ,, g k is -Lipschitz with κ< 1 α k . Then (VRP) has a solution.

Proof Set ϕ(t)=κ and ψ i (t)= for i=1,,k and t[0,) and apply Theorem 3.1. □

Let us now consider the case when the function g is affine, that is, g(y)=Cy+c, where C is a k×m matrix, c is a k vector, and the graph of T is a convex polyhedral set, that is, yT(x) if and only if x and y solve a linear system PyQx+q, where P is a k×m matrix, Q is a k×n matrix and q is a k vector. (VRP) with such linear data is called a linear variational relation problem. It was studied in [14] in which a numerical algorithm based on Delauney’s triangulations is proposed for solving it. It is known that a linear variational problem may have no solutions (see Example 4.1). We wish to apply Theorem 3.1 to derive an existence condition for this model. The rows of the matrix C are denoted C 1 ,, C k , while the components of the vector c are denoted c 1 ,, c k . The best error bound for the system PzQx+q is denoted α .

Corollary 3.2 Assume that the error bounds α and α , the matrices Q and C of the linear (VRP) satisfy

α α Q C 1 2 + + C k 2 <1.

Then (VRP) has a solution.

Proof For every x,yX, by applying (2.1), we have

| g i ( x ) g i ( y ) | = | C i x + c i ( C i y + c i ) | C i d ( x , y ) , h ( T ( x ) , T ( y ) ) α Q x + q ( Q y + q ) α Q d ( x , y ) .

Then, in view of Lemma 3.1, one deduces

| b i (x) b i (y)| α Q C i d(x,y)for i=1,,k.

It remains to apply Theorem 3.1 to conclude that (VRP) has a solution. □

4 An algorithm

In this section, we consider a linear (VRP) as mentioned in the preceding section, namely we wish to find x ¯ R n such that A x ¯ Cy+c for every y solution to the system PyQ x ¯ +q. As we have already noticed, this problem may have no solutions. Here is an example.

Example 4.1 We consider a linear variational relation problem with x[0,2] and yR. The relation R is defined by the system

( 1 1 1 ) x ( 1 0 0 ) y+ ( 1 0 2 )

and T is defined by the system

( 0 0 1 ) x ( 1 1 0 ) y+ ( 4 0 1 ) .

For every x[0,2], yT(x) if and only if 0y4. With y=4, it is clear that R(x,y) does not hold, which means that (VRP) has no solutions.

Let us now describe an algorithm to solve (VRP).

Step 0. Choose any x 0 X, a small tolerance level ϵ>0. Set r=0.

Step 1. For i=1,,k, solve

minimize  C i y + c i subject to  P y Q x r + q , subject to  y Y .

Let b r be the vector whose components are optimal values of the above programs.

Step 2. Solve

minimize  x r z subject to  A z b r , subject to  x X .

Let z be an optimal solution of this program.

Step 3. Check x r zϵ. If yes, stop. The optimal solution z is considered as a solution of (VRP). Otherwise, set r=r+1, x r =z and return to Step 1.

We have the following convergence property of the algorithm.

Proposition 4.1 Assume that the hypothesis of Corollary  3.2 holds. Then, for any ϵ>0, the algorithm terminates after a finite number of iterations. If at some iteration the optimal value of the program in Step 2 is zero, then x r is a solution of (VRP). If not, with ϵ tending to 0, the sequence of x r obtained in Step 3 converges to a solution of (VRP).

Proof It is clear that x r + 1 Γ( x r ). If, for some r, the optimal value of the program in Step 2 is zero, then x r Γ( x r ) which means that x r is a fixed point of Γ, and hence a solution of (VRP). Moreover, since under the hypothesis of Corollary 3.2 the mapping Γ is a contraction, the sequence { x r } r = 0 converges to a fixed point of Γ, which is also a solution of (VRP). By this the algorithm terminates after a finite number of iterations when ϵ is strictly positive. □

Below we give two examples to show how to perform the algorithm. In the first example, the algorithm terminates after two iterations and produces an exact solution. In the second example, we can obtain an approximate solution with any given strictly positive tolerance level.

Example 4.2 Consider a linear (VRP) with the following data: X=[0,2], Y= R 2 and

A = ( 1 1 1 ) , C = ( 1 / 4 1 / 4 0 0 0 0 ) , c = ( 1 0 2 ) , P = ( 1 / 2 1 / 2 1 / 2 1 / 2 1 0 0 1 ) , Q = ( 1 1 0 0 ) , q = ( 0 1 / 2 0 0 ) .

Iteration 1. We start the algorithm with x 0 =0. The feasible set of the programs in Step 1 is given by the system Pyq, or equivalently,

y 1 + y 2 0 , y 1 y 2 1 , y 1 , y 2 0 .

The three functions to minimize are respectively ( y 1 + y 2 )/41, 0 and 2, which implies that b 0 = ( 1 , 0 , 2 ) T . In the next step, we minimize |z| over the set Az b 0 , zX. The optimal solution is z=1. As x 0 z, it is not a solution of (VRP).

Iteration 2. We set x 1 =z=1 and return to Step 1 to solve the three above mentioned functions over the set PyQ x 1 +q. The same optimal solutions are obtained and b 1 = ( 1 , 0 , 2 ) T . In Step 2, we minimize |1z| over the set Az b 1 , zX and obtain the optimal solution z=1. Since x 1 =z, the algorithm terminates and x 1 =1 is a solution of (VRP). It is easy to see that Γ is a (1/2)-contraction mapping, that is, h(Γ(x),Γ( x ))|x x |/2 for every x, x [0,2].

Example 4.3 We solve a linear (VRP) with the data X, Y, A, P, Q and q as given in the preceding example and with

C= ( 1 / 4 7 / 16 0 0 0 0 ) ,c= ( 11 / 16 0 2 ) .

Iteration 1. We start the algorithm with x 0 =0. The feasible set of the programs in Step 1 is given by the system Pyq and consists of one element {0}. Hence the vector b 0 is equal to ( 11 / 16 , 0 , 2 ) T . In the next step, we minimize |z| over the set Az b 0 , zX. The optimal solution is z=11/16. As x 0 z, it is not a solution of (VRP).

Iteration 2. We set x 1 =z=11/16 and return to Step 1 to solve three programs whose feasible set is given by PyQ x 1 +q. We obtain b 1 = ( 109 / 128 , 0 , 2 ) T . In Step 2, we minimize |11/16z| over the set Az b 1 , zX and obtain the optimal solution z=109/128. Since x 1 z, the point x 1 is not a solution of (VRP). If we choose a priori a tolerance level ϵ=0.2, then we may stop the algorithm and consider x 2 =109/128 as an approximate solution of (VRP) because | x 1 x 2 |<ϵ. If not, we continue it with x 2 for restarting the procedure. It can be seen that the algorithm generates the sequence { x r } r = 0 converging to x=2, which is the unique solution of (VRP).