New Trends in General Variational Inequalities

It is well known that general variational inequalities provide us with a unified, natural, novel and simple framework to study a wide class of unrelated problems, which arise in pure and applied sciences. In this paper, we present a number of new and known numerical techniques for solving general variational inequalities and equilibrium problems using various techniques including projection, Wiener-Hopf equations, dynamical systems, auxiliary principle and penalty function. General variational-like inequalities are introduced and investigated. Properties of higher order strongly general convex functions have been discussed. The auxiliary principle technique is used to suggest and analyze some iterative methods for solving higher order general variational inequalities. Some new classes of strongly exponentially general convex functions are introduced and discussed. Our proofs of convergence are very simple as compared with other methods. Our results present a significant improvement of previously known methods for solving variational inequalities and related optimization problems. Since the general variational inequalities include (quasi) variational inequalities and (quasi) implicit complementarity problems as special cases, these results continue to hold for these problems. Some numerical results are included to illustrate the efficiency of the proposed methods. Several open problems have been suggested for further research in these areas.


Introduction
Variational inequalities theory, which was introduced Stampacchia [173] and Ficchera [38] indecently has emerged as an interesting and fascinating branch of applicable mathematics with a wide range of applications in industry, finance, economics, social, pure and applied sciences. Variational inequalities may be viewed as novel generalization of the variational principles, the origin of which can be traced back to Euler, Lagrange and Bernoulli brothers. The variational principles have placed a crucial and important in the development of various fields of sciences and have appeared as a unifying force. The ideas and techniques of variational inequalities are being applied in a variety of diverse areas of sciences and prove to be productive and innovative. Variational inequalities have been extended and generalized in several directions using novel and new techniques. The minimum of a differentiable convex function F on the convex set K is equivalent to finding u ∈ K such that which is known as the variational inequality (1). Here F ′ (u) is the Frechet differential. Stampacchia [173] proved that potential problems associated with elliptic equations can be studied by the variational inequality. This simple fact inspired a great interest in variational inequalities. Lions and Stampacchia [54] studied the existence of a solution of variational inequalities using essentially the auxiliary principle technique coupled with projection idea. Lemke [62] considered the problem of finding u ∈ R n + such that is called the linear complementarity problem. Here A is a linear operator. Lemke [62] proved that the two person game theory problems can be studied in the framework of linear complementarity problem (2). See also Lemke and Howson, Jr. [63] and Cottle al et [27] for the nonlinear complementarity problems. It is worth mentioning that both the problems (1) and (2) are different and have been studied in infinite dimensional spaces and finite dimensional spaces independently using quite different techniques. However. Karmardian [57] established that both problems (1) and (2) are equivalent, if the underlying set K is a convex cone. This equivalent formulation played an important role in developing several techniques for solving these problems.
If the convex set K depends upon the solution explicitly or implicitly, then variational inequality is called the quasi variational inequality. Quasi variational inequalities were introduced and investigated by Bensoussan and Lions [15] in control theory. In fact, for a given operator T : H −→ H, and a point-to-set mapping K : u −→ K(u), which associates a closed convex-valued set K with any element u of H, we consider the problem of finding u ∈ K(u) such that which is known as quasi variational inequality. Chan and Pang [22] considered the generalized quasi variational inequalities for set-valued operators. Noor [84] established the equivalence between the quasi variational inequalities and the fixed point formulation and used this equivalence to suggest some iterative methods for solving (3). This equivalence was used to study the existence of a solution of quasi variational inequalities and develop numerical methods.
Related to the quasi variational inequality, we have the problem of finding u ∈ H such that is called the implicit(quasi) complementarity problem. where m is a point-to-point mapping. Using the technique of Karamardian [57], Pang [155] and Noor [84] established the equivalence between the problems (3) and (4). Noor [85,86] has used the change of variables technique to prove that the implicit complementarity problems are equivalent to the fixed point problem. This alternative formulation played an important part in the development of iterative methods for solving various type of complementarity problems and related optimization problems. It is an interesting problem to extend this technique for solving variational inequalities.
Motivated and inspired by the ongoing research in these fields, Noor [87] introduced and investigated a new class of variational inequalities involving two operators. For given nonlinear operators T, g, consider the problem of finding u ∈ H : g(u) ∈ K, such that Tu, g(v) − g(u) ≥ 0, ∀v ∈ H : g(v) ∈ K, (5) which is known as the general (Noor) variational inequalities. It turned out that odd-order and nonsymmetric obstacle, free, unilateral and moving boundary value problems arising in pure and applied sciences can be studied via the general variational inequalities, see [87,88,89,90,91]. If K is a convex cone, then the implicit complementarity problem (5) is equivalent to finding u ∈ H such that g(u) ≥ 0, Tu ∈ K * , Tu, g(u) = 0, which is known as the general complementarity problem, where K * is the dual (polar) cone. We would like to point out that for appropriate and suitable choice of the operators T, g and the convex sets K, one can obtain several known and new classes of variational inequalities and complementarity problems as special cases of the problem (5).
During the years which have been elapsed since its discovery, a number of numerical methods including projection method and its variant forms, Wiener-Hopf equations, auxiliary principle, dynamical systems have been developed for solving the variational inequalities and related optimization problems. Projection method and its variants forms including the Wiener-Hopf equations represent important tools for finding the approximate solution of variational inequalities, the origin of which can be traced back to Lions and Stampacchia [66]. The main idea in this technique is to establish the equivalence between the variational inequalities and the fixed-point problem by using the concept of projection. This alternative formulation has played a significant part in developing various projection-type methods for solving variational inequalities. It is well known that the convergence of the projection methods requires that the operator must be strongly monotone and Lipschitz continuous. Unfortunately these strict conditions rule out many applications of this method. This fact motivated to modify the projection method or to develop other methods. The extragradient-type methods overcome this difficulty by performing an additional forward step and a projection at each iteration according to the double projection. These methods can be viewed as predictor-corrector methods. Their convergence requires only that a solution exists and the monotone operator is Lipschitz continuous. When the operator is not Lipschitz continuous or when the Lipschitz continuous constant is not known, the extragradient method and its variant forms require an Armijo-like line search procedure to compute the step size with a new projection needed for each trial, which leads to expansive computation. To overcomes these difficulties, several modified projection and extragradient-type methods have been suggested and developed for solving variational inequalities. See [6,7,8,9,10,17,18,24,36,37,39,42,43,44,49,55,63,68,71,79,82,83,90,92,95,100,101,102,103,104,106,107,110,111,112,113,114,116,117,119,120,133,134,135,137,138,143,144,145,146,150,152,153,160] and the references therein.
In Section 4, we have the concept of the general Wiener-Hopf equations, which was introduced by Noor [90]. As a special case, we obtain the original Wiener-Hopf equations, which were considered and studied by Shi [168] and Robinson [165] in conjunction with variational inequalities from different point of views. Using the projection technique, one usually establishes the equivalence between the variational inequalities and the Wiener-Hopf equations. It turned out that the Wiener-Hopf equations are more general and flexible. This approach has played not only an important part in developing various efficient projection-type methods, but also in studying the sensitivity analysis, dynamical systems as well as other concepts of variational inequalities. Noor, Wang and Xiu [152,153] and Noor and Rassias [146] have suggested and analyzed some predictor-corrector type projection methods by modifying the Wiener-Hopf equations. These methods are also known as Forward-Backward methods, see Tseng [179,180]. It has been shown that these predictor-corrector-type methods are efficient and robust. Some numerical examples are given to illustrate the efficiency and implementation of the proposed methods. Consequently, our results represent a refinement and improvement of the known results.
Section 5 is devoted to the concept of projected dynamical system in the context of variational inequalities, which was introduced by Dupuis and Nagurney [35] by using the fixed-point formulation of the variational inequalities. For the recent development and applications of the dynamical systems, see [13,34,35,40,41,56,75,79,109,115,116,122]. In this technique, we reformulate the variational inequality problem as an initial value problem. Using the discretizing of the dynamical systems, we suggest some new iterative methods for solving the general variational inequalities.
It is well known fact that to implement the projection-type methods, one has to evaluate the projection, which is itself a difficult problem. Secondly, the projection and Wiener-Hopf equations techniques can't be extended and generalized for some classes of variational inequalities involving the nonlinear (non)differentiable functions, see [92,94,108]. These facts motivated to use the auxiliary principle technique. This technique deals with finding the auxiliary variational inequality and proving that the solution of the auxiliary problem is the solution of the original problem by using the fixed-point approach. It turned out that this technique can be used to find the equivalent differentiable optimization problems, which enables us to construct gap (merit) functions. Glowinski et al. [47] used this technique to study the existence of a solution of mixed variational inequalities. Noor [93,94,95,100,101,114,121,122] has used this technique to suggest some predictor-corrector methods for solving various classes of variational inequalities. It is well known that a substantial number of numerical methods can be obtained as special cases from this technique. We use this technique to suggest and analyze some explicit predictor-corrector methods for general variational inequalities. In this paper, we give the basic idea of the inertial proximal methods and show that the auxiliary principle technique can be used to construct gap (merit) functions. We use the gap function to consider an optimal control problem governed by the general variational inequalities. The control problem as an optimization problem is also referred as a generalized bilevel programming problem or mathematical programming with equilibrium constraints. These results are mainly due to Deitrich [32,33]. It is an open problem to compare the efficiency of the inertial methods with other methods and this is another direction for future research. In Section 7, we discuss the application of the penalty function method, which was introduced by Lewy and Stampacchia [64] to study the regularity of the solutions of the variational inequalities. It is known that the finite difference and similar numerical methods cannot be applied to find the approximate solutions of the obstacle, free and moving value problems due to the presence of the obstacle and other constraint conditions. However, it is known that if the obstacle is known then these obstacle and unilateral problems can be characterized by a system of differential equations in conjunction with the general variational inequalities using the penalty function technique. Al-Said [3], Noor and Al-Said [112], Noor and Tirmizi [148] and Al-Said et al. [4,5] used this technique to develop some numerical methods for solving these system of differential equations. The main advantage of this technique is its simple applicability for solving system of differential equations. We here give only the main idea of this technique for solving odd-order obstacle and unilateral problems.
In recent years, much attention has been given to study the equilibrium problems, which were considered and studied by Blum and Oettli [19] and Noor and Oettli [142]. It is known that equilibrium problems include variational inequalities and complementarity problems as special cases. It is remarked that there are very few iterative methods for solving equilibrium problems, since the projection method and its variant forms including the Wiener-Hopf equations cannot be extended for these problems. We use the auxiliary principle technique to suggest and analyze some iterative type methods for solving general equilibrium problems, which is considered in Section 8.
Hanson [50] introduced the concept of the invex functions to study the mathematical programming problems, which appeared to be significant generalization of the convex functions. Ben-Israel and Mond [14] considered the concepts of invex sets and preinvex functions. They proved that the differentiable preinvex functions are invex function. Mohan and Neogy [74] proved that the converse is also true under certain conditions. Noor [93] proved that the optimality conditions can be characterized by a class of variational inequalities, which is called variational-like inequality. Due to the inherent nonlinearity, one can not use the projection type iterative methods for considering the existence results and numerical methods for variational-like inequalities. However, one uses the auxiliary principle technique to study the existence and numerical methods for variational-like inequalities. Fulga and previda [43] and Awan et al [11] considered the general invex sets and general preinvex functions involving an arbitrary function and studied their basic properties. We show that the minimum of the differentiable general preinvex functions is characterized by a class of variational-like inequality. This fact motivated us to introduce a the general variational-like inequalities and study its properties. We have used the auxiliary principle technique to analyze some iterative methods for solving the general variational-like inequalities. Several special cases are discussed as applications of the general variational-like inequalities. These aspects are discussed in Section 9.
In section 10, we consider the concept of higher order strongly general convex functions involving an arbitrary function, which can viewed as a novel and innovative extension of the strongly convex functions. Polyak [159] in 1966 introduced the strongly convex functions to study the optimization problems. Zu and Marcotte [201] discussed the role of the strongly convex functions in the analysis of the iterative methods for solving variational inequalities. Mohsen et al [75] introduced the higher order strongly convex functions involving bifunction, which can be viewed as significant refinement of the higher order strongly convex function, which were considered by Lin and Fukushima [65] in mathematical programming with equilibrium constraints. They have shown that parallelogram laws for Banach spaces can be obtained as applications of the higher order strongly convex functions. Parallelogram laws for Banach spaces were analyzed by Bynum [21] and Chen at al [23,24,25], which have applications in prediction theory and information technology. We have investigated some basic properties of the higher 0rder strongly general convex functions and have shown that the optimality conditions of the differentiable higher order strongly general convex functions can be expressed as higher order general variational inequalities. Higher order general variational inequalities are introduced in Section 11. Some iterative methods are suggested and analyzed for solving higher order general variational inequalities. It is shown that general variational inequalities and related optimization problems can be obtained as applications.
Related to the convex functions, we have the concept of exponentially convex(concave) functions, which have important applications in information theory, big data analysis, machine learning and statistic. Exponentially convex functions have appeared significant generalization of the convex functions, the origin of which can be traced back to Bernstein [16]. Avriel [9,10] introduced the concept of rconvex functions, from which one can deduce the exponentially convex functions. Antczak [2] considered the (r, p) convex functions and discussed their applications in mathematical programming and optimization theory. Alirazaie and Mahar [1] investigated the impact of exponentially concave functions in information theory. Zhao et al [200] discussed some characterizations of r-convex functions. Awan et al [5] also investigated some classes of exponentially convex functions. Noor and Noor [132,133,134,135,136,137,138] discussed the characterization of several classes of exponentially convex functions. In Section 12, we introduce the concept of strongly exponentially general convex functions and show that the strongly exponentially enjoy some nice properties, which convex functions have.
Theory of general variational inequalities is quite broad, so we shall content ourselves here to give the flavour of the ideas and techniques involved. The techniques used to analysis the iterative methods and other results for general variational inequalities are a beautiful blend of ideas of pure and applied mathematical sciences. In this paper, we have presented the some results regarding the development of various algorithms, their convergence analysis and penalty computational technique. Although this paper is expository in nature, our choice has been rather to consider some interesting aspects of general variational inequalities. The framework chosen should be seen as a model setting for more general results for other classes of variational inequalities and variational inclusions. One of the main purposes of this expository paper is to demonstrate the close connection among various classes of algorithms for the solution of the general variational inequalities and to point out that researchers in different field of variational inequalities and optimization have been considering parallel paths. We would like to emphasize that the results obtained and discussed in this paper may motivate and bring a large number of novel, innovate and potential applications, extensions and interesting topics in these areas. The comparison of the proposed methods with other techniques needs further efforts and is itself an open interesting problem We have given only a brief introduction of the general variational inequalities in the real Hilbert spaces. For some other aspects of the general variational inequalities, readers are referred to the state-of-art articles of Noor [85,86,87,89,90,91,92,93,94,95,100,101,102,103,104,105,110,118,119,120,121,122,124] and the references therein. The interested reader is advised to explore this field further and discover novel and fascinating applications of this theory in Banach and topological spaces. It is perhaps part of the fascination of the subject that so many branches of pure and applied sciences are involved in the variational inequality theory. The task of becoming conversant with a wide spectrum of knowledge is indeed a real challenge. The general theory is quite technical, so we shall content ourselves here to give the flavour of the main ideas involved. The techniques used to analyze the existence results and iterative algorithms for variational inequalities are a beautiful blend of ideas from different areas of pure and applied mathematical sciences. The framework chosen should be seen as a model setting for more general results. However, by just relying on these special results, interesting problems arising in the applications can be dealt with easily. Our main motivation of this paper to give a summary account of the basic theory of variational inequalities set in the framework of nonlinear operators defined on convex sets in a real Hilbert space. We focus our attention on the iterative methods for solving variational inequalities. The equivalence between the variational inequalities and the Wiener-Hopf equations has been used to suggest some new iterative methods. The auxiliary principle technique is applied to study the existence of the solution and to propose a novel and innovative general algorithm for the general variational inequalities, equilibrium problems and related optimization.

Preliminaries and Basic concepts
Let H be a real Hilbert space, whose inner product and norm are denoted by ·, · and · respectively.
It is well known that a function F is a convex functions, if and only if, it satisfies the inequality which is known as the Hermite-Hadamard type inequality. Such type of the inequalities provide us with the upper and lower bounds for the mean value integral. If the convex function F is differentiable, then u ∈ K is the minimum of the F, if and only if, u ∈ K satisfies the inequality which is called the variational inequality, introduced and studied by Stampacchia [173] in 1964. For the applications, formulation, sensitivity, dynamical systems, generalizations, and other aspects of the variational inequalities, see  and references therein. It is known that a set may not be a convex set. However, a set may be made convex set with respect to an arbitrary function. Motivated by this fact, Youness [197] introduced the concept of a general convex set involving an arbitrary function.
Definition 3. The set K in H is said to be a general convex set, if there exists an arbitrary function g, such that Note that, if g = I, the identity operator, then general convex set reduces to the classical convex set. Clearly every convex set is a general convex set, but the converse is not true. For the sake of simplicity, we always assume that ∀u, v ∈ H : g(u), g(v) ∈ K, unless otherwise.

Definition 4.
A function F is said to a general convex function, if there exists an arbitrary function g such that It is known that every convex function is a general convex function, but the converse is not true. For example, the function F(x) = e x 2 is a general convex function, but it is not convex.
We now define the general convex functions on I g = [g(a), g(b)].
Then F is a general convex function, if and only if, One can easily show that the following are equivalent: 1. F is a general convex function.
We now show that the minimum of a differentiable general convex function on K in H can be characterized by the general variational inequality. This result is mainly due to Noor [110]. Theorem 1. [110] Let F : K −→ H be a differentiable general convex function. Then u ∈ H : g(u) ∈ K is the minimum of a differentiable general convex function F on K, if and only if, u ∈: g(u) ∈ K satisfies the inequality where F ′ is the differential of F at g(u) ∈ K in the direction g(v) − g(u).
Proof. Let u ∈ H : g(u) ∈ K be a minimum of general convex function F on K. Then Dividing the above inequality by t and taking t −→ 0, we have which is the required result (7). Conversely, let u ∈ H, g(u) ∈ K satisfy the inequality (7). Since F is a general convex function, so ∀g(u), which implies that Letting t −→ 0, we have which implies that showing that u ∈ H : g(u)K is the minimum of F on K in H.
Theorem1 implies that general convex programming problem can be studied via the general variational inequality (9) with Tu = F ′ (g(u)). In a similar way, one can show that the general variational inequality is the Fritz-John condition of the inequality constrained optimization problem. In many applications, the general variational inequalities do not arise as the minimization of the differentiable general convex functions. Also, it is known that the variational inequality introduced by Stampacchia [173] can only be used to study the even-order boundary value problems. These facts motivated Noor [87] to introduce a more general variational inequality involving two distinct operators. General variational inequalities is a unified framework to study such type problems.
Let K be a closed convex set in H and T, g : H −→ H be nonlinear operators. We now consider the problem of finding u ∈ H, g(u) ∈ K such that Problem (9) is called the general variational inequality, which was introduced and studied by Noor [87] in 1988. It has been shown that a large class of unrelated oddorder and nonsymmetric obstacle, unilateral, contact, free, moving, and equilibrium problems arising in regional, physical, mathematical, engineering and applied sciences can be studied in the unified and general framework of the general variational inequalities (9). Luc and Noor [69] have studied the local uniqueness of solution of the general variational inequality (9) by using the concept of Frechet approximate Jacobian.
We now discus some special cases of the general variational inequality (9).
(I). For g ≡ I, where I is the identity operator, problem (9) is equivalent to finding u ∈ K such that which is known as the classical variational inequality introduced and studied by Stampacchia [173] in 1964. For recent state-of-the-art in this field, see  and the references therein. From now onward, we assume that g is onto K unless, otherwise specified.
is a normal cone to the convex set K at u, then the general variational inequality (9) is equivalent to finding u ∈ H, g(u) ∈ K such that which are known as the generalized nonlinear equations, see [129,130].
(III). If P tg is the projection of −Tu at g(u) ∈ K, then it has been shown that the general variational inequality problem (9) is equivalent to finding u ∈ H, g(u) ∈ K such that P tg [−Tu] := P tg (u) = 0, which are known as the tangent projection equations. This equivalence has been used to discuss the local convergence analysis of a wide class of iterative methods for solving general variational inequalities (9).
(IV). If K * = {u ∈ H : u, v ≥ 0, ∀v ∈ K} is a polar (dual) cone of a convex cone K in H, then problem (9) is equivalent to finding u ∈ H such that g(u) ∈ K, Tu ∈ K * and Tu, g(u) = 0, which is known as the general complementarity problem, see Noor [87]. For g(u) = m(u) + K, where m is a point-to-point mapping, is called the implicit (quasi) complementarity problem. If g ≡ I, then problem (11) is known as the generalized complementarity problems. Such problems have been studied extensively in recent years.
(V). If K = H, then the general variational inequality (9) is equivalent to finding u ∈ H : g(u) ∈ H such that which is called the weak formulation of the odd-order and nonsymmetric boundary value problems. For suitable and appropriate choice of the operators and spaces, one can obtain several classes of variational inequalities and related optimization problems as special cases of the general variational inequalities (9).
We also need the following result, which plays a key role in the studies of variational inequalities and optimization theory.

Lemma 1.
[59] For a given z ∈ H, u ∈ K satisfies the inequality if and only if u = P K z, where P K is the projection of H onto K.
Also, the projection operator P K is nonexpansive, that is, and satisfies the inequality

Applications
We now discuss some applications of general variational inequalities (9). For this purpose, we consider the functional I [v], defined as which is called the general energy or potential, virtual work functional. We remark that, if g ≡ I, the identity operator, then the functional I[v] reduces to which is known as the standard energy function.
It is known that, if the operator T : H −→ H is linear, symmetric and positive, then the minimum of the functional J[v] on the closed and convex set K in H is equivalent to finding u ∈ K such that Inequality of type (14) is known as variational inequality, which was introduced by Stampacchia [173] in the study of potential theory. See also Fichera [38]. It is clear that the symmetry and positivity of the operator T is must. On the other hand, there are many important problems, which are nonsymmetric and non-positive. For the nonsymmetric and odd-order problems, many methods have developed by several authors including Filippov [39] and Tonti [177] to construct the energy functional of type (13 by introducing the concept of g-symmetry and g-positivity of the operator g. We now recall the following concepts. (c). g-coercive (g-elliptic, if there exists a constant α > 0 such that Note that g-coercivity implies g-positivity, but the converse is not true. It is also worth mentioning that there are operators which are not g-symmetric but g-positive. On the other hand, there are g-positive , but not g-symmetric operators. Furthermore, it is well-known [39,177] that if, for a linear operator T, there exists an inverse operator T −1 operator on R(T ) with R(T ) = H, then one can find an infinite set of auxiliary operators g such that the operator T is both g-symmetric and g-positive.
We now consider the problem of finding the minimum of the functional I[v], defined by (13), on the convex set K in H and this is the main motivation of our next result.
Proof. Let u ∈ H, g(u) ∈ K satisfy (15). Then, using the g-positivity of the operator T, we have Since K is a convex set, so for all t Taking v = v t in (16) and using the fact that g is linear (or convex), we have We now define the function ≥ 0, by (17) .
Thus it follows that h(t) is an increasing function on [0, 1] and so h(0) ≤ h(1) gives us that is, Taking (18), we have Using (13) and the linearity (or convexity ) of g, we obtain from which, as t −→ 0, we have This completes the proof.
We remark that for g = I, the identity operator, Theorem 2 reduces to the following well-known result in variational inequalities, which is due to Stampacchia [173].
Proof. Its proofs follows from Theorem 2.
Example 1. We now show that a wide class of nonsymmetric and odd-order obstacle, unilateral, free, moving and general equilibrium problems arising in pure and applied sciences can be formulated in terms of (13). For simplicity and to illustrate the applications, we consider the obstacle boundary value of third order of the type: Find u such that where Ω = [a, b] is a domain, ψ(x) and f (x) are the given functions. The function ψ is known as the obstacle function. The region, where u(x) = ψ(x), for x ∈ Ω is called the contact region (set).
We not that problem (19) is a generalization of the third-order boundary value problem which arises from a similarity solution of the so-called barotropic quasi-geostrophic potential vorticity equation for one layer ocean circulation. For the formulation of the equation, see [91] and the references therein.
To study the problem (19) in the general framework of the general variational inequality, we define which is closed convex set in H 2 0 (Ω ). For the definition and properties of the spaces H m 0 (Ω ), see [50].
Using the technique of [39,177], we can easily show that the energy functions associated with the problem (19) is where and Here are the linear operators. It is clear that the operator T defined by the relation (21) is g-symmetric, g-positive and linear. Also we note that the operator g = d dx is a linear operator. Consequently all the assumptions of Theorem 2 are satisfied. Thus it follows from Theorem 2 that the minimum of the functional I [v], defined by (20) is equivalent to finding u ∈ H such that g(u) ∈ K and the inequality (9) holds. In fact, we conclude the problems equivalent to (19) are: The Variational Problem. Find u ∈ H 2 0 (Ω ), which gives the minimum value to the functional The Variational Inequality (Weak) Problem. Find u ∈ H 2 0 (Ω ) such that g(u) ∈ K and

Quasi variational inequalities
We now show that quasi variational inequalities are a special case of general variational inequalities (9). If the convex set K depends upon the solution explicitly or implicitly, then variational inequality problem is known as the quasi variational inequality. For a given operator T : H −→ H, and a point-to-set mapping K : u −→ K(u), which associates a closed convex-valued set K(u) with any element u of H, we consider the problem of finding u ∈ K(u) such that Inequality of type (22) is called the quasi variational inequality. For the formulation, applications, numerical methods and sensitivity analysis of the quasi variational inequalities, see [15,22,84,85,102,139,147] and the references therein. We can rewrite the equation (22), for ρ > 0, as which is equivalent ( using Lemma 1) to finding u ∈ K(u) such that In many important applications, the convex-valued set K(u) is of the form where m is a point-to-point mapping and K is a closed convex set.
From (23) and (24), we see that problem (22) is equivalent to which implies that which is equivalent to the general variational inequality (9) by an application of Lemma 1. We have shown that the quasi variational inequalities (22) with the convex-valued set K(u) defined by (24) are equivalent to the general variational inequalities (9). We now recall the well known concepts.
For g ≡ I, Definition 7 reduces to the usual definition of monotonicity and pseudomonotonicity of the operator T. Note that monotonicity implies pseudomonotonicity but the converse is not true, see [35].

Definition 8.
A function F is said to be strongly general convex on the general convex set K with modulus µ > 0, if, for all g(u), g(v) ∈ K g ,t ∈ [0, 1], For differentiable strongly general convex function F, the following statements are equivalent.
It is well known that the general convex functions are not convex function, but they have some nice properties which the convex functions have. Note that, for g = I, the general convex functions are convex functions and definition (8) is the well known result in convex analysis.

Projection Methods
In this section, we use the fixed point formulation to suggest and analyze some new implicit methods for solving the variational inequalities. Using Lemma 1, one can show that the general variational inequalities are equivalent to the fixed point problems.

Lemma 2. [87]
The function u ∈ H : g(u) ∈ K is a solution of the general variational inequalities (9), if and only if, u ∈ H : g(u) ∈ K satisfies the relation where P K is the projection operator and ρ > 0 is a constant.
Lemma 2 implies that the general variational inequality (9) is equivalent to the fixed point problem (25). This equivalent fixed point formulation was used to suggest some implicit iterative methods for solving the general variational inequalities. One uses the equivalent fixed point formulation (25) to suggest the following iterative methods for solving general variational inequalities (9).
Algorithm 1 For a given u 0 ∈ H, compute u n+1 by the iterative scheme which is known as the projection method and has been studied extensively.
which can be viewed as the extragradient method, which was suggested and analyzed by Koperlevich [60] for solving the classical variational inequalities. Noor [114]has proved that the convergence of the extragradient method for pseudomonotone operators.
Algorithm 3 For a given u 0 ∈ H, compute u n+1 by the iterative scheme which is known as the modified projection method and has been studied extensively, see Noor [103].
We can rewrite the equation (25) as: This fixed point formulation was used to suggest the following implicit method for solving variational inequalities, which is due to Noor et al [110,116]. We used this equivalent formulation to suggest implicit methods for general variational inequality (9).
Algorithm 4 For a given u 0 ∈ H, compute u n+1 by the iterative scheme For the implementation of this Algorithm, one can use the predictor-corrector technique to suggest the following two-step iterative method for solving general variational inequalities.
Algorithm 5 For a given u 0 ∈ H, compute u n+1 by the iterative scheme

which is a two-step iterative method:
From the equation (25), we have This fixed point formulation is used to suggest the implicit method for solving the variational inequalities as Algorithm 6 For a given u 0 ∈ H, compute u n+1 by the iterative scheme which is another implicit method, see Noor et al. [149].
To implement this implicit method, one can use the predictor-corrector technique to rewrite Algorithm 6 as equivalent two-step iterative method.
Algorithm 7 For a given u 0 ∈ H, compute u n+1 by the iterative scheme

which is known as the mid-point implicit method for solving general variational inequalities.
For the convergence analysis and other aspects of Algorithm 4, see Noor et al [149]. It is obvious that Algorithm 4 and Algorithm 6 have been suggested using different variant of the fixed point formulations (25). It is natural to combine these fixed point formulation to suggest a hybrid implicit method for solving the general variational inequalities and related optimization problems, which is the main motivation of this paper.
One can rewrite the equation (25) as This equivalent fixed point formulation enables to suggest the following method for solving the general variational inequalities.
Algorithm 8 For a given u 0 ∈ H, compute u n+1 by the iterative scheme

which is an implicit method.
We would like to emphasize that Algorithm 8 is an implicit method. To implement the implicit method, one uses the predictor-corrector technique. We use Algorithm 1 as the predictor and Algorithm 8 as corrector. Thus, we obtain a new two-step method for solving general variational inequalities.
Algorithm 9 For a given u 0 ∈ H, compute u n+1 by the iterative scheme

which is two step method.
For constants λ , ξ ∈ [0.1], we can rewrite the equation (25) as: This equivalent fixed point formulation enables to suggest the following method for solving the general variational inequalities.
Algorithm 10 For a given u 0 ∈ H, compute u n+1 by the iterative scheme

which is an implicit method.
Using the prediction-correction technique, Algorithm 10 can be written in the following form.
Algorithm 11 For a given u 0 ∈ H, compute u n+1 by the iterative scheme.
which is two step method.

Remark 1.
It is worth mentioning that Algorithm 11 is a unified ones. For suitable and appropriate choice of the constant λ and ξ , one can obtain a wide class of iterative methods for solving general variational inequalities and related optimization problems.

Wiener-Hopf Equations Technique
In this Section, we consider the problem of the general Wiener-Hopf equations. To be more precise, let Q K = I − P K , where I is the identity operator and P K is the projection of H onto K. For given nonlinear operators T, g : H → H, consider the problem of finding z ∈ H such that provided g −1 exists. Equations of the type (26) are called the general Wiener-Hopf equations, which were introduced and studied by Noor [90,91]. For g = I, we obtain the original Wiener-Hopf equations, which were introduced and studied by Shi [168] and Robinson [166] in different settings independently. Using the projection operators technique, one can show that the general variational inequalities are equivalent to the general Wiener-Hopf equations. This equivalent alternative formulation has played a fundamental and important role in studying various aspects of variational inequalities. It has been shown that Wiener-Hopf equations are more flexible and provide a unified framework to develop some efficient and powerful numerical techniques for solving variational inequalities and related optimization problems.
Lemma 3. The element u ∈ H : g(u) ∈ K is a solution of the general variational inequality (9), if and only if z ∈ H satisfies the Wiener-Hopf equation (26), where where ρ > 0 is a constant.
From Lemma 3, it follows that the variational inequalities (9) and the Wiener-Hopf equations (26) are equivalent. This alternative equivalent formulation is used to suggest and analyze a wide class of efficient and robust iterative methods for solving general variational inequalities and related optimization problems, see [90,91,99,109] and the references therein. We use the general Wiener-Hopf equations (26) to suggest some new iterative methods for solving the general variational inequalities. From (27) and (28), Thus, we have Consequently, for a constant α > 0, we have where Using (29) and (30), we can suggest the following new predictor-corrector method for solving variational inequalities.
Algorithm 12 For a given u 0 ∈ H, compute u n+1 by the iterative scheme Algorithm 12 can be rewritten in the following equivalent form: Algorithm 13 For a given u 0 ∈ H, compute u n+1 by the iterative scheme which is an explicit iterative method and appears to be a new one. If α n = 1, then Algorithm 13 reduces to Algorithm 14 For a given u 0 ∈ H, compute u n+1 by the iterative scheme which appears to be a new one.
We now consider another Algorithm for solving general variational inequalities (9. We also include some computational experiments of these special cases. See [122,152,153] for further details.
Algorithm 15 For a given u 0 ∈ K, compute where η n = γ m n with m n being the smallest nonnegative integer satisfying To obtain a larger decrease of the distance from the next iterative point to the solution set, we consider the following optimization problem Following the technique of Wang et al [182], one can show that solution to the above optimization problem is just the root, denoted byᾱ n .
If we choose α n as step size in Algorithm 15, then we obtain another convergent algorithm. Obviously, α n guarantees that the distance between the new iterative point and the solution set has a larger decrease, so we call α n the basic step and α n the optimal step. However, in practice, if K does not possess any special structure, it is much expensive to compute α n . That is, we need to find a simple way to compute the projection P K [u n + α n d n ]. Following the proof of Lemma 4.2 in [188], we can show that u n (α n ) = P K∩H n [u n + α n d n ], where Thus, we can obtain our improved double-projection method for solving general variational inequalities.
Algorithm 16 For a given u 0 ∈ K, compute where η n = γ m n with m n being the smallest nonnegative integer satisfying Notice that at each iteration in Algorithm 16, the latter projection region is different from the former. More precisely, the latter projection region is an intersection of the domain set K and a hyperplane, so it does not increase computation cost if K is a polyhedral.
For g = I, we now give some numerical experiments for Algorithms 15 and 16 and some comparison with other double-projection methods. Throughout the computational experiments, the parameters used are set as σ = 0.5, γ = 0.8, and we use R(u n ) ≤ 10 −7 as stopping criteria. All computational results were undertaken on a PC-II by MATLAB. We use symbol e to denote the vector whose components are all ones.
Example 2. Consider the mapping T : R n → R n defined by with the domain set Example 3. This example was tested by Sun [173.174].
It is easy to see that T is strongly monotone on R n .
Again T is strongly monotone on K. The corresponding strongly monotonicity modulus depends on the dimension n and approaches zeros when n tends to infinity. Obviously, x = e is its unique solution. We choose the starting point u 0 = e for Example 2 and choose u = (0, · · · , 0) ⊤ as starting point for Examples 3 and 4 for different dimensions n. For double-projection methods [124,139,153 ], there always exist two step size rules just as in Algorithms 15 and 16. In the following, we give numerical comparison for these methods using two different steps. The numerical results for double-projection methods using the basic step for Examples 2, 3,4 are listed in Table 1, and the numerical results for double projection methods using the optimal step for Examples 2, 3,4 are listed in Table 2 (the symbol "\" denotes the number of iterations exceeds 1000 times).   43 192 Obviously, optimal step α n is better than the basic step α n for any direction. Compared with other double projection methods, Algorithm 16 also shows a better behavior. From Table 1 and Table 2, it is clear that our new methods are as efficient as the methods of Solodov and Svaiter [171,172,174]. This shows that our Algorithm 15 and Algorithm 16 can be considered as practical alternative to the extragradient and other modified projection methods. The comparison of new methods developed in this paper with the recent methods is an interesting problem for future research.

Dynamical Systems Technique
In this section, we consider the projected dynamical systems associated with variational inequalities. We investigate the convergence analysis of these new methods involving only the monotonicity of the operator. We now define the residue vector R(u) by the relation Invoking Lemma 3, one can easily conclude that u ∈ H : g(u) ∈ K is a solution of (9), if and only if, u ∈ H : g(u) ∈ K is a zero of the equation We now consider a projected dynamical system associated with the variational inequalities. Using the equivalent formulation (32), we suggest a class of projected dynamical systems as where λ is a parameter. The system of type (33) is called the projected dynamical system associated with variational inequalities (9). Here the right hand is related to the resolvent and is discontinuous on the boundary. From the definition, it is clear that the solution of the dynamical system always stays in H. This implies that the qualitative results such as the existence, uniqueness and continuous dependence of the solution of (33) can be studied. These projected dynamical systems are associated with the general variational inequalities (9), which have been studied extensively.
We use the projected dynamical system (33) to suggest some iterative for solving variational inequalities (9). These methods can be viewed in the sense of Koperlevich [60] and Noor [122] involving the double resolvent operator. For simplicity, we consider the dynamical system We construct the implicit iterative method using the forward difference scheme. Discretizing the equation (34), we have where h is the step size. Now, we can suggest the following implicit iterative method for solving the variational inequality (9).
Algorithm 17 For a given u 0 ∈ H, compute u n+1 by the iterative scheme This is an implicit method and is quite different from the known implicit method. Using Lemma 1, Algorithm 17 can be rewritten in the equivalent form as: Algorithm 18 For a given u 0 ∈ H, compute u n+1 by the iterative scheme We now study the convergence analysis of algorithm 18 under some mild conditions. Theorem 4. Let u ∈ H : g(v) ∈ K be a solution of the general variational inequality (9). Let u n+1 be the approximate solution obtained from ( 36). If T is g-monotone, then Proof. Let u ∈ H : g(v) ∈ K be a solution of (9). Then since T is a g-monotone operator.
Set v = u n+1 in (38), to have Taking v = u in ( 36), we have From (39) and (40), we have From (41) and using 2 the required result.
Theorem 5. Let u ∈ K be the solution of general variational inequality (9). Let u n+1 be the approximate solution obtained from (36). If T is a g-monotone operator and g −1 exists, then u n+1 converges to u ∈ H satisfying (9).
Proof. Let T be a g-monotone operator. Then, from (37), it follows the sequence is bounded, so there exists a cluster pointû to which the subsequence {u ik } ∞ k=1 converges. Taking limit in (36) and using (43), it follows that u ∈ K satisfies Using this inequality, one can show that the cluster pointû is unique and lim n→∞ u n+1 =û.
We now suggest another implicit iterative method for solving (9). Discretizing (34), we have where h is the step size. This formulation enable us to suggest the following iterative method.

Algorithm 19
For a given u 0 ∈ K, compute u n+1 by the iterative scheme Using lemma 1, Algorithm 19 can be rewritten in the equivalent form as: Algorithm 20 For a given u 0 ∈ K, compute u n+1 by the iterative scheme Again using the dynamical systems, we can suggested some iterative methods for solving the variational inequalities and related optimization problems.
Algorithm 21 For a given u 0 ∈ K, compute u n+1 by the iterative scheme which can be written in the equivalent form as: Algorithm 22 For a given u 0 ∈ K, compute u n+1 by the iterative scheme In a similar way, one can suggest a wide class of implicit iterative methods for solving variational inequalities and related optimization problems. the comparison of these methods with other methods is an interesting problem for future research.

Auxiliary Principle Technique
In the previous sections, we have considered and analyzed several projection-type methods for solving variational inequalities. It is well known that to implement such type of the methods, one has to evaluate the projection, which is itself a difficult problems. Secondly, one can't extend the technique of projection for solving some other classes of variational inequalities. These facts motivate us to consider other methods. One of these techniques is known as the auxiliary principle. This technique is basically due to Lions and Stampacchia [173]. Glowinski et. al [47] used this technique to study the existence of a solution of mixed variational inequalities.
Noor [93,94,95,114,121,122] has used this technique to develop some predictorcorrector methods for solving variational inequalities. It have been be shown that various classes of methods including projection, Wiener-Hopf, decomposition and descent can be obtained from this technique as special cases.
For a given u ∈ H, g(u) ∈ K satisfying (9), consider the problem of finding a unique w ∈ H, g(w) ∈ K such that where ρ > 0 is a constant.
Note that, if w = u, then w is clearly a solution of the general variational inequality (9). This simple observation enables us to suggest and analyze the following predictor-corrector method.
Algorithm 23 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative schemes where ρ > 0, β > 0 and µ > 0 are constants Algorithm 23 can be considered as a three-step predictor-corrector method, which was suggested and studied by Noor [110,122]. If µ = 0, then Algorithm 23 reduces to: Algorithm 24 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative schemes; which is known as the two-step predictor-corrector method, see [110,122]. If µ = 0, β = 0, then Algorithm 23 becomes: Algorithm 25 For a given u 0 ∈ H, compute u n+1 by the iterative scheme Using the projection technique, Algorithm 23 can be written as Algorithm 26 For a given u 0 ∈ H, compute u n+1 by the iterative schemes +ρT g −1 }g(u n ), n = 0, 1, 2, . . . , which is three-step forward-backward method. See also two-step forward-backward splitting method of Tseng [179,180] for solving the classical variational inequalities.
Definition 9. For all u, v, z ∈ H, the operator T : H −→ H is said to be: (i). g-partially relaxed strongly monotone, if there exists a constant α > 0 such that We remark that if z = u, then g-partially relaxed strongly monotonicity is equivalent to monotonicity. For g = I, definition 9 reduces to the standard definition of partially relaxed strongly monotonicity and cocoercivity of the operator. We now show that g-cocoercivity implies g-partially relaxed strongly monotonicity. This result is due to Noor [110,122]. To convey an idea, we include its proof.
which shows that T is a g-partially relaxed strongly monotone operator.
One can easily show that the converse is not true. Thus we conclude that the concept of g-partially relaxed strongly monotonicity is a weaker condition that gcocoercivity.
One can study the convergence criteria of Algorithm 23 using the technique of Noor [104].

Remark 2.
In the implementation of these algorithms, one does not have to evaluate the projection. Our method of convergence is very simple as compared with other methods. Following the technique of Tseng [179], one can obtained new parallel and decomposition algorithms for solving a number of problems arising in optimization and mathematical programming.

Remark 3.
We note that, if the operator g is linear or convex, then the auxiliary problem (9) is equivalent to finding the minimum of the functional I[w] on the convex set K, where It can be easily shown that the optimal solution of (48) is the projection of the point (g(u) − ρTu) onto the convex set K, that is, which is the fixed-point characterization of the general variational inequality (9).
Based on the above observations, one can show that the general variational inequality (9) is equivalent to finding the minimum of the functional where g(w) = g(w(u)). The function N[u] defined by (50) is known as the gap (merit) function associated with the general variational inequality (9). This equivalence has been used to suggest and analyze a number of methods for solving variational inequalities and nonlinear programming, see, for example, Patriksson [156].
In this direction, we have: Algorithm 27 For a given u 0 ∈ H, compute the sequence {u n } by the iterative scheme It is worth to note the sequence {u n } generated by is very much similar to that generated by the projection-type Algorithm 3. Based on the above observations and discussion, it is clear that the auxiliary principle approach is quite general and flexible. This approach can be used not only to study the existence theory but also to suggest and analyze various iterative methods for solving variational inequalities. Using the technique of Fukushima [42], one can easily study the convergence analysis of Algorithm 27.
We have shown that the auxiliary principle technique can be sued to construct gap (merit) functions for general variational inequalities (9). We use the gap function to consider an optimal control problem governed by the general variational inequalities (9). The control problem is an optimization problem, which is also referred as a generalized bilevel programming problem or mathematical programming with equilibrium constraints. It is known that the techniques of the classical optimal control problems cannot be extended for variational inequalities, see Dietrich [22]. This has motivated us to develop some other techniques including the notion of conical derivatives, the penalty method and formulating the variational inequality as operator equation with a set-valued operator. Furthermore, one can construct a so called gap function associated with a variational inequality, so that the variational inequality is equivalent to a scalar equation of the gap function. Under suitable conditions such a gap function is Frechet differentiable and one may use a penalty method to approximate the optimal control problem and calculate a regularized gap function in the sense of Fukushima [42] to the general variational inequality (9) and determine their Frechet derivative. Dietrich [32,33] has developed the similar results for the general variational inequalities. We only give the basic properties of the optimal control problem and the associated gap functions to give an idea of the approach.
We now consider the following problem of optimal control for the general variational inequalities (9), that is, to find u ∈ H : g(u) ∈ K, z ∈ U such that where H and U are Hilbert spaces. The sets K and E are closed convex sets in H and U respectively. Here H is the space of state and K ⊂ H is the set of state constraints for the problem. U is the space of control and closed convex set E ⊂ U is the set of control constraints. T (., .) : H ×U −→ H is a an operator which is Frechet differentiable. The functional I(., .) : H × U −→ R ∪ {+∞} is a proper, convex and lower-semicontinuous function. Also we assume that the problem P has at least one optimal solution denoted by (u * , z * ) ∈ H × U.
Related to the optimization problem (P), we consider the regularized gap (merit) We remark that the regularized function (51) is a natural generalization of the regularized gap function (50) for variational inequalities. It can be shown that the regularized gap function h ρ (., .) defined by (51) has the following properties. The analysis is in the spirit of Dietrich [33].
Theorem 6. The gap function h ρ (., .) defined by (51) is well-defined and where d K is the distance to K and Proof. It is well-known that Take v = u in (51). Then clearly (i) is satisfied.
Theorem 7. If the set K is g-convex in H, then the following are equivalent.
Let u ∈ H and z ∈ U be a solution of Then we have From these above inequalities, we have (i), that is, h ρ (u, z) = 0. Conversely, let (i) hold. Then Since K is a g-convex set, so for all g(w), the required (ii). Thus we conclude that (i) and (ii) are equivalent. Applying Lemma 1, we have (ii) = (iii).
From Theorem 6 and Theorem 7, we conclude that the optimization problem P is equivalent to where h ρ (u, z) is C 1 -differentiable in the sense of Frechet, but is not convex.
If the operators T, g are Frechet differentiable, then the gap function h ρ (u, z) defined by (51) is also Frechet differentiable. In fact, where [T ′ (u, z)] ′ is the adjoint operator of T ′ (u, z). This implies the following connection at a point (u 1 , z 1 ) which is a solution of the general variational inequality (9), that is, for For the optimal problem P, we have We now consider a simple example of optimal control problem to illustrate min(P 1 ) : = min First, we solve the general variational inequality (9) We obtain as the unique optimal solution of P 1 (u opt , z opt ) = ( 1 2 √ 2, 1 2 ) with min(P 1 ) = 3 4 . Next, we calculate the gap function of the general variational inequality problem (9).
This shows that equivalence between these problems.

Penalty Function Method
In this section, we consider a system of third-order boundary value problems, where the solution is required to satisfy some extra continuity conditions on the subintervals in addition to the usual boundary conditions. Such type of systems of boundary value arise in the study of obstacle, free, moving and unilateral problems and has important applications in various branches of pure and applied sciences. Despite of their importance, little attention has been given to develop efficient numerical methods for solving numerically these systems except for special cases. In particular, it is known that if the obstacle function is known then the general variational in-equalities can be characterized by a system of odd-order boundary value problems by using the penalty method. This technique is called the penalty function method and was used by Lewy and Stampacchia [64] to study the regularity of a solution of variational inequalities. The computational advantage of this technique is its simple applicability for solving the system of differential equations. This technique has been explored and developed by Noor and its co-worker to solve the systems of differential equations associated with even and odd-order obstacle problems. Our approach to these problems is to consider them in a general manner and specialize them later on. To convey an idea of the technique involved, we first introduce two numerical schemes for solving a system of third boundary value problems using the splines. An example involving the odd-order obstacle is given.
For simplicity, we consider a system of obstacle third-order boundary value problem of the type with the boundary conditions We usually resort to numerical methods for obtaining the approximate solutions of (53). Here we use cubic spline functions to derive some consistency relations which are then used to develop a numerical technique for solving system of third-order boundary value problems. Without loss of generality, we take c = 3a+b 4 and d = a+3b 4 in order to derive a numerical method for approximating the solution of the system (53). For this purpose, we divide the interval [a,b] into n equal subintervals using the grid point where n is a positive integer chosen such that both n+1 4 and 3(n+1) 4 also positive integer. Also, let u(x) be the exact solution of (53) and s i be an approximation to u i = u(x i ) obtained by the cubic P i (x) passing through the points (x i , s i ) and (x i+1 , s i+1 ). We write P i (x) in the form for i = 0, 1, 2, . . . , n − 1. Then the cubic spline is defined by We now develop explicit expressions for the four coefficients in (55). To do this, we first design and where Using the above discussion, we obtain the following relations Now from the continuity of the cubic spline s(x) and its derivatives up to order two at the point (x i , s i ) where the two cubic P i 1 (x) and P i (x) join, we can have From the above relations, one can easily obtain the following consistency relations From (61) and (62), we obtain Eliminating D i from (63). (62) and (61), we have for i = 2, 3, . . . , n − 1. The recurrence relations (64) gives (n − 2) linear equations in the unknowns s i , i = 1, 2, . . . , n. We need two more equations one at each end of the range of integration. These two equations are: The cubic spline solution of (55) is based on the linear equations given by (63)-(64). The local truncation errors T i , i = 1, 2, . . . associated with the cubic spline method are given by which indicates that this method is a second order convergent process.
To illustrate the applications of the numerical methods developed above, we consider the third order-order obstacle boundary value problem (19). Following the penalty function technique of Levy and Stampacchia [64], the variational inequality (65) can be written as (67) where ν{t} is the discontinuous function defined by is known as the penalty function and ψ < 0 on the boundary is the called the obstacle function. It is clear that problem (19) can be written in the form with where ν{t} is defined by (68). If the obstacle function ψ is known and is given by the relation then problem (19) is equivalent to the following system of third-order differential equations with the boundary conditions and the conditions of continuity of u, u ′ and u ′′ at x = 1 4 and 3 4 . It is obvious that problem (71) is a special case of problem (53) with p(x) = 1 and r = −1.
Note that for f = 0, the system of differential equations (71) reduces to with the boundary condition (72).
The analytical solution for this problem is To find the constants a i , i = 1, 2, . . . , 6, we apply the continuity conditions of u, u ′ and u ′′ at x = 1 4 and 3 4 , which leads to the following system of linear equations and SC 2 = sin 3 √ 3 8 .
One can find the exact solution of this system of linear equations by using Gaussian elimination method.
For various values of h, the system of third-order of boundary value problem defined by (71) and (72) was solved using the numerical method developed in this section. A detailed comparison is given in Table 3. From Table 3, it is clear that quartic spline method gives better results than cubic and quintic splines methods developed earlier for solving system of third-order boundary value systems.
For more details for solving various classes of obstacle boundary value problems using the penalty technique, see [3,4,5,114,150] and the references therein. In recent years, homotopy (analysis) perturbation method, Adomonian decomposition , Laplace transformation and variational iteration techniques are being used to find the analytical solutions of fractional unilateral and obstacle boundary value problems.

General Equilibrium Problems
In this section, we introduce and consider a class of equilibrium problems known as general equilibrium problems. It is known that equilibrium problems [19,142] include variational and complementarity problems as special cases. We note that the projection and its variant forms including the Wiener-Hopf equations cannot be extended to equilibrium problems, since it is not possible to find the projection of the bifunction F(., .). Noor [94,95,121] used the auxiliary principle technique to analyse some iterative methods for equilibrium problems. In this chapter, we introduce and study a class of equilibrium problems involving the arbitrary function, which is called the general equilibrium problem. We show that the auxiliary principle technique can be used to suggest and analyze some iterative methods for solving general equilibrium problems. We also study the convergence analysis of these iterative methods and discuss some special cases. For given nonlinear function F(., .) : H × H −→ R and operator g : H −→ R, we consider the problem of finding u ∈ H, g(u) ∈ K such that which is known as the general equilibrium problem. We now discuss some special cases of general equilibrium problem (75), (I). For g ≡ I, the identity operator, problem (75) is equivalent to finding u ∈ K such that which is called the equilibrium problem, which was introduced and studied by Blum and Oettli [19]. For the recent applications and development, see [102,109,153] and the reference therein.
(II). If F(u, g(v)) = Tu, η(g(v), g(u)) and the set K gη is an invex set in H, then problem (75) is equivalent to finding u ∈ K gη such that Inequality of type (77) is known as the general variational-like inequality, which arises as a minimum of general preinvex functions on the general invex set K gη .
(III). We note that, for F(u, g(v) ≡ Tu, g(v) − g(u) , problem (75) reduces to problem (9), that is, find u ∈ K, g(u) ∈ K such that which is exactly the general variational inequality (9). Thus we conclude that general equilibrium problems (75) are quite general and unifying one.
We now use the auxiliary principle technique as developed in Section 6 to suggest and analyze some iterative methods for solving general equilibrium problems (75). For a given u ∈ H, g(u) ∈ K satisfying (75), consider the auxiliary equilibrium problem of finding w ∈ H, g(w) ∈ K such that Obviously, if w = u, then w is a solution of the general equilibrium problem (75). This fact allows us to suggest the following iterative method for solving (75).
Algorithm 28 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative scheme: where ρ > 0 and β > 0 are constants.
Algorithm 28 is called the predictor-corrector method for solving general equilibrium problem (75). For g = I, where I is the identity operator, Algorithm 28 reduces to: Algorithm 29 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative schemes Algorithm 29 is also a predictor-corrector method for solving equilibrium problem and appears to be a new one.
If F(u, g(v)) = Tu, g(v) − g(u) , then Algorithm 28 becomes: Algorithm 30 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative scheme which is a two-step method for solving general variational inequalities (9).
In brief, for suitable and appropriate choice of the functions F(., .) and the operators T, g, one can obtain various algorithms developed in the previous sections.
For the convergence analysis of Algorithm 28, we need the following concepts.
Definition 10. The function F(., .) : H × H −→ H is said to be: (iii). g-partially relaxed strongly monotone, if there exists a constant α > 0 such that Note that for u = z, g-partially relaxed strongly monotonicity reduces to g-monotonicity of F(., .). For g = I, Definition 10 coincides with the standard definition of monotonicity, pseudomonotonicity of the function F(., .).
We now consider the convergence analysis of Algorithm 28.
Proof. Its proofs is very much similar to that of Noor [122].
We again use the auxiliary principle technique to suggest an inertial proximal method for solving general equilibrium problem (75). It is noted that inertial proximal method include the proximal method as a special case.
For a given u ∈ H, g(u) ∈ K satisfying (75), consider the auxiliary general equilibrium problem of finding w ∈ H, g(w) ∈ K such that (93) where ρ > 0 and α n > 0 are constants. It is clear that if w = u, then w is a solution of the general equilibrium problem (75). This fact enables us to suggest an iterative method for solving (75) as: Algorithm 31 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative scheme where ρ > 0 and α n > 0 are constants.
Algorithm 31 is called the inertial proximal point method. For α n = 0, Algorithm 31 reduces to: Algorithm 32 For a given u 0 ∈ H, find the approximate solution u n+1 by the iterative schemes which is known as the proximal method and appears to be a new one. Note that for g ≡ I, the identity operator, one can obtain inertial proximal method for solving equilibrium problems (76). In a similar way, using the technique of Noor [122], one can suggest and analyze several new inertial type methods for solving general equilibrium problems. It is an challenging problem to compare the efficiency of these methods with other techniques for solving general equilibrium problems.

General variational-like inequalities
It is well known that the minimum of the (non) differentiable preinvex functions on the invex set can be characterized by a class of variational inequalities, called variational-like inequalities. For the applications and numerical methods of variational-like inequalities, see [18,93,95,106] and the references therein. In this section, we introduce the general variational-like inequalities with respect to an arbitrary function. Due the structure of the general variational inequalities, the projection method and its variant forms cannot be used to study the problem of the existence of the solution. This implies that the variational-like inequalities are not equivalent to the projection (resolvent) fixed-point problems. We use the auxiliary principle technique to suggest and analyze some implicit and explicit iterative methods for solving variational-like inequalities. We also show that the general variationallike inequalities are equivalent to the optimization problems, which can be used to study the associated optimal control problem. Such type of the problems have been not studied for general variational-like inequalities and this is another direction for future research.
We recall some known basic concepts and results. Let F : K η → R be a continuous function and let η(., .) : K η × K η → R be an arbitrary continuous bifunction. Let g(.) be a non-negative function.

Definition 11. [14]
The set K η in H is said to be invex set with respect to an arbitrary bifunction η(·, ·), if The invex set K η is also called η-connected set. Note that the innvex set with η(v, u) = v − u is a convex set, but the converse is not true.
From now onward, K η is a nonempty closed invex set in H with respect to the bifunction η(·, ·), unless otherwise specified.

Definition 12. [14]
The set K gη in H is said to be general invex set with respect to an arbitrary bifunction η(·, ·) and the function g,if The invex set K gη is also called gη-connected set. Note that the general invex set with η(g(v), g(u)) = g(v) − g(u) is a general convex set, but the converse is not true. See Youness [197]. We now the concept of the general preinvex function.
Definition 13. . Let K gη ⊆ H be a general invex set with respect to η(., .) : K η × K η −→ R n and g : H −→ H. A function F : K gη −→ R is said to be general preinvex function, if, Note that for g ≡ I, the general preinvex functions are called the preinvex functions. For η(v, u) = g(v) − g(u), general preinvex function are known as general convex functions. Every convex function is a general convex function and every general convex function is a general preinvex function, but the converse is not true. see [11,43].
From now onward, we assume that the set K gη is a general invex set with respect to the functions η(., .) : K gη × K gη −→ H, g : K gη −→ H, unless otherwise specified.

Definition 14. The function F is said to be general semi preinvex, if
For g ≡ I, and t = 1, we have Definition 15. The function F is called general quasi preinvex, if The function F is called the strictly general quasi preinvex, if strict inequality holds for all g(u), g(v) ∈ K gη , g(u) = g(v). The function F is said to be general quasi preconcave, if and only if,−F is general quasi preinvex. A function which is both general quasi preinvex and general quasi preconcave is called the general quasimonotone Definition 16. The function F is said to be general logarithmic preinvex on the general invex set K gη with respect to the bifunction η(., .) and the function g, if where F(.) > 0.
Clearly for t = 1, and g = I, we have It follows that general logarithmic preinvexity =⇒ general preinvexity =⇒ general quasi preinvexity.
For appropriate and suitable choice of the operators and spaces, one can obtain several classes of generalized preinvexity.
In this section, we prove that the minimum of a differentiable general preinvex function on the general invex sets can be characterized by a class of variational-like inequalities, which is called the general variational-like inequality.. Theorem 10. Let F be a differentiable general preinvex. Then u ∈ H : g(u) ∈ K gη is a minimum of F on K η , if and only if, u ∈ H : g(u) ∈ K gη satisfies where F ′ is the Frechet derivative of F at g(u) ∈ K gη .
Dividing the above inequality by t and taking limit as t −→ 0, we have the required (94). Conversely, let u ∈ H : g(u) ∈ K gη satisfy the inequality (94). Then, using the fact that function F is a general preinvex function, we have Dividing the above inequality by t and letting t −→ 0, we have (94) which implies that which shows that u ∈ H : g(u) ∈ K gη is a minimum of the general preinvex function on the general invex set K gη in H.
Inequalities of the type (94) are called the general variational-like inequalities. For g ≡ I, where I is the identity operator, Theorem 10 is mainly due to Noor [93]. From Theorem 10 it follows that the general variational-like inequalities (94) arise naturally in connection with the minimum of general preinvex function over general invex sets. In many applications, problems like (94) do not arise as a result of minimization. This fact motivated us to consider a problem of finding a solution of a more general variational-like inequality of which (94) is a special case.
Given (nonlinear) operator T : H −→ H, and η : K gη × K gη −→ R, where K gη is a nonempty general invex set in H, we consider the problem of finding u ∈ H : g(u) ∈ K gη such that which is known as the general variational-like inequality.
, then the general invex set K gη becomes a general convex set K g . In this case, problem (96) is equivalent to finding u ∈ H : g(u) ∈ K g such that which is exactly the general variational inequality (9). For formulation, numerical methods, sensitivity analysis, dynamical system and other aspects of general variational inequalities, see [109,110,122,144] and the references therein.
For suitable and appropriate choice of the operators T, η, and the general invex set, one may derive a wide class of known and new variational inequalities as special cases of problem (96). It is clear that general variational-like inequalities provide us a framework to study a wide class of unrelated problems in a unified setting.
We now use the auxiliary principle technique to suggest and analyze some iterative methods for general variational-like inequalities (96).
For a given u ∈ H; g(u) ∈ K gη satisfying (96), consider the problem of finding a solution w ∈ H : g(w) ∈ K gη satisfying the auxiliary variational-like inequality (97) where ρ > 0 is a constant and E ′ is the differential of a strongly general preinvex function E. The inequality (97) is called the auxiliary general variational-like inequality.
Note that if w = u, then clearly w is solution of the general variational-like inequality (96). This observation enables us to suggest the following algorithm for solving (96).
Algorithm 33 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative scheme Algorithm 33 is called the proximal point algorithm for solving the general variationallike inequalities (96).
, general preinvex function E is equivalent to the convex function and the invex set K gη becomes the general convex set. Consequently Algorithm 33 reduces to: Algorithm 34 For a given u 0 ∈ K g , compute the approximate solution u n+1 by the iterative scheme which is known as the proximal point algorithm for solving general variational inequalities.

Remark 4. The function
associated with the preinvex functions E(u) is called the general Bregman function. We note that,if η(g(v), g(u)) = g(v) − g(u), then is the well known Bregman function. For the applications of Bregman function in solving variational inequalities and related optimization problems, see [201].
We now study the convergence analysis of Algorithm 33. For this purpose, we recall the following concepts.
Definition 17. ∀u, v, z ∈ H, an operator T : H −→ H is said with respect to an arbitrary function g : H −→ H to be: (ii). general gη-Lipschitz continuous, if there exists a constant β > 0 such that (iii). general gη-cocoercive, if there exists a constant µ > 0 such that (iv). general gη-partially relaxed strongly monotone, if there exists a constant α > 0 such that For η(g(v), g(u)) = g(v) − g(u), definition 17 reduces to the definition of general monotonicity, general Lipschitz continuity, general co-coercivity and partially relaxed general strongly monotonicity of the operator T. We note that for g(z) = g(u), partially strongly monotonicity reduces to monotonicity. One can easily show that general gη-cocoercivity implies general gη-partially relaxed general strongly monotonicity, but the converse is not true.

Definition 18.
A function F is said to be strongly general preinvex function on K gη with respect to the function η(., .) with modulus µ > 0 and function g, if, ∀u, v ∈ H : g(u), g(v) ∈ K gη ,t ∈ [0, 1], such that We note that the differentiable strongly general preinvex function F implies the strongly general invex function, that is, but the converse is also true under some conditions. Assumption 1 ∀u, v, z ∈ H, the operator η : H × H −→ H and the function g satisfy the condition η(g(u), g(v)) = η(g(u), g(z)) + η(g(z), g(v)).
In particular, from Assumption 1, we obtain Assumption 1 has been used to study the existence of a solution of general variationallike inequalities.
Theorem 11. Let T be a general η-pseudomonotone operator. Let E be a strongly differentiable general preinvex function with modulus β and Assumption 1 hold. Then the approximate solution u n+1 obtained from Algorithm 33 converges to a solution of (96).
To implement the proximal method, one has to calculate the solution implicitly, which is in itself a difficult problem. We again use the auxiliary principle technique to suggest another iterative method, the convergence of which requires only the gηpartially relaxed strongly general monotonicity.
For a given u ∈ H : g(u) ∈ K gη , satisfying (96), find a solution w ∈ H : g(w) ∈ K gη such that (104) which is called the auxiliary general variational-like inequality, where E(u) is a differentiable strongly general preinvex function. It is clear that, if w = u, then w is a solution of the general variational-like inequality (96). This fact allows us to suggest and analyze the following iterative method for solving (96).

Algorithm 35 For a given u 0 ∈ H, compute the approximate solution u n+1 by the iterative scheme
Note that for η(g(v), g(u)) = g(v) − g(u), Algorithm 35 reduces to: Algorithm 36 For a given u 0 ∈ H g , find the approximate solution u n+1 by the iterative scheme Algorithm 4 for solving general variational inequalities appears to be a new one. In a similar way, one can obtain a number of new and known iterative methods for solving various classes of variational inequalities and complementarity problems.
We now study the convergence analysis of Algorithm 35. The analysis is in the spirit of Theorem 11. We only give the main points.
Theorem 12. Let T be a partially relaxed strongly general gη-monotone with a constant α > 0. Let E be a differentiable strongly general preinvex function with modulus β and Assumption 1 hold. If 0 < ρ < β α, then the approximate solution u n+1 obtained from Algorithm 35 converges to a solution of (96).
Proof. Since the function E is strongly general preinvex , so the solution u n+1 of (96) is unique. Let u ∈ H : g(u) ∈ K gη be a solution of the general variational-like inequality (96). Then Taking v = u n+1 in the above inequality, we have ρTu, η(g(u n+1 ), g(u)) ≥ 0.
We now show that the solution of the auxiliary general variational-like inequality (104) is the minimum of the functional I[g(w)] on the general invex set K gη , where is known as the auxiliary energy functional associated with the auxiliary general variational-like inequality (104), where B(g(w), g(u)) is a general Bregman function. We now prove that the minimum of the functional I[w], defined by (107), can be characterized by the general variational-like inequality (104). (96).
Dividing both sides by t and letting t → 0, we have the required inequality (104).

Thus it follows that I[g(w)] ≤ I[g(v)]
, showing that g(v) ∈ K η is the minimum of the functional I[g(w)] on K η , the required result.

Higher order strongly general convex functions
We would like to point out that the strongly convex functions were introduced and studied by Polyak [159], which play an important part in the optimization theory and related areas. For example, Karmardian [57] used the strongly convex functions to discuss the unique existence of a solution of the nonlinear complementarity problems. Strongly convex functions also played important role in the convergence analysis of the iterative methods for solving variational inequalities and equilibrium problems, see Zu and Marcotte [201]. Lin and Fukushima [65] introduced the concept of higher order strongly convex functions and used it in the study of mathematical program with equilibrium constraints. These mathematical programs with equilibrium constraints are defined by a parametric variational inequality or com-plementarity system and play an important role in many fields such as engineering design, economic equilibrium and multilevel game. These facts and observations inspired Mohsen et al [75]to consider higher order strongly convex function involving an arbitrary bifunction. Noor and Noor [139,140] have introduced the higher order strongly general convex functions, which include the higher order strongly convex functions [65,75] as special cases.
In this chapter, we introduce concept of higher order strongly general convex functions. Several new concepts of monotonicity are introduced. Our results represent the refinement and improvement of the results of Lin and Fukushima [65]. Higher order strongly general convex functions are used to obtain new characterizations of the uniformly reflex Banach spaces by the parallelogram laws. It is worth mentioning that the parallelogram laws have been discussed in [21,22,23,24,190].
We now define the concept of higher order strongly general convex functions, which have been investigated in [139,140].

Definition 19.
A function F on the convex set K is said to be higher order strongly general convex with respect to the function g, if there exists a constant µ > 0 such that A function F is said to higher order strongly general concave, if and only if, −F is higher order strongly general convex.
The function F is said to be higher order strongly general J-convex function.
We now discuss some special cases.
I. If p = 2, then the higher order strongly convex function becomes strongly convex functions, that is, For the properties of the strongly convex functions in variational inequalities and equilibrium problems, see Noor [95,122,129].
II. If g = I, then Definition 19 reduces to Definition 20. A function F on the convex set K is said to be higher order strongly convex, if there exists a constant µ > 0, such that which appears to be a new one.
For appropriate and suitable choice of the unction g and p, one can obtain various new and known classes of strongly convex functions. This show that the higher order strongly convex functions involving the function g is quite general and unifying one. One can explore the applications of the higher order strongly general convex functions, this is another direction of further research.

Definition 21.
A function F on the convex set K is said to be a higher order strongly affine general convex with respect to the function g, if there exists a constant µ > 0, such that Note that if a functions is both higher order strongly convex and higher order strongly concave, then it is higher order strongly affine convex function.

Definition 22.
A function F is called higher order strongly quadratic equation with respect to the function g, if there exists a constant µ > 0, such that This function F is also called higher order strongly affine general J-convex function.
Definition 23. A function F on the convex set K is said to be higher order strongly quasi convex, if there exists a constant µ > 0 such that

Definition 24.
A function F on the convex set K is said to be higher order strongly log-convex, if there exists a constant µ > 0 such that where F(·) > 0.
From the above definitions, we have This shows that every higher order strongly general log-convex function is a higher order strongly general convex function and every higher order strongly general convex function is a higher order strongly general quasi-convex function. However, the converse is not true. For appropriate and suitable choice of the arbitrary bifunction g, one can obtain several new and known classes of strongly convex functions and their variant forms as special cases of generalized strongly convex functions. This shows that the class of higher order strongly general convex functions is quite broad and unifying one.
Definition 25. An operator T : K → H is said to be: (i). higher order strongly monotone, if and only if, there exists a constant α > 0 such that (ii). higher order strongly pseudomonotone, if and only if, there exists a constant ν > 0 such that (iii). higher order strongly relaxed pseudomonotone, if and only if, there exists a constant µ > 0 such that Definition 26. A differentiable function F on the convex set K g is said to be higher order strongly pseudoconvex function, if and only if, if there exists a constant µ > 0 such that We now consider some basic properties of higher order strongly general convex functions.
Theorem 14. Let F be a differentiable function on the convex set K g . Then the function F is higher order strongly general convex function, if and only if, Proof. Let F be a higher order strongly general convex function on the convex set K g . Then which can be written as Taking the limit in the above inequality as t → 0,, we have which is (112), the required result.
Conversely, let (112) hold. Then, ∀g(u), g(v) ∈ K g ,t ∈ [0, 1], g(v t ) = g()u + t(g(v) − g(u)) ∈ K g , we have In a similar way, we have Multiplying (113) by t and (114) by (1 − t) and adding the resultants, we have showing that F is a higher order strongly general convex function.
Theorem 15. Let F be a differentiable higher order strongly convex function on the convex set K g . Then Proof. Let F be a higher order strongly general convex function on the convex set K g . Then, from Theorem 14. we have Changing the role of u and v in (116), we have Adding (116) and (117), we have which shows that F ′ (.) is a higher order strongly general monotone operator.
We remark that the converse of Theorem 15 is not true. In this direction, we have the following result.
Theorem 16. If the differential operator F ′ (.) of a differentiable higher order strongly general convex function F is higher order strongly monotone operator, then Proof. Let F ′ be a higher order strongly monotone operator. Then, from (118), we have Since K is an convex set, ∀g(u), g(v) ∈ K g , t ∈ [0, 1], g(v t ) = g(u) + t(g(v) − g(u)) ∈ K g . Taking g(v) = g(v t ) in (120), we have which implies that Consider the auxiliary function from which, we have F(g(u)).
Then, from (122), we have Integrating (123) between 0 and 1, we have Thus it follows that which is the required (119).
We note that, if p = 2, then Theorem 16 can be viewed as the converse of Theorem 15.
We now give a necessary condition for higher order strongly general pseudoconvex function.
Theorem 17. Let F ′ (.) be a higher order strongly relaxed pseudomonotone operator. Then F is a higher order strongly pseudo-convex function.

Proof.
Let F ′ (.) be a higher order strongly relaxed general pseudomonotone operator. Then, from(112), we have Since K g is an convex set, ∀g(u), Consider the auxiliary function which is differentiable, since F is differentiable function. Then, using (126), we have Integrating the above relation between 0 to 1, we have that is, showing that F is a higher order strongly general pseudoconvex function.

Definition 27.
A function F is said to be sharply higher order strongly general pseudoconvex, if there exists a constant µ > 0 such that Theorem 18. Let F be a sharply higher order strongly general pseudoconvex function on the general convex set K g with a constant µ > 0. Then

Proof.
Let F be a sharply higher order strongly general pseudoconvex function on the general convex set K g . Then from which, we have Taking limit in the above inequality, as t → 0, we have the required result.

Definition 28.
A function F is said to be a pseudoconvex function with respect to a strictly positive bifunction B(., .), such that Theorem 19. If the function F is higher order strongly convex function such that F(g(v)) < F(g(u)), then the function F is higher order strongly pseudoconvex.

Proof.
Since F(g(v)) < F(g(u)) and F is higher order strongly convex function, then ∀g(u), g(v) ∈ K g , t ∈ [0, 1], we have We now discuss the optimality for the differentiable strongly general convex functions, which is the main motivation of our next result.
Theorem 20. Let F be a differentiable higher order strongly general convex function with modulus µ > 0. If u ∈ H : g(u) ∈ K g is the minimum of the function F, then Proof.
Let u ∈ H : g(u) ∈ K g be a minimum of the function F. Then Since K is a general convex set, so, ∀g(u), Taking Since F is differentiable higher order strongly general convex function, so from which, using(129), we have the required result (127).

Remark:
We would like to mention that, if u ∈ H : g(u) ∈ K g satisfies the inequality then u ∈ K g is the minimum of the function F. The inequality of the type (130) is called the higher order general variational inequality.
Theorem 21. Let f be a higher order strongly affine general convex function. Then F is a higher order strongly general convex function, if and only if, H = F − f is a general convex function. Proof.
Let f be a higher order strongly affine general convex function, Then From the higher order strongly general convexity of F, we have From (131 ) and (132), we have from which it follows that which show that H = F − f is a convex function.The inverse implication is obvious.
It is worth mentioning that the higher order strongly convex function is also higher order strongly Wright general convex function. From the definition, we have which is called the higher order strongly Wright general convex function. One can studies the properties and applications of the Wright higher order strongly convex functions in optimization operations research.
Bynum [21] and Chen et al [22,23,24,25] have studied the properties and applications of the parallelogram laws for the Banach spaces. Xi [190] obtained new characteristics of p-uniform convexity and q-uniform smoothness of a Banach space via the functionals . p and . q , respectively. These results can be obtained from the concepts of higher order strongly general convex(concave) functions, which can be viewed as novel application. Setting F(u) = u p in Definition 21, we have Taking t = 1 2 in (134), we have which implies that which is known as the lower parallelogram for the l p -spaces. In a similar way, one can obtain the upper parallelogram law as Similarly from Definition 2.3, we can have which is known as the parallelogram for the l p -spaces. For the applications of the parallelogram laws for Banach spaces in prediction theory and applied sciences, see [21,22,23,24] and the references therein.
In this section, we have introduced and studied a new class of convex functions, which is called higher order strongly convex function. We have the improved the results of Lin and Fukushima [65]. It is shown that several new classes of strongly convex functions can be obtained as special cases of these higher order strongly general convex functions. We have studied the basic properties of these functions. We have shown that one can derive the parallelogram laws in Banach spaces, which have applications in prediction theory and stochastic analysis. These parallelogram laws can be used to characterize the p-uniform convexity and q-uniform smoothness of a Banach spaces. The interested readers may explore the applications and other properties for the higher order strongly convex functions in various fields of pure and applied sciences. This is an interesting direction of future research.

Higher order general variational inequalities
In this section, we consider a more general variational inequality of which (130) is a special case.
For given two operators T, g, we consider the problem of finding u ∈ K for a constant µ > 0, such that which is called the higher order general variational inequality, see [140]. We note that, if µ = 0, then (139) is equivalent to finding u ∈ K, such that which is known as the general variational inequality (9), which was introduced and studied by Noor [87] in 2008. For suitable and appropriate choice of the parameter µ and p, one can obtain several new and known classes of variational inequalities, see [87,88,90,110,122,123] and the references therein. We note that the projection method and its variant forms can be used to study the higher order strongly general variational inequalities (139) due to its inherent structure. These facts motivated us to consider the auxiliary principle technique, which is mainly due to Glowinski et al [47] and Lions and Stampacchia [66] as developed by Noor [122]. We use this technique to suggest some iterative methods for solving the higher order general variational inequalities (139). For given u ∈ K satisfying (139), consider the problem of finding w ∈ K, such that where ρ > 0 is a parameter. The problem (141) is called the auxiliary higher order strongly general variational inequality. It is clear that the relation (141) defines a mapping connecting the problems (139) and (141). We not that, if w = u, then w is a solution of problem (139). This simple observation enables to suggest an iterative method for solving (139).
Algorithm 37 . For given u ∈ K, find the approximate solution u n+1 by the scheme The Algorithm 37 is known as an implicit method. Such type of methods have been studied extensively for various classes of variational inequalities. See [11,18,19] and the reference therein. If ν = 0, then Algorithm 37 reduces to: Algorithm 38 For given u 0 ∈ K, find the approximate solution u n+1 by the scheme which appears to new ones even for solving the general variational inequalities (9). To study the convergence analysis of Algorithm 37, we need the following concept.
Definition 29. The operator T is said to be pseudo g-monotone with respect to µ g If µ = 0, then Definition 29 reduces to: Definition 30. The operator T is said to be pseudo g-monotone, if which appears to be a new one. We now study the convergence analysis of Algorithm 37.
Theorem 22. Let u ∈ K be a solution of (139) and u n+1 be the approximate solution obtained from Algorithm 37. If T is a pseudo g-monotone operator, then Proof. Let u ∈ K be a solution of (139), then Now taking v = u n+1 in (144), we have Taking v = u in (139), we have Combining (145) and (146), we have Using the inequality the required result (143).
Theorem 23. Let the operator T be a pseudo g-monotone. If u n+1 be the approximate solution obtained from Algorithm 37 and u ∈ K is the exact solution (139), then lim n→∞ u n = u.
Proof. Let u ∈ K be a solution of (139). Then, from (143), it follows that the sequence { u − u n } is nonincreasing and consequently {u n } is bounded. From (143), we have from which, it follows that lim n→∞ u n+1 − u n = 0.
Letû be a cluster point of {u n } and the subsequence {u n j } of the sequence u n converge toû ∈ H. Replacing u n by u n j in (138), taking the limit n j → 0 and from (147), we have This implies thatû ∈ K K satisfies and Thus it follows from the above inequality that the sequence u n has exactly one cluster pointû and lim n→∞ (u n ) =û.
In order to implement the implicit Algorithm 37, one uses the predictor-corrector technique. Consequently, Algorithm 37 is equivalent to the following iterative method for solving the higher order strongly general variational inequality (139).
Algorithm 39 For a given u 0 ∈ K, find the approximate solution u n+1 by the schemes Algorithm 39 is called the two-step iterative method and appears to be a new one.
Using the auxiliary principle technique, on can suggest several iterative methods for solving the higher order strongly general variational inequalities and related optimization problems. We have only given a glimpse of the higher order strongly general variational inequalities. It is an interesting problem to explore the applications of such type variational inequalities in various fields of pure and applied sciences.

Strongly exponentially general convex functions
Convexity theory describes a broad spectrum of very interesting developments involving a link among various fields of mathematics, physics, economics and engineering sciences. The development of convexity theory can be viewed as the simultaneous pursuit of two different lines of research. On the one hand, it is related to integral inequalities. It has been shown that a function is a convex function, if and only if, it satisfies the Hermite-Hadamard type inequality. These inequalities help us to derive the upper and lower bounds of the integrals. On the other hand, the minimum of the differentiable convex functions on the convex set can be characterized by the variational inequalities. the origin of which can be traced back to Bernoullis brothers, Euler and Lagrange. Variational inequalities provide us a powerful tool to discuss the behaviour of solutions (regarding its existence, uniqueness and regularity) to important classes of problems. Variational inequality theory also enables us to develop highly efficient powerful new numerical methods to solve nonlinear problems. Recently various extensions and generalizations of convex functions and convex sets have been considered and studied using innovative ideas and techniques. It is known that more accurate and inequalities can be obtained using the logarithmically convex functions than the convex functions. Closely related to the log-convex functions, we have the concept of exponentially convex(concave) functions, which have important applications in information theory, big data analysis, machine learning and statistic. Exponentially convex functions have appeared significant generalization of the convex functions, the origin of which can be traced back to Bernstein [16]. Avriel [9,10] introduced the concept of r-convex functions, from which one can deduce the exponentially convex functions. Antczak [2] considered the (r, p) convex functions and discussed their applications in mathematical programming and optimization theory. It is worth mentioning that the exponentially convex functions have important applications in information sciences, data mining and statistics, see, for example, [1,2,9,10,132,133,134,135,136,137,138,154] and the references therein.
We would like to point out that the general convex functions and exponentially general convex functions are two distinct generalizations of the convex functions, which have played a crucial and significant role in the development of various branches of pure and applied sciences. It is natural to unify these concepts. Motivated by these facts and observations, we now introduce a new class of convex functions, which is called exponentially general convex functions in involving an arbitrary function. We discuss the basic properties of the exponentially general convex functions. It is has been shown that the exponentially general convex(concave) have nice properties which convex functions enjoy. Several new concepts have been introduced and investigated. We prove that the local minimum of the exponentially general convex functions is also the global minimum. Noor and Noor [132,133,134,135,136,137,138,154] studied and investigated some classes of strongly exponentially convex functions. Inspired by the work of Noor and Noor [138], we introduce and consider some new classes of higher order strongly exponentially convex functions. We establish the relationship between these classes and derive some new results. We have also investigated the optimality conditions for the higher order strongly exponentially convex functions. It is shown that the difference of strongly exponentially convex functions and strongly exponentially affine functions is again an exponentially convex function. The optimal conditions of the differentiable exponentially convex functions can be characterized by a class of variational inequalities, called the exponentially general variational inequality, which is itself an interesting problem.
We now define the exponentially convex functions, which are mainly due to Noor and Noor [132,133,134,135,136,137,138,154].
Definition 31. [132,133,134,138] A function F is said to be exponentially convex function, if It is worth mentioning that Avriel [9,10] and Antczak [2] introduced the following concept.
Definition 32. [3,4] A function F is said to be exponentially convex function, if Avriel [9,10] and Antczak [2] discussed the application of the 1-convex functions in the mathematical programming. We note that the Definitions 7 and 8 are equivalent. A function is called the exponentially concave function f , if − f is exponentially convex function.
For the applications of the exponentially convex(concave) functions in the mathematical programming and information theory. For application in communication theory and information theory, see Alirezaei and Mathar [1].

Example 5. [1] The error function
becomes an exponentially concave function in the form er f ( √ x), x ≥ 0, which describes the bit/symbol error probability of communication systems depending on the square root of the underlying signal-to-noise ratio. This shows that the exponentially concave functions can play important part in communication theory and information theory.
For the properties, generalizations and applications of the various classes of exponentially convex functions, see [1,2,9,10,132,133,134,135,136,137,138,154] It is clear that the exponentially convex functions and general convex functions are two distinct generalizations of the convex functions. It is natural to unify these concepts. Motivated by this fact, we Noor and Noor [138] introduced some new concepts of exponentially general convex functions. We include these results for the sake of completeness and for the convenience of the interested readers.
Definition 33. A function F is said to be exponentially strongly general convex function with respect to an arbitrary non-negative function g, if or equivalently Definition 34. A function F is said to be exponentially general convex function with respect to an arbitrary non-negative function g, if, A function is called the exponentially general concave function f , if − f is exponentially general convex function.

Definition 35.
A function F is said to be exponentially affine general convex function with respect to an arbitrary non-negative function g, if If g = I, the identity operator, then exponentially general convex functions reduce to the exponentially convex functions.
From the above definitions, we have g(v))) }.
This shows that every exponentially general log-convex function is a exponentially general convex function and every exponentially general convex function is a exponentially general quasi-convex function. However, the converse is not true. Let K g = I g = [g(a), g(b)] be the interval. We now define the exponentially general convex functions on I g . g(b)]. Then F is exponentially general convex function, if and only if, One can easily show that the following are equivalent: 1. F is exponentially general convex function.
Theorem 24. Let F be a strictly exponentially general convex function. Then, any local minimum of F is a global minimum.
Proof. Let the strictly exponentially convex function F have a local minimum at g(u) ∈ K g . Assume the contrary, that is, F(g(v)) < F(g(u)) for some g(v) ∈ K g . Since F is strictly exponentially general convex function, so Thus for arbitrary small t > 0, contradicting the local minimum.
Theorem 25. If the function F on the general convex set K g is exponentially general convex, then the level set L α = {g(u) ∈ K g : e F(g(u)) ≤ α, α ∈ R} is a general convex set.

Theorem 26. The function F is exponentially general convex function, if and only if,
is a general convex set.
Proof. Assume thatF is exponentially general convex function. Let Then it follows that e F(g(u)) ≤ α and e F(g(v)) ≤ β . Hence, we have Thus epi(F) is a general convex set. Conversely, let epi(F) be a general convex set. Let g(u), g(v) ∈ K g . Then (g(u), e F (g(u) ) ∈ epi(F) and (g(v, e F(g(v)) ) ∈ epi(F). Since epi(F) is a general convex set, we must have which implies that This shows that F is an exponentially general convex function.

Theorem 27. The function F is exponentially general quasi-convex, if and only if, the level set
is a general convex set.
Then, from the definition of the level set L α , it follows that e F(g(u)+t(g(v)−g(u)) ≤ max (e F(g(u)) , e F(g(v)) ) ≤ α.
Thus F is an exponentially general quasi convex function. This completes the proof.
Theorem 28. Let F be an exponentially general convex function. Let µ = inf u∈K F(u). Then the set E = {g(u) ∈ K g : e F(g(u)) = µ} is a general convex set of K g . If F is strictly exponentially general convex function , then E is a singleton.
). Since F is a exponentially general convex function, then which implies g(w) ∈ E. and hence E is a general convex set. For the second part, assume to the contrary that F(g(u)) = F(g(v)) = µ. Since K is a general convex set, then for 0 < t < 1, g(u) + t(g(v) − g(u)) ∈ K g . Since F is strictly exponentially general convex function, so This contradicts the fact that µ = inf g(u)∈Kg F(u) and hence the result follows.
We now introduce the concept of the strongly exponentially general convex functions, which is the main motivation of this chapter.
Definition 39. A positive function F on the general convex set K g is said to be a strongly exponentially general convex with respect to an arbitrary non-negative function g, if there exists a constant µ > 0, such that The function F is said to be strongly exponentially general concave with respect to an arbitrary non-negative function g, if and only if, −F is strongly exponentially general convex function with respect to an arbitrary non-negative function g.
If t = 1 2 and µ = 1, then The function F is called the strongly exponentially general J-convex function with respect to an arbitrary non-negative function g.

Definition 40.
A positive function is said to be strongly exponentially affine general convex with respect to an arbitrary non-negative function g, if there exists a constant µ > 0, such that then we say that the function F is strongly exponentially affine general J-convex function with respect to an arbitrary non-negative function g, For the properties of the strongly exponentially general convex functions in optimization, inequalities and equilibrium problems, see [4,5,6,7,8,10,11,12,13,14,15,16,17,18,19,20,21] and the references therein..

Definition 41.
A positive function F on the convex set K g is said to be strongly exponentially general quasi-convex, if there exists a constant µ > 0 such that Definition 42. A positive function F on the general convex set K g is said to be strongly exponentially general log-convex, if there exists a constant µ > 0 such that where F(·) > 0.
From this Definition, we have e F(g(u)+t(g(v)−g(u))) ≤ e (F(g(u))) 1−t e (F(g(v))) t − µ{t(1 − t)} g(v) − g(u) 2 , This shows that every strongly exponentially general log-convex function is a strongly exponentially general convex function, but the converse is not true.
From above concepts, we have e F(g(u)+t(g(v)−g(u))) ≤ (e (F(g(u))) ) 1−t (e (F(g(v))) ) t − µ{t( This shows that every strongly exponentially general log-convex function is a strongly exponentially convex function and every strongly exponentially general convex function is a strongly exponentially general quasi-convex function. However, the converse is not true.

Definition 43.
A differentiable function F on the convex set K g is said to be a strongly exponentially general pseudoconvex function with respect to an arbitrary non-negative function g, if and only if, if there exists a constant µ > 0 such that e F(g(u)) F ′ (g(u)), g(v) − g(u) + µ g(v) − g(u) 2 ≥ 0 ⇒ e F(g(v)) − e F(g(u)) ≥ 0, ∀g(u), g(v) ∈ K g .
Multiplying (148) by t and (150) by (1 − t) and adding the resultant, we have showing that F is a strongly exponentially general convex function.
We point out the converse of Theorem 30 is not true expect for p = 2. In fact, we have the following result.
Theorem 29 and Theorem 30 enable us to introduce the following new concepts.
We now give a necessary condition for strongly exponentially pseudo-convex function.
Theorem 32. Let F ′ be a strongly exponentially pseudomonotone operator. Then F is a strongly exponentially general pseudo-invex function.
Integrating the above relation between 0 to 1, we have that is, showing that F is a strongly exponentially general pseudo-convex function.
We now discuss the optimality condition for the differentiable strongly exponentially convex functions.
Theorem 34. Let F be a differentiable strongly exponentially convex function. If u ∈ K is the minimum of the function F, then e F(g(v)) − e F(g(u)) ≥ µ g(v) − g(u) p , ∀g(u), g(v) ∈ K g .

Remark 5.
We would like to mention that, if u ∈ H : g(u) ∈ K g satisfies e F(g(u)) F ′ (g(u)), g(v) − g(u) + µ g(v) − g(u) 2 ≥ 0, ∀g(u), g(v) ∈ K g , then u ∈ H : g(u) ∈ K g is the minimum of the function F. The inequality of the type (165) is called the strongly exponentially variational inequality. It is an interesting problem to study the existence of the inequality (165) and to develop numerical methods for solving the strongly exponentially variational inequalities.
We would like to point that the strongly exponentially convex is also a strongly Wright general convex functions. From the definition 39, we have e F((1−t)g(u)+tg(v)) + e F(tg(u)+(1−t)g(v)) ≤ e F (g(u)) + e F (g(v)) −2µt(1 − t) g(v) − g(u) 2 , ∀g(u), g(v) ∈ K g ,t ∈ [0, 1], which is called the strongly Wright exponentially convex function. It is an interesting problem to study the properties and applications of the strongly Wright exponentially general convex functions.

Generalizations and extensions
We would like to mention that some of the results obtained and presented in this paper can be extended for more strongly general variational inequalities. To be more precise, for a given nonlinear operator T, A, g, consider the problem of finding u ∈ H : g(u) ∈ K such that which is called the strongly general variational inequality. If µ = 0, then problem (166) reduces to is called the general strongly variational inequalities.
We would like to mention that one can obtain various classes of general variational inequalities for appropriate and suitable choices of the operators T, A, g.
Using Lemma 1, one can show that the problem (167) is equivalent to finding u ∈ H : g(u) ∈ K such that These alternative formulations can be used to suggest and analyze similar techniques for solving general strongly variational inequalities (167) as considered in this paper under certain extra conditions. A complete study of these algorithms for problem (167) is the subject of subsequent research. Development of efficient and implementable algorithms for problems (167) need further research efforts.
(I). For given nonlinear operators T, A, g, consider the problem of finding u ∈ H : g(u) ∈ K such that which is also called the strongly general variational inequality.
(III). For given nonlinear operators T, A, g, consider the problem of finding u ∈ H : g(u) ∈ K such that is also known as the strongly general variational inequality.

Remark 6.
We would like to point out that the problems (167), (169) and (170) are quite distinct and different from each other and have significant applications in various branches of pure and applied sciences. They are open and interesting problems for future research. We would like to emphasize that problems (167), (169) and (170) are equivalent in many respect and share the basic and fundamental properties. In particular, they have the same equivalent fixed-point formulations.
Consequently, most of the results obtained in this paper continue to hold for these problems with minor modifications.
(IV). If K = H, then the problem (166) is equivalent to finding u ∈ H : g(u) ∈ H, such that Tu, g(v) = A(u), g(v) , ∀v ∈ H : g(v) ∈ H, (171) which can be viewed as the representation theorem for the nonlinear functions involving an arbitrary function g. For more details, see Noor and Noor [141]. (V). If A(u) = |u|, then the problem (171) is equivalent to finding u ∈ H : g(u) ∈ H, such that is known as the generalized absolute value equation. See Batool et al [13] more details.
The theory of general variational inequalities does not appear to have developed to an extent that it provides a complete framework for studying these problems. Much more research is needed in all these areas to develop a sound basis for applications. We have not treated variational inequalities for time-dependent problems and the spectrum analysis of the variational inequalities. In fact, this field has been continuing and will continue to foster new, innovative and novel applications in various branches of pure and applied sciences. We have given only a brief introduction of this fast growing field. The interested reader is advised to explore this field further. It is our hope that this brief introduction may inspire and motivate the reader to discover new and interesting applications of the general variational inequalities and related optimization problems in other areas of sciences.