## Abstract

The optimality system of a quasi-variational inequality can be reformulated as a non-smooth equation or a constrained equation with a smooth function. Both reformulations can be exploited by algorithms, and their convergence to solutions usually relies on the nonsingularity of the Jacobian, or the fact that the merit function has no nonoptimal stationary points. We prove new sufficient conditions for the absence of nonoptimal constrained or unconstrained stationary points that are weaker than some known ones. All these conditions exploit some properties of a certain matrix, but do not require the nonsingularity of the Jacobian. Further, we present new necessary and sufficient conditions for the nonsingularity of the Jacobian that are based on the signs of certain determinants. Additionally, we consider generalized Nash equilibrium problems that are a special class of quasi-variational inequalities. Exploiting their structure, we also prove some new sufficient conditions for stationarity and nonsingularity results.

### Similar content being viewed by others

## 1 Introduction

We consider quasi-variational inequalities (QVIs). This problem class was introduced in the paper series [1,2,3]. It generalizes the classical variational inequality by allowing dependence of the feasible set from the point under consideration. We will define the problem at the beginning of the next section. For a detailed discussion of QVIs, we refer to [4, 5]. Here, we will discuss the central conditions to ensure that general algorithms for QVIs are well defined and converge. These are conditions guaranteeing that stationary points of a merit function are indeed solutions, and conditions guaranteeing nonsingularity of some Jacobian matrix in order to compute the descent direction. Generalized Nash equilibrium problems (GNEPs) that can be used, for example, to model markets with several players sharing common constraints and that were first formally introduced in [6] can be reformulated as QVIs under mild conditions, see [7], and are hence one of the applications of QVIs. There are a growing number of GNEP applications, and we refer to the appendix of [8] for a test library and references. Further applications can be found, for example, in biological [9], mechanical [4, 5], economic [10, 11] or transportation problems [12, 13]. Since several algorithms for GNEPs have been introduced in the recent years, we also discuss some of our issues for special cases of GNEPs, which are typically in a subclass of QVIs that is not easy to solve in practice.

Algorithms for QVIs or GNEPs have to deal with the fact that solutions are typically nonunique and often nonisolated, resulting in infinitely many solutions and singularity of the Jacobian matrix at the solutions, which in turn make the application of classical Newton methods difficult. Thus, the numerical solution of these problems is mathematically challenging. To prove uniqueness of the solution one usually has restrictive assumptions; see, for example, [14, 15]. However, to get well-defined algorithms one usually only requires nonsingularity at the iterates, and hence, interior point methods are promising approaches since they restrict the iterates to a certain set. For general Newton methods, nonsingularity at the solution is often exploited to get superlinear local convergence. In the case of nonisolated solutions, which we typically have for QVIs, one can still construct local fast convergent algorithms by exploiting some error bound condition. This has been done in the GNEP setting in [16, 17].

A promising approach to solve QVIs is to characterize the corresponding Karushâ€“Kuhnâ€“Tucker (KKT) points as zeros of a nonsmooth function. Then, one can apply a semismooth Newton method, as it was done in [18]. For the semismooth Newton approach in the context of GNEPs, see also [19]. For well-defined local methods, one requires nonsingularity of an element of the generalized Jacobian of the nonsmooth function at an iterate. For global convergence, one can use a linesearch procedure that is based on the squared norm of the nonsmooth function as a merit function. A convergence result for semismooth Newton methods usually proves that accumulation points are stationary points of the merit function. Hence, results guaranteeing that stationary points of the merit function are indeed KKT points are required. Also, if one uses a Levenbergâ€“Marquardt method, see, for example, [20], to overcome the nonsingularity problems, one minimizes the squared norm of the merit function and relies on stationarity results.

A different approach, also based on the solution of the KKT conditions, is an interior point method, where one solves a constrained equation with a smooth function. A potential reduction algorithm for QVIs or GNEPs is presented in [21] and [19], respectively. This algorithm requires very similar nonsingularity properties as the semismooth Newton approach at all iterates, which lie in the interior of some suitable set here.

A further approach is a trust region method for constrained equations as introduced in [22, 23]. This can be applied directly to the smooth constrained equation reformulation of the KKT system, which was done for GNEPs in [19, 24]. At each iteration, one has to solve a linear equation system, which is the same as for the interior point approach, and one gets convergence to stationary points of the merit function. Hence, also in this approach nonsingularity and nonoptimal stationarity are the main issues.

Moreover, let us mention that there is an approach based on the KKT conditions, namely the LP-Newton method proposed in [25] that avoids the singularity issues by solving a linear program instead of a linear equation system. For this approach, there is also a globalization technique through a suitable nondifferentiable merit function given in [26]. This avoids the problem to accumulate at possibly nonoptimal stationary points of the merit function at least for (piecewise) smooth problems. The price one has to pay is that each iteration becomes more expensive, since one has to solve a linear program instead of a linear equation system at each iteration. Even these methods converge to stationary points of some constrained equation reformulations of the KKT conditions of the QVI, and it is fundamental to have conditions guaranteeing that these points are actual solutions.

For all these methods, the results of this paper are directly applicable and are the theoretical base of convergence to a solution of the QVI.

For a short overview on methods to solve QVIs, not only via their KKT conditions, we refer to the introduction in [21] and the references cited there and to [27, 28]. Finally, there are some newer augmented Lagrangian or multiplier-penalty methods to solve QVIs proposed in [29, 30] that are built on the sequential solution of simpler variational inequalities, as first proposed in [31].

In our paper, we consider the constrained equation reformulation of the KKT conditions of the QVI and the nonsmooth unconstrained reformulation. We will present several conditions that are necessary and sufficient for the nonsingularity required by many of the algorithms mentioned above. In the context of QVIs, these are, to the best of our knowledge, the first necessary and sufficient conditions for nonsingularity of the Jacobians. Furthermore, we will present sufficient conditions that avoid the existence of nonoptimal stationary points for different constrained or unconstrained stationarity concepts. For the constrained reformulation, this is, again to the best of our knowledge, the first time that such kind of conditions are given in the QVI context. Furthermore, through our analysis we obtain hints about which specific KKT reformulation is preferable to be used with projected gradient-type methods or with interior point methods. We also extend existing results for QVIs and GNEPs from [18, 21, 32]. In particular, in [21] one can find several nonsingularity results for a number of special structured QVIs.

The paper is organized as follows: In Sect.Â 2, we introduce the problem formulation and define the relevant functions and matrices. Further, we present our assumptions and we state the main facts we want to prove in this paper: Figs. 1 and 2 anticipate and summarize the results proved in Sects.Â 3 and 4.

In Sect.Â 3, we derive new *unconstrained and constrained stationarity results for QVIs* and show their applicability in some examples: In Proposition 3.1, we generalize [19, Theorem 3.1] about unconstrained stationarity; in Proposition 3.2 and 3.3 we give conditions to get unconstrained stationarity results, and in Example 3.1 we show that without these conditions the stationarity results do not hold; in Theorem 3.1 and 3.2 we give new assumptions to get constrained stationarity results; Example 3.2 shows a QVI satisfying all assumptions of Theorem 3.2, while Example 3.3 and 3.4 show that without the conditions of Theorem 3.2 the constrained stationarity results may not hold.

Thereafter, in Sect.Â 4, we present *nonsingularity results for QVIs* and the main theorem provides a new necessary and sufficient condition: In Proposition 4.1, we state the equivalence between the nonsingularity of the Jacobians of the smooth and the nonsmooth reformulations; in Proposition 4.2 and 4.3 we report some technical conditions for nonsingularity results; in Proposition 4.4 we give sufficient conditions for nonsingularity results; our main Theorem 4.1 gives new necessary and sufficient conditions for the nonsingularity of the Jacobian of the smooth reformulation in the case of linear constraints with variable right-hand side; in the same framework, Corollary 4.1 shows that the conditions given in [21, Corollary 2] are not only sufficient, but also necessary, for the nonsingularity of the Jacobian of the smooth reformulation, while Corollary 4.2 and 4.3 state similar nonsingularity results for the nonsmooth reformulation.

We consider the special case of *QVIs with box constraints* in Sect.Â 5 and show that box constraints are useful to weaken our assumptions: Proposition 5.1 provides sufficient conditions to guarantee all assumptions of Proposition 3.2 and 3.3; Theorem 5.1 weakens the conditions of Theorem 3.2 for constrained stationarity results of the smooth reformulation; Corollary 5.1 and 5.2 deal with the nonsingularity issue by assuming less restrictive conditions.

Section 6 contains *nonsingularity and stationarity results for GNEPs* that are not straightforward applications of the corresponding QVI results, but exploit the special structure of GNEPs: Corollary 6.1 significantly improves [32, Theorem 4.1] for linear problems; Lemma 6.1 defines a class of GNEPs for which many nonsingularity and stationarity results hold, while Example 6.2 shows an application satisfying the conditions given in Lemma 6.1; in Lemma 6.2 we identify some GNEPs with box constraints that have nonsingularity and stationarity properties; Proposition 6.1 provides conditions for a GNEP with box constraints in order to obtain constrained stationarity properties for the smooth reformulation.

We close the paper with some final remarks in conclusions section.

## 2 Problem Definition, Assumptions and Motivation

We consider the following QVI: Find a vector \( x^* \in K(x^*) \) such that

where \(F:{\mathbb {R}}^n \rightarrow {\mathbb {R}}^n\) is a (point-to-point) mapping and \(K:{\mathbb {R}}^n \rightrightarrows {\mathbb {R}}^n\) is a point-to-set mapping with closed and convex images. We assume that *F* is continuously differentiable and that the feasible set mapping *K* is given by a parametric set of inequality constraints:

where \(g: {\mathbb {R}}^n \times {\mathbb {R}}^n \rightarrow {\mathbb {R}}^m\) is twice continuously differentiable and, for each \(i=1,\ldots , m\), \(g_i(\cdot , x)\) is convex on \({\mathbb {R}}^n\), for each \(x\in {\mathbb {R}}^n\). We could also include linear equations, which lead to additional smooth equations in the following optimality systems, but for simplicity of notation we drop them from our subsequent analysis. We set

and hence, \(h(x) \le 0\) means that \(x\in K(x)\). We say that a point \(x\in {\mathbb {R}}^n \) satisfies the KKT conditions, if multipliers \(\lambda \in {\mathbb {R}}^m\) exist such that

Note that \(\nabla _y g(x,x)\) is the partial transposed Jacobian of *g*(*y*,Â *x*) with respect to *y* evaluated at \(y=x\).

Our aim is to find a solution \(x^*\) of the QVI by solving the KKT conditions (1). It is known, see, for example, [21, Theorem 1], that under the above convexity and differentiability assumptions, KKT points are solutions of the QVI. Further, under a suitable constraint qualification, like the Slater condition, all solutions of the QVI are KKT points. We can use the Fischerâ€“Burmeister function as a complementarity function and define

to obtain: \( (x^*, \lambda ^*) \) is a KKT point of the QVI, if and only if \( (x^*, \lambda ^*) \) solves the nonlinear equation system

Equivalently, \( (x^*, \lambda ^*) \) is a KKT point of the QVI, if and only if \( (x^*, \lambda ^*, w^*), \) with \( w^* = - h(x^*), \) is a solution of the nonlinear constrained equation system

Herein, \(\lambda \circ w\) is the vector with the componentwise product, and \({\mathbb {R}}^m_+\) is the set of vectors with nonnegative components. Later on, we will also use \({\mathbb {R}}^m_{++}\) for the set of vectors with positive components. If *F* is continuously differentiable and *g* is twice continuously differentiable, it is well known that the Jacobian *JH* of *H* is given by

see, for example, [18, 19, 21]. Here, \(J_xL(x,\lambda )\) denotes the partial Jacobian of \(L(x,\lambda )\) with respect to *x*, and *Jh*(*x*) is the Jacobian of *h*. Further, it is also known, see, for example, [19], that the generalized Jacobian of \(\varPsi \) is given by

where the matrices \(D_h\) and \(D_\lambda \) are \(m \, \times \, m\) diagonal matrices, with entries given by

for all \( i = 1, \ldots , m,\) and \(\mathbb {B}_1\) denotes the closed unit ball in \({\mathbb {R}}^2\).

If the matrix \(J_x L(x,\lambda )\) is nonsingular, we can define the reduced matrix

that is related to both Jacobian matrices described above. We will use the notation \(M_{i \bullet }\) to denote the *i*th row of a matrix *M*. Let us recall some known matrix properties we will use, cf. [33].

A matrix \(M \in {\mathbb {R}}^{n \times n}\) is called a \(P_0\) matrix, if for all \(x \in {\mathbb {R}}^n \setminus \{0\}\) there is an index \(i \in \{1,\ldots ,n\}\) such that \(x_i \ne 0\) and \(x_i \, M_{i \bullet } \, x \ge 0\).

A matrix \(M \in {\mathbb {R}}^{n \times n}\) is a \(P_0\) matrix if all its principal minors are nonnegative.

A matrix \(M \in {\mathbb {R}}^{n \times n}\) is called a

*P*matrix, if for all \(x \in {\mathbb {R}}^n \setminus \{0\}\) there is an index \(i \in \{1,\ldots ,n\}\) such that \(x_i \, M_{i \bullet } \, x > 0\).A matrix \(M \in {\mathbb {R}}^{n \times n}\) is a

*P*matrix if all its principal minors are positive.A matrix \(M \in {\mathbb {R}}^{n \times n}\) is called a semimonotone matrix, if for all \(x \in {\mathbb {R}}^n_{+} \setminus \{0\}\) there is an index \(i \in \{1,\ldots ,n\}\) such that \(x_i > 0\) and \(M_{i \bullet } \, x \ge 0\).

Every \(P_0\) matrix is semimonotone.

Every nonnegative matrix is semimonotone.

The sum of a semimonotone matrix and a nonnegative one is semimonotone.

A matrix \(M \in {\mathbb {R}}^{n \times n}\) is an

*S*matrix if \(x \in {\mathbb {R}}^n_{++}\) exists such that \(Mx > 0\).

An important practical issue in the design of algorithms consists in understanding if all stationary points of the merit function \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\), with \((x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{+}^m \times {\mathbb {R}}_{+}^m\), and of the merit function \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\), possibly with \((x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{+}^m\), are zeros of *H* and \(\varPsi \), respectively.

Moreover, the main topic in order to guarantee that interior point algorithms are well defined and converge to a solution of the QVI is the nonsingularity of \(JH(x, \lambda , w)\) for \((x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m\) or of the elements of \(\partial \varPsi (x,\lambda )\) for \((x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m.\) Note that for \(\lambda \in {\mathbb {R}}_{++}^m\) the set \(\partial \varPsi (x,\lambda )\) is by definition single valued and contains only the Jacobian \(J \varPsi (x,\lambda ).\)

More specifically, and summarizing, we are interested in giving conditions guaranteeing the following facts:

- (F1)
all stationary points of the merit function \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\) are zeros of \(\varPsi \);

- (F2)
all stationary points of the merit function \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\), with \(\lambda \ge 0\) and \(w \ge 0\), are zeros of

*H*; - (F3)
all stationary points of the merit function \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\), with \(\lambda \ge 0\) and \(h(x) \le 0\), are zeros of \(\varPsi \);

- (F4)
all constrained stationary points of

$$\begin{aligned} \min \frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2 \qquad \text {s.t.} \qquad \lambda \ge 0 \end{aligned}$$are zeros of \(\varPsi \);

- (F5)
all constrained stationary points of

$$\begin{aligned} \min \frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2 \qquad \text {s.t.} \qquad \lambda \ge 0, w \ge 0 \end{aligned}$$are zeros of

*H*; - (F6)
\(JH(x, \lambda , w)\) is nonsingular for \((x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m\);

- (F7)
\(J \varPsi (x,\lambda )\) is nonsingular for \((x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m\) with \(h(x)<0\).

To the best of our knowledge, the literature of QVIs and GNEPs has only investigated conditions guaranteeing facts (F1) and (F6) so far; see [19, 21]. Constrained stationary points have been considered in the case of variational inequalities in [34], but not for QVIs yet.

The results we will prove are of great practical interest. On the one hand, for the first time, we give conditions guaranteeing that finding a stationary point of the merit function \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\) subject to \(\lambda \ge 0\) and \(w \ge 0\), or the merit function \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\) subject to \(\lambda \ge 0\), we are actually computing solutions of the QVI. On the other hand, an important hint given by this study is that if we want to solve the QVI by adopting a projected gradient-type method designed to minimize the merit functions described above, then reformulation (2) is the best option. While if we want to resort to a second order interior point algorithm, reformulation (3) seems to be preferable.

Before stating the assumptions we use in our analysis, let us introduce some notation. For an index set \(J \subseteq \{1,\ldots ,m\},\) we write \(\nabla _y g_{\bullet J}(x,x)\) for the matrix with all columns of \(\nabla _y g(x,x)\) that belong to *J*,Â and by \(Jh_{J \bullet }(x)\) we denote the matrix with all rows of *Jh*(*x*) with indices in *J*.

These are our main assumptions:

- (A1)
For all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m\), the matrix \(J_x L(x,\lambda )\) is nonsingular and the matrix \(M(x,\lambda )\) is a \(P_0\) matrix.

- (A1+)
For all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+\), the matrix \(J_x L(x,\lambda )\) is nonsingular and the matrix \(M(x,\lambda )\) is a \(P_0\) matrix.

- (A1++)
For all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}\), the matrix \(J_x L(x,\lambda )\) is nonsingular and the matrix \(M(x,\lambda )\) is a \(P_0\) matrix.

- (A2)
For all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+\), the matrix \(J_x L(x,\lambda )\) is nonsingular and the matrix \(M(x,\lambda )\) is a semimonotone matrix.

- (A3)
For all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+\), the matrix \(J_x L(x,\lambda )\) is nonsingular and \(M(x,\lambda )\) is a positive semidefinite

*S*matrix. - (A4)
All the constraints are of the form \(g(y,x) = A y + c(x)\), with a matrix \(A \in {\mathbb {R}}^{m \times n}\) and a function \(c:{\mathbb {R}}^n \rightarrow {\mathbb {R}}^m.\)

- (A5)
Let (A4) hold and let an \(x \in {\mathbb {R}}^n\) be given. Set

$$\begin{aligned} d_J:=\det \begin{pmatrix} J_x F(x) &{}\quad \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{}\quad 0 \end{pmatrix}. \end{aligned}$$We assume for some index set \(J \subseteq \{1,\ldots ,m\}\) that \(d_J \ne 0\) and either \(d_J \ge 0\) for all \(J \subseteq \{1,\ldots ,m\},\) or \(d_J \le 0\) for all \(J \subseteq \{1,\ldots ,m\}\).

Note that the first three assumptions only differ in the range of \(\lambda \), and (A1) implies (A1+), which again implies (A1++). Since every positive semidefinite matrix is also a \(P_0\) matrix, (A3) implies (A1+). Since every \(P_0\) matrix is a semimonotone matrix, (A1+) implies (A2). In (A4), we refer to a class of QVIs for which the constraints are separable in *x* and *y* and where the *y*-part is linear. This is, in particular, satisfied for linear constraints. In [21], constraints of QVIs for which (A4) holds are termed linear with variable right-hand side. In this case, we obtain

and we see that both \(J_x L(x,\lambda )\) and \(M(x,\lambda )\) become independent of the multipliers \(\lambda \).

In Figs.Â 1 and 2, we summarize the relations we show in the next two sections.

## 3 Stationarity Results for QVIs

In this section, we want to see how (A1)â€“(A3) imply (F1)â€“(F5). For (unconstrained) stationarity results, it is clear that nonsingularity of the Jacobian, or of all the elements of the generalized Jacobian, implies that these points must be zeros of the function. Conditions ensuring this will be shown in the next section. Here, we present conditions not requiring this nonsingularity. First, we generalization the GNEP result [19, Theorem 3.1] to QVIs.

### Proposition 3.1

If *(A1)* holds, every stationary point of the merit function \(\frac{1}{2} \Vert \varPsi (x{,}\lambda )\Vert ^2\) is a zero of \(\varPsi \), i.e., *(F1)* is true.

### Proof

Let \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m\) be a stationary point, meaning

where \(\psi \in \partial \varPsi (x,\lambda )\). Using nonsingularity of \(J_x L(x,\lambda )\) we get

With the definition of \(M(x,\lambda )\) from (4), we obtain

Both \(D_h\) and \(D_\lambda \) can be assumed to be negative diagonal matrices, because they are nonpositive and \(D_h^i\) or \(D_\lambda ^i\) can be zero only if \(\varphi (\lambda _i,-h_i(x)) = 0\). With \(M(x,\lambda )\) also its transposed is a \(P_0\) matrix by (A1). From [35], we get that \(M(x,\lambda )^\top \) being a \(P_0\) matrix and \(D_h, D_\lambda \) being negative definite implies \(M(x,\lambda )^\top \, D_h + D_\lambda \) being nonsingular. This yields \(\varphi (\lambda ,-h(x)) = 0\), and then, \(L(x,\lambda )=0\), which completes the proof. \(\square \)

The following results are useful to guarantee that unconstrained stationary points, whose dual variables are nonnegative, are solutions of the QVI.

### Proposition 3.2

If *(A2)* holds, every stationary point of \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\), for which \(\lambda \ge 0\) and \(w \ge 0\) holds, is a zero of *H*, i.e., *(F2)* is true.

### Proof

Let \((x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+ \times {\mathbb {R}}^m_+\) be a stationary point, meaning

This yields

and, using nonsingularity of \(J_x L(x,\lambda )\) from (A2),

Inserting this in

we obtain

Let us assume for contradiction that \(\lambda \circ w \ne 0\). With \(\text {diag}(\lambda ) \in {\mathbb {R}}^m_{+}\) and \(M(x,\lambda )\) being a semimonotone matrix, also \(M(x,\lambda )^\top \) and \(M(x,\lambda )^\top \, \text {diag}(\lambda )\) are semimonotone matrices. By definition, an index \(j \in \{1,\ldots ,m\}\) exists such that

This implies the contradiction

Thus, we must have \(\lambda \circ w = 0\), and the proof holds by observing that we also get \(h(x)+w=0\) and \(L(x,\lambda )=0\). \(\square \)

For illustration, we present the following example where nonoptimal stationary points, with nonnegative \(\lambda \) and *w*, are present.

### Example 3.1

Define \(F:{\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2\) and \(g:{\mathbb {R}}^2 \times {\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2\) by

Here, every stationary point of \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\) satisfies

Now one can check that the point \(x=(1,1),\lambda =(1,1)\) and \(w=(1,1)\) is a stationary point, but it is certainly not a zero of *H*. In this example, we have nonsingularity of the matrix \(J_xL(x,\lambda )=\begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix},\) but the matrix

is not semimonotone.

Assumption (A2) can also be used for the nonsmooth reformulation to get (F3), as we show next.

### Proposition 3.3

Let *(A2)* hold. Then, every stationary point of \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\) satisfying \(\lambda \ge 0\) and \(h(x) \le 0\) is a zero of \(\varPsi \), i.e., *(F3)* is true.

### Proof

As in Proposition 3.1, we obtain (5) that in turn can be written as

By (A2), \(-D_h\) being a nonnegative diagonal matrix implies that the matrix \(M(x{,}\lambda )^\top \, ({-}D_h)\) is semimonotone. Let us assume by contradiction that \({-}\varphi (\lambda {,}{-}h(x))\ne 0\). Observing that \({-}\varphi (\lambda {,}{-}h(x)) \ge 0\) for all \(\lambda \ge 0\) and \(h(x) \le 0\), the definition of semimonotonicity yields the existence of an index \( j \in \{1,\ldots ,m\}\) such that

This implies the contradiction

where the last inequality follows since \(\varphi (\lambda _{j},-h_{ j}(x)) < 0\) implies \(\lambda _{ j} > 0\) and \(h_{ j}(x) < 0\). Thus, we must have \(\varphi (\lambda ,-h(x)) = 0\), and then, nonsingularity of \(J_x L(x,\lambda )\) implies \(L(x,\lambda )=0\) and \(\varPsi (x,\lambda )=0.\)\(\square \)

In the following part, we consider constrained stationary points. For the merit function \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\), we obtain (F4) under (A1+).

### Theorem 3.1

Let *(A1+)* hold. Then, every constrained stationary point of \(\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2\) subject to \(\lambda \ge 0\) is a zero of \(\varPsi \), i.e., *(F4)* is true.

### Proof

Let \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+\) be a constrained stationary point, meaning

The nonsingularity of \(J_x L(x,\lambda )\) from (A1+) implies

Hence, if we show that \(\varphi (\lambda ,-h(x)) = 0,\) we also get \(L(x,\lambda )=0.\) Let us recall that whenever \(\varphi (\lambda _i,-h_i(x)) \ne 0,\) the diagonal elements \(D_h^i\) and \(D_\lambda ^i\) are negative. Now, we replace \(L(x,\lambda )\) in the constrained stationarity conditions by the above formula, use the definition of \(M(x,\lambda )\) from (4) and consider five possible cases:

- (a)
For \(\lambda _i=0,h_i(x) \le 0\), we have \(\varphi (\lambda _i,-h_i(x))=0\) and \(D_h^i \varphi (\lambda _i,-h_i(x))=0.\)

- (b)
For \(\lambda _i>0, h_i(x)=0\), we have \(\varphi (\lambda _i,-h_i(x))=0\) and \(D_h^i \varphi (\lambda _i,-h_i(x))=0.\)

- (c)
For \(\lambda _i=0, h_i(x) > 0\), we have \(\varphi (\lambda _i,-h_i(x))>0\) and the stationarity conditions imply

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) \ge -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) >0. \end{aligned}$$ - (d)
For \(\lambda _i>0, h_i(x) < 0\) we have \(\varphi (\lambda _i,-h_i(x))<0\) and the stationarity conditions imply

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) = -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) <0. \end{aligned}$$ - (e)
For \(\lambda _i>0, h_i(x) > 0\), we have \(\varphi (\lambda _i,-h_i(x))>0\) and the stationarity conditions imply

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) = -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) >0. \end{aligned}$$

Together, we have for all \(i \in \{1,\ldots ,m\}\) either \(D_h^i \varphi (\lambda _i,-h_i(x))=0\) or

With \(M(x,\lambda )\) also its transposed matrix is a \(P_0\) matrix. If \(D_h \varphi (\lambda ,-h(x)) \ne 0,\) the definition of a \(P_0\) matrix implies that for at least one index \(j \in \{1,\ldots ,m\}\) we have \(D_h^j \varphi (\lambda _j,-h_j(x))\ne 0\) and

This condition excludes cases (c) to (e), and we must have \(\varphi (\lambda ,-h(x))=0.\) This completes the proof. \(\square \)

In order to obtain (F5), we use the stronger assumption (A3).

### Theorem 3.2

If *(A3)* holds, every constrained stationary point of the merit function \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\) subject to \(\lambda \ge 0\) and \(w \ge 0\) is a zero of *H*, i.e., *(F5)* is true.

### Proof

Let \((x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+ \times {\mathbb {R}}^m_+\) be a constrained stationary point, meaning:

Nonsingularity of \(J_x L(x,\lambda )\) implies \( L(x,\lambda ) = - \nabla _x L(x,\lambda )^{-1} \nabla h(x) (h(x) + w). \) We must consider four possible cases:

- (a)
For \(\lambda _i> 0, w_i > 0\), we get

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) = w_i^2 \lambda _i > 0, \quad h_i(x) + w_i = - w_i \lambda _i^2 < 0. \end{aligned}$$ - (b)
For \(\lambda _i>0, w_i=0\), we get

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) = w_i^2 \lambda _i = 0, \quad h_i(x) + w_i \ge - w_i \lambda _i^2 = 0. \end{aligned}$$ - (c)
For \(\lambda _i=0, w_i > 0\), we get

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) \le w_i^2 \lambda _i = 0, \quad h_i(x) + w_i = - w_i \lambda _i^2 = 0. \end{aligned}$$ - (d)
For \(\lambda _i=0, w_i = 0\), we get

$$\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) \le w_i^2 \lambda _i = 0, \quad h_i(x) + w_i \ge - w_i \lambda _i^2 = 0. \end{aligned}$$

Thus, in all the cases we obtain

Since \(M(x,\lambda ),\) and hence also its transposed matrix, is positive semidefinite, we get

This implies that for all \(i=1,\ldots ,m\) we must have

Thus, case (a) cannot hold, and we have \(\lambda \circ w = 0.\) By (A3), \(M(x,\lambda )\) is an *S* matrix, and hence, the system \(M(x,\lambda ) \, z> 0, z>0\) is solvable. The alternatives theorem of Ville, see [36], gives that the system

is not solvable. From the cases (b) to (d), we have that

Hence, we must have \(h(x)+w=0.\) This, in turn implies \(L(x,\lambda )=0,\) and therefore, we have \(H(x,\lambda ,w)=0.\)\(\square \)

Inspecting the proof starting from equation (7), one could also finalize it by assuming that \(M(x,\lambda )\) is a *P* matrix. Unfortunately, the *P* property of \(M(x,\lambda )\) for all \(\lambda \in {\mathbb {R}}^m_+\) is a very strong condition, since it implies nonsingularity of the matrix \(M(x,\lambda )\). This, in turn, can only hold if \(\nabla h(x)\) and \(J_y g(x,x)\) are full rank matrices and hence, in particular, requires that we have linear independent constraints.

Let us give an example, where we can apply Theorem 3.2.

### Example 3.2

We consider the QVI defined through

Here, the matrix \(J_x L(x,\lambda )=J_xF\) is nonsingular. Further, we have

which is symmetric positive semidefinite. Since all matrix entries of *M* are positive, the system \(M \, z> 0, \, z > 0\) can be solved with a vector having all elements equal to 1, and *M* is an *S* matrix. Thus, Theorem 3.2 guarantees that (F5) holds.

With the next example, we see that, in contrast to (F4), (A1+) or positive semi-definiteness of \(M(x,\lambda )\) alone is not sufficient to get (F5).

### Example 3.3

Let us consider the QVI defined through

This QVI does not even have a feasible point, but a constrained stationary point \({\bar{x}}=(0,0), {\bar{\lambda }}=(0,0), {\bar{w}}=(0,0)\), since

imply the constrained stationarity conditions

However, this point is not feasible, hence not a solution of \(H(x,\lambda ,w)=0.\) Furthermore, in this example we have

which is a \(P_0\) matrix and positive semidefinite. This shows that the \(P_0\) property of \(M(x,\lambda )\), or its positive semidefiniteness, is not sufficient for a constrained stationary point of *H* to be a solution of \(H(x,\lambda ,w)=0.\)

We can also give an example that has feasible points and where *M* is positive semidefinite, but nonoptimal constrained stationary points exist.

### Example 3.4

Consider the QVI defined through

This QVI has feasible points, for example (2,Â 2). Here, the point \({\bar{x}}=(0,0),\)\({\bar{\lambda }}=(0,0), {\bar{w}}=(0,0)\) is a constrained stationary point, since

Obviously the point is not feasible and hence not a zero of *H*. We have in this example

which is a symmetric positive semidefinite matrix. Hence, this condition is not sufficient for (F5). Moreover, we can see that (A3) is violated, since the system \(M \, z>0, z>0\) is not solvable.

## 4 Nonsingularity Results for QVIs

In this section, we deal with the nonsingularity facts (F6) and (F7). Let us recall that for \(\lambda \in {\mathbb {R}}^m_{++}\) the set \(\partial \varPsi (x,\lambda )=\{J \varPsi (x,\lambda )\}\) is single valued. First, we show that the two facts (F6) and (F7) are equivalent for all *x* with \(h(x) < 0\).

### Proposition 4.1

Let an arbitrary \(x \in {\mathbb {R}}^n\) with \(h(x) < 0\) be given. Then, \(J\varPsi (x,\lambda )\) is nonsingular for all \(\lambda \in {\mathbb {R}}^m_{++}\) if and only if \(JH(x,\lambda ,w)\) is nonsingular for all \((\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}\). This means *(F7)* is equivalent to *(F6)* for all \(x \in {\mathbb {R}}^n\) with \(h(x) < 0\).

### Proof

For \(h(x) < 0\) and \(\lambda \in {\mathbb {R}}^m_{++}\), the matrices \(D_h\) and \(D_\lambda \) are negative definite, with \( D_h^i =\frac{-h_i(x)}{\sqrt{h_i(x)^2+\lambda _i^2}}-1 <0\) and \(D_\lambda ^i =\frac{\lambda _i}{\sqrt{h_i(x)^2+\lambda _i^2}}-1 <0.\) Thus, the matrix \( J\varPsi (x,\lambda )= \begin{pmatrix} J_x L(x,\lambda ) &{} \nabla _y g(x,x) \\ - D_h Jh(x) &{} D_\lambda \end{pmatrix} \) is nonsingular, if and only if \( \begin{pmatrix} J_x L(x,\lambda ) &{} \nabla _y g(x,x) \\ Jh(x) &{} - D_h^{-1} \, D_\lambda \end{pmatrix} \) is nonsingular. For any fixed \(x \in {\mathbb {R}}^n\) with \(h_i(x) < 0\) and \(\lambda _i>0,\) the function

is continuous with \( \lim _{\lambda _i \downarrow 0} (D_h^i)^{-1} \, (D_\lambda ^i) = + \infty \) and \( \lim _{\lambda _i \rightarrow \infty } (D_h^i)^{-1} \, (D_\lambda ^i) = 0. \) Thus, nonsingularity of \(J\varPsi (x,\lambda )\) for all \(\lambda \in {\mathbb {R}}^m_{++}\) is equivalent to nonsingularity of the matrices

for all positive definite diagonal matrices \(D \in {\mathbb {R}}^{m \times m}_{++}\).

Now, since for some \((\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}\) we can write any positive definite diagonal matrix in the form \(D=\text {diag}(\lambda )^{-1} \, \text {diag}(w),\) nonsingularity of (9) for all positive definite diagonal matrices \(D \in {\mathbb {R}}^{m \times m}_{++}\) is equivalent to nonsingularity of

for all \((\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}.\) This, in turn, is equivalent to

being nonsingular for all \((\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}\), and the proof is complete. \(\square \)

The following general necessary and sufficient condition for the nonsingularity of *JH* was proved in [21, Theorem 3].

### Proposition 4.2

Let \((x, \lambda , w ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}\) be given. Then, the matrix

is nonsingular, if and only if \( JH (x,\lambda ,w) \) is nonsingular.

In the same spirit, we obtain a similar result for \(J\varPsi \).

### Proposition 4.3

Let \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}\) with \(h(x) \ne 0\) be given. Then, the matrix

is nonsingular, if and only if \(J \varPsi (x,\lambda )\) is nonsingular.

### Proof

Since \(h(x) \ne 0\) and \(\lambda \in {\mathbb {R}}^m_{++}\), we have a negative definite diagonal matrix \(D_\lambda \). For a vector \(v=(v^1,v^2) \in {\mathbb {R}}^{n+m},\) we get that

is equivalent to

Thus, nonsingularity of \( J\varPsi (x,\lambda )\) is equivalent to the nonsingularity of the matrix \(J_xL(x,\lambda ) + \nabla _y g(x,x) \, D_\lambda ^{-1} \, D_h \, J h(x).\)\(\square \)

We remark that the assumption \(h(x) \ne 0\) in Proposition 4.3 is mandatory to exploit the reduced matrix given in the proposition. In fact, if \(h_i(x) = 0\) for even one single element, the matrix \(D_\lambda \) becomes singular.

The first part of the following result was proved in [21, Corollary 2]. The second one follows from the first one with Proposition 4.1.

### Proposition 4.4

Let *(A1++)* hold. Then,

- (a)
\(JH(x,\lambda ,w)\) is nonsingular for all \((x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m\), i.e.,

*(F6)*holds; - (b)
\(J\varPsi (x,\lambda )\) is nonsingular for all \((x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m\) with \(h(x) < 0\), i.e.,

*(F7)*holds.

For checking nonsingularity of \(JH(x, \lambda , w)\) for all \((\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++},\) it would be nicer to have a condition that is independent of the choice of \((\lambda ,w).\) In this perspective, if (A4) holds (meaning that we have linear constraints with variable right-hand side), we obtain exactly what we want, that is, (A1) does not involve the variables \(\lambda \), because now, the matrices \(J_xL(x)\) and *M*(*x*) depend exclusively on *x*. This also means that (A1), (A1+) and (A1++) coincide. Moreover, under (A4), we derive sharper nonsingularity results.

Now, we state our main theorem, a necessary and sufficient condition for the matrix \(JH(x,\lambda ,w)\) to be nonsingular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++}.\) This theorem and the following corollary are a nontrivial and fundamental generalization of the results of [21], where only sufficiency was proved.

### Theorem 4.1

Let \(x \in {\mathbb {R}}^n\) be given, and let *(A4)* hold. Then, \(JH(x,\lambda ,w)\) is non-singular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) if and only if *(A5)* holds, i.e., for some \(J \subseteq \{1,\ldots ,m\},\)\(d_J:=\det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{} 0 \end{pmatrix} \ne 0\) and either \(d_J \ge 0\) for all \(J \subseteq \{1,\ldots ,m\},\) or \(d_J \le 0\) for all \(J \subseteq \{1,\ldots ,m\}\).

### Proof

We have \( \det (JH(x,\lambda ,w)) = \det \begin{pmatrix} J_x F(x) &{} \nabla _y g &{} 0 \\ Jh(x) &{} 0 &{} I \\ 0 &{} \text {diag}(w) &{} \text {diag}(\lambda ) \end{pmatrix}. \)

Developing this determinant by the last *m* rows, we obtain a sum over the index sets \(J \subseteq \{1,\ldots ,m\},\) where every addend contains products of \(\lambda _i, i \in J\) and \(w_i, i \in J^c := \{1,\ldots ,m\} \setminus J\). For each \(i \in J\), we drop the corresponding columns of the last block. For each \({\tilde{i}} \in J^c\), we drop the corresponding column \(\nabla _y g_{\bullet {\tilde{i}}}\) in the second block and, since we also drop the row with \(\lambda _{{\tilde{i}}}\), we can develop the remaining matrix after the corresponding column \({\tilde{i}}\) in the middle right block. This results in dropping the row \(Jh_{{\tilde{i}} \bullet }(x)\). Further, note that in the determinant formula, we get for every \(i \in J\) a positive sign, and for every \(i \in J^c\) we get one negative and one positive sign for the two developments. Thus, we obtain for the determinant of \(JH(x,\lambda ,w)\)

If \(d_J \ne 0\) for some \(J \subseteq \{1,\ldots ,m\},\) and, either \(d_J \ge 0\) for all \(J \subseteq \{1,\ldots ,m\}\), or \(d_J \le 0\) for all \(J \subseteq \{1,\ldots ,m\}\), i.e., if (A5) holds, we have \(\det (JH(x,\lambda ,w)) \ne 0\) for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) and hence, \(JH(x,\lambda ,w)\) is nonsingular.

Now, assume (A5) does not hold. Then, we either have \(d_J=0\) for all \(J \subseteq \{1,\ldots ,m\},\) which immediately implies that \(\det (JH(x,\lambda ,w)) = 0\), or we have index sets \(J_1,J_2 \subseteq \{1,\ldots ,m\}\) such that

In this case, let us define

Now we set

With these settings \(f:{\mathbb {R}}_{++} \rightarrow {\mathbb {R}},\)

is a continuous function. To analyze its behavior for \(z \rightarrow +\infty ,\) we have to consider the coefficients for the terms with the largest power of *z*, which is \({|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|}\). Since \(d_{J_1}>0\), we obtain for the index set \(J_1\) the positive addend

The remaining index sets \({\tilde{J}} \ne J_1\) leading to the largest power of *z* must satisfy \(J_1 \setminus J_2 \subseteq {\tilde{J}}\) and \(J_1^c \setminus J_2^c \subseteq {\tilde{J}}.\) By construction, we have natural numbers \(\alpha \le |J_1 \cap J_2| + |J_1^c \cap J_2^c|-1\) and \(\beta \ge 1,\) such that the corresponding addend satisfies

Since the total number of possible index sets \({\tilde{J}} \ne J_1\) is not greater than \(2^m-1,\) and

the coefficient of \(z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|}\) is positive. Thus, we have

Using analogous arguments for the index set \(J_2\) with \(d_{J_2}<0,\) we can show that the coefficient for the largest power of \(\frac{1}{z}\) is negative and hence

Now, by continuity of *f*,Â it has a zero. Thus, we can find \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) such that \(\det (JH(x,\lambda ,w)) =0,\) and hence, \(JH(x,\lambda ,w)\) is singular, which completes the proof. \(\square \)

This theorem shows that under (A4) the nonsingularity condition (F6) is equivalent to (A5) holding for all \(x \in {\mathbb {R}}^n.\) Next, we present a new corollary showing that under (A4) and nonsingularity of \(J_xF(x)\), (A1) is necessary and sufficient for (F6).

### Corollary 4.1

Let \(x \in {\mathbb {R}}^n\) be given, *(A4)* hold, and \(J_xF(x)\) be nonsingular. Then, the matrix \(JH(x,\lambda ,w)\) is nonsingular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++} \), if and only if *M*(*x*) is a \(P_0\) matrix.

### Proof

By Theorem 4.1, \(JH(x,\lambda ,w)\) being nonsingular means that for some \(J \subseteq \{1,\ldots ,m\}\) we have \(d_J=\det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{} 0 \end{pmatrix} \ne 0\) and either \(d_J \ge 0\) for all \(J \subseteq \{1,\ldots ,m\}\) or \(d_J \le 0\) for all \(J \subseteq \{1,\ldots ,m\}\). By the assumption on \(J_xF(x),\) we have \(d_\emptyset = \det J_xF(x) \ne 0.\) Further, for \(J \ne \emptyset \) we can use the determinant formula for block matrices to obtain

Thus, the nonsingularity of *JH* is equivalent to requiring that

This coincides with the fact that *M*(*x*) is a \(P_0\) matrix, and proves the equivalence. \(\square \)

Let us present an example that illustrates the applicability of the corollary, in particular, if we have a linear function *F* yielding that *M*(*x*) becomes independent of *x*.

### Example 4.1

Define \(F:{\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2\) and \(g:{\mathbb {R}}^2 \times {\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2\) by

The QVI has the unique solution \(\left( \frac{1}{3}, \frac{1}{3} \right) .\) We have (A4), and \(J_xF=\begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix}\) is nonsingular. But, the matrix

is not a \(P_0\) matrix for any \(x \in {\mathbb {R}}^n.\) Hence, by Corollary 4.1, \(JH(x,\lambda ,w)\) is for all \(x \in {\mathbb {R}}^n\) singular for some \(\lambda ,w \in {\mathbb {R}}^{m}_{++} \). For example, one can set \(\lambda _1=\lambda _2=w_1=w_2\).

Let us further mention that *M* is semimonotone, and hence, (A2) is satisfied. Thus, Proposition 3.2 yields the stationarity result (F2) in this case, where nonsingularity of \(JH(x,\lambda ,w)\) is violated.

From Proposition 4.1, we obtain the following corollaries for nonsingularity of \(J\varPsi (x,\lambda )\):

### Corollary 4.2

Let \(x \in {\mathbb {R}}^n\) with \(h(x) <0\) be given, and let *(A4)* hold. Then, the matrix \(J \varPsi (x,\lambda )\) is nonsingular for all \(\lambda \in {\mathbb {R}}^{m}_{++},\) if and only if *(A5)* holds.

### Corollary 4.3

Let \(x \in {\mathbb {R}}^n\) with \(h(x) <0\) be given, *(A4)* hold and \(J_xF(x)\) be nonsingular. Then, the matrix \(J \varPsi (x,\lambda )\) is nonsingular for all \(\lambda \in {\mathbb {R}}^{m}_{++} \), if and only if *M*(*x*) is a \(P_0\) matrix.

## 5 QVIs with Box Constraints

Now, we discuss the case of QVIs with additional box constraints, i.e.,

The KKT conditions become

The constrained nonlinear system in the box constrained case is to solve

We show that the presence of the box constraints can be exploited in order to obtain the facts (F1)â€“(F7) under weakened assumptions. First, we will give some sufficient conditions for (A2) that entails (F2) and (F3).

### Proposition 5.1

Let a QVI with box constraints be given and all entries in *Jh*(*x*) be nonnegative. If \(J_x L(x,\lambda )\) is a positive diagonal matrix, the matrix \(M(x,\lambda )\) is semimonotone, i.e., *(A2)* is satisfied.

### Proof

To show that

is semimonotone, let a vector \(v=(v^1,v^2,v^3) \in {\mathbb {R}}^{n+n+m}_+ \setminus \{0\} \) be given.

First, assume \(v^1_j>0\) for some \(j \in \{1,\ldots ,n\}.\) Then, we must either have \(M_{j \bullet }(x,\lambda ) v \ge 0\) (and hence \(M(x,\lambda )\) is semimonotone), or we get from the *j*th row \(v^2_j > 0,\) since \(J_x L(x,\lambda )^{-1}\) is a positive diagonal matrix and with *Jh*(*x*) also all entries in \(\nabla _y g(x,x)\) are nonnegative. Since the *j*th row is the negative of the \((n+j)\)th row, either \(M_{j \bullet }(x,\lambda ) v \ge 0\) or \(M_{(n+j) \bullet }(x,\lambda ) v \ge 0\) must hold, and hence, \(M(x,\lambda )\) must be semimonotone.

Now assume \(v^1=0,\) and define the set

If \(J = \emptyset \), we have \(v^2=0\). Since the matrix \(J h(x)\, J_x L(x,\lambda )^{-1} \, \nabla _y g(x,x)\) has only nonnegative entries, it is semimonotone, and hence, \(M(x,\lambda )\) is semimontone. Thus, assume \(J \ne \emptyset \). We are done, if one of the components

is nonnegative. Hence, assume they are all negative. For \(J \ne \emptyset \) the fact that \(J_x L(x,\lambda )^{-1}\) is a positive diagonal matrix immediately implies that at least one component of \(v^3\) must be positive. Since all entries in \(J_x L(x,\lambda )^{-1}\nabla _y g(x,x) \, v^3\) are nonnegative, we have

Multiplying with the nonpositive matrix \(-Jh(x)\) results in

Since this corresponds to the last *m* rows of \(M(x,\lambda ),\) the matrix must be semimonotone since one of the components of \(v^3\) is positive. \(\square \)

The assumption (A3) is restrictive in the case without box constraints and does not hold, for example, if \(M(x,\lambda )\) consists of two rows, where one is the negative of the other. A different condition can be obtained for QVIs with additional box constraints. Let us denote

Here, we can use a positive linear independence assumption, rather than (A3), to obtain a constrained stationarity result.

### Theorem 5.1

Let a QVI with box constraints be given. Assume that for all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{+}\), the matrix \(J_xL(x,\lambda )\) is nonsingular and that the matrix \(M(x,\lambda ) = J {\widetilde{h}}(x) J_x L(x,\lambda )^{-1} \nabla _y {\widetilde{g}}(x,x)\) is positive semidefinite. Further, assume the gradients \(\nabla {\widetilde{h}}_j(x)\) for \(j \in J:=\{j \in \{1,\ldots ,m\} : {\widetilde{h}}_j(x)>0\}\) are positive linear independent. Then, every constrained stationary point of the merit function \(\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2\) subject to \(\lambda \ge 0\) and \(w \ge 0\) is a zero of *H*, i.e., *(F5)* is true.

### Proof

First, we can follow the lines of the proof of Theorem 3.2 to obtain (8) and thus to exclude the case (a), which was \(\lambda _i >0\) and \(w_i>0\), through the assumed positive semidefiniteness of \(M(x,\lambda )\). Thus, we must have \(\lambda \circ w=0\). Now, we inspect the constrained stationarity conditions for the box constraints. Then, we obtain for every \(i=1,\ldots ,n\):

If \(\lambda _i^1>0\) or \(\lambda _i^2>0\), the first two lines imply \(L_i(x,\lambda )=0.\) If both \(\lambda _i^1=0\) and \(\lambda _i^2=0\), the lines three and four imply again \(L_i(x,\lambda )=0.\) Since this holds for all \(i=1,\ldots , n,\) we have \(L(x,\lambda )=0.\)

It remains to show that \({\widetilde{h}}(x)+w=0.\) From the constrained stationarity conditions, we get with \(\lambda \circ w =0\):

This together with \(L(x,\lambda )=0\) and \(J:=\{j \in \{1,\ldots ,m\} : {\widetilde{h}}_j(x)>0\}\) implies that the remaining stationarity condition yields

Since \({\widetilde{h}}_J(x)>0\) and the gradients \(\nabla {\widetilde{h}}_j(x), j \in J\) are positively linearly independent, we must have \(J=\emptyset \), which means \({\widetilde{h}}(x)+w=0,\) and completes the proof. \(\square \)

In the box constrained case, the Jacobian \(J H(x, \lambda ,w)\) is given by

Let \(\lambda ,w \in {\mathbb {R}}^{2n+m}_{++}\) be given, and let us define the matrix

It is not difficult to see that \(JH(x,\lambda ,w)\) is nonsingular, if and only if

is nonsingular. In this setting, we can replace the nonsingularity of \(J_xL(x,\lambda )\) from (A1++) by positive semidefiniteness of \(J_xL(x,\lambda )\) in Proposition 4.4 and in Corollary 4.1. In fact, in this case \(\widetilde{J_x L}(x,\lambda ,w)\) becomes nonsingular for all \(\lambda , w \in {\mathbb {R}}^m_{++}\), and the developments follow as in the case without box constraints. We observe that this also holds, if we have either a lower or an upper bound for each variable, and we do not need both of them. Defining

we get the following results.

### Corollary 5.1

Assume box constraints. Let the matrix \(J_x L(x,\lambda )\) be positive semi-definite and the matrix \(\widetilde{M}(x)\) be a \(P_0\) matrix for all \((x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}\). Then,

- (a)
\(JH(x,\lambda ,w)\) is nonsingular for all \((x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m\), i.e.,

*(F6)*holds; - (b)
\(J \varPsi (x,\lambda )\) is nonsingular for all \((x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m\) with \(h(x) < 0\), i.e.,

*(F7)*holds.

### Corollary 5.2

Assume we have box constraints, *(A4)* holds, and \(J_xF(x)\) is positive semidefinite for all \(x \in {\mathbb {R}}^n\). Then,

- (a)
*(F6)*is equivalent to \(\widetilde{M}(x)\) being a \(P_0\) matrix; - (b)
*(F7)*is equivalent to \(\widetilde{M}(x)\) being a \(P_0\) matrix for all \(x \in {\mathbb {R}}^n\) with \(h(x) < 0\).

Moreover, if \(J_xF(x)\) is positive definite for all \(x \in {\mathbb {R}}^n\), then

- (c)
\(\widetilde{M}(x)\) is a \(P_0\) matrix if and only if

*M*(*x*) is a \(P_0\) matrix.

### Proof

Point (c) is due to point (a) and Corollary 4.1. \(\square \)

## 6 GNEPs

It was first observed in [7] that under mild assumptions a GNEP can be reformulated as an equivalent QVI. Hence, we can obtain GNEP results by applying the QVI results from the last sections. We will not state all of them here, but we will focus on special cases, where we can obtain new results.

We consider GNEPs with linear constraints and quadratic cost functions. Let matrices \(Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu }\) and \(A_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_\nu },\) and vectors \(c^\nu \in {\mathbb {R}}^{n_\nu }\) and \(b^\nu \in {\mathbb {R}}^{m_\nu }\) be given. Further, \(X_\nu \) is a closed convex subset of \({\mathbb {R}}^{n_\nu }.\) Each player \(\nu =1,\ldots , N\) tries to find \(x^\nu \in {\mathbb {R}}^{n_\nu }\) to solve the problem

Notice that the GNEPs considered here have objective functions that depend only on private variables, while the constraints may couple all the variables. We underline that the coupling constraints are not necessarily shared by all the players, and therefore, these GNEPs are, in general, no generalized potential games as defined in [37, 38]. The QVI reformulation of the GNEP is made up by

By the linearity of the constraints, we have (A4).

For completeness, let us first mention that the case of linear cost functions, i.e., if we have \(Q_\nu =0\) for all \(\nu =1,\ldots ,N,\) is discussed in [32]. In particular, we have in [32, Theorem 4.1] the following result:

### Lemma 6.1

Let \(Q_\nu =0\) and \(X_\nu = {\mathbb {R}}^{n_\nu }\) for all \(\nu \in \{1,\ldots ,N\}\). Assume that \(\det (\nabla _y g_{\bullet J} Jh_{J \bullet }) \ge 0\) for all \(J \subseteq \{1,\ldots ,m\}\) with \(|J|=n\), and that the determinant is positive for one of these sets. Then, \(JH(\lambda ,w)\) is nonsingular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++}.\)

Exploiting Theorem 4.1, we can extend Lemma 6.1 and get necessary and sufficient conditions for the nonsingularity.

### Corollary 6.1

Let \(Q_\nu =0\) and \(X_\nu = {\mathbb {R}}^{n_\nu }\) for all \(\nu \in \{1,\ldots ,N\}\). \(JH(\lambda ,w)\) is nonsingular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) if and only if all determinants \(\det (\nabla _y g_{\bullet J} Jh_{J \bullet })\) with \(J \subseteq \{1,\ldots ,m\}\) and \(|J|=n\), are either all nonnegative or all nonpositive, and at least one of them is not 0.

### Proof

By Theorem 4.1, we consider the signs of \( d_J:=\det \begin{pmatrix} 0 &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet } &{} 0 \end{pmatrix}. \) We have \(d_J = 0\) for every \(J \subseteq \{1,\ldots ,m\}\) with \(|J| \ne n\). For \(|J| = n\) we have

Now, the assertion is a consequence of Theorem 4.1. \(\square \)

In the next example, the matrix \(JH(x,\lambda ,w)\) is always singular.

### Example 6.1

Consider the linear 2-player GNEP

From Lemma 6.1, we get that \(JH(x,\lambda ,w)\) is singular for some \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) since for \(J=\{1,2\}\) we get \(\det (\nabla _y g_{\bullet J} Jh_{J \bullet }) = \det \left( \begin{pmatrix} 1&{}0 \\ 0 &{} 1 \end{pmatrix} \begin{pmatrix} 1&{}1 \\ 1 &{} 1 \end{pmatrix} \right) = 0.\) But even more, in this example \(JH(x,\lambda ,w)\) is singular for all \(\lambda ,w \in {\mathbb {R}}^{m}_{++},\) since

Next, we consider the more general quadratic setting. However, we restrict ourselves to GNEPs, where all players have the same number of variables, i.e., we assume \(n_1=\ldots =n_N.\) For some special structures, we get the facts (F1)â€“(F7).

### Theorem 6.1

Let \(n_1=\ldots =n_N, Q_\nu = \beta _\nu T\) with \(\beta _\nu \in {\mathbb {R}}_{++}\) and a positive definite matrix \(T \in {\mathbb {R}}^{n_1 \times n_1}\), and \(X_\nu = {\mathbb {R}}^{n_1}\) for all \(\nu \in \{1,\ldots ,N\}\). Assume \(A_{\nu \mu } = \alpha _\mu (B_{\nu } + E_{\nu \mu })\) with \(\alpha _\mu \in {\mathbb {R}}\) and \(B_{\nu }, E_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_1}\), \(E_{\nu \nu } = 0\), for all \(\mu , \nu \in \{1,\ldots ,N\}.\)

- 1.
Let all entries of either \(E_{\nu \mu } T^{-1} B_\mu \) or \((B_\nu + E_{\nu \mu }) T^{-1} B_\mu \) be nonnegative for all \(\mu , \nu \in \{1,\ldots ,N\}\). Then,

*M*is a semimonotone matrix, i.e.,*(A2)*holds, implying*(F2)*and*(F3)*. - 2.
Let \(E_{\nu \mu } = 0\) for all \(\mu , \nu \in \{1,\ldots ,N\}\). Then,

*M*is a \(P_0\)-matrix, i.e.,*(A1)*holds, implying*(F1)*-*(F4)*,*(F6)*and*(F7)*. - 3.
Let \(E_{\nu \mu } = 0\), let \(B_{\nu }\) be a positive matrix, let \(T^{-1}\) be a nonnegative matrix with a positive diagonal, and let \(\frac{\alpha _\nu ^2}{\beta _\nu } = \frac{\alpha _\mu ^2}{\beta _\mu } > 0\) for all \(\mu , \nu \in \{1,\ldots ,N\}\). Then,

*M*is a positive semidefinite*S*matrix, i.e.,*(A3)*holds, implying that all the facts*(F1)*-*(F7)*are true.

### Proof

We start with case 2: \( J_xL(x)= \begin{pmatrix} \beta _1 T &{}&{}\\ &{} \ddots &{} \\ &{}&{} \beta _N T \end{pmatrix}, \) is nonsingular, by the assumed positive definiteness of *T* and since \(\beta _i>0\) for all \(i=1,\ldots ,N.\) Further, we have

where \( {\overline{B}} := \begin{pmatrix} B_1 \\ \vdots \\ B_N \end{pmatrix}\) and \( D := \begin{pmatrix} \frac{\alpha _1^2}{\beta _1} I_{m_1} &{}&{}\\ &{} \ddots &{} \\ &{}&{} \frac{\alpha _N^2}{\beta _N} I_{m_N} \end{pmatrix}. \) For \(J \subseteq \{1, \ldots , m\}\), we have

because \({\overline{B}} \, T^{-1} \, {\overline{B}}^\top \) is positive semidefinite, hence a \(P_0\) matrix, and *D* is positive semidefinite and diagonal. Therefore, *M* is a \(P_0\) matrix, that is, (A1) is true. As a consequence, (F1) holds by Proposition 3.1 and (F4) by Theorem 3.1. Further, Proposition 4.4 gives (F6) and (F7). Finally, since (A1) implies (A2), we have (F2) from Proposition 3.2 and (F3) from Proposition 3.3.

Now, let us consider case 1: We have

If the first condition holds, *M* is the sum of a \(P_0\) matrix and a nonnegative one; thus, it is semimonotone. If the second condition holds, *M* is nonnegative and then also semimonotone. In any case, (A2) holds. Now, (F2) follows from Proposition 3.2 and (F3) from Proposition 3.3.

In case 3, the matrix \(M = {\overline{B}} \, T^{-1} \, {\overline{B}}^\top D\), with \(D = \frac{\alpha _1^2}{\beta _1} I\), is positive semidefinite. The assumptions on \(T^{-1}\) and \(B_\nu \) imply that \(M z = \frac{\alpha _1^2}{\beta _1} {\overline{B}} \, T^{-1} \, {\overline{B}}^\top z > 0\), for any vector \(z > 0\). Thus, (A3) is true. Fact (F5) comes from Theorem 3.2, while all the others follow as in case 2, since (A3) implies (A1). \(\square \)

### Example 6.2

Let us consider *N* firms operating in a specific area. Each firm produces its goods in \(n_1\) spots. Let \(x^\nu \in {\mathbb {R}}^{n_1}_+\) be the quantities of production of the \(\nu \)th firm, and let its cost be given by

where \(m^\nu \in {\mathbb {R}}^{n_1}_-\) are the marginal costs and \(\beta _\nu > 0\). Let \(\alpha _\nu > 0\) be a number giving the degree of pollution of the \(n_1\) plants of the \(\nu \)th firm. Any firm \(\nu \) must control the level of pollution in \(m_\nu \) different spots of the area:

where: \(B_\nu \in {\mathbb {R}}^{m_\nu \times n_1}_{++}\) represents the absolute pollution impact of the \(n_1\) spots of production of the \(\nu \)th firm on the \(m_\nu \) spots, where the pollution must be controlled by the firm; \(E_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_1}\) represents some deviations from \(B_\nu \), and \((B_\nu + E_{\nu \mu })\) is the absolute pollution impact of the \(\mu \)th firm on the \(m_\nu \) spots controlled by the \(\nu \)th firm; \(b^\nu \in {\mathbb {R}}^{m_\nu }_{++}\) are the maximum levels of pollution allowed in the \(m_\nu \) spots controlled by the \(\nu \)th firm. Any firm \(\nu \) minimizes \(cost_\nu \) subject to its pollution constraints.

Referring to Lemma 6.1, we can prove the following assertions.

- 1.
Let all \((B_\nu + E_{\nu \mu }) \ge 0\), i.e., the absolute pollution impacts of the firms are nonnegative. Then, (A2) holds and the facts (F2) and (F3) are true.

- 2.
Let \(E_{\nu \mu } = 0\) for all \(\nu \) and \(\mu \), i.e., the plants of all the firms in any specific spot impact equally on the controlled spots. Then, (A1) holds and facts (F1)â€“(F4), (F6) and (F7) are true.

- 3.
Let \(E_{\nu \mu } = 0\) and \(\frac{\alpha _\nu ^2}{\beta _\nu } = \frac{\alpha _\mu ^2}{\beta _\mu }\) for all \(\nu \) and \(\mu \), which is, for example, true if the degrees of pollution \(\alpha _\nu \) and the parameters \(b_\nu \) are the same for all the firms. Then, (A3) holds and all the facts (F1)â€“(F7) are true.

Now we consider a different framework involving box constraints.

### Theorem 6.2

Assume \(n_1=\ldots =n_N, Q_\nu \in {\mathbb {R}}^{n_1 \times n_1}_{+}\) are nonnegative diagonal matrices, \(X_\nu = [l^\nu , u^\nu ]\) for all \(\nu =1,\ldots ,N,\) and \(A_{\nu \mu } = D_{\nu } {\overline{D}}_\mu + E_{\nu \mu },\) with diagonal matrices \(D_{\nu }, {\overline{D}}_\mu \in {\mathbb {R}}^{n_1 \times n_1}\) and matrices \(E_{\nu \mu } \in {\mathbb {R}}^{n_1 \times n_1}\) for all \(\nu , \mu \in \{1,\ldots ,N\}\).

- 1.
Let \(Q_\nu \) be positive diagonal matrices for all \( \nu \in \{1,\ldots ,N\}\), and \(A_{\nu \mu }\) be non-negative for all \(\mu , \nu \in \{1,\ldots ,N\}\). Then,

*(A2)*holds and the facts*(F2)*and*(F3)*are true. - 2.
Let \(E_{\nu \mu } = 0\) for all \(\mu , \nu \in \{1,\ldots ,N\}\). Then, \({\widetilde{M}}\) is a \(P_0\)-matrix, and the facts

*(F6)*and*(F7)*are true. If, moreover, \(Q_\nu \) is positive definite for all \(\nu =1,\ldots ,N\),*(A1)*holds and facts*(F1)*-*(F4)*,*(F6)*and*(F7)*are true.

### Proof

Point 1 is due to Proposition 5.1.

Let us consider point 2. Let arbitrary \(\lambda ,w \in {\mathbb {R}}^{3n}_{++}\) be given. By assumption \(J_xF\) is positive semidefinite. We set

with positive definite diagonal matrices \({\widetilde{Q}}_\nu \in {\mathbb {R}}^{n_1 \times n_1}, \nu =1,\ldots ,N.\) We get

where \( C := \begin{pmatrix} D_1 \\ \vdots \\ D_N \end{pmatrix}\) and \( {\check{D}} :=\begin{pmatrix} {\overline{D}}_1 {\widetilde{Q}}_1^{-1} {\overline{D}}_1&{} &{} \\ &{} \ddots &{} \\ &{} &{} {\overline{D}}_N {\widetilde{Q}}_N^{-1} {\overline{D}}_N \end{pmatrix}. \)

For arbitrary \(J \subseteq \{1, \ldots , m\},\) we have

because \(C \, C^\top \) is positive semidefinite, and \({\check{D}}\) is positive semidefinite and diagonal. Therefore, \({\widetilde{M}}\) is a \(P_0\) matrix. Facts (F6) and (F7) follow from Corollary 5.1.

Moreover, if matrices \(Q_\nu \) are positive definite, then we can exploit the point (c) of Corollary 5.2 to say that assumption (A1) holds. Then, the assertion can be obtained as in point 2 of Lemma 6.1. \(\square \)

### Remark 6.1

Let us give an example class, where Lemma 6.2 can be applied for GNEPs. Suppose, each player has one variable, and, beside box constraints, we have exactly one constraint, that is shared by all players. Then, this constraint can be stated using \(A_{\nu \mu } = {\overline{D}}_\mu \in {\mathbb {R}}\) and \(D_\nu =1\) for all \(\nu ,\mu =1,\ldots ,N.\) Now, if all \(Q_\nu \in {\mathbb {R}}_{++}\), the second case of Lemma 6.2 guarantees (F1)â€“(F4), (F6) and (F7).

### Remark 6.2

Unfortunately, if we have GNEPs with two or more linearly independent shared constraints, the matrix \({\widetilde{M}}\) can easily fail to be a \(P_0\)-matrix. For example, consider a 2 player game, where each player has one variable and we have, beside box constraints, two shared constraints \(a_1 x^1+a_2 x^2 \le b_1,\)\( a_3 x^1+ a_4 x^2 \le b_2.\) Then, if \({\tilde{Q}}_1,{\tilde{Q}}_2 > 0\) we have that \(\widetilde{J_xL}= \begin{pmatrix} {\tilde{Q}}_1 &{} 0 \\ 0 &{} {\tilde{Q}}_2 \end{pmatrix}\) is nonsingular. Further,

For the index sets \(I=\{2,3\}\) and \(J=\{1,4\},\) we obtain

Now we see that \({\widetilde{M}}\) is not a \(P_0\) matrix, whenever \(sign(a_1a_4)=sign(a_2a_3) \ne 0.\) In this case, by point (c) of Corollary 5.2 we obtain that the matrix *M* is not a \(P_0\) matrix. Then, Corollary 4.1 shows that \(JH(x,\lambda ,w)\) fails to be nonsingular for all \((\lambda ,w) \in {\mathbb {R}}^{2}_{++} \times {\mathbb {R}}^{2}_{++}.\)

Finally, we want to present some GNEPs with box constraints, where (F5) can be shown. We discuss an example class that often occurs in applications.

### Proposition 6.1

Let \(X_\nu = [l^\nu , u^\nu ]\) for all \(\nu =1,\ldots ,N,\), and further, only upper constraints are present, i.e., all coefficients of the remaining constraints are positive. Let the feasible set be nonempty. If all \(Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu },\nu =1,\ldots ,N \) are nonsingular and *M* is positive semidefinite, *(F5)* holds.

### Proof

Since all \(Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu },\nu =1,\ldots ,N \) are nonsingular, \(J_xL=J_xF\) is nonsingular. Since *M* is positive semidefinite by assumption, we can apply Theorem 5.1, if we have shown that the columns of the matrix \(\nabla h_{\bullet J}(x)\) with \(J=\{j \in \{1,\ldots ,m\} : h_j(x)>0\}\) are positively linearly independent for all \(x \in {\mathbb {R}}^n\). Assume we have a vector \(v \in {\mathbb {R}}^{|J|}_{+}\) with \(\nabla h_{\bullet J}(x) v = 0.\) We want to show that \(v=0\) must hold.

Whenever *J* contains an upper box constraint with index \(j_0\), the corresponding lower box constraint is not in *J*. Hence, the matrix \(\nabla h_{\bullet J}(x)\) has a corresponding row, with only nonnegative entries, and we must have \(v_{j_0}=0\). Thus, we can assume without loss of generality that *J* does not contain any upper box constraint.

Since the feasible set is nonempty, the set \(J=\{j \in \{1,\ldots ,m\} : h_j(x)>0\}\) cannot contain all lower box constraints and an additional upper constraint. Hence, the matrix \(\nabla h_{\bullet J}(x)\) can either contain only lower box constraints, resulting in \(v=0,\) or it must contain at least one row where all entries are nonnegative. But then, the components of *v* corresponding to the positive coefficients of the upper constraints must all be zero. The remaining matrix has only lower box constraints, again resulting in \(v=0.\) Thus, we must have \(v=0,\) meaning that the columns of \(\nabla h_{\bullet J}(x)\) are positively linear independent for all \(x \in {\mathbb {R}}^n\). By Theorem 5.1, fact (F5) holds. \(\square \)

## 7 Conclusions

In this paper, we presented a number of necessary and sufficient conditions for the main issues, when solving QVIs via a smooth constrained or a nonsmooth unconstrained equation reformulation of their KKT conditions. We have seen that the matrix *M* plays a crucial role. The first issue is to guarantee that stationary points of the merit function are indeed solutions. We considered several concepts of constrained or unconstrained stationary points, and the issue was solved under a \(P_0\) property for the nonsmooth reformulation. For the constrained, but smooth, reformulation we required stronger assumptions, namely positive semidefiniteness of *M* together with the *S* property of this matrix. Further, we gave an example, showing that the \(P_0\) property is not sufficient in the smooth reformulation. Considering only stationary points within the feasible set, which is important for interior point methods, we showed the stationarity result for both reformulations, if *M* is a semimonotone matrix.

For the second issue, the nonsingularity of the Jacobian at feasible iterates, we considered the QVI class, where the constraints are linear with variable right-hand side. Here, we proved equivalence of the nonsingularity with a sign condition for determinants of certain matrices, which are independent of the dual variables. Furthermore, under an additional nonsingularity assumption on \(J_x F(x),\) we showed equivalence to the \(P_0\) property of *M*. In the context of QVIs, these are, to the best of our knowledge, the first necessary and sufficient conditions for nonsingularity of the Jacobians. We have also seen that, for QVIs with box constraints, some of the assumptions can be weakened.

We further presented some subclasses of GNEPs with linear constraints and quadratic cost functions, for which the Jacobian is always nonsingular, and the absence of nonoptimal stationary points can be shown. Moreover, we extended a known sufficient nonsingularity condition for GNEPs with linear constraints and cost functions to a necessary and sufficient condition.

## References

Bensoussan, A., Goursat, M., Lions, J.L.: ContrÃ´le impulsionnel et inÃ©quations quasi-variationnelles stationnaires. CR Acad. Sci. Paris SÃ©r. A

**276**, 1279â€“1284 (1973)Bensoussan, A., Lions, J.L.: Nouvelle formulation de problÃ¨mes de contrÃ´le impulsionnel et applications. CR Acad. Sci. Paris SÃ©r. A

**276**, 1189â€“1192 (1973)Bensoussan, A., Lions, J.L.: Nouvelles methodes en contrÃ´le impulsionnel. Appl. Math. Optim.

**1**(4), 289â€“312 (1975)Baiocchi, C., Capelo, A.: Variational and Quasivariational Inequalities: Applications to Free Boundary Problems. Wiley, New York (1984)

Mosco, U.: Implicit variational problems and quasi variational inequalities. In: Gossez, J.P., LamiÂ Dozo, E.J., Mawhin, J., Waelbroeck, L. (eds.) Nonlinear Operators and the Calculus of Variations, pp. 83â€“156. Springer, Berlin (1976)

Debreu, G.: A social equilibrium existence theorem. Proc. Natl. Acad. Sci. USA

**38**(10), 886â€“893 (1952)Bensoussan, A.: Points de Nash dans le cas de fonctionnelles quadratiques et jeux differentiels lineaires a N personnes. SIAM J. Control

**12**(3), 460â€“499 (1974)Facchinei, F., Kanzow, C.: Penalty methods for the solution of generalized nash equilibrium problems. SIAM J. Optim.

**20**, 2228â€“2253 (2010)Hammerstein, P., Selten, R.: Game theory and evolutionary biology. In: Aumann, R.J., Hart, S. (eds.) Handbook of Game Theory with Economic Applications, vol. 2, pp. 929â€“993. Elsevier, North-Holland, Amsterdam (1994)

Ichiishi, T.: Game Theory for Economic Analysis. Academic Press, New York (1983)

Jing-Yuan, W., Smeers, Y.: Spatial oligopolistic electricity models with cournot generators and regulated transmission prices. Oper. Res.

**47**(1), 102â€“112 (1999)Bliemer, M.C., Bovy, P.H.: Quasi-variational inequality formulation of the multiclass dynamic traffic assignment problem. Transp. Res. Part B Methodol.

**37**(6), 501â€“519 (2003)Scrimali, L.: Quasi-variational inequalities in transportation networks. Math. Models Methods Appl. Sci.

**14**(10), 1541â€“1560 (2004)Dreves, A.: Uniqueness for quasi-variational inequalities. Set Valued Var. Anal.

**24**(2), 285â€“297 (2016)Nesterov, Y., Scrimali, L.: Solving strongly monotone variational and quasi-variational inequalities. CORE Discussion Papers 2006107, UniversitÃ© catholique de Louvain, Center for Operations Research and Econometrics (CORE) (2006)

Dreves, A., Facchinei, F., Fischer, A., Herrich, M.: A new error bound result for generalized Nash equilibrium problems and its algorithmic application. Comput. Optim. Appl.

**59**(1), 63â€“84 (2014)Izmailov, A.F., Solodov, M.V.: On error bounds and Newton-type methods for generalized Nash equilibrium problems. Comput. Optim. Appl.

**59**(1), 201â€“218 (2014)Facchinei, F., Kanzow, C., Karl, S., Sagratella, S.: The semismooth Newton method for the solution of quasi-variational inequalities. Comput. Optim. Appl.

**62**(1), 85â€“109 (2015)Dreves, A., Facchinei, F., Kanzow, C., Sagratella, S.: On the solution of the KKT conditions of generalized Nash equilibrium problems. SIAM J. Optim.

**21**(3), 1082â€“1108 (2011)Yamashita, N., Fukushima, M.: On the rate of convergence of the Levenbergâ€“Marquardt method. In: Alefeld, G., Chen, X. (eds.) Topics in Numerical Analysis, pp. 239â€“249. Springer, Vienna (2001)

Facchinei, F., Kanzow, C., Sagratella, S.: Solving quasi-variational inequalities via their KKT conditions. Math. Program.

**144**(1), 369â€“412 (2014)Bellavia, S., Macconi, M., Morini, B.: An affine scaling trust-region approach to bound-constrained nonlinear systems. Appl. Numer. Math.

**44**(3), 257â€“280 (2003)Bellavia, S., Macconi, M., Morini, B.: STRSCNE: a scaled trust-region solver for constrained nonlinear equations. Comput. Optim. Appl.

**28**(1), 31â€“50 (2004)Galli, L., Kanzow, C., Sciandrone, M.: A nonmonotone trust-region method for generalized Nash equilibrium and related problems with strong convergence properties. Comput. Optim. Appl.

**69**(3), 629â€“652 (2018)Facchinei, F., Fischer, A., Herrich, M.: An LP-Newton method: nonsmooth equations, KKT systems, and nonisolated solutions. Math. Program.

**146**(1), 1â€“36 (2014)Fischer, A., Herrich, M., Izmailov, A., Solodov, M.: A globally convergent LP-Newton method. SIAM J. Optim.

**26**(4), 2012â€“2033 (2016)Aussel, D., Sagratella, S.: Sufficient conditions to compute any solution of a quasivariational inequality via a variational inequality. Math. Methods Oper. Res.

**85**(1), 3â€“18 (2017)Latorre, V., Sagratella, S.: A canonical duality approach for the solution of affine quasi-variational inequalities. J. Global Optim.

**64**(3), 433â€“449 (2016)Kanzow, C.: On the multiplier-penalty-approach for quasi-variational inequalities. Math. Program.

**160**(1), 33â€“63 (2016)Kanzow, C., Steck, D.: Augmented Lagrangian and exact penalty methods for quasi-variational inequalities. Comput. Optim. Appl.

**69**(3), 801â€“824 (2018)Pang, J.S., Fukushima, M.: Quasi-variational inequalities, generalized Nash equilibria, and multi-leader-follower games. Comput. Manag. Sci.

**2**(1), 21â€“56 (2005)Dreves, A., Sudermann-Merx, N.: Solving linear generalized Nash equilibrium problems numerically. Optim. Methods Softw.

**31**(5), 1036â€“1063 (2016)Cottle, R., Pang, J.S., Stone, R.: The Linear Complementarity Problem. Academic Press, New York (1992)

Facchinei, F., Fischer, A., Kanzow, C., Peng, J.M.: A simply constrained optimization reformulation of KKT systems arising from variational inequalities. Appl. Math. Optim.

**40**(1), 19â€“37 (1999)Sandberg, I.W., Willson, A.N.: Some theorems on properties of DC equations of nonlinear networks. Bell Syst. Tech. J.

**48**(1), 1â€“34 (1969)Mangasarian, O.: Nonlinear Programming. McGraw-Hill, New York (1969)

Facchinei, F., Piccialli, V., Sciandrone, M.: Decomposition algorithms for generalized potential games. Comput. Optim. Appl.

**50**(2), 237â€“262 (2011)Sagratella, S.: Algorithms for generalized potential games with mixed-integer variables. Comput. Optim. Appl.

**68**(3), 689â€“717 (2017)

## Acknowledgements

Open Access funding provided by Projekt DEAL.

## Author information

### Authors and Affiliations

### Corresponding author

## Ethics declarations

### Conflict of interest

The authors declare that they have no conflict of interest.

## Additional information

Communicated by Roland Herzog.

### Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Rights and permissions

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the articleâ€™s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the articleâ€™s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

## About this article

### Cite this article

Dreves, A., Sagratella, S. Nonsingularity and Stationarity Results for Quasi-Variational Inequalities.
*J Optim Theory Appl* **185**, 711â€“743 (2020). https://doi.org/10.1007/s10957-020-01678-x

Received:

Accepted:

Published:

Issue Date:

DOI: https://doi.org/10.1007/s10957-020-01678-x

### Keywords

- Quasi-variational inequality
- Generalized Nash equilibrium problem
- Nonsingularity
- Nonoptimal stationary points