## Introduction

We consider quasi-variational inequalities (QVIs). This problem class was introduced in the paper series [1,2,3]. It generalizes the classical variational inequality by allowing dependence of the feasible set from the point under consideration. We will define the problem at the beginning of the next section. For a detailed discussion of QVIs, we refer to [4, 5]. Here, we will discuss the central conditions to ensure that general algorithms for QVIs are well defined and converge. These are conditions guaranteeing that stationary points of a merit function are indeed solutions, and conditions guaranteeing nonsingularity of some Jacobian matrix in order to compute the descent direction. Generalized Nash equilibrium problems (GNEPs) that can be used, for example, to model markets with several players sharing common constraints and that were first formally introduced in  can be reformulated as QVIs under mild conditions, see , and are hence one of the applications of QVIs. There are a growing number of GNEP applications, and we refer to the appendix of  for a test library and references. Further applications can be found, for example, in biological , mechanical [4, 5], economic [10, 11] or transportation problems [12, 13]. Since several algorithms for GNEPs have been introduced in the recent years, we also discuss some of our issues for special cases of GNEPs, which are typically in a subclass of QVIs that is not easy to solve in practice.

Algorithms for QVIs or GNEPs have to deal with the fact that solutions are typically nonunique and often nonisolated, resulting in infinitely many solutions and singularity of the Jacobian matrix at the solutions, which in turn make the application of classical Newton methods difficult. Thus, the numerical solution of these problems is mathematically challenging. To prove uniqueness of the solution one usually has restrictive assumptions; see, for example, [14, 15]. However, to get well-defined algorithms one usually only requires nonsingularity at the iterates, and hence, interior point methods are promising approaches since they restrict the iterates to a certain set. For general Newton methods, nonsingularity at the solution is often exploited to get superlinear local convergence. In the case of nonisolated solutions, which we typically have for QVIs, one can still construct local fast convergent algorithms by exploiting some error bound condition. This has been done in the GNEP setting in [16, 17].

A promising approach to solve QVIs is to characterize the corresponding Karush–Kuhn–Tucker (KKT) points as zeros of a nonsmooth function. Then, one can apply a semismooth Newton method, as it was done in . For the semismooth Newton approach in the context of GNEPs, see also . For well-defined local methods, one requires nonsingularity of an element of the generalized Jacobian of the nonsmooth function at an iterate. For global convergence, one can use a linesearch procedure that is based on the squared norm of the nonsmooth function as a merit function. A convergence result for semismooth Newton methods usually proves that accumulation points are stationary points of the merit function. Hence, results guaranteeing that stationary points of the merit function are indeed KKT points are required. Also, if one uses a Levenberg–Marquardt method, see, for example, , to overcome the nonsingularity problems, one minimizes the squared norm of the merit function and relies on stationarity results.

A different approach, also based on the solution of the KKT conditions, is an interior point method, where one solves a constrained equation with a smooth function. A potential reduction algorithm for QVIs or GNEPs is presented in  and , respectively. This algorithm requires very similar nonsingularity properties as the semismooth Newton approach at all iterates, which lie in the interior of some suitable set here.

A further approach is a trust region method for constrained equations as introduced in [22, 23]. This can be applied directly to the smooth constrained equation reformulation of the KKT system, which was done for GNEPs in [19, 24]. At each iteration, one has to solve a linear equation system, which is the same as for the interior point approach, and one gets convergence to stationary points of the merit function. Hence, also in this approach nonsingularity and nonoptimal stationarity are the main issues.

Moreover, let us mention that there is an approach based on the KKT conditions, namely the LP-Newton method proposed in  that avoids the singularity issues by solving a linear program instead of a linear equation system. For this approach, there is also a globalization technique through a suitable nondifferentiable merit function given in . This avoids the problem to accumulate at possibly nonoptimal stationary points of the merit function at least for (piecewise) smooth problems. The price one has to pay is that each iteration becomes more expensive, since one has to solve a linear program instead of a linear equation system at each iteration. Even these methods converge to stationary points of some constrained equation reformulations of the KKT conditions of the QVI, and it is fundamental to have conditions guaranteeing that these points are actual solutions.

For all these methods, the results of this paper are directly applicable and are the theoretical base of convergence to a solution of the QVI.

For a short overview on methods to solve QVIs, not only via their KKT conditions, we refer to the introduction in  and the references cited there and to [27, 28]. Finally, there are some newer augmented Lagrangian or multiplier-penalty methods to solve QVIs proposed in [29, 30] that are built on the sequential solution of simpler variational inequalities, as first proposed in .

In our paper, we consider the constrained equation reformulation of the KKT conditions of the QVI and the nonsmooth unconstrained reformulation. We will present several conditions that are necessary and sufficient for the nonsingularity required by many of the algorithms mentioned above. In the context of QVIs, these are, to the best of our knowledge, the first necessary and sufficient conditions for nonsingularity of the Jacobians. Furthermore, we will present sufficient conditions that avoid the existence of nonoptimal stationary points for different constrained or unconstrained stationarity concepts. For the constrained reformulation, this is, again to the best of our knowledge, the first time that such kind of conditions are given in the QVI context. Furthermore, through our analysis we obtain hints about which specific KKT reformulation is preferable to be used with projected gradient-type methods or with interior point methods. We also extend existing results for QVIs and GNEPs from [18, 21, 32]. In particular, in  one can find several nonsingularity results for a number of special structured QVIs.

The paper is organized as follows: In Sect. 2, we introduce the problem formulation and define the relevant functions and matrices. Further, we present our assumptions and we state the main facts we want to prove in this paper: Figs. 1 and 2 anticipate and summarize the results proved in Sects. 3 and 4.

In Sect. 3, we derive new unconstrained and constrained stationarity results for QVIs and show their applicability in some examples: In Proposition 3.1, we generalize [19, Theorem 3.1] about unconstrained stationarity; in Proposition 3.2 and 3.3 we give conditions to get unconstrained stationarity results, and in Example 3.1 we show that without these conditions the stationarity results do not hold; in Theorem 3.1 and 3.2 we give new assumptions to get constrained stationarity results; Example 3.2 shows a QVI satisfying all assumptions of Theorem 3.2, while Example 3.3 and 3.4 show that without the conditions of Theorem 3.2 the constrained stationarity results may not hold.

Thereafter, in Sect. 4, we present nonsingularity results for QVIs and the main theorem provides a new necessary and sufficient condition: In Proposition 4.1, we state the equivalence between the nonsingularity of the Jacobians of the smooth and the nonsmooth reformulations; in Proposition 4.2 and 4.3 we report some technical conditions for nonsingularity results; in Proposition 4.4 we give sufficient conditions for nonsingularity results; our main Theorem 4.1 gives new necessary and sufficient conditions for the nonsingularity of the Jacobian of the smooth reformulation in the case of linear constraints with variable right-hand side; in the same framework, Corollary 4.1 shows that the conditions given in [21, Corollary 2] are not only sufficient, but also necessary, for the nonsingularity of the Jacobian of the smooth reformulation, while Corollary 4.2 and 4.3 state similar nonsingularity results for the nonsmooth reformulation.

We consider the special case of QVIs with box constraints in Sect. 5 and show that box constraints are useful to weaken our assumptions: Proposition 5.1 provides sufficient conditions to guarantee all assumptions of Proposition 3.2 and 3.3; Theorem 5.1 weakens the conditions of Theorem 3.2 for constrained stationarity results of the smooth reformulation; Corollary 5.1 and 5.2 deal with the nonsingularity issue by assuming less restrictive conditions.

Section 6 contains nonsingularity and stationarity results for GNEPs that are not straightforward applications of the corresponding QVI results, but exploit the special structure of GNEPs: Corollary 6.1 significantly improves [32, Theorem 4.1] for linear problems; Lemma 6.1 defines a class of GNEPs for which many nonsingularity and stationarity results hold, while Example 6.2 shows an application satisfying the conditions given in Lemma 6.1; in Lemma 6.2 we identify some GNEPs with box constraints that have nonsingularity and stationarity properties; Proposition 6.1 provides conditions for a GNEP with box constraints in order to obtain constrained stationarity properties for the smooth reformulation.

We close the paper with some final remarks in conclusions section.

## Problem Definition, Assumptions and Motivation

We consider the following QVI: Find a vector $$x^* \in K(x^*)$$ such that

\begin{aligned} F(x^*)^\top (y-x^*) \ge 0 \quad \forall y \in K(x^*), \end{aligned}

where $$F:{\mathbb {R}}^n \rightarrow {\mathbb {R}}^n$$ is a (point-to-point) mapping and $$K:{\mathbb {R}}^n \rightrightarrows {\mathbb {R}}^n$$ is a point-to-set mapping with closed and convex images. We assume that F is continuously differentiable and that the feasible set mapping K is given by a parametric set of inequality constraints:

\begin{aligned} K(x) := \{ y \in {\mathbb {R}}^n : g(y,x) \le 0 \}, \end{aligned}

where $$g: {\mathbb {R}}^n \times {\mathbb {R}}^n \rightarrow {\mathbb {R}}^m$$ is twice continuously differentiable and, for each $$i=1,\ldots , m$$, $$g_i(\cdot , x)$$ is convex on $${\mathbb {R}}^n$$, for each $$x\in {\mathbb {R}}^n$$. We could also include linear equations, which lead to additional smooth equations in the following optimality systems, but for simplicity of notation we drop them from our subsequent analysis. We set

\begin{aligned} h(x):=g(x,x), \end{aligned}

and hence, $$h(x) \le 0$$ means that $$x\in K(x)$$. We say that a point $$x\in {\mathbb {R}}^n$$ satisfies the KKT conditions, if multipliers $$\lambda \in {\mathbb {R}}^m$$ exist such that

\begin{aligned} \begin{array}{l} \displaystyle {L(x, \lambda ) := F(x) + \nabla _y g(x,x) \lambda = 0,} \\ \displaystyle {\lambda _i \ge 0, \, g_i(x,x) \le 0, \, \lambda _i g_i(x,x) = 0 \quad \forall i = 1, \ldots , m.} \end{array} \end{aligned}
(1)

Note that $$\nabla _y g(x,x)$$ is the partial transposed Jacobian of g(yx) with respect to y evaluated at $$y=x$$.

Our aim is to find a solution $$x^*$$ of the QVI by solving the KKT conditions (1). It is known, see, for example, [21, Theorem 1], that under the above convexity and differentiability assumptions, KKT points are solutions of the QVI. Further, under a suitable constraint qualification, like the Slater condition, all solutions of the QVI are KKT points. We can use the Fischer–Burmeister function as a complementarity function and define

\begin{aligned} \varphi :{\mathbb {R}}^m \times {\mathbb {R}}^m \rightarrow {\mathbb {R}}^m, \qquad \varphi (a,b):=\begin{pmatrix} \sqrt{a_1^2+b_1^2} -a_1 -b_1 \\ \vdots \\ \sqrt{a_m^2+b_m^2} -a_m -b_m \end{pmatrix} \end{aligned}

to obtain: $$(x^*, \lambda ^*)$$ is a KKT point of the QVI, if and only if $$(x^*, \lambda ^*)$$ solves the nonlinear equation system

\begin{aligned} 0=\varPsi (x,\lambda ):= \begin{pmatrix} L(x, \lambda ) \\ \varphi (\lambda ,-h(x))\end{pmatrix}. \end{aligned}
(2)

Equivalently, $$(x^*, \lambda ^*)$$ is a KKT point of the QVI, if and only if $$(x^*, \lambda ^*, w^*),$$ with $$w^* = - h(x^*),$$ is a solution of the nonlinear constrained equation system

\begin{aligned} 0= H(x, \lambda , w) := \left( \begin{array}{c} L(x, \lambda ) \\ h(x) + w \\ \lambda \circ w \end{array} \right) \qquad (x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{+} \times {\mathbb {R}}^m_{+}. \end{aligned}
(3)

Herein, $$\lambda \circ w$$ is the vector with the componentwise product, and $${\mathbb {R}}^m_+$$ is the set of vectors with nonnegative components. Later on, we will also use $${\mathbb {R}}^m_{++}$$ for the set of vectors with positive components. If F is continuously differentiable and g is twice continuously differentiable, it is well known that the Jacobian JH of H is given by

\begin{aligned} J H(x, \lambda , w) = \left( \begin{array}{ccccc} J_x L(x,\lambda ) &{} \quad &{}\nabla _y g(x,x) &{} \quad &{}\quad 0 \\ Jh(x) &{}\quad &{}\quad 0 &{}\quad &{}\quad I \\ 0 &{}\quad &{}\quad \text {diag}(w)&{}\quad &{}\quad \text {diag}(\lambda ) \end{array} \right) ; \end{aligned}

see, for example, [18, 19, 21]. Here, $$J_xL(x,\lambda )$$ denotes the partial Jacobian of $$L(x,\lambda )$$ with respect to x, and Jh(x) is the Jacobian of h. Further, it is also known, see, for example, , that the generalized Jacobian of $$\varPsi$$ is given by

\begin{aligned} \partial \varPsi (x,\lambda )= \begin{pmatrix} J_x L(x,\lambda ) &{} \quad \nabla _y g(x,x) \\ - D_h Jh(x) &{}\quad D_\lambda \end{pmatrix}, \end{aligned}

where the matrices $$D_h$$ and $$D_\lambda$$ are $$m \, \times \, m$$ diagonal matrices, with entries given by

\begin{aligned} (D_h^i,D_\lambda ^i) \left\{ \begin{array}{lcl} = \frac{(-h_i(x),\lambda _i)}{\sqrt{h_i(x)^2+\lambda _i^2}} - (1,1), &{}\quad &{}\quad \text {if } (-h_i(x),\lambda _i) \ne (0,0),\\ \in \mathbb {B}_1 - (1,1), &{}\quad &{}\quad \text {if } (-h_i(x),\lambda _i) = (0,0), \end{array} \right. \end{aligned}

for all $$i = 1, \ldots , m,$$ and $$\mathbb {B}_1$$ denotes the closed unit ball in $${\mathbb {R}}^2$$.

If the matrix $$J_x L(x,\lambda )$$ is nonsingular, we can define the reduced matrix

\begin{aligned} M(x,\lambda ) := Jh(x) \, J_x L(x,\lambda )^{-1} \, \nabla _y g(x,x) \end{aligned}
(4)

that is related to both Jacobian matrices described above. We will use the notation $$M_{i \bullet }$$ to denote the ith row of a matrix M. Let us recall some known matrix properties we will use, cf. .

• A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is called a $$P_0$$ matrix, if for all $$x \in {\mathbb {R}}^n \setminus \{0\}$$ there is an index $$i \in \{1,\ldots ,n\}$$ such that $$x_i \ne 0$$ and $$x_i \, M_{i \bullet } \, x \ge 0$$.

A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is a $$P_0$$ matrix if all its principal minors are nonnegative.

• A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is called a P matrix, if for all $$x \in {\mathbb {R}}^n \setminus \{0\}$$ there is an index $$i \in \{1,\ldots ,n\}$$ such that $$x_i \, M_{i \bullet } \, x > 0$$.

A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is a P matrix if all its principal minors are positive.

• A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is called a semimonotone matrix, if for all $$x \in {\mathbb {R}}^n_{+} \setminus \{0\}$$ there is an index $$i \in \{1,\ldots ,n\}$$ such that $$x_i > 0$$ and $$M_{i \bullet } \, x \ge 0$$.

Every $$P_0$$ matrix is semimonotone.

Every nonnegative matrix is semimonotone.

The sum of a semimonotone matrix and a nonnegative one is semimonotone.

• A matrix $$M \in {\mathbb {R}}^{n \times n}$$ is an S matrix if $$x \in {\mathbb {R}}^n_{++}$$ exists such that $$Mx > 0$$.

An important practical issue in the design of algorithms consists in understanding if all stationary points of the merit function $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$, with $$(x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{+}^m \times {\mathbb {R}}_{+}^m$$, and of the merit function $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$, possibly with $$(x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{+}^m$$, are zeros of H and $$\varPsi$$, respectively.

Moreover, the main topic in order to guarantee that interior point algorithms are well defined and converge to a solution of the QVI is the nonsingularity of $$JH(x, \lambda , w)$$ for $$(x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m$$ or of the elements of $$\partial \varPsi (x,\lambda )$$ for $$(x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m.$$ Note that for $$\lambda \in {\mathbb {R}}_{++}^m$$ the set $$\partial \varPsi (x,\lambda )$$ is by definition single valued and contains only the Jacobian $$J \varPsi (x,\lambda ).$$

More specifically, and summarizing, we are interested in giving conditions guaranteeing the following facts:

1. (F1)

all stationary points of the merit function $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$ are zeros of $$\varPsi$$;

2. (F2)

all stationary points of the merit function $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$, with $$\lambda \ge 0$$ and $$w \ge 0$$, are zeros of H;

3. (F3)

all stationary points of the merit function $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$, with $$\lambda \ge 0$$ and $$h(x) \le 0$$, are zeros of $$\varPsi$$;

4. (F4)

all constrained stationary points of

\begin{aligned} \min \frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2 \qquad \text {s.t.} \qquad \lambda \ge 0 \end{aligned}

are zeros of $$\varPsi$$;

5. (F5)

all constrained stationary points of

\begin{aligned} \min \frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2 \qquad \text {s.t.} \qquad \lambda \ge 0, w \ge 0 \end{aligned}

are zeros of H;

6. (F6)

$$JH(x, \lambda , w)$$ is nonsingular for $$(x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m$$;

7. (F7)

$$J \varPsi (x,\lambda )$$ is nonsingular for $$(x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m$$ with $$h(x)<0$$.

To the best of our knowledge, the literature of QVIs and GNEPs has only investigated conditions guaranteeing facts (F1) and (F6) so far; see [19, 21]. Constrained stationary points have been considered in the case of variational inequalities in , but not for QVIs yet.

The results we will prove are of great practical interest. On the one hand, for the first time, we give conditions guaranteeing that finding a stationary point of the merit function $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$ subject to $$\lambda \ge 0$$ and $$w \ge 0$$, or the merit function $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$ subject to $$\lambda \ge 0$$, we are actually computing solutions of the QVI. On the other hand, an important hint given by this study is that if we want to solve the QVI by adopting a projected gradient-type method designed to minimize the merit functions described above, then reformulation (2) is the best option. While if we want to resort to a second order interior point algorithm, reformulation (3) seems to be preferable.

Before stating the assumptions we use in our analysis, let us introduce some notation. For an index set $$J \subseteq \{1,\ldots ,m\},$$ we write $$\nabla _y g_{\bullet J}(x,x)$$ for the matrix with all columns of $$\nabla _y g(x,x)$$ that belong to J,  and by $$Jh_{J \bullet }(x)$$ we denote the matrix with all rows of Jh(x) with indices in J.

These are our main assumptions:

1. (A1)

For all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m$$, the matrix $$J_x L(x,\lambda )$$ is nonsingular and the matrix $$M(x,\lambda )$$ is a $$P_0$$ matrix.

2. (A1+)

For all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+$$, the matrix $$J_x L(x,\lambda )$$ is nonsingular and the matrix $$M(x,\lambda )$$ is a $$P_0$$ matrix.

3. (A1++)

For all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}$$, the matrix $$J_x L(x,\lambda )$$ is nonsingular and the matrix $$M(x,\lambda )$$ is a $$P_0$$ matrix.

4. (A2)

For all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+$$, the matrix $$J_x L(x,\lambda )$$ is nonsingular and the matrix $$M(x,\lambda )$$ is a semimonotone matrix.

5. (A3)

For all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+$$, the matrix $$J_x L(x,\lambda )$$ is nonsingular and $$M(x,\lambda )$$ is a positive semidefinite S matrix.

6. (A4)

All the constraints are of the form $$g(y,x) = A y + c(x)$$, with a matrix $$A \in {\mathbb {R}}^{m \times n}$$ and a function $$c:{\mathbb {R}}^n \rightarrow {\mathbb {R}}^m.$$

7. (A5)

Let (A4) hold and let an $$x \in {\mathbb {R}}^n$$ be given. Set

\begin{aligned} d_J:=\det \begin{pmatrix} J_x F(x) &{}\quad \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{}\quad 0 \end{pmatrix}. \end{aligned}

We assume for some index set $$J \subseteq \{1,\ldots ,m\}$$ that $$d_J \ne 0$$ and either $$d_J \ge 0$$ for all $$J \subseteq \{1,\ldots ,m\},$$ or $$d_J \le 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$.

Note that the first three assumptions only differ in the range of $$\lambda$$, and (A1) implies (A1+), which again implies (A1++). Since every positive semidefinite matrix is also a $$P_0$$ matrix, (A3) implies (A1+). Since every $$P_0$$ matrix is a semimonotone matrix, (A1+) implies (A2). In (A4), we refer to a class of QVIs for which the constraints are separable in x and y and where the y-part is linear. This is, in particular, satisfied for linear constraints. In , constraints of QVIs for which (A4) holds are termed linear with variable right-hand side. In this case, we obtain

\begin{aligned} Jh(x)=A+J c(x), \quad \nabla _y g(x,x)=\nabla _y g=A^\top \quad \text {and} \qquad J_x L(x,\lambda )=J_x F(x), \end{aligned}

and we see that both $$J_x L(x,\lambda )$$ and $$M(x,\lambda )$$ become independent of the multipliers $$\lambda$$.

In Figs. 1 and 2, we summarize the relations we show in the next two sections.

## Stationarity Results for QVIs

In this section, we want to see how (A1)–(A3) imply (F1)–(F5). For (unconstrained) stationarity results, it is clear that nonsingularity of the Jacobian, or of all the elements of the generalized Jacobian, implies that these points must be zeros of the function. Conditions ensuring this will be shown in the next section. Here, we present conditions not requiring this nonsingularity. First, we generalization the GNEP result [19, Theorem 3.1] to QVIs.

### Proposition 3.1

If (A1) holds, every stationary point of the merit function $$\frac{1}{2} \Vert \varPsi (x{,}\lambda )\Vert ^2$$ is a zero of $$\varPsi$$, i.e., (F1) is true.

### Proof

Let $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m$$ be a stationary point, meaning

\begin{aligned} 0 = \psi ^\top \varPsi (x,\lambda ) = \begin{pmatrix} \nabla _x L(x,\lambda ) &{} - \nabla h(x) D_h \\ J_y g(x,x) &{} D_\lambda \end{pmatrix} \begin{pmatrix} L(x, \lambda ) \\ \varphi (\lambda ,-h(x)) \end{pmatrix}, \end{aligned}

where $$\psi \in \partial \varPsi (x,\lambda )$$. Using nonsingularity of $$J_x L(x,\lambda )$$ we get

\begin{aligned} L(x,\lambda )= \nabla _x L(x,\lambda )^{-1} \, \nabla h(x) D_h \varphi (\lambda ,-h(x)). \end{aligned}

With the definition of $$M(x,\lambda )$$ from (4), we obtain

\begin{aligned} \left( M(x,\lambda )^\top \, D_h + D_\lambda \right) (\varphi (\lambda ,-h(x))) = 0. \end{aligned}
(5)

Both $$D_h$$ and $$D_\lambda$$ can be assumed to be negative diagonal matrices, because they are nonpositive and $$D_h^i$$ or $$D_\lambda ^i$$ can be zero only if $$\varphi (\lambda _i,-h_i(x)) = 0$$. With $$M(x,\lambda )$$ also its transposed is a $$P_0$$ matrix by (A1). From , we get that $$M(x,\lambda )^\top$$ being a $$P_0$$ matrix and $$D_h, D_\lambda$$ being negative definite implies $$M(x,\lambda )^\top \, D_h + D_\lambda$$ being nonsingular. This yields $$\varphi (\lambda ,-h(x)) = 0$$, and then, $$L(x,\lambda )=0$$, which completes the proof. $$\square$$

The following results are useful to guarantee that unconstrained stationary points, whose dual variables are nonnegative, are solutions of the QVI.

### Proposition 3.2

If (A2) holds, every stationary point of $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$, for which $$\lambda \ge 0$$ and $$w \ge 0$$ holds, is a zero of H, i.e., (F2) is true.

### Proof

Let $$(x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+ \times {\mathbb {R}}^m_+$$ be a stationary point, meaning

\begin{aligned} 0 = JH(x,\lambda ,w)^\top H(x,\lambda ,w) = \begin{pmatrix} \nabla _x L(x,\lambda ) &{}\quad \nabla h(x) &{}\quad 0 \\ J_y g(x,x) &{}\quad 0 &{} \text {diag}(w) \\ 0 &{}\quad I &{}\quad \text {diag}(\lambda ) \end{pmatrix} \begin{pmatrix} L(x, \lambda ) \\ h(x) + w \\ \lambda \circ w \end{pmatrix}. \end{aligned}

This yields

\begin{aligned} h(x)+w=-\text {diag}(\lambda ) (\lambda \circ w), \end{aligned}

and, using nonsingularity of $$J_x L(x,\lambda )$$ from (A2),

\begin{aligned} L(x,\lambda )=-\nabla _x L(x,\lambda )^{-1} \, \nabla h(x) (h(x)+w). \end{aligned}

Inserting this in

\begin{aligned} J_y g(x,x) \, L(x,\lambda ) + \text {diag}(w)(\lambda \circ w) = 0, \end{aligned}

we obtain

\begin{aligned} \left( M(x,\lambda )^\top \, \text {diag}(\lambda ) + \text {diag}(w) \right) (\lambda \circ w) = 0. \end{aligned}
(6)

Let us assume for contradiction that $$\lambda \circ w \ne 0$$. With $$\text {diag}(\lambda ) \in {\mathbb {R}}^m_{+}$$ and $$M(x,\lambda )$$ being a semimonotone matrix, also $$M(x,\lambda )^\top$$ and $$M(x,\lambda )^\top \, \text {diag}(\lambda )$$ are semimonotone matrices. By definition, an index $$j \in \{1,\ldots ,m\}$$ exists such that

\begin{aligned} \lambda _{ j} w_{j} > 0, \quad \left( M(x,\lambda )^\top \, \text {diag}(\lambda ) (\lambda \circ w)\right) _{ j} \ge 0. \end{aligned}

\begin{aligned} 0 =&\left( \left( M(x,\lambda )^\top \, \text {diag}(\lambda ) + \text {diag}(w) \right) (\lambda \circ w) \right) _{ j} \\ =&\left( M(x,\lambda )^\top \, \text {diag}(\lambda ) (\lambda \circ w)\right) _{ j} + \lambda _{ j} w_{ j}^2 > 0. \end{aligned}

Thus, we must have $$\lambda \circ w = 0$$, and the proof holds by observing that we also get $$h(x)+w=0$$ and $$L(x,\lambda )=0$$. $$\square$$

For illustration, we present the following example where nonoptimal stationary points, with nonnegative $$\lambda$$ and w, are present.

### Example 3.1

Define $$F:{\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2$$ and $$g:{\mathbb {R}}^2 \times {\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2$$ by

\begin{aligned} F(x)=\begin{pmatrix} 2x_2-\frac{9}{2} \\ 2x_1-\frac{9}{2} \end{pmatrix} \qquad \text {and} \qquad g(y,x)=\begin{pmatrix} -3x_1+2y_2-1 \\ -3x_2+2y_1-1 \end{pmatrix}. \end{aligned}

Here, every stationary point of $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$ satisfies

\begin{aligned} 0= & {} JH(x,\lambda ,w)^\top H(x,\lambda ,w)\\= & {} \begin{pmatrix} 0 &{}\quad 2 &{}\quad -3&{}\quad 2&{}\quad 0&{}\quad 0\\ 2 &{}\quad 0 &{}\quad 2&{}\quad -3&{}\quad 0&{}\quad 0\\ 0&{}\quad 2&{}\quad 0&{}\quad 0&{}\quad w_1&{}\quad 0\\ 2&{}\quad 0&{}\quad 0&{}\quad 0&{}\quad 0&{}\quad w_2\\ 0&{}\quad 0&{}\quad 1&{}\quad 0&{}\quad \lambda _1&{}\quad 0\\ 0&{}\quad 0&{}\quad 0&{}\quad 1&{}\quad 0&{}\quad \lambda _2 \end{pmatrix} \begin{pmatrix} 2x_2-\frac{9}{2}+2\lambda _2\\ 2x_1-\frac{9}{2}+2\lambda _1\\ -3x_1+2x_2-1+w_1\\ 2x_1-3x_2-1+w_2\\ \lambda _1 w_1\\ \lambda _2 w_2 \end{pmatrix}. \end{aligned}

Now one can check that the point $$x=(1,1),\lambda =(1,1)$$ and $$w=(1,1)$$ is a stationary point, but it is certainly not a zero of H. In this example, we have nonsingularity of the matrix $$J_xL(x,\lambda )=\begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix},$$ but the matrix

\begin{aligned} M(x,\lambda ) = \begin{pmatrix} -3&{}\quad 2 \\ 2&{}\quad -3 \end{pmatrix} \begin{pmatrix} 0&{}\quad \frac{1}{2} \\ \frac{1}{2}&{}\quad 0 \end{pmatrix} \begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix} =\begin{pmatrix} -3&{}\quad 2 \\ 2&{}\quad -3 \end{pmatrix} \end{aligned}

is not semimonotone.

Assumption (A2) can also be used for the nonsmooth reformulation to get (F3), as we show next.

### Proposition 3.3

Let (A2) hold. Then, every stationary point of $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$ satisfying $$\lambda \ge 0$$ and $$h(x) \le 0$$ is a zero of $$\varPsi$$, i.e., (F3) is true.

### Proof

As in Proposition 3.1, we obtain (5) that in turn can be written as

\begin{aligned} \left( M(x,\lambda )^\top \, (-D_h) + (-D_\lambda ) \right) (-\varphi (\lambda ,-h(x))) = 0. \end{aligned}

By (A2), $$-D_h$$ being a nonnegative diagonal matrix implies that the matrix $$M(x{,}\lambda )^\top \, ({-}D_h)$$ is semimonotone. Let us assume by contradiction that $${-}\varphi (\lambda {,}{-}h(x))\ne 0$$. Observing that $${-}\varphi (\lambda {,}{-}h(x)) \ge 0$$ for all $$\lambda \ge 0$$ and $$h(x) \le 0$$, the definition of semimonotonicity yields the existence of an index $$j \in \{1,\ldots ,m\}$$ such that

\begin{aligned} -\varphi (\lambda _{ j},-h_{ j}(x)) > 0, \quad \left( M(x,\lambda )^\top \, (-D_h) (-\varphi (\lambda ,-h(x))) \right) _{ j} \ge 0. \end{aligned}

\begin{aligned} 0&= \left( \left( M(x,\lambda )^\top \, (-D_h) + (-D_\lambda ) \right) (-\varphi (\lambda ,-h(x))) \right) _{ j} \\&= \left( M(x,\lambda )^\top \, (-D_h) (-\varphi (\lambda ,-h(x))) \right) _{ j} \\&\quad + \left( \frac{\lambda _{ j}}{\sqrt{\lambda ^2_{ j} + h_{j}(x)^2}} - 1\right) \varphi (\lambda _{ j},-h_{ j}(x)) > 0, \end{aligned}

where the last inequality follows since $$\varphi (\lambda _{j},-h_{ j}(x)) < 0$$ implies $$\lambda _{ j} > 0$$ and $$h_{ j}(x) < 0$$. Thus, we must have $$\varphi (\lambda ,-h(x)) = 0$$, and then, nonsingularity of $$J_x L(x,\lambda )$$ implies $$L(x,\lambda )=0$$ and $$\varPsi (x,\lambda )=0.$$$$\square$$

In the following part, we consider constrained stationary points. For the merit function $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$, we obtain (F4) under (A1+).

### Theorem 3.1

Let (A1+) hold. Then, every constrained stationary point of $$\frac{1}{2} \Vert \varPsi (x,\lambda )\Vert ^2$$ subject to $$\lambda \ge 0$$ is a zero of $$\varPsi$$, i.e., (F4) is true.

### Proof

Let $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+$$ be a constrained stationary point, meaning

\begin{aligned}&\nabla _x L(x,\lambda ) L(x,\lambda ) - \nabla h(x) D_h \varphi (\lambda ,-h(x)) = 0, \\ \lambda _i>0 \quad \Rightarrow \quad&\nabla _y g_i(x,x)^\top L(x,\lambda ) + D_\lambda ^i \varphi (\lambda _i,-h_i(x)) = 0,\\ \lambda _i=0 \quad \Rightarrow \quad&\nabla _y g_i(x,x)^\top L(x,\lambda ) + D_\lambda ^i \varphi (\lambda _i,-h_i(x)) \ge 0. \end{aligned}

The nonsingularity of $$J_x L(x,\lambda )$$ from (A1+) implies

\begin{aligned} L(x,\lambda ) = \nabla _x L(x,\lambda )^{-1} \, \nabla h(x) \, D_h \, \varphi (\lambda ,-h(x)). \end{aligned}

Hence, if we show that $$\varphi (\lambda ,-h(x)) = 0,$$ we also get $$L(x,\lambda )=0.$$ Let us recall that whenever $$\varphi (\lambda _i,-h_i(x)) \ne 0,$$ the diagonal elements $$D_h^i$$ and $$D_\lambda ^i$$ are negative. Now, we replace $$L(x,\lambda )$$ in the constrained stationarity conditions by the above formula, use the definition of $$M(x,\lambda )$$ from (4) and consider five possible cases:

1. (a)

For $$\lambda _i=0,h_i(x) \le 0$$, we have $$\varphi (\lambda _i,-h_i(x))=0$$ and $$D_h^i \varphi (\lambda _i,-h_i(x))=0.$$

2. (b)

For $$\lambda _i>0, h_i(x)=0$$, we have $$\varphi (\lambda _i,-h_i(x))=0$$ and $$D_h^i \varphi (\lambda _i,-h_i(x))=0.$$

3. (c)

For $$\lambda _i=0, h_i(x) > 0$$, we have $$\varphi (\lambda _i,-h_i(x))>0$$ and the stationarity conditions imply

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) \ge -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) >0. \end{aligned}
4. (d)

For $$\lambda _i>0, h_i(x) < 0$$ we have $$\varphi (\lambda _i,-h_i(x))<0$$ and the stationarity conditions imply

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) = -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) <0. \end{aligned}
5. (e)

For $$\lambda _i>0, h_i(x) > 0$$, we have $$\varphi (\lambda _i,-h_i(x))>0$$ and the stationarity conditions imply

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) = -D_\lambda ^i \varphi (\lambda _i,-h_i(x)) >0. \end{aligned}

Together, we have for all $$i \in \{1,\ldots ,m\}$$ either $$D_h^i \varphi (\lambda _i,-h_i(x))=0$$ or

\begin{aligned} (\varphi (\lambda _i,-h_i(x)) D_h^i) (M_{\bullet i}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) < 0. \end{aligned}

With $$M(x,\lambda )$$ also its transposed matrix is a $$P_0$$ matrix. If $$D_h \varphi (\lambda ,-h(x)) \ne 0,$$ the definition of a $$P_0$$ matrix implies that for at least one index $$j \in \{1,\ldots ,m\}$$ we have $$D_h^j \varphi (\lambda _j,-h_j(x))\ne 0$$ and

\begin{aligned} (\varphi (\lambda _j,-h_j(x)) D_h^j) (M_{\bullet j}(x,\lambda ))^\top D_h \varphi (\lambda ,-h(x)) \ge 0. \end{aligned}

This condition excludes cases (c) to (e), and we must have $$\varphi (\lambda ,-h(x))=0.$$ This completes the proof. $$\square$$

In order to obtain (F5), we use the stronger assumption (A3).

### Theorem 3.2

If (A3) holds, every constrained stationary point of the merit function $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$ subject to $$\lambda \ge 0$$ and $$w \ge 0$$ is a zero of H, i.e., (F5) is true.

### Proof

Let $$(x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+ \times {\mathbb {R}}^m_+$$ be a constrained stationary point, meaning:

\begin{aligned}&\nabla _x L(x,\lambda ) L(x,\lambda ) + \nabla h(x) (h(x)+w) = 0, \\ \lambda _i>0 \quad \Rightarrow \quad&\nabla _y g_i(x,x)^\top L(x,\lambda ) + w_i^2 \lambda _i = 0,\\ \lambda _i=0 \quad \Rightarrow \quad&\nabla _y g_i(x,x)^\top L(x,\lambda ) + w_i^2 \lambda _i \ge 0, \\ w_i>0 \quad \Rightarrow \quad&h_i(x) + w_i + w_i \lambda _i^2 = 0,\\ w_i=0 \quad \Rightarrow \quad&h_i(x) + w_i + w_i \lambda _i^2 \ge 0. \end{aligned}

Nonsingularity of $$J_x L(x,\lambda )$$ implies $$L(x,\lambda ) = - \nabla _x L(x,\lambda )^{-1} \nabla h(x) (h(x) + w).$$ We must consider four possible cases:

1. (a)

For $$\lambda _i> 0, w_i > 0$$, we get

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) = w_i^2 \lambda _i > 0, \quad h_i(x) + w_i = - w_i \lambda _i^2 < 0. \end{aligned}
2. (b)

For $$\lambda _i>0, w_i=0$$, we get

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) = w_i^2 \lambda _i = 0, \quad h_i(x) + w_i \ge - w_i \lambda _i^2 = 0. \end{aligned}
3. (c)

For $$\lambda _i=0, w_i > 0$$, we get

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) \le w_i^2 \lambda _i = 0, \quad h_i(x) + w_i = - w_i \lambda _i^2 = 0. \end{aligned}
4. (d)

For $$\lambda _i=0, w_i = 0$$, we get

\begin{aligned} (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) \le w_i^2 \lambda _i = 0, \quad h_i(x) + w_i \ge - w_i \lambda _i^2 = 0. \end{aligned}

Thus, in all the cases we obtain

\begin{aligned} (h_i(x) + w_i) (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) \le 0. \end{aligned}
(7)

Since $$M(x,\lambda ),$$ and hence also its transposed matrix, is positive semidefinite, we get

\begin{aligned} 0\ge & {} \sum _{i=1}^m (h_i(x) + w_i) (M_{\bullet i}(x,\lambda ))^\top (h(x) + w)\\= & {} (h(x) + w) (M(x,\lambda ))^\top (h(x) + w) \ge 0. \end{aligned}

This implies that for all $$i=1,\ldots ,m$$ we must have

\begin{aligned} (h_i(x) + w_i) (M_{\bullet i}(x,\lambda ))^\top (h(x) + w) = 0. \end{aligned}
(8)

Thus, case (a) cannot hold, and we have $$\lambda \circ w = 0.$$ By (A3), $$M(x,\lambda )$$ is an S matrix, and hence, the system $$M(x,\lambda ) \, z> 0, z>0$$ is solvable. The alternatives theorem of Ville, see , gives that the system

\begin{aligned} M(x,\lambda )^\top v \le 0, \qquad v \ge 0, \qquad v \ne 0 \end{aligned}

is not solvable. From the cases (b) to (d), we have that

\begin{aligned} M(x,\lambda )^\top (h(x)+w) \le 0, \qquad h(x)+w \ge 0. \end{aligned}

Hence, we must have $$h(x)+w=0.$$ This, in turn implies $$L(x,\lambda )=0,$$ and therefore, we have $$H(x,\lambda ,w)=0.$$$$\square$$

Inspecting the proof starting from equation (7), one could also finalize it by assuming that $$M(x,\lambda )$$ is a P matrix. Unfortunately, the P property of $$M(x,\lambda )$$ for all $$\lambda \in {\mathbb {R}}^m_+$$ is a very strong condition, since it implies nonsingularity of the matrix $$M(x,\lambda )$$. This, in turn, can only hold if $$\nabla h(x)$$ and $$J_y g(x,x)$$ are full rank matrices and hence, in particular, requires that we have linear independent constraints.

Let us give an example, where we can apply Theorem 3.2.

### Example 3.2

We consider the QVI defined through

\begin{aligned} F(x):=\begin{pmatrix} 2x_1 \\ 2x_2 \\ 2x_3 \end{pmatrix}, \qquad g(y,x):=\begin{pmatrix} 2y_1+x_2+2x_3+1 \\ x_1+2y_2+x_3+2\\ 2x_1+x_2+2y_3+3 \end{pmatrix}. \end{aligned}

Here, the matrix $$J_x L(x,\lambda )=J_xF$$ is nonsingular. Further, we have

\begin{aligned} M=\begin{pmatrix} 2&{}\quad 1 &{}\quad 2\\ 1&{}\quad 2&{}\quad 1 \\ 2 &{}\quad 1 &{}\quad 2 \end{pmatrix} \begin{pmatrix} 1/2&{}\quad 0&{}\quad 0 \\ 0&{}\quad 1/2&{}\quad 0 \\ 0&{}\quad 0&{}\quad 1/2 \end{pmatrix} \begin{pmatrix} 2&{}\quad 0&{}\quad 0 \\ 0&{}\quad 2&{}\quad 0 \\ 0&{}\quad 0&{}\quad 2 \end{pmatrix} = \begin{pmatrix} 2&{}\quad 1 &{}\quad 2\\ 1&{}\quad 2&{}\quad 1 \\ 2 &{}\quad 1 &{}\quad 2 \end{pmatrix}, \end{aligned}

which is symmetric positive semidefinite. Since all matrix entries of M are positive, the system $$M \, z> 0, \, z > 0$$ can be solved with a vector having all elements equal to 1, and M is an S matrix. Thus, Theorem 3.2 guarantees that (F5) holds.

With the next example, we see that, in contrast to (F4), (A1+) or positive semi-definiteness of $$M(x,\lambda )$$ alone is not sufficient to get (F5).

### Example 3.3

Let us consider the QVI defined through

\begin{aligned} F(x):=\begin{pmatrix} x_1 \\ x_2 \end{pmatrix}, \qquad g(y,x):=\begin{pmatrix} y_1+x_2+2 \\ -2x_1-2y_2+1 \end{pmatrix}. \end{aligned}

This QVI does not even have a feasible point, but a constrained stationary point $${\bar{x}}=(0,0), {\bar{\lambda }}=(0,0), {\bar{w}}=(0,0)$$, since

\begin{aligned} L({\bar{x}},{\bar{\lambda }})=\begin{pmatrix} 0 \\ 0 \end{pmatrix}, \quad h({\bar{x}})= \begin{pmatrix} 2 \\ 1 \end{pmatrix},\quad \text {and} \quad \nabla h({\bar{x}})=\begin{pmatrix} 1&{}-2 \\ 1&{}-2 \end{pmatrix}, \end{aligned}

imply the constrained stationarity conditions

\begin{aligned}&\nabla _x L({\bar{x}},{\bar{\lambda }}) L({\bar{x}},{\bar{\lambda }}) + \nabla h({\bar{x}}) (h({\bar{x}})+{\bar{w}}) = 0, \\ \lambda _i=0 \quad \Rightarrow \quad&\nabla _y g_i({\bar{x}},{\bar{x}})^\top L({\bar{x}},{\bar{\lambda }}) + {\bar{w}}_i^2 {\bar{\lambda }}_i \ge 0, \\ {\bar{w}}_i=0 \quad \Rightarrow \quad&h_i({\bar{x}}) + {\bar{w}}_i + {\bar{w}}_i {\bar{\lambda }}_i^2 \ge 0. \end{aligned}

However, this point is not feasible, hence not a solution of $$H(x,\lambda ,w)=0.$$ Furthermore, in this example we have

\begin{aligned} M=\begin{pmatrix} 1&{}\quad 1 \\ -2&{}\quad -2 \end{pmatrix} \begin{pmatrix} 1&{}\quad 0 \\ 0&{}\quad 1 \end{pmatrix} \begin{pmatrix} 1&{}\quad 0 \\ 0&{}\quad -2 \end{pmatrix} = \begin{pmatrix} 1&{}\quad -2 \\ -2&{}\quad 4\end{pmatrix}, \end{aligned}

which is a $$P_0$$ matrix and positive semidefinite. This shows that the $$P_0$$ property of $$M(x,\lambda )$$, or its positive semidefiniteness, is not sufficient for a constrained stationary point of H to be a solution of $$H(x,\lambda ,w)=0.$$

We can also give an example that has feasible points and where M is positive semidefinite, but nonoptimal constrained stationary points exist.

### Example 3.4

Consider the QVI defined through

\begin{aligned} F(x):=\begin{pmatrix} x_1 \\ x_2+3 \end{pmatrix}, \qquad g(y,x):=\begin{pmatrix} y_1-2x_2+2 \\ -2y_1+x_2+1 \end{pmatrix}. \end{aligned}

This QVI has feasible points, for example (2, 2). Here, the point $${\bar{x}}=(0,0),$$$${\bar{\lambda }}=(0,0), {\bar{w}}=(0,0)$$ is a constrained stationary point, since

\begin{aligned}&h(x)+w = \begin{pmatrix} 2 \\ 1 \end{pmatrix}, \qquad L= \begin{pmatrix} 0 \\ 3 \end{pmatrix}, \qquad \nabla _y g^\top \, L = \begin{pmatrix} 0 \\ 0 \end{pmatrix}, \\&\quad \nabla _x L({\bar{x}},{\bar{\lambda }}) L({\bar{x}},{\bar{\lambda }}) + \nabla h({\bar{x}}) \, (h({\bar{x}})+w)=\begin{pmatrix} 0 \\ 3 \end{pmatrix} + \begin{pmatrix} 1&{}\quad -2 \\ -2&{}\quad 1 \end{pmatrix} \begin{pmatrix} 2 \\ 1 \end{pmatrix} = \begin{pmatrix} 0 \\ 0 \end{pmatrix}. \end{aligned}

Obviously the point is not feasible and hence not a zero of H. We have in this example

\begin{aligned} M=\begin{pmatrix} 1&{}\quad -2 \\ -2&{}\quad 1 \end{pmatrix}\begin{pmatrix} 1&{}\quad -2 \\ 0&{}\quad 0 \end{pmatrix} =\begin{pmatrix} 1&{}\quad -2 \\ -2&{}\quad 4 \end{pmatrix}, \end{aligned}

which is a symmetric positive semidefinite matrix. Hence, this condition is not sufficient for (F5). Moreover, we can see that (A3) is violated, since the system $$M \, z>0, z>0$$ is not solvable.

## Nonsingularity Results for QVIs

In this section, we deal with the nonsingularity facts (F6) and (F7). Let us recall that for $$\lambda \in {\mathbb {R}}^m_{++}$$ the set $$\partial \varPsi (x,\lambda )=\{J \varPsi (x,\lambda )\}$$ is single valued. First, we show that the two facts (F6) and (F7) are equivalent for all x with $$h(x) < 0$$.

### Proposition 4.1

Let an arbitrary $$x \in {\mathbb {R}}^n$$ with $$h(x) < 0$$ be given. Then, $$J\varPsi (x,\lambda )$$ is nonsingular for all $$\lambda \in {\mathbb {R}}^m_{++}$$ if and only if $$JH(x,\lambda ,w)$$ is nonsingular for all $$(\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}$$. This means (F7) is equivalent to (F6) for all $$x \in {\mathbb {R}}^n$$ with $$h(x) < 0$$.

### Proof

For $$h(x) < 0$$ and $$\lambda \in {\mathbb {R}}^m_{++}$$, the matrices $$D_h$$ and $$D_\lambda$$ are negative definite, with $$D_h^i =\frac{-h_i(x)}{\sqrt{h_i(x)^2+\lambda _i^2}}-1 <0$$ and $$D_\lambda ^i =\frac{\lambda _i}{\sqrt{h_i(x)^2+\lambda _i^2}}-1 <0.$$ Thus, the matrix $$J\varPsi (x,\lambda )= \begin{pmatrix} J_x L(x,\lambda ) &{} \nabla _y g(x,x) \\ - D_h Jh(x) &{} D_\lambda \end{pmatrix}$$ is nonsingular, if and only if $$\begin{pmatrix} J_x L(x,\lambda ) &{} \nabla _y g(x,x) \\ Jh(x) &{} - D_h^{-1} \, D_\lambda \end{pmatrix}$$ is nonsingular. For any fixed $$x \in {\mathbb {R}}^n$$ with $$h_i(x) < 0$$ and $$\lambda _i>0,$$ the function

\begin{aligned} \lambda _i \mapsto (D_h^i)^{-1} \, (D_\lambda ^i) = \frac{\lambda _i-\sqrt{h_i(x)^2+\lambda _i^2}}{-h_i(x)-\sqrt{h_i(x)^2+\lambda _i^2}}. \end{aligned}

is continuous with $$\lim _{\lambda _i \downarrow 0} (D_h^i)^{-1} \, (D_\lambda ^i) = + \infty$$ and $$\lim _{\lambda _i \rightarrow \infty } (D_h^i)^{-1} \, (D_\lambda ^i) = 0.$$ Thus, nonsingularity of $$J\varPsi (x,\lambda )$$ for all $$\lambda \in {\mathbb {R}}^m_{++}$$ is equivalent to nonsingularity of the matrices

\begin{aligned} \begin{pmatrix} J_x L(x,\lambda ) &{} \nabla _y g(x,x) \\ Jh(x) &{} - D \end{pmatrix} \end{aligned}
(9)

for all positive definite diagonal matrices $$D \in {\mathbb {R}}^{m \times m}_{++}$$.

Now, since for some $$(\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}$$ we can write any positive definite diagonal matrix in the form $$D=\text {diag}(\lambda )^{-1} \, \text {diag}(w),$$ nonsingularity of (9) for all positive definite diagonal matrices $$D \in {\mathbb {R}}^{m \times m}_{++}$$ is equivalent to nonsingularity of

\begin{aligned} \begin{pmatrix} J_x L(x,\lambda ) &{}\quad \nabla _y g(x,x) \\ Jh(x) &{}\quad - \text {diag}(\lambda )^{-1} \, \text {diag}(w) \end{pmatrix} \end{aligned}

for all $$(\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}.$$ This, in turn, is equivalent to

\begin{aligned} JH(x,\lambda ,w) = \begin{pmatrix} J_x F(x) &{}\quad \nabla _y g &{}\quad 0 \\ Jh(x) &{}\quad 0 &{}\quad I \\ 0 &{}\quad \text {diag}(w) &{}\quad \text {diag}(\lambda ) \end{pmatrix} \end{aligned}

being nonsingular for all $$(\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}$$, and the proof is complete. $$\square$$

The following general necessary and sufficient condition for the nonsingularity of JH was proved in [21, Theorem 3].

### Proposition 4.2

Let $$(x, \lambda , w ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++}$$ be given. Then, the matrix

\begin{aligned} J_xL(x,\lambda ) + \nabla _y g(x,x) \, \text {diag} \big ( w^{-1} \circ \lambda \big ) \, J h(x) \end{aligned}

is nonsingular, if and only if $$JH (x,\lambda ,w)$$ is nonsingular.

In the same spirit, we obtain a similar result for $$J\varPsi$$.

### Proposition 4.3

Let $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}$$ with $$h(x) \ne 0$$ be given. Then, the matrix

\begin{aligned} J_xL(x,\lambda ) +\nabla _y g(x,x) \, D_\lambda ^{-1} \, D_h \, J h(x) \end{aligned}

is nonsingular, if and only if $$J \varPsi (x,\lambda )$$ is nonsingular.

### Proof

Since $$h(x) \ne 0$$ and $$\lambda \in {\mathbb {R}}^m_{++}$$, we have a negative definite diagonal matrix $$D_\lambda$$. For a vector $$v=(v^1,v^2) \in {\mathbb {R}}^{n+m},$$ we get that

\begin{aligned} 0 = J\varPsi (x,\lambda ) v= \begin{pmatrix} J_x L(x,\lambda ) &{} \quad \nabla _y g(x,x) \\ - D_h Jh(x) &{}\quad D_\lambda \end{pmatrix}\begin{pmatrix} v^1\\ v^2 \end{pmatrix} \end{aligned}

is equivalent to

\begin{aligned} v^2= & {} D_\lambda ^{-1} \, D_h \, Jh(x) \, v^1,\\ 0= & {} \left( J_x L(x,\lambda ) + \nabla _y g(x,x) \, D_\lambda ^{-1} \, D_h \, J h(x) \right) v^1. \end{aligned}

Thus, nonsingularity of $$J\varPsi (x,\lambda )$$ is equivalent to the nonsingularity of the matrix $$J_xL(x,\lambda ) + \nabla _y g(x,x) \, D_\lambda ^{-1} \, D_h \, J h(x).$$$$\square$$

We remark that the assumption $$h(x) \ne 0$$ in Proposition 4.3 is mandatory to exploit the reduced matrix given in the proposition. In fact, if $$h_i(x) = 0$$ for even one single element, the matrix $$D_\lambda$$ becomes singular.

The first part of the following result was proved in [21, Corollary 2]. The second one follows from the first one with Proposition 4.1.

### Proposition 4.4

Let (A1++) hold. Then,

1. (a)

$$JH(x,\lambda ,w)$$ is nonsingular for all $$(x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m$$, i.e., (F6) holds;

2. (b)

$$J\varPsi (x,\lambda )$$ is nonsingular for all $$(x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m$$ with $$h(x) < 0$$, i.e., (F7) holds.

For checking nonsingularity of $$JH(x, \lambda , w)$$ for all $$(\lambda ,w) \in {\mathbb {R}}^m_{++} \times {\mathbb {R}}^m_{++},$$ it would be nicer to have a condition that is independent of the choice of $$(\lambda ,w).$$ In this perspective, if (A4) holds (meaning that we have linear constraints with variable right-hand side), we obtain exactly what we want, that is, (A1) does not involve the variables $$\lambda$$, because now, the matrices $$J_xL(x)$$ and M(x) depend exclusively on x. This also means that (A1), (A1+) and (A1++) coincide. Moreover, under (A4), we derive sharper nonsingularity results.

Now, we state our main theorem, a necessary and sufficient condition for the matrix $$JH(x,\lambda ,w)$$ to be nonsingular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++}.$$ This theorem and the following corollary are a nontrivial and fundamental generalization of the results of , where only sufficiency was proved.

### Theorem 4.1

Let $$x \in {\mathbb {R}}^n$$ be given, and let (A4) hold. Then, $$JH(x,\lambda ,w)$$ is non-singular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ if and only if (A5) holds, i.e., for some $$J \subseteq \{1,\ldots ,m\},$$$$d_J:=\det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{} 0 \end{pmatrix} \ne 0$$ and either $$d_J \ge 0$$ for all $$J \subseteq \{1,\ldots ,m\},$$ or $$d_J \le 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$.

### Proof

We have $$\det (JH(x,\lambda ,w)) = \det \begin{pmatrix} J_x F(x) &{} \nabla _y g &{} 0 \\ Jh(x) &{} 0 &{} I \\ 0 &{} \text {diag}(w) &{} \text {diag}(\lambda ) \end{pmatrix}.$$

Developing this determinant by the last m rows, we obtain a sum over the index sets $$J \subseteq \{1,\ldots ,m\},$$ where every addend contains products of $$\lambda _i, i \in J$$ and $$w_i, i \in J^c := \{1,\ldots ,m\} \setminus J$$. For each $$i \in J$$, we drop the corresponding columns of the last block. For each $${\tilde{i}} \in J^c$$, we drop the corresponding column $$\nabla _y g_{\bullet {\tilde{i}}}$$ in the second block and, since we also drop the row with $$\lambda _{{\tilde{i}}}$$, we can develop the remaining matrix after the corresponding column $${\tilde{i}}$$ in the middle right block. This results in dropping the row $$Jh_{{\tilde{i}} \bullet }(x)$$. Further, note that in the determinant formula, we get for every $$i \in J$$ a positive sign, and for every $$i \in J^c$$ we get one negative and one positive sign for the two developments. Thus, we obtain for the determinant of $$JH(x,\lambda ,w)$$

\begin{aligned}&\det (JH(x,\lambda ,w)) \\&\quad = \sum _{J \subseteq \{1,\ldots ,m\}} (-1)^{|J^c|} \left( \prod _{i \in J^c} w_i \right) \left( \prod _{i \in J} \lambda _i \right) \det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ Jh_{J \bullet }(x) &{} 0 \end{pmatrix}\\&\quad = \sum _{J \subseteq \{1,\ldots ,m\}} (-1)^{|J^c|} (-1)^{|J|}\left( \prod _{i \in J^c} w_i \right) \left( \prod _{i \in J} \lambda _i \right) \det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{} 0 \end{pmatrix}\\&\quad = (-1)^{m}\sum _{J \subseteq \{1,\ldots ,m\}} \left( \prod _{i \in J^c} w_i \right) \left( \prod _{i \in J} \lambda _i \right) d_J. \end{aligned}

If $$d_J \ne 0$$ for some $$J \subseteq \{1,\ldots ,m\},$$ and, either $$d_J \ge 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$, or $$d_J \le 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$, i.e., if (A5) holds, we have $$\det (JH(x,\lambda ,w)) \ne 0$$ for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ and hence, $$JH(x,\lambda ,w)$$ is nonsingular.

Now, assume (A5) does not hold. Then, we either have $$d_J=0$$ for all $$J \subseteq \{1,\ldots ,m\},$$ which immediately implies that $$\det (JH(x,\lambda ,w)) = 0$$, or we have index sets $$J_1,J_2 \subseteq \{1,\ldots ,m\}$$ such that

\begin{aligned} d_{J_1}> 0 > d_{J_2}. \end{aligned}

In this case, let us define

\begin{aligned} d_{\max }:= \max _{J \subseteq \{1,\ldots ,m\}} |d_J| \qquad \text {and} \qquad d_{\min }:= \min _{J \subseteq \{1,\ldots ,m\}, |d_J| \ne 0} |d_J|. \end{aligned}

Now we set

\begin{aligned} \lambda _i = \left\{ \begin{array}{ll} 2^m, &{} \text {if } i \in J_1 \cap J_2,\\ z, &{} \text {if } i \in J_1 \setminus J_2,\\ \frac{1}{z}, &{} \text {if } i \in J_2 \setminus J_1,\\ \frac{d_{\min }}{d_{\max }}, &{} \text {else}, \end{array} \right. \quad \text {and} \quad w_i=\left\{ \begin{array}{lllclll} 2^m, &{} \text {if } i \in J_1^c \cap J_2^c,\\ z, &{} \text {if } i \in J_1^c \setminus J_2^c,\\ \frac{1}{z}, &{} \text {if } i \in J_2^c \setminus J_1^c,\\ \frac{d_{\min }}{d_{\max }}, &{} \text {else}. \end{array} \right. \end{aligned}

With these settings $$f:{\mathbb {R}}_{++} \rightarrow {\mathbb {R}},$$

\begin{aligned} f(z):=\sum _{J \subseteq \{1,\ldots ,m\}} \left( \prod _{i \in J^c} w_i \right) \left( \prod _{i \in J} \lambda _i \right) d_J \end{aligned}

is a continuous function. To analyze its behavior for $$z \rightarrow +\infty ,$$ we have to consider the coefficients for the terms with the largest power of z, which is $${|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|}$$. Since $$d_{J_1}>0$$, we obtain for the index set $$J_1$$ the positive addend

\begin{aligned} z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|} \, (2^m)^{|J_1 \cap J_2| + |J_1^c \cap J_2^c|} \, d_{J_1} \end{aligned}

The remaining index sets $${\tilde{J}} \ne J_1$$ leading to the largest power of z must satisfy $$J_1 \setminus J_2 \subseteq {\tilde{J}}$$ and $$J_1^c \setminus J_2^c \subseteq {\tilde{J}}.$$ By construction, we have natural numbers $$\alpha \le |J_1 \cap J_2| + |J_1^c \cap J_2^c|-1$$ and $$\beta \ge 1,$$ such that the corresponding addend satisfies

\begin{aligned}&z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|} \, (2^m)^{\alpha } \, \left( \frac{d_{\min }}{d_{\max }}\right) ^\beta \, d_{{\tilde{J}}} \\&\quad \le z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|} \, (2^m)^{|J_1 \cap J_2| + |J_1^c \cap J_2^c|-1} \, \frac{d_{\min }}{d_{\max }} \, |d_{{\tilde{J}}} |\\&\quad \le z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|} \, (2^m)^{|J_1 \cap J_2| + |J_1^c \cap J_2^c|-1} \, d_{J_1} \end{aligned}

Since the total number of possible index sets $${\tilde{J}} \ne J_1$$ is not greater than $$2^m-1,$$ and

\begin{aligned} (2^m-1) (2^m)^{|J_1 \cap J_2| + |J_1^c \cap J_2^c|-1} \, d_{J_1} < (2^m)^{|J_1 \cap J_2| + |J_1^c \cap J_2^c|} \, d_{J_1}, \end{aligned}

the coefficient of $$z^{|J_1 \setminus J_2| + |J_1^c \setminus J_2^c|}$$ is positive. Thus, we have

\begin{aligned} \lim _{z \rightarrow \infty } f(z) = + \infty . \end{aligned}

Using analogous arguments for the index set $$J_2$$ with $$d_{J_2}<0,$$ we can show that the coefficient for the largest power of $$\frac{1}{z}$$ is negative and hence

\begin{aligned} \lim _{z \rightarrow 0} f(z) = -\infty . \end{aligned}

Now, by continuity of f,  it has a zero. Thus, we can find $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ such that $$\det (JH(x,\lambda ,w)) =0,$$ and hence, $$JH(x,\lambda ,w)$$ is singular, which completes the proof. $$\square$$

This theorem shows that under (A4) the nonsingularity condition (F6) is equivalent to (A5) holding for all $$x \in {\mathbb {R}}^n.$$ Next, we present a new corollary showing that under (A4) and nonsingularity of $$J_xF(x)$$, (A1) is necessary and sufficient for (F6).

### Corollary 4.1

Let $$x \in {\mathbb {R}}^n$$ be given, (A4) hold, and $$J_xF(x)$$ be nonsingular. Then, the matrix $$JH(x,\lambda ,w)$$ is nonsingular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++}$$, if and only if M(x) is a $$P_0$$ matrix.

### Proof

By Theorem 4.1, $$JH(x,\lambda ,w)$$ being nonsingular means that for some $$J \subseteq \{1,\ldots ,m\}$$ we have $$d_J=\det \begin{pmatrix} J_x F(x) &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet }(x) &{} 0 \end{pmatrix} \ne 0$$ and either $$d_J \ge 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$ or $$d_J \le 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$. By the assumption on $$J_xF(x),$$ we have $$d_\emptyset = \det J_xF(x) \ne 0.$$ Further, for $$J \ne \emptyset$$ we can use the determinant formula for block matrices to obtain

\begin{aligned} d_J=\det (J_xF(x)) \, \det ( Jh_{J \bullet }(x) \, J_x F(x)^{-1} \, \nabla _y g_{\bullet J}) =d_\emptyset \, \det ( M_{JJ}(x) ). \end{aligned}

Thus, the nonsingularity of JH is equivalent to requiring that

\begin{aligned} \det ( M_{JJ}(x) ) \ge 0, \qquad \forall \, J \subseteq \{1,\ldots ,m\}, \; J \ne \emptyset . \end{aligned}

This coincides with the fact that M(x) is a $$P_0$$ matrix, and proves the equivalence. $$\square$$

Let us present an example that illustrates the applicability of the corollary, in particular, if we have a linear function F yielding that M(x) becomes independent of x.

### Example 4.1

Define $$F:{\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2$$ and $$g:{\mathbb {R}}^2 \times {\mathbb {R}}^2 \rightarrow {\mathbb {R}}^2$$ by

\begin{aligned} F(x)=\begin{pmatrix} 2x_2-1 \\ 2x_1-1 \end{pmatrix} \qquad \text {and} \qquad g(y,x)=\begin{pmatrix} x_1+2y_2-1 \\ x_2+2y_1-1 \end{pmatrix}. \end{aligned}

The QVI has the unique solution $$\left( \frac{1}{3}, \frac{1}{3} \right) .$$ We have (A4), and $$J_xF=\begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix}$$ is nonsingular. But, the matrix

\begin{aligned} M=M(x) = \begin{pmatrix} 1&{}\quad 2 \\ 2&{}\quad 1 \end{pmatrix} \begin{pmatrix} 0&{}\quad \frac{1}{2} \\ \frac{1}{2}&{}\quad 0 \end{pmatrix} \begin{pmatrix} 0&{}\quad 2 \\ 2&{}\quad 0 \end{pmatrix} =\begin{pmatrix} 1&{}\quad 2 \\ 2&{}\quad 1\end{pmatrix} \end{aligned}

is not a $$P_0$$ matrix for any $$x \in {\mathbb {R}}^n.$$ Hence, by Corollary 4.1, $$JH(x,\lambda ,w)$$ is for all $$x \in {\mathbb {R}}^n$$ singular for some $$\lambda ,w \in {\mathbb {R}}^{m}_{++}$$. For example, one can set $$\lambda _1=\lambda _2=w_1=w_2$$.

Let us further mention that M is semimonotone, and hence, (A2) is satisfied. Thus, Proposition 3.2 yields the stationarity result (F2) in this case, where nonsingularity of $$JH(x,\lambda ,w)$$ is violated.

From Proposition 4.1, we obtain the following corollaries for nonsingularity of $$J\varPsi (x,\lambda )$$:

### Corollary 4.2

Let $$x \in {\mathbb {R}}^n$$ with $$h(x) <0$$ be given, and let (A4) hold. Then, the matrix $$J \varPsi (x,\lambda )$$ is nonsingular for all $$\lambda \in {\mathbb {R}}^{m}_{++},$$ if and only if (A5) holds.

### Corollary 4.3

Let $$x \in {\mathbb {R}}^n$$ with $$h(x) <0$$ be given, (A4) hold and $$J_xF(x)$$ be nonsingular. Then, the matrix $$J \varPsi (x,\lambda )$$ is nonsingular for all $$\lambda \in {\mathbb {R}}^{m}_{++}$$, if and only if M(x) is a $$P_0$$ matrix.

## QVIs with Box Constraints

Now, we discuss the case of QVIs with additional box constraints, i.e.,

\begin{aligned} {\overline{K}}(x) := \{ y \in [l,u] \subseteq {\mathbb {R}}^n : g(y,x) \le 0 \}. \end{aligned}

The KKT conditions become

\begin{aligned}&\displaystyle {L(x,\lambda ):=F(x) + \lambda ^1 - \lambda ^2 + \nabla _y g(x,x) \lambda ^3 = 0,} \\&\quad \displaystyle { \lambda ^1_i \ge 0, \ x_i - u_i \le 0, \ \lambda ^1_i (x_i - u_i) = 0 \,} \quad \forall i = 1, \ldots , n, \\&\quad \displaystyle {\lambda ^2_i \ge 0, \ l_i-x_i \le 0, \ \lambda ^2_i ( l_i-x_i) = 0 \,} \quad \forall i = 1, \ldots , n, \\&\quad \displaystyle {\lambda _i^3 \ge 0, \ g_i(x,x) \le 0,\ \lambda _i^3 g_i(x,x) = 0 \,} \quad \forall i = 1, \ldots , m. \end{aligned}

The constrained nonlinear system in the box constrained case is to solve

\begin{aligned} H(x, \lambda , w) := \left( \begin{array}{c} L(x, \lambda ) \\ x - u + w^1 \\ -x + l + w^2 \\ h(x) + w^3 \\ \lambda ^1 \circ w^1 \\ \lambda ^2 \circ w^2 \\ \lambda ^3 \circ w^3 \end{array} \right) = 0 \qquad (x,\lambda ,w) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_+ \times {\mathbb {R}}^m_+. \end{aligned}

We show that the presence of the box constraints can be exploited in order to obtain the facts (F1)–(F7) under weakened assumptions. First, we will give some sufficient conditions for (A2) that entails (F2) and (F3).

### Proposition 5.1

Let a QVI with box constraints be given and all entries in Jh(x) be nonnegative. If $$J_x L(x,\lambda )$$ is a positive diagonal matrix, the matrix $$M(x,\lambda )$$ is semimonotone, i.e., (A2) is satisfied.

### Proof

To show that

\begin{aligned}&M(x,\lambda )= \begin{pmatrix}I \\ -I \\ J h(x) \end{pmatrix} J_x L(x,\lambda )^{-1} \begin{pmatrix} I&-I&\nabla _y g(x,x) \end{pmatrix}\\&\quad = \begin{pmatrix}J_x L(x,\lambda )^{-1} &{} -J_x L(x,\lambda )^{-1} &{} J_x L(x,\lambda )^{-1}\, \nabla _y g(x,x) \\ -J_x L(x,\lambda )^{-1} &{}J_x L(x,\lambda )^{-1} &{} -J_x L(x,\lambda )^{-1} \nabla _y g(x,x) \\ J h(x) \,J_x L(x,\lambda )^{-1} &{} - J h(x) \, J_x L(x,\lambda )^{-1} &{} J h(x)\, J_x L(x,\lambda )^{-1} \, \nabla _y g(x,x) \end{pmatrix} \end{aligned}

is semimonotone, let a vector $$v=(v^1,v^2,v^3) \in {\mathbb {R}}^{n+n+m}_+ \setminus \{0\}$$ be given.

First, assume $$v^1_j>0$$ for some $$j \in \{1,\ldots ,n\}.$$ Then, we must either have $$M_{j \bullet }(x,\lambda ) v \ge 0$$ (and hence $$M(x,\lambda )$$ is semimonotone), or we get from the jth row $$v^2_j > 0,$$ since $$J_x L(x,\lambda )^{-1}$$ is a positive diagonal matrix and with Jh(x) also all entries in $$\nabla _y g(x,x)$$ are nonnegative. Since the jth row is the negative of the $$(n+j)$$th row, either $$M_{j \bullet }(x,\lambda ) v \ge 0$$ or $$M_{(n+j) \bullet }(x,\lambda ) v \ge 0$$ must hold, and hence, $$M(x,\lambda )$$ must be semimonotone.

Now assume $$v^1=0,$$ and define the set

\begin{aligned} J:=\{j \in \{1,\ldots ,n\} : v^2_j > 0\}. \end{aligned}

If $$J = \emptyset$$, we have $$v^2=0$$. Since the matrix $$J h(x)\, J_x L(x,\lambda )^{-1} \, \nabla _y g(x,x)$$ has only nonnegative entries, it is semimonotone, and hence, $$M(x,\lambda )$$ is semimontone. Thus, assume $$J \ne \emptyset$$. We are done, if one of the components

\begin{aligned} J_x L_{J \bullet }(x,\lambda )^{-1} \, v^2- J_x L_{J \bullet }(x,\lambda )^{-1}\nabla _y g(x,x) \, v^3 \end{aligned}

is nonnegative. Hence, assume they are all negative. For $$J \ne \emptyset$$ the fact that $$J_x L(x,\lambda )^{-1}$$ is a positive diagonal matrix immediately implies that at least one component of $$v^3$$ must be positive. Since all entries in $$J_x L(x,\lambda )^{-1}\nabla _y g(x,x) \, v^3$$ are nonnegative, we have

\begin{aligned} J_x L(x,\lambda )^{-1} \, v^2- J_x L(x,\lambda )^{-1}\nabla _y g(x,x) \, v^3 \le 0 \end{aligned}

Multiplying with the nonpositive matrix $$-Jh(x)$$ results in

\begin{aligned} -Jh(x) \, J_x L(x,\lambda )^{-1} \, v^2+Jh(x) \, J_x L(x,\lambda )^{-1}\nabla _y g(x,x) \, v^3 \ge 0. \end{aligned}

Since this corresponds to the last m rows of $$M(x,\lambda ),$$ the matrix must be semimonotone since one of the components of $$v^3$$ is positive. $$\square$$

The assumption (A3) is restrictive in the case without box constraints and does not hold, for example, if $$M(x,\lambda )$$ consists of two rows, where one is the negative of the other. A different condition can be obtained for QVIs with additional box constraints. Let us denote

\begin{aligned} {\widetilde{g}}(y,x) := \left( \begin{array}{c} y-u \\ l-y \\ g(y,x) \end{array}\right) . \end{aligned}

Here, we can use a positive linear independence assumption, rather than (A3), to obtain a constrained stationarity result.

### Theorem 5.1

Let a QVI with box constraints be given. Assume that for all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{+}$$, the matrix $$J_xL(x,\lambda )$$ is nonsingular and that the matrix $$M(x,\lambda ) = J {\widetilde{h}}(x) J_x L(x,\lambda )^{-1} \nabla _y {\widetilde{g}}(x,x)$$ is positive semidefinite. Further, assume the gradients $$\nabla {\widetilde{h}}_j(x)$$ for $$j \in J:=\{j \in \{1,\ldots ,m\} : {\widetilde{h}}_j(x)>0\}$$ are positive linear independent. Then, every constrained stationary point of the merit function $$\frac{1}{2} \Vert H(x,\lambda ,w)\Vert ^2$$ subject to $$\lambda \ge 0$$ and $$w \ge 0$$ is a zero of H, i.e., (F5) is true.

### Proof

First, we can follow the lines of the proof of Theorem 3.2 to obtain (8) and thus to exclude the case (a), which was $$\lambda _i >0$$ and $$w_i>0$$, through the assumed positive semidefiniteness of $$M(x,\lambda )$$. Thus, we must have $$\lambda \circ w=0$$. Now, we inspect the constrained stationarity conditions for the box constraints. Then, we obtain for every $$i=1,\ldots ,n$$:

\begin{aligned} \begin{array}{lcrcr} \lambda _i^{1}>0 &{}\quad \Rightarrow \quad &{} L_i(x,\lambda ) + (w_i^1)^2 \lambda _i^1 = 0 &{} \quad \Rightarrow \quad &{} L_i(x,\lambda )=0,\\ \lambda _i^{2}>0 &{} \quad \Rightarrow \quad &{} - L_i(x,\lambda ) + (w_i^2)^2 \lambda _i^2 = 0 &{} \quad \Rightarrow \quad &{}L_i(x,\lambda )=0,\\ \lambda _i^{1}=0 &{}\quad \Rightarrow \quad &{} L_i(x,\lambda ) + (w_i^1)^2 \lambda _i^1 \ge 0 &{}\quad \Rightarrow \quad &{}L_i(x,\lambda ) \ge 0. \\ \lambda _i^{2}=0 &{}\quad \Rightarrow \quad &{} - L_i(x,\lambda ) + (w_i^2)^2 \lambda _i^2 \ge 0 &{} \quad \Rightarrow \quad &{} - L_i(x,\lambda ) \ge 0, \end{array} \end{aligned}

If $$\lambda _i^1>0$$ or $$\lambda _i^2>0$$, the first two lines imply $$L_i(x,\lambda )=0.$$ If both $$\lambda _i^1=0$$ and $$\lambda _i^2=0$$, the lines three and four imply again $$L_i(x,\lambda )=0.$$ Since this holds for all $$i=1,\ldots , n,$$ we have $$L(x,\lambda )=0.$$

It remains to show that $${\widetilde{h}}(x)+w=0.$$ From the constrained stationarity conditions, we get with $$\lambda \circ w =0$$:

\begin{aligned} \begin{array}{lcrcr} w_i>0 &{}\quad \Rightarrow \quad &{} {\widetilde{h}}_i(x) + w_i + w_i \lambda _i^2 = 0 &{}\quad \Rightarrow \quad &{} {\widetilde{h}}_i(x) + w_i=0,\\ w_i=0 &{}\quad \Rightarrow \quad &{} {\widetilde{h}}_i(x) + w_i + w_i \lambda _i^2 \ge 0 &{}\quad \Rightarrow \quad &{} {\widetilde{h}}_i(x) \ge 0. \end{array} \end{aligned}

This together with $$L(x,\lambda )=0$$ and $$J:=\{j \in \{1,\ldots ,m\} : {\widetilde{h}}_j(x)>0\}$$ implies that the remaining stationarity condition yields

\begin{aligned} 0 = \nabla _x L(x,\lambda ) L(x,\lambda ) + \nabla {\widetilde{h}}(x) ({\widetilde{h}}(x)+w)=\nabla {\widetilde{h}}_{\bullet J}(x) {\widetilde{h}}_J(x) \end{aligned}

Since $${\widetilde{h}}_J(x)>0$$ and the gradients $$\nabla {\widetilde{h}}_j(x), j \in J$$ are positively linearly independent, we must have $$J=\emptyset$$, which means $${\widetilde{h}}(x)+w=0,$$ and completes the proof. $$\square$$

In the box constrained case, the Jacobian $$J H(x, \lambda ,w)$$ is given by

\begin{aligned} \begin{pmatrix} J_x L(x) &{}\quad I &{}\quad -I &{}\quad \nabla _y g(x,x) &{}\quad 0 &{}\quad 0 &{}\quad 0 \\ I &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad I &{}\quad 0 &{}\quad 0 \\ -I &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad I &{}\quad 0 \\ Jh(x) &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad 0 &{}\quad I \\ 0 &{}\quad \text {diag}(w^1) &{}\quad 0 &{}\quad 0 &{}\quad \text {diag}(\lambda ^1) &{}\quad 0 &{}\quad 0 \\ 0 &{}\quad 0 &{}\quad \text {diag}(w^2) &{}\quad 0 &{}\quad 0 &{}\quad \text {diag}(\lambda ^2) &{}\quad 0 \\ 0 &{}\quad 0 &{}\quad 0 &{}\quad \text {diag}(w^3) &{}\quad 0 &{}\quad 0 &{}\quad \text {diag}(\lambda ^3) \end{pmatrix}. \end{aligned}

Let $$\lambda ,w \in {\mathbb {R}}^{2n+m}_{++}$$ be given, and let us define the matrix

\begin{aligned} \widetilde{J_x L}(x,\lambda ,w):=J_x L(x) +\text {diag}(\lambda ^1 \circ (w^1)^{-1})+\text {diag}(\lambda ^2 \circ (w^2)^{-1}). \end{aligned}

It is not difficult to see that $$JH(x,\lambda ,w)$$ is nonsingular, if and only if

\begin{aligned} \begin{pmatrix} \widetilde{J_x L}(x,\lambda ,w)&{}\quad \nabla _y g(x,x) &{}\quad 0 \\ Jh(x) &{}\quad 0 &{}\quad I \\ 0 &{}\quad \text {diag}(w^3) &{}\quad \text {diag}(\lambda ^3) \end{pmatrix} \end{aligned}

is nonsingular. In this setting, we can replace the nonsingularity of $$J_xL(x,\lambda )$$ from (A1++) by positive semidefiniteness of $$J_xL(x,\lambda )$$ in Proposition 4.4 and in Corollary 4.1. In fact, in this case $$\widetilde{J_x L}(x,\lambda ,w)$$ becomes nonsingular for all $$\lambda , w \in {\mathbb {R}}^m_{++}$$, and the developments follow as in the case without box constraints. We observe that this also holds, if we have either a lower or an upper bound for each variable, and we do not need both of them. Defining

\begin{aligned} \widetilde{M}(x):=Jh(x) \, \widetilde{J_x L}(x,\lambda ,w)^{-1} \nabla _y g(x,x), \end{aligned}
(10)

we get the following results.

### Corollary 5.1

Assume box constraints. Let the matrix $$J_x L(x,\lambda )$$ be positive semi-definite and the matrix $$\widetilde{M}(x)$$ be a $$P_0$$ matrix for all $$(x,\lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}^m_{++}$$. Then,

1. (a)

$$JH(x,\lambda ,w)$$ is nonsingular for all $$(x, \lambda , w) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m \times {\mathbb {R}}_{++}^m$$, i.e., (F6) holds;

2. (b)

$$J \varPsi (x,\lambda )$$ is nonsingular for all $$(x, \lambda ) \in {\mathbb {R}}^n \times {\mathbb {R}}_{++}^m$$ with $$h(x) < 0$$, i.e., (F7) holds.

### Corollary 5.2

Assume we have box constraints, (A4) holds, and $$J_xF(x)$$ is positive semidefinite for all $$x \in {\mathbb {R}}^n$$. Then,

1. (a)

(F6) is equivalent to $$\widetilde{M}(x)$$ being a $$P_0$$ matrix;

2. (b)

(F7) is equivalent to $$\widetilde{M}(x)$$ being a $$P_0$$ matrix for all $$x \in {\mathbb {R}}^n$$ with $$h(x) < 0$$.

Moreover, if $$J_xF(x)$$ is positive definite for all $$x \in {\mathbb {R}}^n$$, then

1. (c)

$$\widetilde{M}(x)$$ is a $$P_0$$ matrix if and only if M(x) is a $$P_0$$ matrix.

### Proof

Point (c) is due to point (a) and Corollary 4.1. $$\square$$

## GNEPs

It was first observed in  that under mild assumptions a GNEP can be reformulated as an equivalent QVI. Hence, we can obtain GNEP results by applying the QVI results from the last sections. We will not state all of them here, but we will focus on special cases, where we can obtain new results.

We consider GNEPs with linear constraints and quadratic cost functions. Let matrices $$Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu }$$ and $$A_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_\nu },$$ and vectors $$c^\nu \in {\mathbb {R}}^{n_\nu }$$ and $$b^\nu \in {\mathbb {R}}^{m_\nu }$$ be given. Further, $$X_\nu$$ is a closed convex subset of $${\mathbb {R}}^{n_\nu }.$$ Each player $$\nu =1,\ldots , N$$ tries to find $$x^\nu \in {\mathbb {R}}^{n_\nu }$$ to solve the problem

\begin{aligned} \min _{x^\nu } \frac{1}{2} (x^\nu )^\top Q_\nu x^\nu + (c^\nu )^\top x^\nu \qquad \text {s.t.} \qquad \sum _{\mu = 1}^N A_{\nu \mu } \, x^\mu \le b^\nu , \quad x^\nu \in X_\nu . \end{aligned}

Notice that the GNEPs considered here have objective functions that depend only on private variables, while the constraints may couple all the variables. We underline that the coupling constraints are not necessarily shared by all the players, and therefore, these GNEPs are, in general, no generalized potential games as defined in [37, 38]. The QVI reformulation of the GNEP is made up by

\begin{aligned} F(x) = \begin{pmatrix} Q_1 &{}&{}\\ &{} \ddots &{} \\ &{}&{} Q_N \end{pmatrix} x{+} \begin{pmatrix} c^1 \\ \vdots \\ c^N \end{pmatrix}{,} \, g(y,x) {=} \begin{pmatrix} A_{11} \, y^1 {+} \sum _{\mu \ne 1} A_{1 \mu } \, x^\mu {-} b^1 \\ \vdots \\ A_{NN} \, y^N {+} \sum _{\mu \ne N} A_{N \mu } \, x^\mu {-} b^N \end{pmatrix}. \end{aligned}

By the linearity of the constraints, we have (A4).

For completeness, let us first mention that the case of linear cost functions, i.e., if we have $$Q_\nu =0$$ for all $$\nu =1,\ldots ,N,$$ is discussed in . In particular, we have in [32, Theorem 4.1] the following result:

### Lemma 6.1

Let $$Q_\nu =0$$ and $$X_\nu = {\mathbb {R}}^{n_\nu }$$ for all $$\nu \in \{1,\ldots ,N\}$$. Assume that $$\det (\nabla _y g_{\bullet J} Jh_{J \bullet }) \ge 0$$ for all $$J \subseteq \{1,\ldots ,m\}$$ with $$|J|=n$$, and that the determinant is positive for one of these sets. Then, $$JH(\lambda ,w)$$ is nonsingular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++}.$$

Exploiting Theorem 4.1, we can extend Lemma 6.1 and get necessary and sufficient conditions for the nonsingularity.

### Corollary 6.1

Let $$Q_\nu =0$$ and $$X_\nu = {\mathbb {R}}^{n_\nu }$$ for all $$\nu \in \{1,\ldots ,N\}$$. $$JH(\lambda ,w)$$ is nonsingular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ if and only if all determinants $$\det (\nabla _y g_{\bullet J} Jh_{J \bullet })$$ with $$J \subseteq \{1,\ldots ,m\}$$ and $$|J|=n$$, are either all nonnegative or all nonpositive, and at least one of them is not 0.

### Proof

By Theorem 4.1, we consider the signs of $$d_J:=\det \begin{pmatrix} 0 &{} \nabla _y g_{\bullet J} \\ -Jh_{J \bullet } &{} 0 \end{pmatrix}.$$ We have $$d_J = 0$$ for every $$J \subseteq \{1,\ldots ,m\}$$ with $$|J| \ne n$$. For $$|J| = n$$ we have

\begin{aligned} d_J= & {} (-1)^n \det \begin{pmatrix} -Jh_{J \bullet } &{} 0 \\ 0 &{} \nabla _y g_{\bullet J} \end{pmatrix} =(-1)^{2n} \det (Jh_{J \bullet }) \det (\nabla _y g_{\bullet J})\\= & {} \det (\nabla _y g_{\bullet J} Jh_{J \bullet }). \end{aligned}

Now, the assertion is a consequence of Theorem 4.1. $$\square$$

In the next example, the matrix $$JH(x,\lambda ,w)$$ is always singular.

### Example 6.1

Consider the linear 2-player GNEP

\begin{aligned} \min -x^1&\quad \text {s.t.} \quad&x^1+x^2 \le 1,\\ \min -x^2&\quad \text {s.t.} \quad&x^1+x^2 \le 1. \end{aligned}

From Lemma 6.1, we get that $$JH(x,\lambda ,w)$$ is singular for some $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ since for $$J=\{1,2\}$$ we get $$\det (\nabla _y g_{\bullet J} Jh_{J \bullet }) = \det \left( \begin{pmatrix} 1&{}0 \\ 0 &{} 1 \end{pmatrix} \begin{pmatrix} 1&{}1 \\ 1 &{} 1 \end{pmatrix} \right) = 0.$$ But even more, in this example $$JH(x,\lambda ,w)$$ is singular for all $$\lambda ,w \in {\mathbb {R}}^{m}_{++},$$ since Next, we consider the more general quadratic setting. However, we restrict ourselves to GNEPs, where all players have the same number of variables, i.e., we assume $$n_1=\ldots =n_N.$$ For some special structures, we get the facts (F1)–(F7).

### Theorem 6.1

Let $$n_1=\ldots =n_N, Q_\nu = \beta _\nu T$$ with $$\beta _\nu \in {\mathbb {R}}_{++}$$ and a positive definite matrix $$T \in {\mathbb {R}}^{n_1 \times n_1}$$, and $$X_\nu = {\mathbb {R}}^{n_1}$$ for all $$\nu \in \{1,\ldots ,N\}$$. Assume $$A_{\nu \mu } = \alpha _\mu (B_{\nu } + E_{\nu \mu })$$ with $$\alpha _\mu \in {\mathbb {R}}$$ and $$B_{\nu }, E_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_1}$$, $$E_{\nu \nu } = 0$$, for all $$\mu , \nu \in \{1,\ldots ,N\}.$$

1. 1.

Let all entries of either $$E_{\nu \mu } T^{-1} B_\mu$$ or $$(B_\nu + E_{\nu \mu }) T^{-1} B_\mu$$ be nonnegative for all $$\mu , \nu \in \{1,\ldots ,N\}$$. Then, M is a semimonotone matrix, i.e., (A2) holds, implying (F2) and (F3).

2. 2.

Let $$E_{\nu \mu } = 0$$ for all $$\mu , \nu \in \{1,\ldots ,N\}$$. Then, M is a $$P_0$$-matrix, i.e., (A1) holds, implying (F1)-(F4), (F6) and (F7).

3. 3.

Let $$E_{\nu \mu } = 0$$, let $$B_{\nu }$$ be a positive matrix, let $$T^{-1}$$ be a nonnegative matrix with a positive diagonal, and let $$\frac{\alpha _\nu ^2}{\beta _\nu } = \frac{\alpha _\mu ^2}{\beta _\mu } > 0$$ for all $$\mu , \nu \in \{1,\ldots ,N\}$$. Then, M is a positive semidefinite S matrix, i.e., (A3) holds, implying that all the facts (F1)-(F7) are true.

### Proof

We start with case 2: $$J_xL(x)= \begin{pmatrix} \beta _1 T &{}&{}\\ &{} \ddots &{} \\ &{}&{} \beta _N T \end{pmatrix},$$ is nonsingular, by the assumed positive definiteness of T and since $$\beta _i>0$$ for all $$i=1,\ldots ,N.$$ Further, we have

\begin{aligned} M =&\begin{pmatrix} \alpha _1 B_1 &{} \cdots &{} \alpha _N B_1 \\ \vdots &{} &{} \vdots \\ \alpha _1 B_N &{} \cdots &{} \alpha _N B_N \end{pmatrix} \begin{pmatrix} \frac{1}{\beta _1} T^{-1} &{}&{}\\ &{} \ddots &{} \\ &{}&{} \frac{1}{\beta _N} T^{-1} \end{pmatrix} \begin{pmatrix} \alpha _1 B_1^\top &{} &{} \\ &{} \ddots &{} \\ &{} &{} \alpha _N B_N^\top \end{pmatrix}\\ =&\begin{pmatrix} \frac{\alpha _1^2}{\beta _1} B_{1} T^{-1} B_{1}^\top &{} \cdots &{} \frac{\alpha _N^2}{\beta _N} B_{1} T^{-1} B_{N}^\top \\ \vdots &{} &{} \vdots \\ \frac{\alpha _1^2}{\beta _1} B_{N} T^{-1} B_{1}^\top &{} \cdots &{} \frac{\alpha _N^2}{\beta _N} B_{N} T^{-1} B_{N}^\top \end{pmatrix}\\ =&{\overline{B}} \, T^{-1} \, {\overline{B}}^\top D, \end{aligned}

where $${\overline{B}} := \begin{pmatrix} B_1 \\ \vdots \\ B_N \end{pmatrix}$$ and $$D := \begin{pmatrix} \frac{\alpha _1^2}{\beta _1} I_{m_1} &{}&{}\\ &{} \ddots &{} \\ &{}&{} \frac{\alpha _N^2}{\beta _N} I_{m_N} \end{pmatrix}.$$ For $$J \subseteq \{1, \ldots , m\}$$, we have

\begin{aligned} \det (M_{JJ}) = \det (({\overline{B}} \, T^{-1} \, {\overline{B}}^\top )_{JJ} D_{JJ}) = \det (({\overline{B}} \, T^{-1} \, {\overline{B}}^\top )_{JJ}) \, \det (D_{JJ}) \ge 0, \end{aligned}

because $${\overline{B}} \, T^{-1} \, {\overline{B}}^\top$$ is positive semidefinite, hence a $$P_0$$ matrix, and D is positive semidefinite and diagonal. Therefore, M is a $$P_0$$ matrix, that is, (A1) is true. As a consequence, (F1) holds by Proposition 3.1 and (F4) by Theorem 3.1. Further, Proposition 4.4 gives (F6) and (F7). Finally, since (A1) implies (A2), we have (F2) from Proposition 3.2 and (F3) from Proposition 3.3.

Now, let us consider case 1: We have

\begin{aligned}&M = {\overline{B}} \, T^{-1} \, {\overline{B}}^\top D + \\&\quad { \begin{pmatrix} 0 &{} \frac{\alpha _2^2}{\beta _2} E_{1 \, 2} T^{-1} B_{2}^\top &{} \cdots &{} \frac{\alpha _N^2}{\beta _N} E_{1 \, N} T^{-1} B_{N}^\top \\ \frac{\alpha _1^2}{\beta _1} E_{2 \, 1} T^{-1} B_{1}^\top &{} 0 &{} &{} \vdots \\ \vdots &{} &{} \ddots &{} \frac{\alpha _{N}^2}{\beta _{N}} E_{N-1 \, N} T^{-1} B_{N}^\top \\ \frac{\alpha _1^2}{\beta _1} E_{N \, 1} T^{-1} B_{1}^\top &{} \cdots &{} \frac{\alpha _{N-1}^2}{\beta _{N-1}} E_{N \, N-1} T^{-1} B_{N-1}^\top &{} 0 \end{pmatrix}}. \end{aligned}

If the first condition holds, M is the sum of a $$P_0$$ matrix and a nonnegative one; thus, it is semimonotone. If the second condition holds, M is nonnegative and then also semimonotone. In any case, (A2) holds. Now, (F2) follows from Proposition 3.2 and (F3) from Proposition 3.3.

In case 3, the matrix $$M = {\overline{B}} \, T^{-1} \, {\overline{B}}^\top D$$, with $$D = \frac{\alpha _1^2}{\beta _1} I$$, is positive semidefinite. The assumptions on $$T^{-1}$$ and $$B_\nu$$ imply that $$M z = \frac{\alpha _1^2}{\beta _1} {\overline{B}} \, T^{-1} \, {\overline{B}}^\top z > 0$$, for any vector $$z > 0$$. Thus, (A3) is true. Fact (F5) comes from Theorem 3.2, while all the others follow as in case 2, since (A3) implies (A1). $$\square$$

### Example 6.2

Let us consider N firms operating in a specific area. Each firm produces its goods in $$n_1$$ spots. Let $$x^\nu \in {\mathbb {R}}^{n_1}_+$$ be the quantities of production of the $$\nu$$th firm, and let its cost be given by

\begin{aligned} cost_\nu (x^\nu ) := (m^\nu )^\top x^\nu + \frac{1}{2} \beta _\nu \sum _{i=1}^{n_1} (x^\nu )^2, \end{aligned}

where $$m^\nu \in {\mathbb {R}}^{n_1}_-$$ are the marginal costs and $$\beta _\nu > 0$$. Let $$\alpha _\nu > 0$$ be a number giving the degree of pollution of the $$n_1$$ plants of the $$\nu$$th firm. Any firm $$\nu$$ must control the level of pollution in $$m_\nu$$ different spots of the area:

\begin{aligned} \alpha _\nu B_\nu \, x^\nu + \sum _{\mu \ne \nu } \alpha _\mu (B_\nu + E_{\nu \mu }) \, x^\mu \le b^\nu , \end{aligned}

where: $$B_\nu \in {\mathbb {R}}^{m_\nu \times n_1}_{++}$$ represents the absolute pollution impact of the $$n_1$$ spots of production of the $$\nu$$th firm on the $$m_\nu$$ spots, where the pollution must be controlled by the firm; $$E_{\nu \mu } \in {\mathbb {R}}^{m_\nu \times n_1}$$ represents some deviations from $$B_\nu$$, and $$(B_\nu + E_{\nu \mu })$$ is the absolute pollution impact of the $$\mu$$th firm on the $$m_\nu$$ spots controlled by the $$\nu$$th firm; $$b^\nu \in {\mathbb {R}}^{m_\nu }_{++}$$ are the maximum levels of pollution allowed in the $$m_\nu$$ spots controlled by the $$\nu$$th firm. Any firm $$\nu$$ minimizes $$cost_\nu$$ subject to its pollution constraints.

Referring to Lemma 6.1, we can prove the following assertions.

1. 1.

Let all $$(B_\nu + E_{\nu \mu }) \ge 0$$, i.e., the absolute pollution impacts of the firms are nonnegative. Then, (A2) holds and the facts (F2) and (F3) are true.

2. 2.

Let $$E_{\nu \mu } = 0$$ for all $$\nu$$ and $$\mu$$, i.e., the plants of all the firms in any specific spot impact equally on the controlled spots. Then, (A1) holds and facts (F1)–(F4), (F6) and (F7) are true.

3. 3.

Let $$E_{\nu \mu } = 0$$ and $$\frac{\alpha _\nu ^2}{\beta _\nu } = \frac{\alpha _\mu ^2}{\beta _\mu }$$ for all $$\nu$$ and $$\mu$$, which is, for example, true if the degrees of pollution $$\alpha _\nu$$ and the parameters $$b_\nu$$ are the same for all the firms. Then, (A3) holds and all the facts (F1)–(F7) are true.

Now we consider a different framework involving box constraints.

### Theorem 6.2

Assume $$n_1=\ldots =n_N, Q_\nu \in {\mathbb {R}}^{n_1 \times n_1}_{+}$$ are nonnegative diagonal matrices, $$X_\nu = [l^\nu , u^\nu ]$$ for all $$\nu =1,\ldots ,N,$$ and $$A_{\nu \mu } = D_{\nu } {\overline{D}}_\mu + E_{\nu \mu },$$ with diagonal matrices $$D_{\nu }, {\overline{D}}_\mu \in {\mathbb {R}}^{n_1 \times n_1}$$ and matrices $$E_{\nu \mu } \in {\mathbb {R}}^{n_1 \times n_1}$$ for all $$\nu , \mu \in \{1,\ldots ,N\}$$.

1. 1.

Let $$Q_\nu$$ be positive diagonal matrices for all $$\nu \in \{1,\ldots ,N\}$$, and $$A_{\nu \mu }$$ be non-negative for all $$\mu , \nu \in \{1,\ldots ,N\}$$. Then, (A2) holds and the facts (F2) and (F3) are true.

2. 2.

Let $$E_{\nu \mu } = 0$$ for all $$\mu , \nu \in \{1,\ldots ,N\}$$. Then, $${\widetilde{M}}$$ is a $$P_0$$-matrix, and the facts (F6) and (F7) are true. If, moreover, $$Q_\nu$$ is positive definite for all $$\nu =1,\ldots ,N$$, (A1) holds and facts (F1)-(F4), (F6) and (F7) are true.

### Proof

Point 1 is due to Proposition 5.1.

Let us consider point 2. Let arbitrary $$\lambda ,w \in {\mathbb {R}}^{3n}_{++}$$ be given. By assumption $$J_xF$$ is positive semidefinite. We set

\begin{aligned} \widetilde{J_x L}(\lambda ,w)=J_x F +\text {diag}(\lambda ^2 \circ (w^2)^{-1})+\text {diag}(\lambda ^3 \circ (w^3)^{-1}) =:\begin{pmatrix} {\widetilde{Q}}_1 &{}&{}\\ &{} \ddots &{} \\ &{}&{} {\widetilde{Q}}_N \end{pmatrix}, \end{aligned}

with positive definite diagonal matrices $${\widetilde{Q}}_\nu \in {\mathbb {R}}^{n_1 \times n_1}, \nu =1,\ldots ,N.$$ We get

\begin{aligned} {\widetilde{M}}= & {} \begin{pmatrix} D_1 {\overline{D}}_1 &{} \cdots &{} D_1 {\overline{D}}_N \\ \vdots &{} &{} \vdots \\ D_N {\overline{D}}_1 &{} \cdots &{} D_N {\overline{D}}_N \end{pmatrix} \begin{pmatrix} {\widetilde{Q}}_1^{-1} &{}&{}\\ &{} \ddots &{} \\ &{}&{} {\widetilde{Q}}_N^{-1} \end{pmatrix} \begin{pmatrix} D_1 {\overline{D}}_1 &{} &{} \\ &{} \ddots &{} \\ &{} &{} D_N {\overline{D}}_N \end{pmatrix}\\= & {} \begin{pmatrix} D_1 {\overline{D}}_1 {\widetilde{Q}}_1^{-1} D_1 {\overline{D}}_1&{} \cdots &{} D_1 {\overline{D}}_N {\widetilde{Q}}_N^{-1} D_N {\overline{D}}_N\\ \vdots &{} &{} \vdots \\ D_N {\overline{D}}_1 {\widetilde{Q}}_1^{-1} D_1 {\overline{D}}_1 &{} \cdots &{} D_N {\overline{D}}_N {\widetilde{Q}}_N^{-1} D_N {\overline{D}}_N \end{pmatrix}\\= & {} C \, C^\top {\check{D}}, \end{aligned}

where $$C := \begin{pmatrix} D_1 \\ \vdots \\ D_N \end{pmatrix}$$ and $${\check{D}} :=\begin{pmatrix} {\overline{D}}_1 {\widetilde{Q}}_1^{-1} {\overline{D}}_1&{} &{} \\ &{} \ddots &{} \\ &{} &{} {\overline{D}}_N {\widetilde{Q}}_N^{-1} {\overline{D}}_N \end{pmatrix}.$$

For arbitrary $$J \subseteq \{1, \ldots , m\},$$ we have

\begin{aligned} \det ({\widetilde{M}}_{JJ}) = \det ((C\, C^\top )_{JJ} {\check{D}}_{JJ}) = \det ((C\, C^\top )_{JJ}) \, \det ({\check{D}}_{JJ}) \ge 0, \end{aligned}

because $$C \, C^\top$$ is positive semidefinite, and $${\check{D}}$$ is positive semidefinite and diagonal. Therefore, $${\widetilde{M}}$$ is a $$P_0$$ matrix. Facts (F6) and (F7) follow from Corollary 5.1.

Moreover, if matrices $$Q_\nu$$ are positive definite, then we can exploit the point (c) of Corollary 5.2 to say that assumption (A1) holds. Then, the assertion can be obtained as in point 2 of Lemma 6.1. $$\square$$

### Remark 6.1

Let us give an example class, where Lemma 6.2 can be applied for GNEPs. Suppose, each player has one variable, and, beside box constraints, we have exactly one constraint, that is shared by all players. Then, this constraint can be stated using $$A_{\nu \mu } = {\overline{D}}_\mu \in {\mathbb {R}}$$ and $$D_\nu =1$$ for all $$\nu ,\mu =1,\ldots ,N.$$ Now, if all $$Q_\nu \in {\mathbb {R}}_{++}$$, the second case of Lemma 6.2 guarantees (F1)–(F4), (F6) and (F7).

### Remark 6.2

Unfortunately, if we have GNEPs with two or more linearly independent shared constraints, the matrix $${\widetilde{M}}$$ can easily fail to be a $$P_0$$-matrix. For example, consider a 2 player game, where each player has one variable and we have, beside box constraints, two shared constraints $$a_1 x^1+a_2 x^2 \le b_1,$$$$a_3 x^1+ a_4 x^2 \le b_2.$$ Then, if $${\tilde{Q}}_1,{\tilde{Q}}_2 > 0$$ we have that $$\widetilde{J_xL}= \begin{pmatrix} {\tilde{Q}}_1 &{} 0 \\ 0 &{} {\tilde{Q}}_2 \end{pmatrix}$$ is nonsingular. Further,

\begin{aligned} \widetilde{M}= & {} \begin{pmatrix} a_1^2 {\tilde{Q}}_1^{-1} &{}\quad a_1a_3 {\tilde{Q}}_1^{-1} &{}\quad a_2^2 {\tilde{Q}}_2^{-1} &{}\quad a_2 a_4 {\tilde{Q}}_2^{-1} \\ a_1 a_3 {\tilde{Q}}_1^{-1} &{}\quad a_3^2 {\tilde{Q}}_1^{-1} &{}\quad a_2 a_4 {\tilde{Q}}_2^{-1} &{}\quad a_4^2 {\tilde{Q}}_2^{-1} \\ a_1^2 {\tilde{Q}}_1^{-1} &{}\quad a_1a_3 {\tilde{Q}}_1^{-1} &{}\quad a_2^2 {\tilde{Q}}_2^{-1} &{}\quad a_2 a_4 {\tilde{Q}}_2^{-1} \\ a_1 a_3 {\tilde{Q}}_1^{-1} &{}\quad a_3^2 {\tilde{Q}}_1^{-1} &{}\quad a_2 a_4 {\tilde{Q}}_2^{-1} &{}\quad a_4^2 {\tilde{Q}}_2^{-1}\end{pmatrix}. \end{aligned}

For the index sets $$I=\{2,3\}$$ and $$J=\{1,4\},$$ we obtain

\begin{aligned} \det (\widetilde{M}_{II})= & {} -a_2a_3 (a_1a_4-a_2a_3) {\tilde{Q}}_1^{-1} {\tilde{Q}}_2^{-1},\\ \det (\widetilde{M}_{JJ})= & {} a_1a_4 (a_1a_4-a_2a_3) {\tilde{Q}}_1^{-1} {\tilde{Q}}_2^{-1}. \end{aligned}

Now we see that $${\widetilde{M}}$$ is not a $$P_0$$ matrix, whenever $$sign(a_1a_4)=sign(a_2a_3) \ne 0.$$ In this case, by point (c) of Corollary 5.2 we obtain that the matrix M is not a $$P_0$$ matrix. Then, Corollary 4.1 shows that $$JH(x,\lambda ,w)$$ fails to be nonsingular for all $$(\lambda ,w) \in {\mathbb {R}}^{2}_{++} \times {\mathbb {R}}^{2}_{++}.$$

Finally, we want to present some GNEPs with box constraints, where (F5) can be shown. We discuss an example class that often occurs in applications.

### Proposition 6.1

Let $$X_\nu = [l^\nu , u^\nu ]$$ for all $$\nu =1,\ldots ,N,$$, and further, only upper constraints are present, i.e., all coefficients of the remaining constraints are positive. Let the feasible set be nonempty. If all $$Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu },\nu =1,\ldots ,N$$ are nonsingular and M is positive semidefinite, (F5) holds.

### Proof

Since all $$Q_\nu \in {\mathbb {R}}^{n_\nu \times n_\nu },\nu =1,\ldots ,N$$ are nonsingular, $$J_xL=J_xF$$ is nonsingular. Since M is positive semidefinite by assumption, we can apply Theorem 5.1, if we have shown that the columns of the matrix $$\nabla h_{\bullet J}(x)$$ with $$J=\{j \in \{1,\ldots ,m\} : h_j(x)>0\}$$ are positively linearly independent for all $$x \in {\mathbb {R}}^n$$. Assume we have a vector $$v \in {\mathbb {R}}^{|J|}_{+}$$ with $$\nabla h_{\bullet J}(x) v = 0.$$ We want to show that $$v=0$$ must hold.

Whenever J contains an upper box constraint with index $$j_0$$, the corresponding lower box constraint is not in J. Hence, the matrix $$\nabla h_{\bullet J}(x)$$ has a corresponding row, with only nonnegative entries, and we must have $$v_{j_0}=0$$. Thus, we can assume without loss of generality that J does not contain any upper box constraint.

Since the feasible set is nonempty, the set $$J=\{j \in \{1,\ldots ,m\} : h_j(x)>0\}$$ cannot contain all lower box constraints and an additional upper constraint. Hence, the matrix $$\nabla h_{\bullet J}(x)$$ can either contain only lower box constraints, resulting in $$v=0,$$ or it must contain at least one row where all entries are nonnegative. But then, the components of v corresponding to the positive coefficients of the upper constraints must all be zero. The remaining matrix has only lower box constraints, again resulting in $$v=0.$$ Thus, we must have $$v=0,$$ meaning that the columns of $$\nabla h_{\bullet J}(x)$$ are positively linear independent for all $$x \in {\mathbb {R}}^n$$. By Theorem 5.1, fact (F5) holds. $$\square$$

## Conclusions

In this paper, we presented a number of necessary and sufficient conditions for the main issues, when solving QVIs via a smooth constrained or a nonsmooth unconstrained equation reformulation of their KKT conditions. We have seen that the matrix M plays a crucial role. The first issue is to guarantee that stationary points of the merit function are indeed solutions. We considered several concepts of constrained or unconstrained stationary points, and the issue was solved under a $$P_0$$ property for the nonsmooth reformulation. For the constrained, but smooth, reformulation we required stronger assumptions, namely positive semidefiniteness of M together with the S property of this matrix. Further, we gave an example, showing that the $$P_0$$ property is not sufficient in the smooth reformulation. Considering only stationary points within the feasible set, which is important for interior point methods, we showed the stationarity result for both reformulations, if M is a semimonotone matrix.

For the second issue, the nonsingularity of the Jacobian at feasible iterates, we considered the QVI class, where the constraints are linear with variable right-hand side. Here, we proved equivalence of the nonsingularity with a sign condition for determinants of certain matrices, which are independent of the dual variables. Furthermore, under an additional nonsingularity assumption on $$J_x F(x),$$ we showed equivalence to the $$P_0$$ property of M. In the context of QVIs, these are, to the best of our knowledge, the first necessary and sufficient conditions for nonsingularity of the Jacobians. We have also seen that, for QVIs with box constraints, some of the assumptions can be weakened.

We further presented some subclasses of GNEPs with linear constraints and quadratic cost functions, for which the Jacobian is always nonsingular, and the absence of nonoptimal stationary points can be shown. Moreover, we extended a known sufficient nonsingularity condition for GNEPs with linear constraints and cost functions to a necessary and sufficient condition.