Now it is time to present the main results of the paper.
Theorem 1
Let \(\mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_{n+1}\) be a (minimal) positive basis for \(\mathbb {R}^n\). Then the following holds:
-
1.
Its cosine measure is bounded above by 1 / n.
-
2.
The positive bases of unit vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) with the property that
$$\begin{aligned} \mathbf {v}_i^T \mathbf {v}_j=-1/n \end{aligned}$$
for \(i\ne j\) are the only positive bases of unit vectors that has cosine measure 1 / n.
Proof
We first prove (1). Without loss of generality we assume that the vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) are unit vectors. Form the Gram matrix
$$\begin{aligned} \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1})= \begin{pmatrix} \mathbf {v}_1^T \mathbf {v}_1 &{}\quad \dots &{}\quad \mathbf {v}_1^T \mathbf {v}_{n+1} \\ \vdots &{}\quad &{}\quad \vdots \\ \mathbf {v}_{n+1}^T \mathbf {v}_1 &{}\quad \dots &{}\quad \mathbf {v}_{n+1}^T \mathbf {v}_{n+1} \end{pmatrix}. \end{aligned}$$
The Gram matrix \(\mathbf {G}\) is a positive semidefinite matrix with rank n with ones on the diagonal. Since \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) is a positive spanning set there exists \(\alpha _1,\dots ,\alpha _{n+1} \ge 1\) such that \(\sum _{i=1}^{n+1} \alpha _i \mathbf {v}_i=0\). This implies that the vector \(\left( {\begin{matrix} \alpha _1 \dots \alpha _{n+1} \end{matrix}}\right) ^T\) is an eigenvector of \(\mathbf {G}\) with eigenvalue 0. Without loss of generality we can, if neccesary, reorder the vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) and scale the \(\alpha \)’s such that \(1=\alpha _1 \le \alpha _i\), \(i >1\).
Let \(\mathbf {u} \in \mathbb {R}^n\) be a unit vector and extend the Gram matrix \(\mathbf {G} \in \mathbb {R}^{(n+1)\times (n+1)}\) to a Gram matrix
$$\begin{aligned} \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1},\mathbf {u})= \begin{pmatrix} \mathbf {v}_1^T \mathbf {v}_1 &{} \quad \dots &{}\quad \mathbf {v}_1^T \mathbf {v}_{n+1} &{} \quad \mathbf {v}_1^T \mathbf {u} \\ \mathbf {v}_2^T \mathbf {v}_1 &{}\quad \dots &{} \quad \mathbf {v}_2^T \mathbf {v}_{n+1} &{} \quad \mathbf {v}_2^T \mathbf {u} \\ \vdots &{}\quad &{} \quad \vdots &{}\quad \vdots \\ \mathbf {v}_{n+1}^T \mathbf {v}_1 &{}\quad \dots &{}\quad \mathbf {v}_{n+1}^T \mathbf {v}_{n+1} &{}\quad \mathbf {v}_{n+1}^T \mathbf {u} \\ \mathbf {u}^T \mathbf {v}_1 &{} \quad \dots &{}\quad \mathbf {u}^T \mathbf {v}_{n+1} &{}\quad \mathbf {u}^T \mathbf {u} \\ \end{pmatrix}. \end{aligned}$$
(2)
The matrix \(\mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1},\mathbf {u})\) is a positive semidefinite matrix, again with rank n. Since \(\mathbf {v}_2,\dots ,\mathbf {v}_{n+1}\) is a basis, there exists a vector \(\mathbf {u}\) as described above where
$$\begin{aligned} \mathbf {v}_2^T \mathbf {u}= \mathbf {v}_3^T \mathbf {u}= \dots = \mathbf {v}_{n+1}^T \mathbf {u}= \gamma >0. \end{aligned}$$
Since \(\gamma >0\), it follows that \(\beta =\mathbf {v}_1^T \mathbf {u} <0\) since \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) is a positive basis. Since \(\mathbf {v}_2,\dots ,\mathbf {v}_{n+1}\) form a basis for \(\mathbb {R}^n\) the latter statement is true for any \(\gamma \), but \(\mathbf {u}\) will be a unit vector only for one positive value of \(\gamma \). We can obtain a bound for \(\gamma \) by noting that the matrix
$$\begin{aligned} \begin{pmatrix} \mathbf {v}_1^T \mathbf {v}_1 &{}\quad \mathbf {v}_1^T \mathbf {v}_2&{}\quad \dots &{}\quad \mathbf {v}_1^T \mathbf {v}_{n+1} &{}\quad \beta \\ \mathbf {v}_2^T \mathbf {v}_1 &{}\quad \mathbf {v}_2^T\mathbf {v}_2 &{}\quad \dots &{}\quad \mathbf {v}_2^T \mathbf {v}_{n+1} &{}\quad \gamma \\ \vdots &{}\quad &{} \quad \vdots &{}\quad \vdots \\ \mathbf {v}_{n+1}^T \mathbf {v}_1 &{}\quad \mathbf {v}_{n+1}^T \mathbf {v}_2&{}\quad \dots &{}\quad \mathbf {v}_{n+1}^T \mathbf {v}_{n+1} &{}\quad \gamma \\ \beta &{}\quad \gamma &{}\quad \dots &{}\quad \gamma &{}\quad 1 \\ \end{pmatrix} \end{aligned}$$
is a positive semidefinite matrix only if the vector \(\bigl ( {\begin{matrix} \beta&\gamma&\dots&\gamma \end{matrix}}\bigr )^T\) is orthogonal to the kernel of \(\mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1})\). This means that \(\beta \alpha _1+\sum _{j=2}^{n+1} \gamma \alpha _j=0\). Since \(\alpha _1=1\), solving for \(\gamma \) gives \(\gamma =\frac{-\beta }{\sum _{j=2}^{n+1} \alpha _j}\). Since \(0> \beta \ge -1\) and \(\alpha _j \ge 1\) we obtain the bound \(\gamma \le \frac{1}{n}\) which shows that the cosine measure of \(\mathbf {v}_1,\dots ,\mathbf {v}_{n+1}\) is bounded above by 1 / n. The fact that \(\beta \ge -1\) follows for instance by applying the Cauchy–Schwarz inequality to \(\mathbf {v}_1^T \mathbf {u}\).
Proof of (2): \(\gamma =1/n\) only if \(\alpha _2=\dots =\alpha _n=1\) and \(\beta =-1\). In the first part of the proof, the vectors \(\mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_{n+1}\) was, if necessary, reordered such that \(\mathbf {v}_1\) was chosen such that \(\alpha _1=1\). It is clear that in this case any of the vectors \(\mathbf {v}_i\) can be interchanged with \(\mathbf {v}_1\). Moreover, since all principal submatrices of a positive semidefinite matrix are again positive semidefinite this can only hold if \(\mathbf {v}_i^T \mathbf {v}_j=-1/n\) for all \(i\ne j\). This follows by considering the principal submatrix
$$\begin{aligned} \mathbf {G}(\mathbf {v}_i,\mathbf {v}_j,\mathbf {u})= \begin{pmatrix} 1 &{}\quad \mathbf {v}_i^T \mathbf {v}_j &{}\quad -1 \\ \mathbf {v}_j^T \mathbf {v}_i &{}\quad 1 &{}\quad 1/n \\ -1 &{} \quad 1/n &{} \quad 1 \end{pmatrix} \end{aligned}$$
of the matrix given in (2). Since we have, \(\mathbf {u}^T \mathbf {v}_i=-1\) it follows that \(\mathbf {v}_i=-\mathbf {u}\). But then \(\mathbf {u}^T \mathbf {v}_j=1/n\) implies that \(\mathbf {v}_j^T \mathbf {v}_i^T=-1/n\). (The fact that \(\det \mathbf {G}(\mathbf {v}_i,\mathbf {v}_j,\mathbf {u}) = - (\mathbf {v}_i^T \mathbf {v}_j)^2 - (2/n)(\mathbf {v}_i^T \mathbf {v}_j) - (1/n^2) = - (\mathbf {v}_i^T \mathbf {v}_j + (1/n))^2\) must be nonnegative, since the determinant of a positive semidefinite matrix is always nonnegative, also provides the result.)
That there indeed exists a positive basis of unit vectors \(\mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_{n+1} \in \mathbb {R}^n\) such that \(\mathbf {v}_i^T \mathbf {v}_j = -1/n\) for all i, j is proven in [1, 5]. \(\square \)
Since every positive spanning set of \(\mathbb {R}^n\) of size \(n+1\) must be a positive basis, we also can state that:
Corollary 1
The maximal cosine measure for any positive spanning set of size \(n+1\) in \(\mathbb {R}^n\) is 1 / n.
Now we proceed to show the result for maximal positive bases. The maximal positive bases in \(\mathbb {R}^n\) has size 2n and it consists of a set of n linearly independent vectors and their negatives (which can be multiplied by an arbitrary positive constant). The structure for maximal positive bases is conveniently stated in [12, Theorem 6.3]. They can be listed as
$$\begin{aligned} \mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_n,-\delta _1 \mathbf {v}_1,-\delta _2 \mathbf {v}_2,\dots ,-\delta _n \mathbf {v}_n \end{aligned}$$
where \(\delta _i>0\) for \(1\le i \le n\).
Theorem 2
Let
$$\begin{aligned} \mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_n,-\delta _1 \mathbf {v}_1,-\delta _2 \mathbf {v}_2,\dots ,-\delta _n \mathbf {v}_n \end{aligned}$$
where \(\delta _i>0\) for \(1\le i \le n\) be a (maximal) positive basis for \(\mathbb {R}^n\). Then the following holds:
-
1.
The cosine measure is bounded above by \(1/\sqrt{n}\).
-
2.
The bound is attained if and only if \(\mathbf {v}_i^T \mathbf {v}_j=0\) for \(i\ne j\) and \(1\le i,j \le n\).
Proof
By the definition of the cosine measure we will need the scaled positive basis vector \(\frac{\mathbf {v}_i}{\Vert \mathbf {v}_i\Vert }\). Therefore we can without loss of generality we assume that \(\mathbf {v}_1,\dots ,\mathbf {v}_n\) are unit vectors and \(\delta _i=1\) for \(1 \le i \le n\). The proof proceeds by showing that by making a proper selection of \(\mathbf {v}_i\) or \(-\mathbf {v}_i\) for \(1\le i \le n\) there exist a unit vector \(\mathbf {u}\) with the property that \(\mathbf {u}^T \mathbf {v}_1=\mathbf {u}^T \mathbf {v}_2= \dots = \mathbf {u}^T \mathbf {v}_n \le 1/\sqrt{n}\).
That there exists a vector \(\mathbf {u}\) such that \(\mathbf {u}^T \mathbf {v}_1=\mathbf {u}^T \mathbf {v}_2 = \dots = \mathbf {u}^T \mathbf {v}_n = \beta \) for any \(\beta \) follows immediately since the vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_n\) are linearly independent. If \(\beta =0\), the vector \(\mathbf {u}=\mathbf {0}\). Otherwise, \(\mathbf {u}\ne \mathbf {0}\), and scaling \(\mathbf {u}\) to be a unit vector of course also scales \(\beta \). The choice of \(\beta \) that provides a solution \(\mathbf {u}\) that is a unit vector can be found by considering the Gram matrix \(\mathbf {G}(\mathbf {v}_1,\mathbf {v}_2,\dots ,\mathbf {v}_n,\mathbf {u})\). This matrix has the required properties if \(\mathbf {u}^T \mathbf {u}=1\) and \(\mathbf {u}^T \mathbf {v}_1=\mathbf {u}^T \mathbf {v}_2 \dots = \mathbf {u}^T \mathbf {v}_n =\beta \) if \(\beta =1/\sqrt{\mathbf {e}_n^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_n)^{-1}\mathbf {e}_n}\) where \(\mathbf {e}_n\) is the vector of length n with all its entries equal to one. The required upper bound will follow from Lemma 1 if we can select the vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_n\) such that \(\mathbf {e}_n^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_n) \mathbf {e}_n \le n\) by replacing some \(\mathbf {v}_i\)’s by \(-\mathbf {v}_i\) if necessary.
We will prove by induction that this is possible for all \(k\le n\) to construct \(\mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_k)\) such that \(\mathbf {e}_k^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_k) \mathbf {e}_k \le k\). For \(k=1\) there is nothing to prove, since \(\mathbf {G}(\mathbf {v}_1)=1\). For \(k=2\) the result follows by replacing \(\mathbf {v}_2\) with \(-\mathbf {v}_2\) if \(\mathbf {v}_1^T \mathbf {v}_2>0\). Let us assume that \(\mathbf {e}_{k-1}^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{k-1}) \mathbf {e}_{k-1} \le k-1\). Since
$$\begin{aligned} \mathbf {e}_k^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_k) \mathbf {e}_k = \mathbf {e}_{k-1}^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_{k-1}) \mathbf {e}_{k-1} + 2\left( \mathbf {v}_k^T \mathbf {v}_1 + \dots + \mathbf {v}_k^T \mathbf {v}_{k-1}\right) +1, \end{aligned}$$
we see that the induction step is completed by replacing \(\mathbf {v}_k\) with \(-\mathbf {v}_k\) if \((\mathbf {v}_k^T \mathbf {v}_1 + \dots + \mathbf {v}_k^T \mathbf {v}_{k-1}) = \mathbf {v}_k^T( \mathbf {v}_1 + \dots + \mathbf {v}_{k-1}) >0\), but keeping \(\mathbf {v}_k\) otherwise.
To prove the second part of the theorem, one needs to do the minor change in the provided construction that one numbers the vectors such that \(\mathbf {v}_1^T \mathbf {v}_2 \ne 0\), if no such pair of vectors exists, the proof is done. Otherwise, select \(\mathbf {v}_1\) and \(\mathbf {v}_2\) such that \(\mathbf {v}_1^T \mathbf {v}_2 < 0\). This leads to \(\mathbf {e}_2^T \mathbf {G}(\mathbf {v}_1,\mathbf {v}_2) \mathbf {e}_2 < 2\). Now we can repeat the induction step shown above, but we will have \(\mathbf {e}_k^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_k) \mathbf {e}_k < k\), which in the end leads to \(e_n^T \mathbf {G}(\mathbf {v}_1,\dots ,\mathbf {v}_n)e_n < n\), and therefore we get \(\beta < 1/\sqrt{n}\) by Lemma 1. \(\square \)