Abstract
Using the connection between ellipsoids and positive semidefinite matrices we provide alternative proofs to some recently proven inequalities concerning the volume of \(L_2\) zonoids as consequences of classical inequalities for matrices.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction and background
Many connections exist between the theory of matrices and the theory of convex bodies. Within the realm of convex geometry, the fundamental Brunn–Minkowski inequality [14, 22, Section 7.1], and the Aleksandrov–Fenchel inequality [22, Section 7.3] have versions for positive semidefinite matrices. Appropriate analogue versions of the classical Bergstrom and Ky-Fan inequalities within the theory of matrices have been studied for convex bodies, as well as inequalities for mixed volumes have been investigated in the context of mixed discriminants (see e.g. [1, 12, 13, 15, 19]). The \(L_2\) Brunn–Minkowski theory of convex bodies provides us with a correspondence between positive semidefinite matrices and ellipsoids, which takes the usual sum of matrices to the so-called \(L_2\) sum of ellipsoids into account.
This note aims to observe that some recently obtained inequalities for the particular family of convex bodies known as \(L_2\) zonoids correspond to classical matrix inequalities. On the other hand, a characterization result included in (one of) the mentioned recently proven geometrical inequalities allows to establish directly the equality conditions in one of the classical matrix inequalities, which, to the best of the authors’ knowledge, seem not to have been explicitly documented in the literature.
Let \(\mathcal {M}^{n}\) be the vector space of real square symmetric matrices, and let the set of positive semidefinite symmetric ones be denoted by \(\mathcal {S}^{n}_{+}\). It is well-known (see e.g. [18]) that \(\mathcal {S}^{n}_{+}\) is a closed, convex cone. Next, we recall the Bergstrom and Ky Fan classical matrix inequalities.
Theorem 1.1
(Bergstrom’s inequality) [5,6,7] Let A and B be two \(n\times n\) positive definite real symmetric matrices, and denote by \(A_i\) and \(B_i\) the two \((n-1)\times (n-1)\) matrices resulting from A and B by deleting the i-th row and the i-th column. Then we have
for every \(i\in \{1,\ldots ,n\}\).
Theorem 1.2
(Ky Fan’s inequality) [5, 10] Let A and B be two \(n\times n\) positive definite real symmetric matrices, and denote by \(A_{(k)}\) and \(B_{(k)}\) the principal \(k\times k\) matrices of A and B obtained by taking the first k rows and k columns from A and B, respectively. Then we have
for every \(k\in \{1,\ldots ,n-1\}\).
Inequalities (1.1) and (1.2) have motivated a number of questions concerning quotients of sums of quermassintegrals of convex bodies, in particular, volume and surface area, see e.g. [12, 13, 15], and the references therein.
The next result is Brunn–Minkowski’s (or Minkowski’s) inequality for positive semidefinite symmetric matrices.
Theorem 1.3
[19, Theorem 7.8.21] Let \(A,B\in \mathcal {M}^{n}\) be positive definite matrices. Then
for any \(\lambda \in [0,1]\), with equality if and only if \(A=cB\), for some \(c>0\).
Next, we state the geometric analogue of the latter, namely, the Brunn–Minkowski inequality for the volume of convex bodies. For that, we first need to introduce some further notation.
In the n-dimensional Euclidean space \(\mathbb {R}^n\), endowed with the standard inner product \(\langle \cdot ,\cdot \rangle \) and the associated Euclidean norm \(||\cdot ||\), we denote by \(\mathcal {K}^n\) the set of all convex bodies, i.e., compact convex sets in \(\mathbb {R}^n\). We set \(B_n\) to be the n-dimensional unit ball, and \({\mathbb S}^{n-1}\) to be its boundary, the unit sphere of \(\mathbb {R}^n\). Let \(1\le k\le n\), we denote by \(\mathcal {L}^n_k\) the set of all k-dimensional linear subspace of \(\mathbb {R}^n\). If \(L\in \mathcal {L}^n_k\) and \(K\in \mathcal {K}^n\), then we denote by \(P_L(K)\subset L\) the orthogonal projection of K onto L, which is also a convex body. The subset of \(\mathcal {K}^n\) consisting of all convex bodies containing the origin will be denoted by \(\mathcal {K}_0^n\), meanwhile, the subset of \(\mathcal {K}_0^n\) consisting of all convex bodies containing the origin in their interior is denoted by \(\mathcal {K}^n_{(o)}\). The volume of a measurable set \(M\varsubsetneq \mathbb {R}^n\), i.e., its n-dimensional Lebesgue measure, is denoted by \(\textrm{vol}_n(M)\). If M is contained in a k-dimensional affine subspace of \(\mathbb {R}^n\), we will write \(\textrm{vol}_k(K)\) to denote its k-dimensional volume.
The Minkowski sum of the convex bodies K, H is defined as \(K+H:=\{x+y : x\in K, y\in H\}\). Moreover, if \(\alpha \ge 0\), then \(\alpha K:=\{\alpha x : x\in K\}\). For every \(K,H\in \mathcal {K}^n\) and \(\alpha ,\beta \ge 0\), we have that \(\alpha K+\beta H\) is again a convex body.
The Brunn–Minkowski inequality provides us with the concavity of the n-th root of the volume with respect to the Minkowski sum. The Brunn–Minkowski inequality is the content of the following theorem, which is a cornerstone of the classical Brunn–Minkowski theory.
Theorem 1.4
[22, Theorem 7.1.1] Let \(K,H\in \mathcal {K}^n\) be two convex bodies. Then, for \(\lambda \in [0,1]\)
Equality for some \(\lambda \in (0,1)\) holds if and only if K and H either lie in parallel hyperplanes or are homothetic.
We refer the interested reader to [14, 22] for details and a wealth of results and contributions to the Brunn–Minkowski theory. Considering other additions of convex bodies, other than the vectorial one, far-reaching extensions of the classical Brunn–Minkowski theory have emerged [22, Chapter 9]. An example of those is the \(L_p\) Brunn–Minkowski theory.
The mentioned \(L_2\) sum of convex bodies -which contain the origin- is just a particular case of the more general \(L_p\) sum, defined via the support function of a convex body, within the \(L_p\) Brunn–Minkowski theory. With the aim of introducing the latter precisely, we need some further background on the theory of convex bodies.
Given a convex body \(K\in \mathcal {K}^n\), the support function of K in the direction \(x\in \mathbb {R}^n\) is defined as \(h(K,x)=\max \{ \langle x,y\rangle :\ y\in K \}\), and it describes K uniquely. If \(f:\mathbb {R}^n \rightarrow \mathbb {R}\) is a positively 1-homogeneous and sub-additive function in \(\mathbb {R}^n\), then there exists a unique convex body \(K\in \mathcal {K}^n\), such that \(f(x)=h(K,x)\), for every \(x\in \mathbb {R}^n\), see [22, Theorem 1.7.1].
Now, for convex bodies K, H containing the origin, the latter ensures that the p-mean of the support functions of K and H provides us with the support function of a new convex body, \(K+_pL\), called the \(L_p\) sum of K and H. More precisely, for \(x\in \mathbb {R}^n\) and \(K,H\in \mathcal {K}^n_0\) (see [22, Chapter 9]), the function \(f:\mathbb {R}^n\rightarrow \mathbb {R}\), given by
is the support function of \(K+_p H\).
Observe that the origin belongs to \(K\in \mathcal {K}^n\), i.e., \(K\in \mathcal {K}^n_0\), if and only if \(h(K,\cdot )\ge 0\). Let furthermore \(\lambda \cdot _p K:=\lambda ^{1/p}K\), i.e., \(h^p(\lambda \cdot _p K,x)=\lambda h^p(K,x)\), for \(x\in \mathbb {R}^n\). With this notation, the following inequality is known as the \(L_p\) Brunn–Minkowski inequality. Although we will only use it in the case \(p=2\), we establish it in the general case \(p\ge 1\), for completeness. For \(p=1\) we recover the Brunn–Minkowski inequality above.
Theorem 1.5
(\(L_p\) Brunn–Minkowski) [22, Corollary 9.1.5] Let \(K,H\in \mathcal {K}^n_{(0)}\) be two convex bodies containing the origin. Then
for \(\lambda \in [0,1]\) and \(p\ge 1\). Equality holds if and only if K and H are dilates of each other.
When \(p=2\), using the explicit expression of the support function of an ellipsoid, there is a correspondence between positive definite matrices and ellipsoids, which involves the \(L_2\) sum. In the next, we follow [11] to denote \(\mathcal {E}^n\) the set of all ellipsoids centered at the origin in \(\mathbb {R}^n\), i.e., \(E\in \mathcal {E}^n\) if there is a linear map \(T:\mathbb {R}^n\longrightarrow \mathbb {R}^n\), such that \(E=T\left( B_n\right) =:TB_n\). Indeed, if \(E\in \mathcal {E}^n\) and \(T:\mathbb {R}^n \longrightarrow \mathbb {R}^n\) is a linear map such that \(E=TB_n\), then the support function of E is given by \(h(E,x)=h(TB_n,x) =\max \{ \langle y,x \rangle : y\in TB_n \}=\max \{ \langle Tb,x \rangle : b\in B_n \},\) for \(x\in \mathbb {R}^n\). Thus,
The matrix \(A=TT^T\) defines uniquely an element in the space of positive semidefinite real symmetric matrices \(\mathcal {S}^{n}_{+}\). Therefore, for every \(x\in \mathbb {R}^n,\) \(h^2(E,x)=\langle x,Ax\rangle .\) By the latter, a centered ellipsoid \(E\in \mathcal {E}^n\) determines uniquely a matrix \(A\in \mathcal {S}^{n}_{+}\). On the other hand, any matrix \(A\in \mathcal {S}^{n}_{+}\) determines uniquely a centered ellipsoid \(E\in \mathcal {E}^n\) via its support function as follows:
In the next, we use the notation \(E_A\) for the ellipsoid associated with the matrix \(A\in \mathcal {S}^{n}_{+}\). In this setting, we have also \(\dim (E_A)=\textrm{rank}(A)\) and
Moreover, as already mentioned, there is a correspondence of the sum of positive semidefinite matrices to the \(L_2\) sum of ellipsoids, which follows directly from (1.5) and reads
for every \(A,B\in \mathcal {S}^{n}_{+}\) and \(\lambda \in [0,1]\), where \(\lambda \cdot _2 E_A=\sqrt{\lambda }E_A\) (see e.g. [11]).
Our first aim in this note is to use the mentioned correspondence of ellipsoids and positive semidefinite matrices, along with (1.9), to observe that the classical determinantal inequalities of Bergstrom and Ky-Fan provide us directly with alternative proofs of the following two results proven in [12].
For those results, we first need to introduce the notion of \(L_2\) zonoid. An \(L_p\) zonotope is the \(L_p\) sum of centered segments and an \(L_p\) zonoid is the limit of \(L_p\) zonotopes, where the space of convex bodies has been endowed with the usual Hausdorff metric (see e.g. [22, Section 1.8]). For \(p=2\), an \(L_2\) zonoid is a centered ellipsoid. We provide a short argument of this fact, for completeness.
Remark 1.6
Let \(x_1,\dots ,x_n\in \mathbb {R}^n\) be points and \([-x_i,x_i]\) denote the centred segment joining \(-x_i\) and \(x_i\). We denote by X the \(n\times n\) matrix whose columns are \(x_1,\dots ,x_n\). Then, the support function of the \(L_2\) sum of these segments, \([-x_1,x_1]+_2\dots +_2 [-x_n,x_n]\), according to (1.5) satisfies
where X is the matrix having \(x_1,\dots ,x_n\) as columns and \(U=(XX^T)^{1/2}\) is the square root of the positive semidefinite real symmetric matrix \(XX^T\). Hence, \([-x_1,x_1]+_2\dots +_2 [-x_n,x_n]=UB^n\), which is an ellipsoid.
In the case that the sum consists of \(m\ne n\) segments, the same argument proves that the \(L_2\) sum of centered segments is an ellipsoid. We point out that \(L_2\) zonoids may not have interior points. Indeed, if \(E\subset \mathbb {R}^n\) is the sum of \(m<n\) centered segments, then it is an \(L_2\) zonotope, and it has clearly empty interior (see [12] and the references therein for further aspects of \(L_p\) zonotopes and zonoids in this context).
For a vector \(u\in {\mathbb S}^{n-1}\), we denote by \(u^\perp \) the hyperplane orthogonal to u, i.e., the \((n-1)\)-dimensional linear subspace having u as normal vector. Then, as before, for \(K\in \mathcal {K}^n\), the orthogonal projection of K onto \(u^\perp \) is denoted by \(P_{u^\perp }K\).
Theorem 1.7
[12, Theorem 6.2] Let K, H be a pair of \(L_2\) zonoids in \(\mathbb {R}^n\) and let \(u\in {\mathbb S}^{n-1}\). Then
with equality if and only if K and H have parallel tangent hyperplanes at \(\rho _K(u)u\) and \(\rho _H(u)u\), where \(\rho _K(u)=\max \{\lambda :\, \lambda u\in K\}\) is the radial function of the body \(K\in \mathcal {K}^n\) at \(u\in {\mathbb S}^{n-1}\).
The following theorem is a generalization of Theorem 1.7.
Theorem 1.8
[12, Theorem 6.6] Let n be an integer, then for any \(1\le k\le n\) and for every pair of \(L^2\) zonoids K, H in \(\mathbb {R}^n\) and any \((n-k)\)-dimensional subspace L of \(\mathbb {R}^n\), i.e., \(L\in \mathcal {L}^n_k\), one has
In [9], Bergstrom and Ky-Fan inequalities are used to obtain linearized versions of inequalities within the realm of convex geometry. In particular, they are fundamental to obtain a linearized version of the Brunn–Minkowski, and the Aleksandrov–Fenchel inequalities for positive semidefinite matrices, that satisfy certain conditions on the projection onto a subspace. In the last subsection of this note, we investigate connections of some of the results in [9] with other results coming from the context of convex geometry, in the spirit of the previous results.
2 Ellipsoids, positive semidefinite matrices and projections
We start fixing the notation that will be used throughout the paper. Let \(A\in \mathcal {M}^{n}\) be a positive semidefinite matrix, and let \(S=\{e_1,\dots ,e_n\}\) denote the standard orthonormal basis of \(\mathbb {R}^n\). With some abuse of notation, let \(A:\mathbb {R}^n\rightarrow \mathbb {R}^n\), \(x\mapsto A\cdot x\), denote the linear map defined by the matrix A, when considered with respect to the standard basis in \(\mathbb {R}^n\).
Let L be a linear subspace of \(\mathbb {R}^n\), and let \(B_{L}, B_{L^\perp }, B_{L, L^\perp }=B_{L}{\dot{\cup }} B_{L^{\perp }}\) be orthonormal bases of \(L, L^\perp \), and \(\mathbb {R}^n\), respectively. We denote by \(A_L\) the matrix of the linear map associated with A, with respect to the bases \(B_{L,L^\perp }\).
The inclusion of the subspace L into \(\mathbb {R}^n\) will be denoted by \(\iota _L:L\rightarrow \mathbb {R}^n\), and the orthogonal projection of \(\mathbb {R}^n\) onto L will be denoted by \(P_L:\mathbb {R}^n\rightarrow L\). The linear map \(\iota _L\circ P_L:\mathbb {R}^n\rightarrow \mathbb {R}^n\), which embeds the projection onto L into \(\mathbb {R}^n\) will be denoted by \(P^L\).
The following notion of projection of a matrix onto a subspace has been considered in [3, 4], and it is inherited from the definition of the restriction of a quadratic form to a linear subspace (see e.g. [20]).
Let L be a linear subspace of \(\mathbb {R}^n\), \(A\in \mathcal {M}^{n}\), and let \(q_A\) be the quadratic form on \(\mathbb {R}^n\) associated to A, i.e., \(q_A(x)=\langle x, Ax\rangle \). The projection of the matrix A onto L is defined as the matrix associated with the restriction of q to the subspace \(L\subset \mathbb {R}^n\), and denoted by \(P_L(A)\in \mathcal {M}^{\mathrm{dim\,}L}\). The matrix \(P_L(A)\) is well defined, and if A is positive semidefinite, then so is \(P_L(A)\).
Proposition 2.1
[3, 4] Let \(A\in \mathcal {M}^{n}\) be positive semidefinite matrix, and let \(L\subseteq \mathbb {R}^n\) be a linear subspace of \(\mathbb {R}^n\) of dimension \(1\le k\le n\), i.e., \(L\in \mathcal {L}^n_k\). The following statements are equivalent:
-
i)
Let \(q_A:\mathbb {R}^n\rightarrow \mathbb {R}\) be the quadratic form \(x\mapsto x^TA x\). Then, the projection of the matrix A onto the subspace L is the \(k\times k\) positive semidefinite real symmetric matrix of the restriction of \(q_A\) to the subspace L.
-
ii)
The projection of the matrix A onto L is the matrix \(P_L(A)\) given by
$$\begin{aligned} \begin{pmatrix}1 &{}\quad 0 &{}\quad \dots &{}\quad 0&{}\quad 0 &{}\quad \dots &{}\quad 0 \\ 0&{}\quad 1&{}\quad \dots &{}\quad 0&{}\quad 0&{}\quad \dots &{}\quad 0\\ \vdots &{}\quad \vdots &{}\quad \cdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots &{}\quad \vdots \\ 0 &{}\quad 0 &{}\quad \dots &{}\quad 1&{}\quad 0 &{}\quad \dots &{}\quad 0\end{pmatrix}\;\cdot \;A_L\;\cdot \;\begin{pmatrix}1 &{}\quad \dots &{}\quad 0\\ 0&{}\quad \dots &{}\quad 0\\ \vdots &{}\quad \cdots &{}\quad \vdots \\ 0 &{}\quad \dots &{}\quad 1\\ 0 &{}\quad \dots &{}\quad 0\\ \vdots &{}\quad \cdots &{}\quad \vdots \\ 0 &{}\quad \dots &{}\quad 0 \end{pmatrix}, \end{aligned}$$where the identity submatrices in the left and right-hand sides are of size \(k\times k\). We observe that the left and right matrices in the above product are, indeed, the matrix representations, with respect to \(B_{L}, B_{L^\perp }\), and \(B_{L, L^\perp }\), of the inclusion \(\iota _L\) and the projection \(P_L\). We also remark that \(P_L(A)\) is the \(k\times k\) principal submatrix of \(A_L\) given by the first k columns and rows of \(A_L\).
The next lemma provides us with a connection between the projection of a matrix and its principal submatrices.
Lemma 2.2
[9, Lemma 3.4] Let \(A\in \mathcal {S}^{n}_{+}\) be a positive definite matrix and \(1\le i\le n\). Let \(A_i\) denote the \((n-1)\times (n-1)\) matrix obtained from A by removing the i-th row and the i-th column, and let \(A_{(i)}\) denote the \(i\times i\) matrix obtained from A by taking the first i columns and i rows. Then,
-
i)
\(A_i=P_L(A)\) for \(L=e_i^\perp \),
-
ii)
\(A_{(i)}=P_L(A)\) for \(L=\mathrm{lin\,}\{e_1,\dots ,e_i\}\).
The following two remarks will be useful in the next.
Remark 2.3
[3, Proof of Lemma 2.3.1] Let \(A\in \mathcal {M}^{n}\) and let \(u\in {\mathbb S}^{n-1}\). Let O be an orthogonal matrix such that \(\,O\,u=e_i\). As O is orthogonal, we also have \(\,O\left( u^\perp \right) =e_i^\perp \). Consequently, \( \det \left( P_{u^\perp }(A)\right) =\det \left( P_{e_i^\perp }(O A O^T)\right) .\)
In a similar manner, let \(L\in \mathcal {L}^n_k\) be a k-dimensional linear subspace of \(\mathbb {R}^n\), and let O be an orthonormal matrix such that \(O(L)=L_k\), where \(L_k=\mathrm{lin\,}\{e_1,\cdots ,e_k\}\). As before, the orthonormality of O yields \(\,O(L^\perp )=L_k^\perp \) and thus, \(\det \left( P_L(A)\right) =\det \left( P_{L_k}(O A O^T)\right) \).
Next, we will describe the existing connection between the projection of a matrix and the ellipsoids associated with the given matrix, and the projection of that matrix.
For any \(A\in \mathcal {S}^{n}_{+}\), let \(E_A\in \mathcal {E}^n\) be the ellipsoid given by A, i.e., \(h(E_A,x)^2=\langle x,Ax\rangle \), for every \(x\in \mathbb {R}^n\).
Let \(L\in \mathcal {L}^n_k\). Then the projection of A onto L, i.e., \(P_L(A)\), is a \(k\times k\) symmetric and positive semidefinite matrix. Therefore, there exists a unique ellipsoid \(E_{P_L(A)}\in \mathcal {E}^k\), such that \(h(E_{P_L(A)},x)^2=\langle x,P_L(A)x\rangle \), for every \(x\in L\), where the inner product is taken in L and inherited from \(\mathbb {R}^n\).
We consider now the projection of the ellipsoid \(E_A\) onto L, i.e., the ellipsoid \(\iota _L(P_L(E_A))\subseteq \mathbb {R}^n\). We recall the notation \(P^L=\iota _L \circ P_L\) for the projection onto L embedded in \(\mathbb {R}^n\), in contrast to \(P_L\), the projection onto L, where L is the ambient space.
As from the very definition \(\iota _L(P_L(E_A))=P^L(E_A)\subseteq \mathbb {R}^n\) is an ellipsoid in \(\mathbb {R}^n\), there exists a unique matrix \(C\in \mathcal {S}^{n}_{+}\) such that \(h(P^L(E_A)),x)^2=\langle x,Cx\rangle \), for every \(x\in \mathbb {R}^n\).
We point out that we need to distinguish the projected ellipsoid as a subset of L, being L a k-dimensional space, and as a subset of (L embedded into) \(\mathbb {R}^n\), which explains the introduction of the matrix C, of rank k, defining the projected ellipsoid as a subset of \(\mathbb {R}^n\). The next proposition establishes the precise relation between \(E_{P_L(A)}\) and \(E_C\).
Proposition 2.4
Let \(A\in \mathcal {S}^{n}_{+}\), and let \(L\in \mathcal {L}^n_k\). Then
Further,
Proof
Let \(B_{L}, B_{L^\perp }, B_{L, L^\perp }=B_{L}{\dot{\cup }} B_{L^{\perp }}\) be orthonormal bases of \(L, L^\perp \), and \(\mathbb {R}^n\), respectively. Further, let \(A\in \mathcal {S}^{n}_{+}\), and let \(L\in \mathcal {L}^n_k\). We denote by \(\iota _L:\,L\longrightarrow \mathbb {R}^n\) both, the inclusion of L into \(\mathbb {R}^n\), and the matrix of it w.r.t. \(B_L\), and \(B_{L, L^\perp }\), where we are again making some abuse of notation. It is enough to prove that the support functions of \(E_{P_L(A)}\) and \(P_L(E_A)\) coincide. For that, observe first that \(E_{P_L(A)},P_L(E_A)\subset L\), and, moreover, by definition, \(P_L(A)\) is the \(k\times k\) matrix satisfying
for every \(x\in L\). Thus, for \(x\in L\),
For the second equality, we first state the following property of the support function: for \(K\in \mathcal {K}^n\) and \(L\in \mathcal {L}^n_k\), then for every \(x\in L\), we have
Using \( E_{P_L(A)}=P_L(E_A)\) and (2.3), we obtain
for all \(x\in \mathbb {R}^n\). Finally, Eq. (2.2) follows from (2.1). Indeed, as
projecting onto L yields \(P_L(E_C)=P_L \iota _L P_L(E_A)=P_L(E_A)\). Therefore, by the definition of C, one gets \(P_L(A)=P_L(C)\). \(\square \)
3 Main results
3.1 Inequalities for \(L_2\) zonoids via determinantal inequalities
In this section, we provide proofs for Theorems 1.7 and 1.8, alternative to those in [12], based on classical inequalities for matrices. We remark here, that we do not provide a proof of the equality case, stated in [12]. Instead, we use the equality case of Theorem 1.7 proven in [12] to provide Bergstrom’s inequality with a characterization of the equality case.
We prove first Theorem 1.7 for the particular case of \(u=e_i\), \(1\le i\le n\), i.e., when u is one of the vectors of the orthonormal canonical basis of \(\mathbb {R}^n\), as a direct application of the Bergstrom’s inequality.
Theorem 3.1
Let K, H be two \(L_2\) zonoids in \(\mathbb {R}^n\), let \(1\le i\le n\), and let \(e_i\), be the i-th vector of the canonical orthonormal basis of \(\mathbb {R}^n\). Then,
Proof
Let K and H be two \(L_2\) zonoids. From Remark 1.6 we know that K and H are two centered ellipsoids. Hence, there exist \(A,B\in \mathcal {S}^{n}_{+}\) such that \(K=E_A\) and \(H=E_B\). Using (1.8) we have
Taking now Proposition 2.4 into account we have
Moreover, Lemma 2.2 i) yields
From Proposition 2.1 and Lemma 2.2 i), we know that \(A_i\) and \(B_i\) are two \((n-1)\times (n-1)\) positive semidefinite real symmetric matrices, and determine uniquely the ellipsoids \(P_{e_i^{\perp }}(K)\) and \(P_{e_i^{\perp }}(H)\). Hence by (1.8) we have
Using (1.9)
Inserting all the previous equalities in (1.10) yields that (3.1) holds if and only if
holds, which corresponds to Bergstrom’s inequality (1.1) and finishes the proof.
\(\square \)
The proof of Theorem 1.7 is now a direct consequence of Theorem 3.1 and Remark 2.3.
Proof of Theorem 1.7
Let K, H be two \(L_2\) zonoids in \(\mathbb {R}^n\), and let \(u\in {\mathbb S}^{n-1}\). Theorem 1.7 yields the validity of (3.1) for every \(e_i\), \(1\le i\le n\). Using now Remark 2.3, and an orthogonal matrix O such that \(O u=e_i\), we have that \(\det (P_{u^\perp }A)=\det \left( P_{e_i^\perp }(O A O^T)\right) \), and \(\det (P_{u^\perp }B)=\det \left( P_{e_i^\perp }(O B O^T)\right) \). Since, clearly, \(\det (A)=\det (O A O^T)\), \(\det (B)=\det (O B O^T)\), a direct application of Theorem 3.1 for \(O A O^T\), \(OBO^T\), and \((O A O^T)_i\), and \((OBO^T)_i\), yields the result. \(\square \)
As already mentioned, we are not proving the equality case, but we will be using the equality case characterization proven in [12, Theorem 6.2] to state a characterization of the equality case of Bergstrom’s inequality. Equality in the inequality established in [12, Theorem 6.2], i.e., in Theorem 1.7, holds for some \(u\in {\mathbb S}^{n-1}\) if and only if the \(L_2\) zonoids K and H have parallel tangent hyperplanes at the boundary points \(\rho _K(u)u\in K\) and \(\rho _H(u)u\in H\). The following remark is also established in [12], in connection to the equality case of [12, Theorem 6.2].
Remark 3.2
[12, Remark 6.3] Let \(K=T_1 B_n\) and \(H=T_2 B_n\) for\(T_1,T_2\in \mathcal {S}^{n}_{+}\). Then, the condition of equality in Theorem 1.7 is equivalent to the fact that there is \(\lambda >0\) such that \((T_1^{-2}-\lambda T_2^{-2})u=0\), or simply that u is an eigenvector of \(T_1^2T_2^{-2}\).
Proposition 3.3
Let A and B be two \(n\times n\) positive definite real symmetric matrices, let \(1\le i\le n\), and let \(A_i\) and \(B_i\) the \((n-1)\times (n-1)\) matrices given by A and B deleting the i-th row and the i-th column. Then, there is equality in Bergstrom’s inequality (1.1):
if and only if \(\mathrm{lin\,}\{A^{-1} e_i\}=\mathrm{lin\,}\{B^{-1}e_i\}\).
Proof
Let \(A,B\in \mathcal {S}^{n}_{+}\), be two \(n\times n\) positive definite real symmetric matrices, such that equality holds in Bergstrom’s inequality (1.1) for some \(i\in \{1,\cdots ,n\}\). Let further \(T_1\) and \(T_2\) be the unique square root of A and B, respectively, i.e., \(A=T_1^2\) and \(B=T^2_2\). It is well-known that \(T_1,T_2\in \mathcal {S}^{n}_{+}\) (see e.g. [19, Exercise 1.3.P7 and Theorem 2.6.3]). Let us consider \(E_A\) and \(E_B\) the centered ellipsoids defined by A and B, respectively. Thus, by (1.7), we have that \(E_A=T_1 B_n\) and \(E_B=T_2 B_n\). The latter considerations yield equality in (3.1) for \(E_A\) and \(E_B\). Therefore, by Remark 3.2, \(T_1^2T_2^{-2}e_i=\lambda e_i\), which implies \(\mathrm{lin\,}\{A^{-1} e_i\}=\mathrm{lin\,}\{B^{-1}e_i\}\). \(\square \)
3.2 Other inequalities
In the last part of this note, we consider other results within convex geometry, which have found a sort of counterpart in matrix theory, in particular involving the projection of a matrix. We are mostly looking at refinements of inequalities of the type of Brunn–Minkowski inequality (1.4). We start recalling the following result.
Theorem 3.4
[22] Let \(K,H\in \mathcal {K}^n\) be convex bodies such that there exists a direction \(u\in {\mathbb S}^{n-1}\) with \(\textrm{vol}_{n-1}(P_{u^{\perp }}(K))=\textrm{vol}_{n-1}(P_{u^{\perp }}(H))\). Then
for all \(\lambda \in [0,1]\).
We point out that inequality (3.4) refines the Brunn–Minkowski inequality (see [14, Section 10]), as the inequality
holds for every \(K,H\in \mathcal {K}^n\), \(\lambda \in [0,1]\), see [14, Section 10]. We refer to [22, Section 7] and to [8, 16, 17] for results in the direction of Theorem 3.4 within the theory of convex geometry.
In the next result, using the connection of positive semidefinite real symmetric matrices and ellipsoids, we obtain, as a corollary of Theorem 3.4, a refinement of the Brunn–Minkowski inequality for matrices, assuming that the determinants of the projections of the two matrices onto a hyperplane coincide.
Corollary 3.5
Let \(A,B\in \mathcal {S}^{n}_{+}\) be positive definite matrices, and let \(u\in {\mathbb S}^{n-1}\) be such that \(\det \left( P_{u^{\perp }}(A)\right) =\det \left( P_{u^{\perp }}(B)\right) \). Then,
for every \(\lambda \in [0,1]\).
Proof
Let \(A,B\in \mathcal {S}^{n}_{+}\) be positive definite matrices, and let \(E_A\) and \(E_B\) the associated ellipsoids. We observe first, that he assumption \(\det \left( P_{u^{\perp }}(A)\right) =\det \left( P_{u^{\perp }}(B)\right) \) is equivalent to the fact that \(\textrm{vol}_{n-1}(P_{u^{\perp }}(E_A))=\textrm{vol}_{n-1}(P_{u^{\perp }}(E_B))\) by means of (1.8), and Proposition 2.4. Then, taking (1.9) into account, direct application of (1.8) yields the result. \(\square \)
However, under the same assumptions, a sharper inequality is known to hold. In [9], the authors established a linear refinement inequality of the Brunn–Minkowski inequality for the determinant, inequality (1.3), under the assumption that the matrices involved share equal determinant of their projection onto a common hyperplane.
Theorem 3.6
[9, Theorem 5.7] Let \(A,B\in \mathcal {S}^{n}_{+}\) be positive definite matrices, and let \(u\in {\mathbb S}^{n-1}\). Assume that \(\det \left( P_{u^{\perp }}(A)\right) =\det \left( P_{u^{\perp }}(B)\right) \). Then,
for every \(\lambda \in [0,1]\).
Using the latter, we can now write the linear Brunn–Minkowski inequality for matrices, inequality (3.6), by means of (1.8), as a linear refinement of the \(L_2\) Brunn–Minkowski inequality (), in the case of \(L_2\) zonoids. That is the content of the following proposition.
Proposition 3.7
Let K, H be two \(L_2\) zonoids in \(\mathbb {R}^n\) and let \(u\in {\mathbb S}^{n-1}\). If there exists a direction \(u\in {\mathbb S}^{n-1}\), such that \(\textrm{vol}_{n-1}(P_{u^{\perp }}(K))=\textrm{vol}_{n-1}(P_{u^{\perp }}(H))\), then
for all \(\lambda \in [0,1]\).
We point out the following result on the equality case of the linear refinement of Brunn–Minkowski inequality for the determinant.
Theorem 3.8
[21, Theorem 5.1] Equality in inequality (3.6) holds if and only if there exists a matrix R of rank at most 1, such that \(B=A+R\).
We observe that there is also equality in (3.5) in the case \(B=A+R\), namely, when the condition in Theorem 3.8 holds.
Further, we observe that Theorem 3.8 allows us to establish a characterization of equality in ().
Corollary 3.9
Equality holds in (3.7) if and only if there exists a segment S in \(\mathbb {R}^n\), i.e., a 1-dimensional \(L_2\) zonoid, such that \(K=H +_2 S\).
Data availability
Data sharing is not applicable to this article.
References
Artstein-Avidan, S., Floretin, D., Ostrover, Y.: Remarks about mixed discriminants and volumes. In: Communications in Contemporary Mathematics, vol. 16 (2014)
Bapat, R.B.: Mixed discriminants of positive semidefinite matrices. Linear Algebra Appl. 126, 107–124 (1989)
Barvinok, A.: Computing mixed discriminants, mixed volumes, and permanents. Discrete Comput. Geom. 18(2), 205–237 (1997)
Barvinok, A.: Concentration of the mixed discriminant of well-conditioned matrices. Linear Algebra Appl. 493, 120–133 (2016)
Beckenbach, E.F., Bellman, R.: Inequalities. Springer, Berlin (1971)
Bellman, R.: Notes on matrix theory-IV: an inequality due to Bergstrom. Am. Math. Monthly 62, 172–173 (1955)
Bergstrom, H.: A triangle inequality of matrices. Den Elfte Skandinaviski Matimatiker-Kongress, Trondheim, 1949. Oslo: Johan Grundt Tanums Forlag (1952)
Colesanti, A., Saorín Gómez, E., Yepes Nicolás, J.: On a linear refinement of the Prékopa-Leindler inequality. Canad. J. Math. 68, 762–783 (2016)
de Vries, C., Lombardi, N., Gómez, E. Saorín: Notes on mixed discriminant and related linear inequalities (2022)
Fan, K.: Some inequalities concerning positive-definite Hermitian matrices. Math. Proc. Camb. Philos. Soc. 51(3), 414–421 (1955)
Florentin, D., Milman, V.D., Schneider, R.: A characterization of the mixed discriminant. Proc. Am. Math. Soc. 144(5), 2197–2204 (2016)
Fradelizi, M., Madiman, M., Meyer, M., Zvavitch, A.: On the volume of the Minkowski sum of zonoids
Fradelizi, M., Giannopoulos, A., Meyer, M.: Some inequalities about mixed volumes. Israel J. Math. 135, 157–179 (2003)
Gardner, R.J.: The Brunn–Minkowski inequality. Bull. Am. Math. Soc. 39(3), 355–405 (2002)
Giannopoulos, A., Hartzoulaki, M., Paouris, G.: On a local version of the Aleksandrov–Fenchel inequality for the quermassintegrals of a convex body. Proc. Am. Math. Soc. 130, 2403–2412 (2002)
Hernández Cifre, M.A., Yepes Nicolás, J.: Refinements of the Brunn-Minkowski inequality. J. Convex Anal. 3, 1–17 (2014)
Hernández Cifre, M.A., Yepes Nicolás, J.: Brunn–Minkowski and Prékopa–Leindler’s inequalities under projection assumptions. J. Math. Anal. Appl. 445, 1257–1271 (2017)
Hiriart-Urruty, J.B., Lemaréchal, C.: Convex Analysis and Minimization Algorithms I & II. Springer, New York (1993)
Horn, R.A., Johnson, C.R.: Matrix Analysis. Cambridge University Press, Cambridge, New York (1985)
Maddocks, J.H.: Restricted quadratic forms, inertia theorems, and the Schur complement. Linear Algebra Appl. 108, 1–36 (1988)
Saorín Gómez, E., Yepes Nicolás, J.: Linearity of the volume. Looking for a characterization of sausages. J. Math. Anal. Appl. 421(2), 1081–1100 (2015)
Schneider, R.: Convex Bodies: The Brunn–Minkowski Theory. Second expanded edition, Cambridge University Press, Cambridge (2014)
Acknowledgements
The second author is supported by PGC2018-097046-B-I00 “Análisis global en Geometría Diferencial y Convexa” by the Ministerio de Ciencia, Innovación y Universidades (Spain).
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
All authors worked equally on the paper.
Corresponding author
Ethics declarations
Conflicts of interest
There is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lombardi, N., Saorín Gómez, E. Short note on some geometric inequalities derived from matrix inequalities. Positivity 28, 26 (2024). https://doi.org/10.1007/s11117-024-01042-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11117-024-01042-7