1 Introduction

In [8], the authors introduced and studied so-called row-dense, column-dense, and dense matrices. A matrix is dense (row-dense, column-dense) if there are no zeros between two nonzero entries for every line (row, column) of this matrix. In this note, we show that if we specify these notions in a certain way, we obtain a class of matrices which in certain cases forms a subclass closed under multiplication. We now define a matrix to be strongly row-dense, SRD for short, if it is row-dense, does not have a zero row, and, in addition, the union of nonzero positions in any two consecutive rows is also, in a clear way, row-dense. Analogously, strongly column-dense, shortly SCD, and strongly dense, shortly SD, matrices are defined.

In the following examples, \(*\) indicates a nonzero entry. The matrix

$$\begin{aligned} A = \left[ \begin{matrix} 0&{}\quad *&{}\quad 0&{}\quad 0\\ *&{}\quad *&{}\quad *&{}\quad 0\\ 0&{}\quad *&{}\quad *&{}\quad *\\ 0&{}\quad 0&{}\quad *&{}\quad 0 \end{matrix}\right] \end{aligned}$$

is both SRD and SCD, hence SD. The matrix

$$\begin{aligned} A_1 = \left[ \begin{matrix} 0&{}\quad 0&{}\quad 0&{}\quad 0\\ *&{}\quad 0&{}\quad 0&{}\quad 0\\ 0&{}\quad 0&{}\quad *&{}\quad *\\ 0&{}\quad *&{}\quad 0&{}\quad 0 \end{matrix}\right] \end{aligned}$$

is row-dense, but not SRD for two reasons: A contains a zero row and the union of the nonzero positions in rows 2 and 3 is not row-dense. Also, \(A_1\) is column-dense but not SCD, as the union of the nonzero positions in columns 1 and 2 is not column-dense. The matrix

$$\begin{aligned} A_2 = \left[ \begin{matrix} 0&{}\quad *&{}\quad 0&{}\quad 0\\ *&{}\quad *&{}\quad 0&{}\quad 0\\ 0&{}\quad 0&{}\quad *&{}\quad *\\ 0&{}\quad *&{}\quad 0&{}\quad 0 \end{matrix}\right] \end{aligned}$$

is SRD but not SCD (\(A_2\) is not even column-dense), while

$$\begin{aligned} A_2^T = \left[ \begin{matrix} 0&{}\quad *&{}\quad 0&{}\quad 0\\ *&{}\quad *&{}\quad 0&{}\quad *\\ 0&{}\quad 0&{}\quad *&{}\quad 0\\ 0&{}\quad 0&{}\quad *&{}\quad 0 \end{matrix}\right] \end{aligned}$$

is SCD but not SRD.

In this paper, we first give some basic properties of these matrices. In particular, we show that for nonnegative matrices and Boolean matrices, the product of conformable SRD- (SCD-, SD-) matrices is SRD (SCD, SD). We then obtain structural characterizations for the idempotent nonnegative SRD- (SCD-, SD-) matrices, as well as for the idempotent Boolean SD-matrices with a full diagonal. Connections with generalized complementary basic matrices are made.

2 Basic properties

By induction with respect to the number of rows one easily observes that

Lemma 2.1

The union of nonzero positions in any non-void set of consecutive rows of an SRD-matrix is row-dense and nonzero.

We observe that two classes of matrices, namely nonnegative matrices and Boolean matrices, share common combinatorial properties with respect to density. In particular, applying Lemma 2.1, we have the following.

Theorem 2.2

In the classes of nonnegative matrices and Boolean matrices, the product of SRD nonnegative matrices which can be multiplied is SRD.

Proof

For nonnegative matrices and Boolean matrices Lemma 2.1 states: If u is a dense row-vector and P an SRD-matrix, then uP is a dense row-vector.

Let thus A and B be nonnegative SRD-matrices or (both) Boolean SRD-matrices which can be multiplied, and C their product. Since each row of A is dense, each row of C is dense. To show that C is SRD, let \(c_i\) and \(c_{i+1}\) be two consecutive rows of C. Multiply C from the left by the row vector \(v=(0,0,\ldots , 1,1,0,\ldots ,0)\), where the ones are in the ith and st positions. We have \(vC=vAB\) and \(w=vA\) is dense, then \(vC=wB\) is also dense, and thus C is SRD. \(\square \)

Remark 2.3

An analogous result holds for SCD- and SD-matrices. Also, instead of assuming nonnegativity or Boolean in Theorem 2.2 one can assume that the product is intrinsic in the sense introduced in [3].

Example 2.4

Observe that square nonsingular bidiagonal matrices are SRD- as well as SCD- and SD-matrices. If they are nonnegative, it is known that their products generate the class of nonsingular totally nonnegative matrices. This fact opens many new aspects and problems for these matrices.

In the sequel, we find some simple properties of SRD-matrices useful. The following is clear.

Lemma 2.5

If an SRD-matrix has a zero column c, then either all columns before c are zero columns, or all columns behind c are zero columns. Analogously, if an SCD-matrix has a zero row r, then either all rows above r are zero rows, or all rows below r are zero rows.

Lemma 2.6

If an SRD-matrix A has a block diagonal form

(1)

where \(D_i\) are square positive matrices of rank one, then for every permutation matrix P for which \(PAP^T\) is SRD, the matrix \(PAP^T\) also has such form with the same k.

Proof

Every diagonal block \(D_i\) can be written as , where \(X_i\) and \(Y_i\) are positive column vectors of the same mutual dimension. Thus , where

are equally partitioned.

Then is a row-dense matrix. Indeed, suppose that in the mth row of there is a zero between nonzeros. There is a row in PX, say row q, which has a single nonzero entry in the mth column. It follows that in \(PAP^T\) the qth row is not dense, a contradiction. This fact implies that the groups of consecutive indices in are by the permutation P either only permuted, or the groups as a whole are permuted themselves. The same happens with PX. Altogether, \(PAP^T\) has a block form, with k blocks positive and of rank one.\(\square \)

Remark 2.7

We call a square positive matrix of rank one basic. We also call a direct sum of matrices complete if it does not have a zero row. In this sense, matrices (1) and \(PAP^T\) are complete direct sums of basic matrices.

3 Idempotents

In this section, we investigate the structure of SRD-matrices. Since nonnegative matrices form a multiplicative semigroup, it is natural to study idempotent SRD-matrices first. For nonnegative matrices, the basic result of Flor [9], as given in the book [2], can be used for SRD-matrices.

Theorem 3.1

(Flor) Let E be a nonnegative idempotent matrix of rank k. Then there exists a permutation matrix P such that

(2)

where \(J_i\) are idempotent matrices of rank one, i.e., , where are positive column vectors satisfying , and U and V are nonnegative matrices of appropriate sizes.

Conversely, every square matrix of the form (2), with U and V arbitrary nonnegative matrices of appropriate sizes, is idempotent and of rank k.

If specified for matrices without zero rows, Flor’s theorem states

Lemma 3.2

Let E be a nonnegative idempotent matrix of rank k without zero rows. Then there exists a permutation matrix P such that

(3)

where \(J_i\) are idempotent matrices of rank one and V is a nonnegative matrix of appropriate size.

Conversely, every square matrix of the form (3), with V an arbitrary nonnegative matrix of appropriate size, is idempotent and of rank k.

Remark 3.3

If E does not have a zero row, then the same holds for V.

Theorem 3.4

A nonnegative matrix E is an idempotent SRD-matrix of rank r if and only if it has the symmetrically partitioned form

where

are equally partitioned, \(X_i\) and \(Y_j\) are positive column vectors satisfying , \(i = 1,\ldots ,r\), and C and D nonnegative matrices such that

$$\begin{aligned} \left[ \begin{matrix} C\\ I_r\\ D \end{matrix}\right] \end{aligned}$$
(4)

is an SRD-matrix.

Proof

(\(\Leftarrow \)) The SRD property of E follows from Theorem 2.2 since E is the product of the SRD-matrices

$$\begin{aligned} \left[ \begin{matrix} 0&{}\quad C&{}\quad 0\\ 0&{}\quad X&{}\quad 0\\ 0&{}\quad D&{}\quad 0 \end{matrix}\right] \qquad \text {and}\qquad \left[ \begin{matrix} I&{}\quad 0&{}\quad 0\\ 0&{}\quad Y^T&{}\quad 0\\ 0&{}\quad 0&{}\quad I \end{matrix}\right] \!. \end{aligned}$$

(\(\Rightarrow \)) Let E be an idempotent SRD-matrix of rank r. By Lemma 2.5, the zero columns in E can be at most some first columns and at most some last columns. By Lemma 3.2, there exists a permutation matrix P such that

where \(J_i\) are idempotent matrices of rank one. Since P preserves the diagonal of E, E has the block symmetrically partitioned form

where \(E_2\) is a permuted matrix of J. Since E is an SRD-matrix, the diagonal block \(E_2\) is, by Lemma 2.6, also a matrix of the form \(XY^T\) with possibly permuted rows of X and correspondingly permuted columns of . The condition is also satisfied. Since E is idempotent, , the matrix \(E_1X\) plays the role of C and \(E_3X\) plays the role of D. It remains to show that the matrix in (4) is an SRD-matrix.

Since is an SRD-matrix and Y is an SRD-matrix, the matrix is an SRD-matrix. Since \(Y^TY\) is a diagonal matrix with positive diagonal entries, the matrix is an SRD-matrix, and thus also the matrix is SRD. \(\square \)

Corollary 3.5

A complete characterization of idempotent nonnegative SCD-matrices is obtained by transposition of the result in Theorem 3.4.

Recalling definitions in Remark 2.7, we also have

Corollary 3.6

A nonnegative matrix is an idempotent SD-matrix if and only if it is a complete direct sum of idempotent basic matrices. The number of the basic matrices is then the rank of the direct sum.

Let us turn now to idempotents in the class of Boolean matrices. We start with idempotent SD-matrices. The following is clear.

Observation 3.7

A Boolean idempotent upper triangular SD-matrix U is a matrix which is formed by the upper triangular part of a direct sum \(B_U\) of unit (i.e. full of ones) diagonal blocks. Similarly, a Boolean idempotent lower triangular SD-matrix L is a matrix which is formed by the lower triangular part of a direct sum \(B_L\) of unit diagonal blocks. (Of course, the diagonal is considered included to the triangular part.)

For example, if

then

$$\begin{aligned} U = \left[ \begin{matrix} 1&{}\quad 1&{}\quad 0&{}\quad 0&{}\quad 0\\ 0&{}\quad 1&{}\quad 0&{}\quad 0&{}\quad 0\\ 0&{}\quad 0&{}\quad 1&{}\quad 1&{}\quad 1\\ 0&{}\quad 0&{}\quad 0&{}\quad 1&{}\quad 1\\ 0&{}\quad 0&{}\quad 0&{}\quad 0&{}\quad 1 \end{matrix} \right] \end{aligned}$$

is a Boolean idempotent upper triangular SD-matrix.

We can now completely characterize Boolean idempotent SD-matrices with a full diagonal.

Remark 3.8

The example

$$\begin{aligned} \left[ \begin{matrix} 0&{}\quad 1&{}\quad 0\\ 0&{}\quad 1&{}\quad 0\\ 1&{}\quad 1&{}\quad 1 \end{matrix} \right] \end{aligned}$$

shows that not every idempotent SD-matrix has a full diagonal.

Theorem 3.9

A Boolean SD-matrix with a full diagonal is idempotent if and only if both its upper and lower parts are idempotent Boolean SD-matrices.

Proof

Let a Boolean SD-matrix A with a full diagonal be written as \(A=I+L_0+U_0\), where I is the identity matrix, \(L_0\) is the strict lower, and \(U_0\) is the strict upper triangular part of A.

Suppose first that both matrices \(U=I+U_0\) and \(L=I+L_0\) are idempotent SD-matrices. To prove that A is idempotent, it suffices to show that both Boolean products LU and UL are majorized by \(I+L_0+U_0\).

In addition to the notation \(B_U\) and \(B_L\) from Observation 3.7 we use the notation \(L_C\) for the complete Boolean lower triangular matrix and \(U_C\) for the complete Boolean upper triangular matrix with ones in the whole part. We thus have the chain of inequalities in the Boolean sense

$$\begin{aligned} LU\leqslant L_CU=L_C+B_U=L_C+U, \end{aligned}$$

which implies that the upper triangular part of LU is majorized by the upper triangular part of A. Also,

$$\begin{aligned} LU\leqslant LU_C=B_LU_C=B_L+U_C=L+U_C \end{aligned}$$

which implies that the lower triangular part of LU is majorized by the lower triangular part of A. Altogether, LU is majorized by A. Analogously, one shows that UL is majorized by A, so that A is idempotent.

To prove the converse part, suppose that a Boolean SD-matrix A with a full diagonal is idempotent. Write as before \(A=I+L_0+U_0\). For \(L=I+L_0\), \(A=L+U_0\) so that

$$\begin{aligned} L^2 \leqslant (L+U_0)^2=L+U_0. \end{aligned}$$

Thus \(L^2\) is majorized by \(L+U_0\), and as well by L, since \(U_0\) is strict upper triangular. Next, since

$$\begin{aligned} L^2 = L(I+L_0) = L+LL_0 \geqslant L, \end{aligned}$$

L is majorized by \(L^2\). Altogether, \(L^2=L\). Analogously, \(U=I+U_0\) is idempotent. Finally, since A is an SD-matrix, L and U are SD-matrices. \(\square \)

Remark 3.10

We mention the interesting paper [1] by Beasley et al., where it is proved that a Boolean matrix is idempotent if and only if it can represented as a sum of line and rectangular parts of a certain specific structure. However, our Theorem 3.9 specifically characterizes when a Boolean SD-matrix with a full diagonal is idempotent. It is not clear how the structure in [1] would apply to SD-matrices.

4 Connections with GCB-matrices

Let \(A_1, A_2, \dots , A_s\) be real matrices of respective orders \(k_1,k_2,\ldots ,k_s\), \(k_i\geqslant 2\) for all i. Denote \(n=\sum _{i=1}^s k_i -s+1\) and form the block diagonal matrices \(G_1,G_2,\ldots ,G_s\) as follows:

Then, for any permutation \((i_1,i_2,\ldots ,i_s)\) of \((1,2,\ldots ,s)\), we can consider the product

$$\begin{aligned} G_{i_1}G_{i_2}\!\cdots G_{i_s}. \end{aligned}$$

We call products of this form generalized complementary basic matrices, GCB-matrices for short. We use the notation \(\prod G_k\) for these general products. The diagonal blocks \(A_k\) are called distinguished blocks and \(G_k\) are called generators of \(\prod G_k\). GCB-matrices have many striking properties such as permanental, graph theoretic, spectral, and inheritance properties (see for example [57]).

The upper-left and lower-right corner entries of each matrix \(A_k\) will be called intermediate and will be assumed to be nonzero.

Lemma 4.1

Suppose \(n>k>1\). Let

$$\begin{aligned} A_0= \left[ \begin{matrix} a_{11}&{}\quad \cdots &{}\quad a_{1k}\\ \vdots &{}\quad \ddots &{}\quad \vdots \\ a_{k1}&{}\quad \cdots &{}\quad a_{kk} \end{matrix}\right] \end{aligned}$$

be a matrix, and

$$\begin{aligned} B_0= \left[ \begin{matrix} b_{kk}&{}\quad \cdots &{}\quad b_{kn}\\ \vdots &{}\quad \ddots &{}\quad \vdots \\ b_{nk}&{}\quad \cdots &{}\quad b_{nn} \end{matrix}\right] \end{aligned}$$

be an matrix (the sum of orders of \(A_0\) and \(B_0\) thus exceeds n by one). Then, for the matrices

the product AB has the explicit form

$$\begin{aligned} AB=\left[ \begin{matrix} a_{11}&{}\quad \cdots &{}\quad a_{1,k-1}&{}\quad a_{1k}b_{kk}&{}\quad \cdots &{}\quad a_{1k}b_{kn}\\ a_{21}&{}\quad \cdots &{}\quad a_{2,k-1}&{}\quad a_{2k}b_{kk}&{}\quad \cdots &{}\quad a_{2k}b_{kn}\\ \vdots &{}\quad \ddots &{}\quad \vdots &{}\quad \vdots &{}\quad \ddots &{}\quad \vdots \\ a_{k1}&{}\quad \cdots &{}\quad a_{k,k-1}&{}\quad a_{kk}b_{kk}&{}\quad \cdots &{}\quad a_{kk}b_{kn}\\ 0 &{}\quad \cdots &{}\quad 0 &{}\quad b_{k+1,k}&{}\quad \cdots &{}\quad b_{k+1,n}\\ \vdots &{}\quad \ddots &{}\quad \vdots &{}\quad \vdots &{}\quad \ddots &{}\quad \vdots \\ 0 &{}\quad \cdots &{}\quad 0&{}\quad b_{nk} &{}\quad \cdots &{}\quad b_{nn} \end{matrix}\right] \!. \end{aligned}$$
(5)

Theorem 4.2

Suppose that GCB-matrices A and B are as in Lemma 4.1, where \(a_{11}, a_{kk}, b_{kk}, b_{nn}\) are nonzero. If both \(A_0\) and \(B_0\) are SRD (SCD), then both AB and BA are SRD (SCD).

Proof

The proof of the assertion for AB follows by examining the product in (5); the proof for BA is similar. \(\square \)

Corollary 4.3

Assume that all intermediate entries in matrices \(A_k\) are nonzero. If each \(A_k\) is SRD (SCD), then the generalized complementary basic matrix \(\prod G_k\) is SRD (SCD), the order of factors does not matter.

Proof

Let \(A_k\), \(k=1,\dots ,s\), be SRD; the proof for SCD is similar. We proceed by induction. If \(s=2\), the result follows from Theorem 4.2. Suppose that \(s>2\) and that the result holds for \(s-1\) matrices. Observe that matrices \(G_i\) and \(G_k\) commute if \(|i- k|>1\). This means that if 1 is before 2 in the permutation \((i_1,i_2,\ldots ,i_s)\), we can move \(G_1\) into the first position without changing the product. The product \({\Pi }\) of the remaining \(s-1\) matrices \(G_k\) has the form

$$\begin{aligned} \left[ \begin{matrix} I_{k_1-1}&{}\quad 0\\ 0&{}\quad B_0 \end{matrix}\right] \!. \end{aligned}$$

By the induction hypothesis, \(B_0\) is SRD. Hence, by Theorem 4.2, \( \prod G_k = G_1\mathrm{\Pi }\) is SRD.

If 1 is behind 2 in the permutation, we can move \(G_1\) into the last position and then we have a BA product as in Theorem 4.2. \(\square \)

Strongly dense matrices deserve further exploration. This investigation can include minimum ranks of the corresponding sign pattern matrices, as was done in [4] for the dense Alternating Sign Matrices. Connections to the interesting paper [1] might also be made.