1 Introduction

Consider the two nonlinear matrix equations

$$ X^{r} + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} = I,\quad - 1 < \delta_{i} < 0, $$
(1.1)

and

$$ X^{r} - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} = I,\quad - 1 < \delta_{i} < 0, $$
(1.2)

where \(A_{i}\) are \(n \times n\) nonsingular matrices, I is the \(n \times n\) identity matrix, and r, m are positive integers, whereas X is an \(n \times n\) unknown matrix to be determined (\(A_{i}^{*}\) stands for the conjugate transpose of the matrix \(A_{i}\)). The existence and uniqueness, the rate of convergence, and necessary and sufficient conditions for the existence of positive definite solutions of similar kinds of nonlinear matrix equations have been studied by several authors [115]. El-Sayed [16] considered the matrix equation \(X + A^{*}F(X)A = Q\) with positive definite matrix Q and has shown that under some conditions an iteration method converges to the positive definite solution. Dehghan and Hajarian [17] constructed an iterative algorithm to solve the generalized coupled Sylvester matrix equations over reflexive matrices Y, Z. Also, they obtained an optimal approximation reflexive solution pair to a given matrix pair \([Y, Z]\) in the reflexive solution pair set of the generalized coupled Sylvester matrix equations \((AY - ZB,CY - ZD) = (E,F)\).

Dehghan and Hajarian [18] constructed an iterative method to solve the general coupled matrix equations \(\sum_{j = 1}^{p} A_{ij}X_{j}B_{ij} = M_{i}\), \(i = 1,2, \ldots,p\) (including the generalized (coupled) Lyapunov and Sylvester matrix equations as particular cases) over generalized bisymmetric matrix group \((X_{1},X_{2}, \ldots,X_{p})\) by extending the idea of conjugate gradient (CG) method. They determined the solvability of the general coupled matrix equations over generalized bisymmetric matrix group in the absence of roundoff errors. In addition, they obtained the optimal approximation generalized bisymmetric solution group to a given matrix group ( X 1 , X 2 ,, X p ) in Frobenius norm by finding the least Frobenius norm of the generalized bisymmetric solution group of new general coupled matrix equations.

Hajarian [19] derived a simple and efficient matrix algorithm to solve the general coupled matrix equations \(\sum_{j = 1}^{p} A_{ij}X_{j}B_{ij} = C_{i}\), \(i = 1,2, \ldots,p\) (including several linear matrix equations as particular cases) based on the conjugate gradients squared (CGS) method.

Hajarian [20] developed the conjugate gradient squared (CGS) and biconjugate gradient stabilized (Bi-CGSTAB) methods for obtaining matrix iterative methods for solving the Sylvester-transpose matrix equation \(\sum_{i = 1}^{k} (A_{i}XB_{i} + C_{i}X^{T}D_{i}) = E\) and the periodic Sylvester matrix equation ( A j X j B j + C j X j + 1 D j )= E j for \(j = 1,2, \ldots,\lambda\).

Shi [2] considered the matrix equation

$$ X = \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i}, \quad \vert \delta_{i} \vert < 1, $$
(1.3)

and Duan [1] considered the matrix equation

$$ X - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} = Q, \quad 0 < \vert \delta_{i} \vert < 1, \text{with }Q \mbox{ positive definite}, $$
(1.4)

and Duan [6] considered the matrix equation

$$ X - \sum_{i = 1}^{m} A_{i}^{*}X^{r}A_{i} = Q,\quad r \in [ - 1,0) \cup (0,1). $$
(1.5)

The main results of [2] are the following:

  1. (a1)

    The proof of the uniqueness of a positive definite solution of Equation (1.3).

  2. (a2)

    An algorithm for obtaining that solution with an original proof of convergence.

The main results of [1] are the following:

  1. (b1)

    The uniqueness of a positive definite solution of Equation (1.4), depending on Lemma 2.3.

  2. (b2)

    The same algorithm given by [2] is used with the same slightly changed proof.

In this paper, we show that the proof given in [1] can be deduced directly from [2]. In [6], the authors proved that Equation (1.5) always has a unique Hermitian positive definite solution for every fixed r. We would like to assure that Equations (1.1) and (1.2) are a nontrivial generalization of the corresponding problems with \(r = 1\). In fact, Equations (1.1) and (1.2) correspond to infinitely many \((r = 1)\)-problems with nested intervals of \(\delta_{i}\) as \(- \frac{1}{r} < \delta_{i} < 0\), as shown by the following 1-1 transformation: Putting \(X^{r} = Y\), Equations (1.1) and (1.2) become \(Y \pm \sum_{i = 1}^{m} A_{i}^{*}Y^{\delta_{i}}A_{i} = I\) with \(- \frac{1}{r} < \delta_{i} < 0\) for \(r = 1,2,\ldots\) . Our results show that some properties are invariant for all the corresponding problems such as the existence of the positive definite and extremal solutions of Equation (1.1) and the property of uniqueness of a solution of Equation (1.2). Moreover, it is clear that the set of solutions of Equation (1.1) differs with respect to the corresponding problems. Furthermore, in our paper [21], we obtained a sufficient condition for Equation (1.1) to have a unique solution. Since this condition depends on r, the uniqueness may hold for some corresponding problems and fails for others.

This paper is organized as follows. First, in Section 2, we introduce some notation, definitions, lemmas, and theorems that will be needed for this work. In Section 3, the existence of positive definite solutions of Equation (1.1) beside the extremal (maximal and minimal) solutions, which is of a more general form than other existing ones, is proved. In Section 4, two algorithms for obtaining the extremal positive definite solutions of Equation (1.1) are proposed. The merit of the proposed method is of iterative nature, which makes it more efficient. In Section 5, some numerical examples are considered to illustrate the performance and effectiveness of the algorithms. In Section 6, the existence and uniqueness of a positive definite solution of Equation (1.2) is proved. Finally, the algorithm in [1] is adapted for solving this equation. At the end of this paper, in Appendix, we analyze the defined ordering on the positive cone \(\overline{P(n)}\) showing that the ordering is total. This important result shows that problem (1.1) has one maximal and one minimal solution. Also, we explain the effect of the value of α on the number of iterations of the given algorithms, shown in Tables 1 and 2.

Table 1 INVERSE-FREE iterative Algorithm  4.1
Table 2 INVERSE-FREE iterative Algorithm  4.3

2 Preliminaries

The following notation, definitions, lemmas, and theorems will be used herein:

  1. 1.

    For \(A,B \in C^{n \times n}\), we write \(A > 0\) (≥0) if the matrix A is Hermitian positive definite (HPD) (semidefinite). If \(A - B > 0\) (\(A - B \ge 0\)), then we write \(A > B\) (\(A \ge B\)).

  2. 2.

    If a Hermitian positive definite matrix X satisfies \(A \le X \le B\), then we write \(X \in [A,B]\).

  3. 3.

    By a solution we mean a Hermitian positive definite solution.

  4. 4.

    If Equation (1.1) has the maximal solution \(X_{L}\) (minimal solution \(X_{S}\)), then for any solution X, \(X_{S} \le X \le X_{L}\).

  5. 5.

    \(P(n)\) denotes the set of all \(n \times n\) positive definite matrices.

  6. 6.

    Let E be a real Banach space. A nonempty convex closed set \(P \subset E\) is called a cone if:

    1. (i)

      \(x \in P\), \(\lambda \ge 0\) implies \(\lambda x \in P\).

    2. (ii)

      \(x \in P\), \(- x \in P\) implies \(x = \theta\), where θ denotes the zero element.

We denote the set of interior points of P by \(P^{o}\). A cone is said to be solid if \(P^{o} \ne \phi\). Each cone P in E defines a partial ordering in E given by \(x \le y\) if and only if \(y - x \in P\).

In this paper, we consider P to be the cone of \(n \times n\) positive semidefinite matrices, denoted \(\overline{P(n)}\); its interior is the set of \(n \times n\) positive definite matrices \(P(n)\).

Definition 2.1

([1])

A cone \(P \subset E\) is said to be normal if there exists a constant \(M > 0\) such that \(\theta \le x \le y\) implies \(\Vert x \Vert \le M\Vert y \Vert \).

Definition 2.2

([1])

Let P be a solid cone of a real Banach space E, and \(\Gamma:P^{o} \to P^{o}\). Let \(0 \le a < 1\). Then Γ is said to be a-concave if \(\Gamma (tx) \ge t^{a}\Gamma (x)\) \(\forall x \in P^{o}\), \(0 < t < 1\).

Similarly, Γ is said to be \(( - a)\)-convex if \(\Gamma (tx) \le t^{ - a}\Gamma (x)\) \(\forall x \in P^{o}\), \(0 < t < 1\).

Lemma 2.3

([1])

Let P be a normal cone in a real Banach space E, and let \(\Gamma:P^{o} \to P^{o}\) be a-concave and increasing (or \(( - a)\)-convex and decreasing) for an \(a \in [0,1)\). Then Γ has exactly one fixed point x in \(P^{o}\).

Lemma 2.4

([4])

If \(A \ge B > 0\) (or \(A > B > 0\)), then \(A^{\gamma} \ge B^{\gamma} > 0\) (or \(A^{\gamma} > B^{\gamma} > 0\)) for all \(\gamma \in (0,1]\), and \(B^{\gamma} \ge A^{\gamma} > 0\) (or \(B^{\gamma} > A^{\gamma} > 0\)) for all \(\gamma \in [ - 1,0)\).

Definition 2.5

([22])

A function f is said to be matrix monotone of order n if it is monotone with respect to this order on \(n \times n\) Hermitian matrices, that is, if \(A \le B\) implies \(f(A) \le f(B)\). If f is matrix monotone of order n for all n, then we say that f is matrix monotone or operator monotone.

Theorem 2.6

([22])

Every operator monotone function f on an interval I is continuously differentiable.

Definition 2.7

([1])

Let \(D \subset E\). An operator \(f:D \to E\) is said to be an increasing operator if \(y_{1} \ge y_{2}\) implies \(f(y_{1}) \ge f(y_{2})\), where \(y_{1},y_{2} \in D\). Similarly, f is said to be a decreasing operator if \(y_{1} \ge y_{2}\) implies \(f(y_{1}) \le f(y_{2})\), where \(y_{1},y_{2} \in D\).

Theorem 2.8

(Brouwer’s Fixed Point, [23])

Every continuous map of a closed bounded convex set in \(R^{n}\) into itself has a fixed point.

3 On the existence of positive definite solutions of \(X^{r}+\sum_{i=1}^{m}A_{i}^{*}X^{\delta_{i}}A_{i}=I\)

The map F associated with Equation (1.1) is defined by

$$ F(X) = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}}. $$
(3.1)

Theorem 3.1

The mapping F defined by (3.1) is operator monotone.

Proof

Suppose \(X_{1} \ge X_{2} > 0\). Then, \(F(X_{1}) = (I - \sum_{i = 1}^{m} A_{i}^{*}X_{1}^{\delta_{i}}A_{i} )^{\frac{1}{r}}\) and \(F(X_{2}) = (I - \sum_{i = 1}^{m} A_{i}^{*} X_{2}^{\delta_{i}}A_{i} )^{\frac{1}{r}}\).

Since \(X_{1} \ge X_{2}\), we have \(X_{1}^{\delta_{i}} \le X_{2}^{\delta_{i}}\) for all \(i = 1,2,\ldots,m\).

Then \(A_{i}^{*}X_{1}^{\delta_{i}}A_{i} \le A_{i}^{*}X_{2}^{\delta_{i}}A_{i}\) and \(\sum_{i = 1}^{m} A_{i}^{*}X_{1}^{\delta_{i}}A_{i} \le \sum_{i = 1}^{m} A_{i}^{*}X_{2}^{\delta_{i}}A_{i}\).

Therefore, \(I - \sum_{i = 1}^{m} A_{i}^{*}X_{1}^{\delta_{i}}A_{i} \ge I - \sum_{i = 1}^{m} A_{i}^{*}X_{2}^{\delta_{i}}A_{i}\), and since r is a positive integer, \(0 < \frac{1}{r} \le 1\).

Hence, \((I - \sum_{i = 1}^{m} A_{i}^{*}X_{1}^{\delta_{i}}A_{i} )^{\frac{1}{r}} \ge (I - \sum_{i = 1}^{m} A_{i}^{*}X_{2}^{\delta_{i}}A_{i} )^{\frac{1}{r}}\), that is, \(F(X_{1}) \ge F(X_{2})\). Thus, \(F(X)\) is operator monotone. □

The following theorem proves the existence of positive definite solutions for Equation (1.1), based on the Brouwer fixed point theorem.

Theorem 3.2

If a real number \(\beta < 1\) satisfies \((1 - \beta^{r})I \ge \sum_{i = 1}^{m} \beta^{\delta_{i}}A_{i}^{*}A_{i}\), then Equation (1.1) has positive definite solutions.

Proof

It can easily be proved that the condition of the theorem leads to \(\beta I \le (I - \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{1}{r}}\). Let \(D_{1} = [\beta I,(I - \sum_{i = 1}^{m} A_{i}^{*}A_{i})^{\frac{1}{r}} ]\). It is clear that \(D_{1}\) is closed, bounded, and convex. To show \(F:D_{1} \to D_{1}\), let \(X \in D_{1}\). Then \(\beta I \le X\), and thus \(\beta^{\delta_{i}}I \ge X^{\delta_{i}}\), \(- 1 < \delta_{i} < 0\). Therefore, \((I - \sum_{i = 1}^{m} \beta^{\delta_{i}}A_{i}^{*}A_{i} )^{\frac{1}{r}} \le (I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} )^{\frac{1}{r}}\). By the definition of β, \((I - \sum_{i = 1}^{m} \beta^{\delta_{i}}A_{i}^{*}A )^{\frac{1}{r}} \ge \beta I\), and thus \(\beta I \le (I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} )^{\frac{1}{r}} = F(X)\), that is,

$$ \beta I \le F(X). $$
(3.2)

It is clear that \(X \le I\). Then \(X^{\delta_{i}} \ge I\) and \((I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} )^{\frac{1}{r}} \le (I - \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{1}{r}}\), that is,

$$ F(X) \le \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}. $$
(3.3)

From (3.2) and (3.3) we get that \(F(X) \in D_{1}\); therefore, \(F:D_{1} \to D_{1}\). F is continuous since it is operator monotone. Therefore, F has a fixed point in \(D_{1}\), which is a solution of Equation (1.1).

The following remark, in addition to Examples 5.1 and 5.2 in Section 5, assures the validity of this theorem. □

Remark

Let us consider the simple case \(r = m = n = 1\), \(\delta = - \frac{1}{2}\). It is clear that β does not exist for \(a^{2} > \frac{2\sqrt{3}}{9}\) (where \(a^{2}\) is \(A^{*}A\) in this simple case) and also that no solution exists. For \(a^{2} \le \frac{2\sqrt{3}}{9}\), there exist β and solutions (i.e., the theorem holds). For instance, for \(a^{2} = \frac{2\sqrt{3}}{9}\), there exist \(\beta = \frac{1}{3}\) and the solution, namely \(x = \frac{1}{3}\).

Theorem 3.3

The mapping F has the maximal and the minimal elements in \(D_{1} = [\beta I,(I - \sum_{i = 1}^{m} A_{i}^{*}A_{i})^{\frac{1}{r}} ]\), where β is given in Theorem  3.2.

Proof

By Theorems 2.6 and 3.1 the mapping F is continuous and bounded above since \(F(X) < I\). Let \(\sup_{X \in D_{1}}F(X) = Y\). So, there exists an in \(D_{1}\) satisfying \(Y - \varepsilon I < F(\hat{X}) \le Y\). We can choose a sequence \(\{ X_{n} \}\) in \(D_{1}\) satisfying \(Y - (\frac{1}{n})I < F(X_{n}) \le Y\). Since \(D_{1}\) is compact, the sequence \(\{ X_{n} \}\) has a subsequence \(\{ X_{n_{k}} \}\) convergent to \(\overline{X} \in D_{1}\). So, \(Y - (\frac{1}{n})I < F(X_{n_{k}}) \le Y\). Taking the limit as \(n \to \infty\), by the continuity of F we get \(\lim_{n \to \infty} F(X_{n_{k}}) = F(\overline{X}) = Y = \max \{ F(X):X \in D_{1} \}\). Hence, F has a maximal element \(\overline{X} \in D_{1}\).

Similarly, we can prove that F has a minimal element in \(D_{1}\), noting that F is bounded below by the zero matrix. □

4 Two algorithms for obtaining extremal positive definite solution of \(X^{r}+\sum_{i=1}^{m}A_{i}^{*}X^{\delta_{i}}A_{i}=I\)

In this section, we present two algorithms for obtaining the extremal positive definite solutions of Equation (1.1). The main idea of the algorithms is to avoid computing the inverses of matrices.

Algorithm 4.1

(INVERSE-FREE Algorithm)

Consider the iterative algorithm

$$ \begin{aligned} &X_{k + 1} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*} Y_{k}^{\frac{\delta_{i}}{\delta}} A_{i}\Biggr)^{\frac{1}{r}}, \\ &Y_{k + 1} = Y_{k}\bigl[2I - X_{k}Y_{k}^{ - \frac{1}{\delta}} \bigr],\quad k = 0,1,2,\ldots, i = 1,2,\ldots,m, \\ &X_{0} = \alpha I,\qquad Y_{0} = \alpha^{\delta} I,\quad \alpha > 1, \end{aligned} $$
(4.1)

where δ is a negative integer such that \(\frac{\vert \delta \vert }{r} < 1\).

Theorem 4.2

Suppose that Equation (1.1) has a positive definite solution. Then the iterative Algorithm  4.1 generates subsequences \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) that are decreasing and converge to the maximal solution \(X_{L}\).

Proof

Suppose that Equation (1.1) has a solution. We first prove that the subsequences \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are decreasing and the subsequences \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are increasing. Consider the sequence of matrices generated by (4.1).

For \(k = 0\), we have

$$X_{1} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{0}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}} = \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \((\alpha^{r} - 1)I + \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} > 0\), we have \(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} < \alpha^{r}I\). Then

$$ \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}} < \alpha I = X_{0}. $$
(4.2)

So we get \(X_{1} < X_{0}\). Also, \(Y_{1} = Y_{0}[2I - X_{0}Y_{0}^{ - \frac{1}{\delta}} ] = \alpha^{\delta} [2I - \alpha \alpha^{ - 1}I] = \alpha^{\delta} I = Y_{0}\).

For \(k = 1\), we have

$$X_{2} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{1}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}} = \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}} = X_{1}. $$

So we have \(X_{2} < X_{0}\). Also,

$$\begin{aligned}& Y_{2} = Y_{1}\bigl[2I - X_{1}Y_{1}^{ - \frac{1}{\delta}} \bigr] = \alpha^{\delta} \Biggl[2I - \Biggl(I - \sum _{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}\alpha^{ - 1}\Biggr] \\& \quad = 2\alpha^{\delta} I - \alpha^{\delta - 1}\Biggl(I - \sum _{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}. \end{aligned}$$

From (4.2) we get \(2\alpha^{\delta} I - \alpha^{\delta - 1}(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} )^{\frac{1}{r}} > \alpha^{\delta} I\). Thus, \(Y_{2} > Y_{0}\).

For \(k = 2\), we have

$$X_{3} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(Y_{2} > Y_{0}\), we have \(Y_{2}^{\frac{\delta_{i}}{\delta}} > Y_{0}^{\frac{\delta_{i}}{\delta}}\), so we get \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} < (I - \sum_{i = 1}^{m} A_{i}^{*}Y_{0}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}}\).

Thus, \(X_{3} < X_{1}\). Also, \(Y_{3} = Y_{2}[2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} ]\). Since \(Y_{2} > Y_{0}\) and \(X_{2} < X_{0}\), we have \(Y_{2}^{ - \frac{1}{\delta}} > Y_{0}^{ - \frac{1}{\delta}}\) and \(- X_{2} > - X_{0}\), so that \(- X_{2}Y_{2}^{ - \frac{1}{\delta}} > - X_{0}Y_{0}^{ - \frac{1}{\delta}}\) and \(2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} > 2I - X_{0}Y_{0}^{ - \frac{1}{\delta}}\), and we get \(Y_{2}[2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} ] > Y_{0}[2I - X_{0}Y_{0}^{ - \frac{1}{\delta}} ]\). Thus, \(Y_{3} > Y_{1}\).

Similarly, we can prove:

$$\begin{aligned}& X_{0} > X_{2} > X_{4} >\cdots \quad \mbox{and} \quad X_{1} > X_{3} > X_{5} >\cdots, \\& Y_{0} < Y_{2} < Y_{4} < \cdots \quad \mbox{and} \quad Y_{1} < Y_{3} < Y_{5} < \cdots\,. \end{aligned}$$

Hence, \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\), \(k = 0,1,2,\ldots\) , are decreasing, whereas \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\), \(k = 0,1,2,\ldots\) , are increasing.

Now, we show that \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are bounded from below by \(X_{L}\) (\(X_{k} > X_{L}\)) and that \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are bounded from above by \(X_{L}^{\delta}\).

By induction on k we get

$$X_{0} - X_{L} = \alpha I - \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(X_{L}\) is a solution of (1.1), we have that \((I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} )^{\frac{1}{r}} < I\) and \(\alpha I - (I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} )^{\frac{1}{r}} > \alpha I - I = (\alpha - 1)I > 0\), \(\alpha > 1\), that is, \(X_{0} > X_{L}\). Also, \(X_{1} - X_{L} = (I - \sum_{i = 1}^{m} A_{i}^{*}X_{0}^{\delta_{i}}A_{i} )^{\frac{1}{r}} - (I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i})^{\frac{1}{r}}\).

Since \(X_{0} > X_{L}\), we have \(X_{0}^{\delta_{i}} < X_{L}^{\delta_{i}}\), and therefore \((I - \sum_{i = 1}^{m} A_{i}^{*}X_{0}^{\delta_{i}}A_{i} )^{\frac{1}{r}} > (I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i})^{\frac{1}{r}}\). So we get \(X_{1} > X_{L}\). Also, \(X_{L}^{\delta} - Y_{0} = (I - \sum_{i - 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} )^{\frac{\delta}{r}} - \alpha^{\delta} I\).

Since \(X_{0} > X_{L}\), we have \(\alpha I > X_{L}\) and \(\alpha^{\delta_{i}}I < X_{L}^{\delta_{i}}\) \(\forall i = 1,2,\ldots, m\), so we have

$$ \begin{aligned} &I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} > I - \sum _{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} \quad \text{and} \\ & \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} \Biggr)^{\frac{\delta}{r}} < \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} \Biggr)^{\frac{\delta}{r}}. \end{aligned} $$
(4.3)

From (4.2) we get

$$ \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} \Biggr)^{\frac{\delta}{r}} > \alpha^{\delta} I. $$
(4.4)

From (4.3) and (4.4) we get \((I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}}A_{i} )^{\frac{\delta}{r}} > \alpha^{\delta} I\), so we have \(X_{L}^{\delta} > Y_{0}\) and \(X_{L}^{\delta} > Y_{1}\).

Assume that \(X_{2k} > X_{L}\), \(X_{2k + 1} > X_{L}\) at \(k = t\) that is, \(X_{2t} > X_{L}\), \(X_{2t + 1} > X_{L}\). Also, \(Y_{2t} < X_{L}^{\delta}\) and \(Y_{2t + 1} < X_{L}^{\delta}\).

Now, for \(k = t + 1\),

$$X_{2t + 2} - X_{L} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i}\Biggr)^{\frac{1}{r}} - \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i}\Biggr)^{\frac{1}{r}}. $$

Since \(Y_{2t + 1} < X_{L}^{\delta}\), we have \(Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} < X_{L}^{\delta_{i}}\) and thus \(\sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i} < \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i}\). Therefore, \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} > (I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i})^{\frac{1}{r}}\).

Hence, \(X_{2t + 2} - X_{L} > 0\), that is, \(X_{2t + 2} > X_{L}\). Also, \(X_{L}^{\delta} - Y_{2t + 2} = X_{L}^{\delta} - Y_{2t + 1}[2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta}} ]\).

Since \(Y_{2t + 1} < X_{L}^{\delta}\) and \(X_{2t + 1} > X_{L}\), we have \(Y_{2t + 1}^{ - \frac{1}{\delta}} < X_{L}^{ - 1}\) and \(- X_{2t + 1} < - X_{L}\),

$$- X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta_{i}}} < - X_{L}^{ - 1}X_{L} = - I,\quad \mbox{and}\quad 2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta_{i}}} < 2I - I = I. $$

Then \(Y_{2t + 1}[2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta}} ] < X_{L}^{\delta}\). Hence, \(Y_{2t + 2} < X_{L}^{\delta}\), and thus

$$X_{2t + 3} - X_{L} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i}\Biggr)^{\frac{1}{r}} - \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i}\Biggr)^{\frac{1}{r}}. $$

Since, \(Y_{2t + 2} < X_{L}^{\delta}\), we have \(Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} < X_{L}^{\delta_{i}}\) and \(\sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i} < \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i}\). Therefore, \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} > (I - \sum_{i = 1}^{m} A_{i}^{*}X_{L}^{\delta_{i}} A_{i})^{\frac{1}{r}}\).

Hence, \(X_{2t + 3} - X_{L} > 0\), that is, \(X_{2t + 3} > X_{L}\), and thus \(X_{L}^{\delta} - Y_{2t + 3} = X_{L}^{\delta} - Y_{2t + 2}[2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} ]\).

Since \(Y_{2t + 2} < X_{L}^{\delta}\) and \(X_{2t + 2} > X_{L}\), we have \(Y_{2t + 2}^{ - \frac{1}{\delta}} < X_{L}^{ - 1}\) and \(- X_{2t + 2} < - X_{L}\), and thus

$$- X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} < - X_{L}^{ - 1}X_{L} = - I\quad \mbox{and} \quad 2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} < 2I - I = I. $$

Then \(Y_{2t + 2}[2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} ] < X_{L}^{\delta}\). Hence, \(Y_{2t + 3} < X_{L}^{\delta}\).

Since \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are decreasing and bounded from below by \(X_{L}\) and \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are increasing and bounded from above by \(X_{L}^{\delta}\), it follows that \(\lim_{k \to \infty} X_{k} = X\) and \(\lim_{k \to \infty} Y_{k} = Y\) exist.

Taking limits in (4.1) gives \(Y = X^{\delta}\) and \(X = (I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}} A_{i})^{\frac{1}{r}}\), that is, X is a solution. Hence, \(X = X_{L}\). □

Remark

We have proved that the maximal solution is unique; see the Appendix.

Now, we consider the case \(0 < \alpha < 1\).

Algorithm 4.3

(INVERSE-FREE Algorithm)

Consider the iterative (simultaneous) algorithm

$$ \begin{aligned} &X_{k + 1} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*} Y_{k}^{\frac{\delta_{i}}{\delta}} A_{i}\Biggr)^{\frac{1}{r}}, \\ &Y_{k + 1} = Y_{k}\bigl[2I - X_{k}Y_{k}^{ - \frac{1}{\delta}} \bigr],\quad k = 0,1,2,\ldots, i = 1,2,\ldots,m, \\ &X_{0} = \alpha I,\qquad Y_{0} = \alpha^{\delta} I, \quad 0 < \alpha < 1, \end{aligned} $$
(4.5)

where δ is a negative integer such that \(\frac{\vert \delta \vert }{r} < 1\).

Theorem 4.4

Suppose that Equation (1.1) has a positive definite solution such that \(\sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} < (1 - \alpha^{r})I\). Then the iterative Algorithm  4.3 generates the subsequences \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) that are increasing and converge to the minimal solution \(X_{S}\).

Proof

Suppose that Equation (1.1) has a solution. We first prove that the subsequences \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are increasing and the subsequences \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are decreasing. Consider the sequence of matrices generated by (4.5).

For \(k = 0\), we have

$$X_{1} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{0}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}} = \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}. $$

From the condition of the theorem we have

$$ I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} > \alpha^{r}I. $$
(4.6)

Then,

$$ \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}} > \alpha I = X_{0}. $$
(4.7)

So we get \(X_{1} > X_{0}\). Also, \(Y_{1} = Y_{0}[2I - X_{0}Y_{0}^{ - \frac{1}{\delta}} ] = \alpha^{\delta} [2I - \alpha \alpha^{ - 1}I] = \alpha^{\delta} I = Y_{0}\).

For \(k = 1\), we have

$$X_{2} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{1}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}} = \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}} = X_{1}. $$

So we have \(X_{2} > X_{0}\). Also,

$$\begin{aligned}& Y_{2} = Y_{1}\bigl[2I - X_{1}Y_{1}^{ - \frac{1}{\delta}} \bigr] = \alpha^{\delta} \Biggl[2I - \Biggl(I - \sum _{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}\alpha^{ - 1}\Biggr] \\ & \hphantom{Y_{2}} = 2\alpha^{\delta} I - \alpha^{\delta - 1}\Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}}. \end{aligned}$$

From (4.7) we get \(2\alpha^{\delta} I - \alpha^{\delta - 1}(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}}A_{i}^{*}A_{i} )^{\frac{1}{r}} < \alpha^{\delta} I\). Thus, \(Y_{2} < Y_{0}\).

For \(k = 2\), we have

$$X_{3} = \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(Y_{2} < Y_{0}\), we have \(Y_{2}^{\frac{\delta_{i}}{\delta}} < Y_{0}^{\frac{\delta_{i}}{\delta}}\), so we get \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} > (I - \sum_{i = 1}^{m} A_{i}^{*}Y_{0}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}}\). Thus, \(X_{3} > X_{1}\). Also, \(Y_{3} = Y_{2}[2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} ]\).

Since \(Y_{2} < Y_{0}\) and \(X_{2} > X_{0}\), we have \(Y_{2}^{ - \frac{1}{\delta}} < Y_{0}^{ - \frac{1}{\delta}}\) and \(- X_{2} < - X_{0}\), so that \(- X_{2}Y_{2}^{ - \frac{1}{\delta}} < - X_{0}Y_{0}^{ - \frac{1}{\delta}}\) and \(2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} < 2I - X_{0}Y_{0}^{ - \frac{1}{\delta}}\), and we get \(Y_{2}[2I - X_{2}Y_{2}^{ - \frac{1}{\delta}} ] < Y_{0}[2I - X_{0}Y_{0}^{ - \frac{1}{\delta}} ]\). Thus, \(Y_{3} < Y_{1}\).

Similarly, we can prove that

$$\begin{aligned}& X_{0} < X_{2} < X_{4} < \cdots \quad \mbox{and} \quad X_{1} < X_{3} < X_{5} < \cdots, \\ & Y_{0} > Y_{2} > Y_{4} >\cdots \quad \mbox{and} \quad Y_{1} > Y_{3} > Y_{5} >\cdots\,. \end{aligned}$$

Hence \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\), \(k = 0,1,2,\ldots\) , are increasing, whereas \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\), \(k = 0,1,2,\ldots\) , are decreasing.

Now, we show that \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are bounded from above by \(X_{S}\) (\(X_{S} > X_{k}\)), and \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are bounded from below by \(X_{S}^{\delta}\).

By induction on k we obtain

$$X_{S} - X_{0} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} - \alpha I > 0, $$

that is, \(X_{S} > X_{0}\). Also,

$$X_{S} - X_{1} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} - \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X_{0}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(X_{S} > X_{0}\), we have \(X_{S}^{\delta_{i}} < X_{0}^{\delta_{i}}\), and therefore \((I - \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} )^{\frac{1}{r}} > (I - \sum_{i = 1}^{m} A_{i}^{*}X_{0}^{\delta_{i}}A_{i})^{\frac{1}{r}}\). So we get \(X_{S} > X_{1}\). Also, \(Y_{0} - X_{S}^{\delta} = \alpha^{\delta} I - (I - \sum_{i - 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} )^{\frac{\delta}{r}}\).

Since \(X_{S} > X_{0}\), we have \(X_{S} > \alpha I\) and \(X_{S}^{\delta_{i}} < \alpha^{\delta_{i}}I\) \(\forall i = 1,2,\ldots, m\), and we get

$$ \begin{aligned} &I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} < I - \sum _{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \quad \text{and} \\ &\Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} \Biggr)^{\frac{\delta}{r}} > \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \Biggr)^{\frac{\delta}{r}}. \end{aligned} $$
(4.8)

From (4.6) we get

$$ \Biggl(I - \sum_{i = 1}^{m} \alpha^{\delta_{i}} A_{i}^{*}A_{i} \Biggr)^{\frac{\delta}{r}} < \alpha^{\delta} I. $$
(4.9)

From (4.8) and (4.9) we get \(\alpha^{\delta} I > (I - \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} )^{\frac{\delta}{r}}\), and thus \(Y_{0} > X_{S}^{\delta}\) and \(Y_{1} > X_{S}^{\delta}\).

Assume that \(X_{2k} < X_{S}\) and \(X_{2k + 1} < X_{S}\) at \(k = t\), that is, \(X_{2t} < X_{S}\) and \(X_{2t + 1} < X_{S}\). Also, \(Y_{2t} > X_{S}^{\delta}\) and \(Y_{2t + 1} > X_{S}^{\delta}\).

Now, for \(k = t + 1\), we have

$$X_{S} - X_{2t + 2} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} - \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(Y_{2t + 1} > X_{S}^{\delta}\), we have \(Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} > X_{S}^{\delta_{i}}\) and \(\sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i} > \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}} A_{i}\). Therefore, \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 1}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} < (I - \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}} A_{i})^{\frac{1}{r}}\).

Hence, \(X_{S} - X_{2t + 2} > 0\), that is, \(X_{S} > X_{2t + 2}\). Also,

$$Y_{2t + 2} - X_{S}^{\delta} = Y_{2t + 1}\bigl[2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta}} \bigr] - X_{S}^{\delta}. $$

Since \(Y_{2t + 1} > X_{S}^{\delta}\) and \(X_{2t + 1} < X_{S}\), we have \(Y_{2t + 1}^{ - \frac{1}{\delta}} > X_{S}^{ - 1}\) and \(- X_{2t + 1} > - X_{S}\), and then

$$- X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta_{i}}} > - X_{S}X_{S}^{ - 1} = - I\quad \mbox{and}\quad 2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta_{i}}} > 2I - I = I. $$

Then \(Y_{2t + 1}[2I - X_{2t + 1}Y_{2t + 1}^{ - \frac{1}{\delta}} ] > X_{S}^{\delta}\). Hence, \(Y_{2t + 2} > X_{S}^{\delta}\). Consider

$$X_{S} - X_{2t + 3} = \Biggl(I - \sum _{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} - \Biggl(I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i} \Biggr)^{\frac{1}{r}}. $$

Since \(Y_{2t + 2} > X_{S}^{\delta}\), we have \(Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} > X_{S}^{\delta_{i}}\) and \(\sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i} > \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}} A_{i}\). Therefore, \((I - \sum_{i = 1}^{m} A_{i}^{*} Y_{2t + 2}^{\frac{\delta_{i}}{\delta}} A_{i})^{\frac{1}{r}} < (I - \sum_{i = 1}^{m} A_{i}^{*}X_{S}^{\delta_{i}} A_{i})^{\frac{1}{r}}\).

Hence, \(X_{S} - X_{2t + 3} > 0\), that is, \(X_{S} > X_{2t + 3}\). Now consider

$$Y_{2t + 3} - X_{S}^{\delta} = Y_{2t + 2}\bigl[2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} \bigr] - X_{S}^{\delta}. $$

Since \(Y_{2t + 2} > X_{S}^{\delta}\) and \(X_{2t + 2} < X_{S}\), we have \(Y_{2t + 2}^{ - \frac{1}{\delta}} > X_{S}^{ - 1}\) and \(- X_{2t + 2} > - X_{S}\), and thus

$$- X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} > - X_{S}X_{S}^{ - 1} = - I\quad \mbox{and} \quad 2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} > 2I - I = I. $$

Then \(Y_{2t + 2}[2I - X_{2t + 2}Y_{2t + 2}^{ - \frac{1}{\delta}} ] > X_{S}^{\delta}\), and hence \(Y_{2t + 3} > X_{S}^{\delta}\).

Since \(\{ X_{2k} \}\) and \(\{ X_{2k + 1} \}\) are increasing and bounded from above by \(X_{S}\) and \(\{ Y_{2k} \}\) and \(\{ Y_{2k + 1} \}\) are decreasing and bounded from below by \(X_{S}^{\delta}\), it follows that \(\lim_{k \to \infty} X_{k} = X\) and \(\lim_{k \to \infty} Y_{k} = Y\) exist. Taking the limits in (4.5) gives \(Y = X^{\delta}\) and \(X = (I - \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}} A_{i})^{\frac{1}{r}}\), that is, X is a solution of Equation (1.1). Hence, \(X = X_{S}\). □

Remark

We have proved that the minimal solution is unique; see the Appendix.

5 Numerical examples

In this section, we report a variety of numerical examples to illustrate the accuracy and efficiency of the two proposed Algorithms 4.1 and 4.3 to obtain the extremal positive definite solutions of Equation (1.1). The solutions are computed for different matrices \(A_{i}\), \(i = 1,2,\ldots,m\), and different values of α, r, δ, and \(\delta_{i}\), \(i = 1,2,\ldots,m\). All programs are written in MATLAB version 5.3. We denote \(\varepsilon (X_{k}) = \Vert X_{k}^{r} + \sum_{i = 1}^{m} A_{i}^{*}X_{k}^{\delta_{i}}A_{i} - I \Vert _{\infty} = \Vert X_{k + 1} - X_{k} \Vert _{\infty}\) for the stopping criterion, and we use \(\varepsilon (X_{k}) < \mathrm{tol}\) for different chosen tolerances.

Example 5.1

Consider the matrix equation (1.1) with the following two matrices \(A_{1}\) and \(A_{2}\):

$$\begin{aligned}& A_{1} = 0.5\times \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} - 0.1 & 0.2 & - 0.06 & - 0.16 \\ 0.2 & - 0.3 & 0.16 & 0.33 \\ - 0.1 & 0 & 0.02 & 0.1 \\ 0 & 0.1 & 0 & 0.03 \end{array}\displaystyle \right ), \\& A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0.01 & 0.02 & 0.03 & 0.04 \\ 0.01 & 0.225 & 0.12 & 0.02 \\ 0 & 0.09 & 0.07 & 0.03 \\ 0.12 & 0.01 & 0.02 & 0.19 \end{array}\displaystyle \right ), \end{aligned}$$

where \(r = 3\), \(\delta_{1} = - \frac{1}{2}\), \(\delta_{2} = - \frac{1}{3}\), \(\delta = - 2\).

We applied Algorithm 4.1 for different values of the parameter \(\alpha > 1\). The number of iterations needed to satisfy the stopping condition (required accuracy) in order the sequence of positive definite matrices to converge to the maximal solution of Equation (1.1) are listed in Table 1.

$$X_{L} \cong \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0.9897 & 0.0059 & - 0.0045 & - 0.0146 \\ 0.0059 & 0.9670 & - 0.0066 & 0.0080 \\ - 0.0045 & - 0.0066 & 0.9903 & - 0.0090 \\ - 0.0146 & 0.0080 & - 0.0090 & 0.9735 \end{array}\displaystyle \right ). $$

The eigenvalues of \(X_{L}\) are \((0.9556, 0.9984, 0.9695, 0.9969)\).

Note: Theorem 3.2 holds for \(\beta = \frac{1}{3}, \frac{1}{2},\ldots\) , etc.

Example 5.2

Consider the matrix equation (1.1) with the following two matrices \(A_{1}\) and \(A_{2}\):

$$\begin{aligned}& A_{1} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0.00792 & - 0.01424 & - 0.00072 & 0.00152 \\ - 0.01424 & 0.02645 & 0.00162 & - 0.00201 \\ - 0.00072 & 0.00162 & 0.00204 & 0.0003 \\ 0.00152 & - 0.00201 & 0.0003 & 0.00109 \end{array}\displaystyle \right ), \\& A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0.0500 & - 0.0250 & - 0.0750 & 0.1000 \\ 0.1750 & 0.1500 & - 0.1250 & 0.2250 \\ 0.1000 & 0.2000 & 0.2500 & 0.1500 \\ - 0.0750 & 0.1250 & 0.0500 & 0.2000 \end{array}\displaystyle \right ), \end{aligned}$$

which satisfy the conditions of Theorem 4.4, where \(r = 3\), \(\delta_{1} = - \frac{1}{4}\), \(\delta_{2} = - \frac{1}{5}\), \(\delta = - 2\). We applied Algorithm 4.3 for different values of the parameter α (\(0 < \alpha < 1\)). The obtained results are summarized in Table 2, where k is the number of iterations needed to satisfy the stopping condition (required accuracy) in order the sequence of positive definite matrices converge to the minimal solution of Equation (1.1).

$$X_{S} \cong \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c} 0.9829 & - 0.0130 & - 0.0015 & - 0.0164 \\ - 0.0130 & 0.9710 & - 0.0131 & - 0.0318 \\ - 0.0015 & - 0.0131 & 0.9700 & - 0.0103 \\ - 0.0164 & - 0.0318 & - 0.0103 & 0.9549 \end{array}\displaystyle \right ). $$

The eigenvalues of \(X_{S}\) are \((0.9180, 0.9909, 0.9733, 0.9966)\).

Also, Theorem 3.2 holds for \(\beta = \frac{1}{3}, \frac{1}{2},\ldots\) , etc.

Remark

From Table 1 we see that the number of iterations k increases as the value of α (\(\alpha > 1\)) increases, and from Table 2 we see that the number of iterations k decreases as the value of α (\(0 < \alpha < 1\)) increases. For details, see the Appendix in the end of this paper.

Example 5.3

Consider the matrix equation (1.1) with the following two matrices \(A_{1}\) and \(A_{2}\):

$$\begin{aligned}& A_{1} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} - 0.005 & 0.01 &- 0.0006& - 0.005 &0.0115 &-0.008 \\ 0.01& - 0.015 &0.008 &- 0.0005& 0.005& - 0.01 \\ - 0.005& 0.0015& 0.001& - 0.015& 0.01& 0.02 \\ - 0.005 &0.0012 &- 0.0058 &0.01& 0.02 &0.005 \\ - 0.005& 0.005 &0.0015 &0.025 &0.0005 &- 0.01 \\ 0.015 &0.005 &- 0.008 &0.0115 &0.01 &- 0.015 \end{array}\displaystyle \right ), \\& A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} - 0.004& 0.002 &0.006 &0.002 &0.00046 &0.002 \\ 0.002 &0.01 &0.002 &- 0.004& - 0.004 &- 0.0032 \\ 0.004 &0.006 &0.008 &0.002 &0.0064 &0.002 \\ 0.012 &0.0004 &0.01 &0.008 &0.006 &0.002 \\ - 0.002 &- 0.006 &0.002 &0.0012 &0.004 &0.006 \\ 0.004 &- 0.002 &0.006 &0.002 &0.00022 &- 0.004 \end{array}\displaystyle \right ), \end{aligned}$$

where \(r = 4\), \(\delta_{1} = - \frac{1}{3}\), \(\delta_{2} = - \frac{1}{5}\), \(\delta = - 3\). We applied Algorithm 4.1 for different values of the parameter α (\(\alpha > 1\)). The number of iterations needed to satisfy the stopping condition (required accuracy) in order the sequence of positive definite matrices to converge to the maximal solution of Equation (1.1) are listed in Table 3.

$$X_{L} \cong \left( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} 0.99984 & 2.9583 \mathrm{e}{-}05 & - 3.2833 \mathrm{e} {-} 05 & - 4.6466 \mathrm{e} {-} 05 & - 1.7636 \mathrm{e} {-} 05 & 9.2391 \mathrm{e} {-} 05 \\ 2.9583 \mathrm{e} {-} 05 & 0.99986 & 2.5905 \mathrm{e} {-} 05 & - 2.4311 \mathrm{e} {-} 05 & - 2.7122 \mathrm{e} {-} 05 & 1.5525 \mathrm{e} {-} 05 \\ - 3.2833 \mathrm{e} {-} 05 & 2.5905 \mathrm{e} {-} 05 & 0.9999 & 3.4834 \mathrm{e} {-} 06 & 9.1585 \mathrm{e} {-} 06 & - 1.2574 \mathrm{e} {-} 05 \\ - 4.6466 \mathrm{e} {-} 05 & - 2.4311 \mathrm{e} {-} 05 & 3.4834 \mathrm{e} {-} 06 & 0.9997 & - 5.0029 \mathrm{e} {-} 05 & 0.00014755 \\ - 1.7636 \mathrm{e} {-} 05 & - 2.7122 \mathrm{e} {-} 05 & 9.1585 \mathrm{e} {-} 06 & - 5.0029 \mathrm{e} {-} 05 & 0.99978 & - 1.613 \mathrm{e} {-} 05 \\ 9.2391 \mathrm{e} {-} 05 & 1.5525 \mathrm{e} {-} 05 & - 1.2574 \mathrm{e} {-} 05 & 0.00014755 & - 1.613 \mathrm{e} {-} 05 & 0.99975 \end{array}\displaystyle \right) $$
Table 3 INVERSE-FREE iterative Algorithm  4.1

The eigenvalues of \(X_{L}\) are: \((0.99954, 0.99976, 0.99995, 0.99982, 0.99987, 0.99991)\).

Example 5.4

Consider the matrix equation (1.1) with the two matrices \(A_{1}\) and \(A_{2}\) as follows:

$$\begin{aligned}& A_{1} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} - 0.07 &0.03& - 0.04 &0.02 &0.072& - 0.0152 \\ 0.014 &- 0.05 &- 0.0172 &0.023 &- 0.034& - 0.053 \\ 0.072 &- 0.0161 &- 0.024 &- 0.0313 &0.0101 &- 0.0421 \\ 0.0152 & 0.0201 &- 0.0113 &- 0.0109 &0.0712 &- 0.0112 \\ - 0.0144 &0.0172 &0.0302 &- 0.0164 &- 0.012 &- 0.0503 \\ - 0.024 &- 0.0155 &0.082 &- 0.0192 &0.0221 &- 0.0411 \end{array}\displaystyle \right ), \\& A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} 0.01 &- 0.0125& 0.0165 &- 0.0225 &- 0.0375& 0.005 \\ 0.0085 &0.016 &- 0.0205& - 0.0075& - 0.00625 &0.01125 \\ 0.005& - 0.0355& 0.031 &- 0.01 &0.0135& - 0.0075 \\ - 0.0375 &0.017 &0.0325 &0.0385 &0.00625 &- 0.025 \\ - 0.01 &0.0455 &0.009 &- 0.0365 &0.033 &0.0195 \\ 0.021 &- 0.006 &0.017 &0.0335 &- 0.025 &0.0055 \end{array}\displaystyle \right ) \end{aligned}$$

which satisfy the conditions of Theorem 4.4, where \(r = 3\), \(\delta_{1} = - \frac{1}{2}\), \(\delta_{2} = - \frac{1}{3}\), \(\delta = - 2\). Applying Algorithm 4.3, for different values of the parameter α (\(0 < \alpha < 1\)). The obtained results are summarized in Table 4, where k is the number of iterations needed to satisfy the stopping condition (required accuracy) in order the sequence of positive definite matrices converge to the minimal solution of Equation (1.1).

$$X_{S} \cong \left ( \textstyle\begin{array}{c@{\quad}c@{\quad}c@{\quad}c@{\quad}c@{\quad}c} 0.99548 & 0.0016568 & 0.00085578 & 0.0011809 & 0.001863 & 6.999 \mathrm{e} {-} 05 \\ 0.0016568 & 0.99709 & 0.00057721 & 0.00031395 & - 0.0021003 & - 0.001088 \\ 0.00085578 & 0.00057721 & 0.99553 & 0.00047718 & 0.00063772 & 0.0010653 \\ 0.0011809 & 0.00031395 & 0.00047718 & 0.99757 & 0.00057145 & 2.6448 \mathrm{e} {-} 05 \\ 0.001863 & - 0.0021003 & 0.00063772 & 0.00057145 & 0.99478 & 0.00027116 \\ 6.9989 \mathrm{e} {-} 05 & - 0.001088 & 0.0010653 & 2.6448 \mathrm{e} {-} 05 & 0.00027116 & 0.99651 \end{array}\displaystyle \right ). $$
Table 4 INVERSE-FREE iterative Algorithm  4.3

The eigenvalues of \(X_{S}\) are \((0.99181, 0.9944, 0.9992, 0.99882, 0.99596, 0.99676)\).

Remarks

  1. 1.

    The obtained results for Examples 5.3 and 5.4 shown in Tables 3 and 4, respectively, indicate that increasing the dimension of the problem does not affect the efficiency of the proposed algorithms.

  2. 2.

    From Tables 1, 2, 3, and 4, it is clear that we obtained a high accuracy for different values of α after a few numbers of steps; see the number of iterations, which indicate that our algorithms have high efficiency.

6 On the existence and the uniqueness of a positive definite solution of \(X^{r}-\sum_{i=1}^{m}A_{i}^{*}X^{\delta_{i}}A_{i}=I\)

In this section, we prove that Equation (1.2) has a unique positive definite solution and construct an interval that includes that solution. Associated with Equation (1.2) is the operator G defined by

$$ G(X) = \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}}. $$
(6.1)

Theorem 6.1

If Equation (1.2) has a positive definite solution X, then \(X \in [G^{2}(I),G(I)]\).

Proof

Let X be the positive definite solution of Equation (1.2). Then \(X = (I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}} A_{i} )^{\frac{1}{r}}\), and we have

$$ X \ge I, $$
(6.2)

which implies that \(X^{\delta_{i}} \le I\), \(- 1 < \delta_{i} < 0\), \(\forall i = 1,2,\ldots,m\).

Then

$$ X = \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} \le \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} \Biggr)^{\frac{1}{r}} = G(I). $$
(6.3)

Hence, from (6.2) and (6.3) we get

$$ I \le X \le G(I). $$
(6.4)

Then \((I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}} \le X^{\delta_{i}} \le I\), and thus \(\sum_{i = 1}^{m} A_{i}^{*}(I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}}A_{i} \le \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} \le \sum_{i = 1}^{m} A_{i}^{*}A_{i}\).

We have \(\sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} = X^{r} - I\). Then \(\sum_{i = 1}^{m} A_{i}^{*}(I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}}A_{i} \le X^{r} - I \le \sum_{i = 1}^{m} A_{i}^{*}A_{i}\).

Therefore, \((I + \sum_{i = 1}^{m} A_{i}^{*}(I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}}A_{i})^{\frac{1}{r}} \le X \le (I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{1}{r}}\), that is, \(G^{2}(I) \le X \le G(I)\). Hence, \(X \in [G^{2}(I),G(I)]\).

We use the Brouwer fixed point theorem to prove the existence of positive definite solutions of Equation (1.2). Since \(P(n)\) is not complete, we consider the subset \(D_{2} = [G^{2}(I),G(I)] \subset P(n)\), which is compact. □

Theorem 6.2

Equation (1.2) has a Hermitian positive definite solution.

Proof

It is obvious that \(D_{2}\) is closed, bounded, and convex. To show that \(G:D_{2} \to D_{2}\), let \(X \in D_{2}\). Then \(G^{2}(I) \le X \le G(I)\), that is, \(X \le (I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{1}{r}}\), and thus \(X^{\delta_{i}} \ge (I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}}\), \(- 1 < \delta_{i} < 0\). Therefore, \((I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i})^{\frac{1}{r}} \ge (I + \sum_{i = 1}^{m} A_{i}^{*}(I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{\delta_{i}}{r}} A_{i} )^{\frac{1}{r}}\).

Hence,

$$ G(X) \ge G^{2}(I). $$
(6.5)

Similarly, since \(X \ge I\), we have \(X^{\delta_{i}} \le I\) and \((I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i})^{\frac{1}{r}} \le (I + \sum_{i = 1}^{m} A_{i}^{*}A_{i} )^{\frac{1}{r}}\), that is,

$$ G(X) \le G(I). $$
(6.6)

From (6.5) and (6.6) we get \(G(X) \in D_{2}\), that is, \(G:D_{2} \to D_{2}\).

According to Lemma 6.2.37 in [16], \(\sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i}\) is continuous on \(D_{2}\). Hence, \(G(X)\) is continuous on \(D_{2}\). By Brouwer’s fixed point theorem, G has a fixed point in \(D_{2}\), which is a solution of Equation (1.2). □

Remark

Applying Lemma 2.3, in [1], it is proved that Equation (1.2) has a unique positive definite solution for \(r = 1\). The following theorem shows that the uniqueness holds for any r, that is, for Equation (1.2).

Theorem 6.3

Equation (1.2) has a unique positive definite solution.

Proof

We show that G has a unique fixed point. For all \(X,Y \in P(n)\) such that \(X \ge Y\), by Definition 2.7 we have \(G:P(n) \to P(n)\) and

$$G(X) = \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} \le \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}Y^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} = G(Y), $$

that is, the operator G is decreasing. We define \(\theta = \max \{ \vert \delta_{i} \vert ,i = 1,2,\ldots,m \}\). Then \(0 < \theta < 1\). For all \(t \in (0,1)\), we have

$$\begin{aligned} G(tX) =& \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}(tX)^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} \\ =& \Biggl(I + \sum_{i = 1}^{m} t^{\delta_{i}}A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} \\ \le& \Biggl(t^{ - \theta} I + t^{ - \theta} \sum _{i = 1}^{m} t^{\delta_{i}}A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} \\ =& t^{ - \frac{\theta}{r}}\Biggl(I + \sum_{i = 1}^{m} t^{\delta_{i}}A_{i}^{*}X^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}}, \end{aligned}$$

where \(0 < \frac{\theta}{r} < 1\). Let \(\gamma = \frac{\theta}{r}\). Then \(G(tX) \le t^{ - \gamma} G(X)\), which means that the operator G is \(( - \gamma )\)-convex. We obtain that G has a unique fixed point X in \(P(n)\), which is the unique positive definite solution of Equation (1.2). □

Remarks

  1. 1.

    In [2], the authors considered the nonlinear matrix equation

    $$ X = \sum_{i = 1}^{m} A_{i}^{T}X^{\delta_{i}}A_{i},\quad \vert \delta_{i} \vert < 1, i = 1,2,\ldots,m. $$
    (6.7)

    They proved that the recursively defined matrix sequence

    $$ X_{n + m + 1} = \sum_{i = 1}^{m} A_{i}^{T}X_{n + i}^{\delta_{i}}A_{i}, \quad n \ge 0, $$
    (6.8)

    where \(X_{1},X_{2},\ldots,X_{m}\) are arbitrary initial positive definite matrices, converge to the unique positive definite solution of (6.7).

    In [1], the authors considered the nonlinear matrix equation

    $$ X = Q + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i}, \quad 0 < \vert \delta_{i} \vert < 1, i = 1,2,\ldots,m, $$
    (6.9)

    where Q is a known positive definite matrix. They considered the same formula (6.8) as

    $$ X_{n + m + 1} = Q + \sum_{i = 1}^{m} A_{i}^{*}X_{n + i}^{\delta_{i}}A_{i}, \quad n \ge 0. $$
    (6.10)

    They used the algorithm given in [2] and used the main steps for the proof of convergence with slight changes to suit their problem (6.9).

  2. 2.

    Since Q has no effect on the convergence of the algorithm and since in [2] it is proved that both \(X_{n + m + 1}\) and \(X_{n + i}\) converge to the unique solution X, we see that the proof given in [1] is redundant.

    For our Equation (1.2), we use the recursive formula

    $$ X_{n + m + 1} = \Biggl(I + \sum_{i = 1}^{m} A_{i}^{*}X_{n + i}^{\delta_{i}}A_{i} \Biggr)^{\frac{1}{r}} $$
    (6.11)

    according to the following proposition.

Proposition 6.4

The matrix sequence defined by (6.11) converges to the unique positive definite matrix solution X of Equation (1.2) for arbitrary initial positive definite matrices \(X_{1},X_{2},\ldots,X_{m}\), provided that it is valid for \(r = 1\).

Proof

We have that \(\sum_{i = 1}^{m} A_{i}^{*}X_{n + i}^{\delta_{i}}A_{i}\) is continuous and thus \(G_{n}(X) = I + \sum_{i = 1}^{m} A_{i}^{*}X_{n + i}^{\delta_{i}}A_{i}\) is continuous. Define \(F(G_{n}(X)) = (G_{n}(X))^{\frac{1}{r}}\), where r is a positive integer. Then F is also continuous. So, \(\lim_{n \to \infty} F(G_{n}(X)) = F\lim_{n \to \infty} G_{n}(X)\). Taking the limit of (6.11) as \(n \to \infty\) and using (6.9) and (6.10) with \(Q = I\), we obtain \(X = (I + \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} )^{\frac{1}{r}}\). □

7 Numerical examples

Example 7.1

Consider the matrix equation (1.2), with the following two matrices \(A_{1}\) and \(A_{2}\):

$$A_{1} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.9501 & 0.6068 \\ 0.2311 & 0.4860 \end{array}\displaystyle \right ),\quad\quad A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.8913 & 0.4565 \\ 0.7621 & 0.0185 \end{array}\displaystyle \right ), $$

where \(r = 2\), \(\delta_{1} = - \frac{1}{2}\), \(\delta_{2} = - \frac{1}{3}\). Let the initial positive definite matrices be

$$X_{1} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.8214 & 0.6154 \\ 0.4447 & 0.7919 \end{array}\displaystyle \right ), \quad\quad X_{2} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.9218 & 0.1763 \\ 0.7382 & 0.4057 \end{array}\displaystyle \right ). $$

Applying the recursive formula (6.11), after nine iterations of (6.11), we get the unique positive definite solution of (1.2)

$$X = X_{9} = \left ( \textstyle\begin{array}{c@{\quad}c} 1.6662 & 0.2891 \\ 0.2891 & 1.2438 \end{array}\displaystyle \right ). $$

We see that \(X_{9} \in [G^{2}(I),G(I)]\), where G is defined by (6.1).

Example 7.2

Consider the matrix equation (1.2) with three matrices

$$\begin{aligned}& A_{1} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.0579 & 0.8132 \\ 0.3529 & 0.0099 \end{array}\displaystyle \right ),\quad\quad A_{2} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.1389 & 0.1987 \\ 0.2028 & 0.6038 \end{array}\displaystyle \right ),\quad \text{and} \\& A_{3} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.2722 & 0.0153 \\ 0.1988 & 0.7468 \end{array}\displaystyle \right ), \end{aligned}$$

where \(r = 3\), \(\delta_{1} = - \frac{1}{3}\), \(\delta_{2} = - \frac{1}{4}\), \(\delta_{3} = - \frac{1}{2}\). Let the initial positive definite matrices be

$$\begin{aligned}& X_{1} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.8462 & 0.2026 \\ 0.5252 & 0.6721 \end{array}\displaystyle \right ),\quad\quad X_{2} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.8381 & 0.6813 \\ 0.0196 & 0.3795 \end{array}\displaystyle \right ), \quad \text{and} \\& X_{3} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.8318 & 0.7095 \\ 0.5028 & 0.4289 \end{array}\displaystyle \right ). \end{aligned}$$

Applying the recursive formula (6.11), after 14 iterations of (6.11), we get the unique positive definite solution of (1.2)

$$X = X_{14} = \left ( \textstyle\begin{array}{c@{\quad}c} 1.0793 & 0.0693 \\ 0.0693 & 1.3536 \end{array}\displaystyle \right ). $$

We see that \(X_{14} \in [G^{2}(I),G(I)]\), where G is defined by (6.1).

Remarks

  1. 1.

    If we use \(X_{1}\) and \(X_{2}\) from Example 7.2, then we get the same solution \(X = X_{9}\), which proves the uniqueness of the solution of Equation (1.2).

  2. 2.

    If we put

    $$\begin{aligned}& X_{1} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.9355 & 0.4103 \\ 0.9169 & 0.8936 \end{array}\displaystyle \right ),\qquad X_{2} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.3046 & 0.1934 \\ 0.1897 & 0.9822 \end{array}\displaystyle \right ),\quad \mbox{and} \\& X_{3} = \left ( \textstyle\begin{array}{c@{\quad}c} 0.3028 & 0.1509 \\ 0.5417 & 0.6979 \end{array}\displaystyle \right ) \end{aligned}$$

    in Example 7.2, then we get the same solution

    $$X = X_{10} = \left ( \textstyle\begin{array}{c@{\quad}c} 1.0793 & 0.0693 \\ 0.0693 & 1.3536 \end{array}\displaystyle \right ), $$

    which proves the uniqueness of the solution of Equation (1.2).

The above examples show that the recursive formula defined by (6.11) is feasible and effective to compute the unique positive definite solution of Equation (1.2).

8 Conclusion

In this paper we considered two nonlinear matrix equations \(X^{r} \pm \sum_{i = 1}^{m} A_{i}^{*}X^{\delta_{i}}A_{i} = I\). For the first equation (plus case), the proofs of the existence of positive definite solutions beside the extremal solutions are given. Also two algorithms are suggested for computing the extremal solutions. For the second equation (negative case), the existence and uniqueness of a positive definite solution are proved. The algorithm in [1] is adapted for solving this equation. Numerical examples are introduced to illustrate the obtained theoretical results.