Keywords

1 Introduction

Consider the linear system \(Tx=b\) where \(T\in \mathbb {R}^{n\times n}\) is a nonsingular matrix and bT are given. If we consider \(L=I-T\), then

$$\begin{aligned} x=Lx+b. \end{aligned}$$
(1)

The iterative form of (1) is \(x^{(k+1)}=Lx^{(k)}+b, k=0,1,2...\). Let us have now \(x^{0}=0, L^0=I\), we have

$$\begin{aligned} x^{(k+1)}=\varSigma _{m=0}^{k} L^mb. \end{aligned}$$
(2)

If \(\Vert L\Vert <1\), then \(x^{(k)}\) tends to the unique solution x [7]. In fact, the solution of (1) can be obtained by using the iterations

$$\begin{aligned} \lim _{k\rightarrow \infty }x^{(k)}=\lim _{k\rightarrow \infty } \varSigma _{m=0}^{k} L^m b=(I-L)^{-1}b=T^{-1}b=x. \end{aligned}$$
(3)

We consider the stochastic approach. Suppose that we have a Markov chain given by:

$$\alpha _0\rightarrow \alpha _1\rightarrow ...\rightarrow \alpha _k$$

where \(\alpha _i,~i=0,1,2,3...,k\) belongs to the state space \(\{1,2,...,n\}\). Then \(\alpha ,~\beta \in \{1,2...,n\}\), \(p_{\alpha }=P(\alpha _{0}=\alpha )\) is the probability that the Markov chain starts at state \(\alpha \) and \(p_{\alpha \beta }=P(\alpha _{i+1}=\beta |\alpha _{i}=\alpha )\) is the transition probability from state \(\alpha \) to \(\beta \). The set of all probabilities \(p_{\alpha \beta }\) defines a transition probability matrix \([p_{\alpha \beta }]\). We say that the distribution \([p_1,p_2 {\ldots }, p_n]^{t}\) is acceptable for a given vector h, and the distribution \(p_{\alpha \beta }\) is acceptable for matrix L, if \(p_{\alpha }>0\) when \(h_{\alpha }\ne 0\), and \(p_{\alpha }\ge 0\) when \(h_{\alpha }=0\), and \(p_{\alpha \beta }> 0\) when \(l_{\alpha \beta }\ne 0\) and \(p_{\alpha \beta }\ge 0\) when \(l_{\alpha \beta }=0\) respectively. We assume

$$\varSigma _{\alpha =1}^{n}p_{\alpha }=1,~~\varSigma _{\beta =1}^{n}p_{\alpha \beta }=1$$

for all \(\alpha =1,2 {\ldots },n\). The random variable whose mathematical expectation is equal to \(\langle x, h\rangle \) is given by the following expression

$$\begin{aligned} \theta (h)=\frac{h_{\alpha _{0}}}{p_{\alpha _{0}}}\varSigma _{j=0}^{\infty }W_j b_{\alpha _j} \end{aligned}$$
(4)

where \(W_0=1, W_j=W_{j-1}\frac{l_{\alpha _{j-1}\alpha _j}}{p{\alpha _{j-1}\alpha _j}}, j=1,2,3...\). We use the following notation for the partial sum:

$$\begin{aligned} \theta _i(h)=\frac{h_{\alpha _{0}}}{p_{\alpha _{0}}}\varSigma _{j=0}^{i}W_j b_{\alpha _j} \end{aligned}$$
(5)

It is shown that \(E(\theta _i(h))=\langle h, \varSigma _{m=0}^{i} L^m b \rangle =\langle h, x^{(i+1)} \rangle \) and \(E(\theta _i(h))\) tends to \(\langle x, h\rangle \) as \(i\rightarrow \infty \) [7]. To find \(r^{th}\) component of x, we put

$$h=(\underbrace{0,0...,1}_\text {r},0,...,0).$$

It follows that

$$\langle h, x\rangle =x_r.$$

The number of Markov chain is given by \(N\ge (\frac{0.6745}{\epsilon }\frac{\Vert b\Vert }{(1-\Vert L\Vert )})^2\). With considering N paths \({\alpha _0}^{(m)}\rightarrow {\alpha _1}^{(m)}\rightarrow ...\rightarrow {\alpha _k}^{(m)},~m=1, 2, 3..., N\), on the coefficient matrix, we have the Monte Carlo estimated solution by

$$\varTheta _i(h)=\frac{1}{N}\varSigma _{m=1}^{N} \theta _i^{(m)}(h)\simeq \langle h, x^{(i+1)}\rangle .$$

The condition \(\Vert L\Vert \le 1\) is not very strong. In [9, 10], it is shown that, it is possible to consider a Monte Carlo algorithm for which the Neumann series does not converge.

In this paper, we continue research on resolvent Monte Carlo algorithms presented in [4] and developed in [2, 3, 5]. We consider Monte Carlo algorithms for solving linear systems in the case when the corresponding Neumann series does not necessarily converge. We apply a mapping of the spectral parameter \(\lambda \) to obtain a convergent algorithm. First, sufficient conditions for the convergence of the resolvent operator are given. Then the Monte Carlo algorithm is employed.

2 Resolvent Operator Approach

2.1 The Convergence of Resolvent Operator

We study the behaviour of the equation

$$x-\lambda Lx=b$$

depending on the parameter \(\lambda .\) Define nonsingular values of L by

$$\pi (L)=\{\lambda |~x-\lambda Lx=b~\textit{has a unique solution}\}.$$

\(\chi (L)=(\pi (L))^c\) is called the characteristic set of L. Let X and Y be Banach spaces and let \(U=\{x\in X: \Vert x\Vert \le 1\}.\) An operator \(L:X\rightarrow Y\) is called compact if the closure of L(U) is compact. By Theorem 4.18 in [8], if dimension the rang of L is finite, then L is compact. The statement that \(\lambda \in \pi (L)\) is equivalent to asserting the existence of the two-sided inverse operator \((I-\lambda L)^{-1}\). It is shown that for compact operators, \(\lambda \) is a characteristic point of L if and only if \(\frac{1}{\lambda }\) is an eigenvalue of L. Also, it is shown that for every \(r>0\), the disk \(|\lambda |<r\) contains at most a finite number of characteristic values.

The operator \(R_\lambda \) defined by \(R_\lambda =(I-\lambda L)^{-1}\) is called the resolvent of L, and

$$R_\lambda =I+L+\lambda L^2+...+\lambda ^nL^{n+1}+ {\ldots }$$

The radius of convergence r of the series is equal to the distance \(r_0\) from the point \(\lambda =0\) to the characteristic set \(\chi (L)\). Let \(\lambda _1,\lambda _2, {\ldots }\) be the characteristic values of L that \(|\lambda _1|\le |\lambda _2|\le ...\). The systematic error of the above presentation when m terms are used is

$$O((\frac{|\lambda |}{|\lambda _1|})^{m+1}m^{\rho -1})$$

where \(\rho \) is multiplicity of roots \(\lambda _1\). This follows that when \(|\lambda |\ge |\lambda _1|\) the series does not converge. In this case we apply the analytical method in functional analysis. The following theorem for the case of compact operators has been proved in [6].

Theorem 1

Let \(\lambda _0\) be a characteristic value of a compact operator L. Then, in a sufficiently small neighbourhood of \(\lambda _0\), we have the expansion

$$\begin{aligned} R_{\lambda } =...+\frac{L_{-r}}{(\lambda -\lambda _0)^r}+...+\frac{L_{-1}}{(\lambda -\lambda _0)}+L_0+L_1(\lambda -\lambda _0)+...+L_n(\lambda -\lambda _0)^n+... \end{aligned}$$
(6)

Here r is the rank of characteristic value \(\lambda _0\), the operators \(L_{-r},...,L_{-1}\) are finite dimensional and \(L_{-r}\ne 0\). The series on the right-hand side of (6) is convergent in the space of operators B(XX).

3 The Convergence of Monte Carlo Method

Let \(\lambda _1, \lambda _2,..\) be real characteristic values of L such that \(\lambda _k\in (-\infty ,-a]\). In this case we may apply a mapping of the spectral parameter \(\lambda \). We consider a domain \(\varOmega \) lying inside the definition domain of \(R_{\lambda }\) as a function of \(\lambda \) such that all characteristic values are outside of \(\varOmega \), \(\lambda _{*}=1\in \varOmega \), \(0\in \varOmega \). Define \(\psi (\alpha )=\frac{4a\alpha }{(1-\alpha )^2},~(|\alpha |<1)\), which maps \(\{\alpha : |\alpha |<1 \}\) to \(\varOmega \) described in [1]. Therefore the resolvent operator can be written in the form

$$\begin{aligned} R_{\lambda }b\simeq & {} \varSigma _{k=1}^{m}b_k\alpha ^k=\varSigma _{k=1}^{m}\varSigma _{i=1}^{k}d_i^{(k)}c_i\alpha ^k\\ {}= & {} \varSigma _{k=1}^{m}\varSigma _{j=k}^{m} d_k^{(j)}\alpha ^jc_k=\varSigma _{k=1}^{m}g_k^{(m)}c_k \end{aligned}$$

where \(g_k^{(m)}=\varSigma _{j=k}^{m}d_k^{(j)}\alpha ^j\) and \(c_k=L^{k+1}b\). In [1], it is shown that \(d_k^{(j)}=(4a)^kC_{k+j-1}^{2k-1}\). All in all, in the following theorem, it is shown that the random variable whose mathematical expectation is equal to \(\langle h, \varSigma _{k=0}^{m} L^{k}\rangle \), is given by the following expression:

$$\varTheta _m^*(h)=\frac{h_{\alpha _0}}{p_{\alpha _0}}\varSigma _{\nu =0}^{m}g_\nu ^{(m)}W_{\nu }b_{\alpha _{\nu }}$$

where \(W_0=1, W_j=W_{j-1}\frac{l_{\alpha _{j-1}\alpha _j}}{p_{\alpha _{j-1}\alpha _j}}, j=1,2,3...\), \(g_0^{(m)}=1\) and \(\alpha _0,\alpha _1,...\) is a Markov chain with initial probability \(p_{\alpha _0}\) and one step transition probability \(p_{\alpha _{\nu -1}\alpha _{\nu }}\) for choosing the element \(l_{\alpha _{\nu -1}\alpha _{\nu }}\) of the matrix L [1].

Theorem 2

Consider matrix L, whose Neumann series does not necessarily converge. Let \(\psi (\alpha )=\frac{4a\alpha }{(1-\alpha )^2}\) be the required mapping, so that the presentation \(g_k^{(m)}\) exists. Then

$$E\{\lim _{m\rightarrow \infty }\frac{h_{\alpha _0}}{p_{\alpha _0}}\varSigma _{\nu =0}^{m}g_{\nu }^{(m)}W_{\nu }b_{\alpha _{\nu }}\}=\langle h,x\rangle $$

In [5], authors have analysed the robustness of the Monte Carlo algorithm for solving a class of linear algebra problems based on bilinear form of matrix powers \(\langle h,L^{k}b\rangle \). In [5], authors have considered real symmetric matrices with norms smaller than one. In this paper, results are extended considerably compared to cases [3, 5]. We consider singular matrices. For matrices that are stochastic matrices the accuracy of the algorithm is particularly high.

3.1 Numerical Tests

In this section of paper we employed our resolvent Monte Carlo algorithm for solving systems of singular linear algebraic equations. The test matrices are randomly generated. The factor of the improvement of the convergence depends on parameter \(\alpha \). An illustration of this fact is Table 1. We consider randomly generated matrices of order 100, 1000 and 5000. But more precise consideration shows that the error decreases with the increasing of the matrix size.

Table 1. Resolvent Monte Carlo results (number of trajectories \(N = 10^5\))

4 Conclusion

A new Monte Carlo algorithm for solving singular linear systems of equations is presented in this paper. The approach is based on the resolvent operator \(R_{\lambda }\) convergence. In fact we construct an algorithm based on the mapping of the spectral parameter \(\lambda \). The approach is applied to systems with singular matrices. The initial results show that for such matrices a fairly high accuracy can be obtained.