1 Introduction

Consider an unknown compressible discrete signal described by a vector \({\mathbf {x}}\) in the \({\mathbb {C}}^N\) space, where \({\mathbb {C}}\) denotes the complex number set. The support \({T}_x\) of \({\mathbf {x}}\) is defined to be set \(T_x={\hbox {supp}}({\mathbf {x}})=\{i:x_i\ne 0\}\). Due to specific properties of \({\mathbf {x}}\), all the signal samples usually have nonzero amplitudes so the cardinality of \(T_x\) is equal to the signal dimensionality N, i.e., \(|T_x|=N\). The second relevant feature related to the signal compressibility is a rapid decrease in the signal sample amplitudes. For instance, when the power law is considered, the amplitude values of \({\mathbf {x}}\), when sorted in decreasing order, conform to the following inequality [1]:

$$\begin{aligned} |x_{U(i)}|\le Gi^{-1/r}, ~~~~i=1,\dots N, \end{aligned}$$
(1)

where G, r are positive real constants and U(i) indexes the sorted signal amplitudes.

The compressible signal is usually reconstructed from noisy measurement samples. The presence of noise has negative impact on the signal samples recovery, especially these ones with low amplitudes. Therefore, it is suggested to exclude them from an estimation process (use a sparse approximation instead) than to recover their amplitudes with high errors that could result in reduction in overall estimation quality. The quality of the sparse approximation \({\varvec{{{\hat{\mathbf{x}}}}}}\) of the compressible signal depends on both the recovered support \(T_{{\hat{x}}}\) and estimation precision of its corresponding sample amplitudes. For the given support \(T_{{\hat{x}}}\), the best estimate is provided by the least squares (LS) method [10]. The only problem that remains is how to determine the proper signal support. Orthogonal Greedy Algorithms (OGAs) that belong to the family of the Compressed Sensing (CS) methods try to solve this issue [14]. They restore the signal \({\mathbf {x}}\) step-by-step in an iterative manner. Details of the reconstruction process depend on a particular algorithm. The Orthogonal Matching Pursuit (OMP) recovers a single signal element per iteration [15]. Another known method - the Compressive Sampling Matching Pursuit (CoSaMP) estimates several signal samples at once in the single iteration [12]. In both the mentioned methods and their many modifications [16, 21], the first stage of the OGA’s operation is the support \(T_{{\hat{x}}}\) update. A desirable update route of \(T_{{\hat{x}}}\) assumes selection of the consecutive elements of \({\varvec{{{\hat{\mathbf{x}}}}}}\) according to the descending order of their amplitudes in \({\mathbf {x}}\). Unfortunately, system noise and some particular OGA properties usually violate the support recovery due to an incorrect selection of elements to be estimated. Consequently, it can result in reduction in signal reconstruction accuracy with the LS procedure in the second stage of OGA. But it is worth to notice that a recent research report reveals that OMP can trace the proper order of sparse signal reconstruction with high probability if the signal coefficients exhibit decay [13].

Selection of an appropriate moment when the iterative process should be stopped is a challenging task due to lack of a priori knowledge about the signal sparsity which, in case of compressible signals, can be only approximated. Thus, any thresholding based on a predetermined number of iterations is rather ineffective. Therefore, many known stopping rules compare the energy of the OGA residual error to some reference level that is based on a scaled value of the system noise power. Their definitions can also ensure the support recovery, if some additional conditions are met [2, 11, 17, 18], but they are developed on the basis of the strict sparsity assumption. The average properties of the residual error at the moment of perfect support recovery (i.e., the exact sparsity detection) in relation to the stopping rule design are analyzed in [6]. Yet another approach based on a hypothesis test on the presence of signal part in the residual error is proposed in [19].

According to the support definition, some part of the compressible signal must be left unrecognized when sparse signal approximation is applied. How to select the support for the sparse representation if the MSE level of the reconstructed signal can be presupposed in advance? Is it possible to implement such an idea of the support determination to manage the reconstruction task with some OGAs? Attempts to answer these questions are the contribution of this paper. To do so, the following issues are considered:

  • an analysis of the compressible signal substitution with a sparse one. Mathematical formulas that express, both the MSE and an average residual error of a LS-based reconstruction algorithm, in terms of a given sparse approximation are derived using second-order signal properties (see Sect. 2);

  • a validation of the proposed formulas with statistics obtained from numerical experiments;

  • a relation between the MSE and the residual error and its application for developing novel and practical stopping thresholds which control the approximation process of the compressible signal (see Sect. 3). The threshold definitions directly use the MSE properties of the reconstructed signal;

  • experimental tests of a compressible signal estimation with the proposed stopping thresholds (see Sect. 4);

The presented theoretical analysis assumes correct support recovery, but the final result is also empirically tested with OGA even if this assumption can not be met perfectly. To evaluate a quality of the MSE-controlled sparse approximation, numerical experiments of the compressible signal estimation with a modified orthogonal matching pursuit (OMP) algorithm were performed. The classical implementation of the OMP was supplemented with the proposed definitions of the threshold levels. The presented simulation results depict the performance of compressible signal reconstruction using the MSE metric as a quality measure. Comparisons with a case of the correct support recovery and the classical stopping threshold were included.

Similar results but without a solid mathematical background were presented in the conference paper [8] in relation to a channel estimation problem in communication systems. The results are only focused on the stopping thresholds definitions but they did not provide detailed analyses of the MSE and the residual error. It is worth to notice that the problem of accurate sparse approximation concerns a vast spectrum of compressed sensing methods in many different application areas including communications systems, image processing, power systems [3, 4, 7].

In this article, a closer look at the MSE metric will explain the meaning of sparse approximation in practical implementation and will let to propose how to control the OMP method knowing a target MSE level. The link between the MSE and the residual error will be explored in case of use of greedy methods for compressible signal estimation. These methods are considered mathematically tractable approximations of the L0-norm solution (e.g., OMP method). A similar relationship was investigated in [20] but for other signal recovery conditions. In that reference work, a sparse signal recovery was analyzed (no mention of a compressible signal case). The estimation result, that was obtained using an optimization based on the L1-norm objective function, was linked to the residual error. Ultimately, there was suggested a good tradeoff between the signal sparsity and the residual error value. Also in this article, the sparsity of the final results depends on a threshold value of the residual error but this value can be additionally adjusted by the MSE settings.

The following main notation rules are used through the paper. Matrices and vectors are denoted by the calligraphic font (e.g., \({\mathcal {C}}\)) and boldface letters, respectively. Italic letter with a subscript represents the entry of the particular vector or matrix. Superscript in parentheses denotes the current iteration number. \((\cdot )^H\) denotes the conjugate transpose. A vector (matrix) with |T notation as the subscript, e.g., \(\mathbf{c }_{|T}\), denotes the vector (matrix) with the same size and values as \(\mathbf{c }\) but only for those elements whose position corresponds to indices (columns) indicated by numbers belonging to the set T. All the other elements are zero. A notation \(\mathbf{c }_{T}\) denotes a specific vector (reduced in size) that is based on \(\mathbf{c }_{|T}\), but the elements not indicated by T are excluded while the indexing of the remaining ones is unchanged. The similar relation applies to matrices but the columns are deleted.

2 Sparse Approximation of the Compressible Signal

The compressible signal \({\mathbf {x}}\) is estimated based on a measurement vector \(\mathbf{y }\in {\mathbb {C}}^M \), that is linearly related to \({\mathbf {x}}\) according to the following formula:

$$\begin{aligned} \mathbf{y }={\mathcal {A}}\mathbf{x }+\mathbf{e }, \end{aligned}$$
(2)

where \({\mathcal {A}}\) with entries \(a_{m,n}\) is a \(M\times N\) complex measurement matrix and the vector \(\mathbf{e }\) of size \(M\times 1\) with entries \(e_i\) has a general meaning of the measurement error that is caused by an environmental noise, a quantization noise, etc. Assume that the respective elements of the vector \(\mathbf{e }\) and the matrix \({\mathcal {A}}\) are modeled as a realization of the independent and identically distributed Gaussian zero mean stationary processes E and A with \(\sigma _e^2\) and \(\sigma _a^2\) variances, respectively.

An unambiguous least squares solution of \(\mathbf{x }\) can be provided only for the approximated signal with the support \(T_{{\hat{x}}}\) of cardinality equal to or less than M. This result is as follows [10]:

$$\begin{aligned} \hat{\mathbf{x }}_{T_{{\hat{x}}}}=\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H\mathbf{y }, \end{aligned}$$
(3)

where the \(M\times |T_{{\hat{x}}}|\) matrix \({\mathcal {A}}_{T_{{\hat{x}}}}={\mathcal {A}}(T_{{\hat{x}}})\) is the set of columns of \({\mathcal {A}}\) corresponding to the support \(T_{{\hat{x}}}\). Define also the complement set of the set \(T_{{\hat{x}}}\) as \(T'_{{\hat{x}}}=\varOmega \backslash T_{{\hat{x}}}\) where \(\varOmega \) consists of all the possible signal indices, i.e., \(\varOmega =\{0,1,2,\ldots ,N-1\}\).

The residual error of the sparse approximation is defined as follows:

$$\begin{aligned} \mathbf{r }=\mathbf{y }-{\mathcal {A}}_{T_{{\hat{x}}}}\hat{\mathbf{x }}_{T_{{\hat{x}}}}, \end{aligned}$$
(4)

2.1 MSE of the Sparse Approximation

Define a mean square error (MSE) of \(\hat{{\mathbf {x}}}_{|T_{{\hat{x}}}}\) as follows:

$$\begin{aligned} \text {MSE}={\mathbf{E }}\left[ \Vert \hat{\mathbf{x }}(E,A)_ {|T_{{\hat{x}}}}-\mathbf{x }\Vert _2^2\right] =\epsilon _E+\epsilon _A, \end{aligned}$$
(5)

where the operator \({\mathbf{E }}[\ldots ]\) represents the expectation value over two random variables E and A. Here, it is necessary to emphasize that only elements of the submatrix \({\mathcal {A}}_{T'_{{\hat{x}}}}\) are considered random variables. They did not participate in signal estimation but are responsible for an additional noise-like distortion as it is stated later. Because only part of \({\mathbf {x}}\), which is pointed out by the support \(T_{{\hat{x}}}\), is recovered, it is convenient to split the MSE into an estimation error \(\epsilon _E\) and an approximation error \(\epsilon _A\). Next paragraphs provide their definitions in detail.

Using the disjoint sets \(T_{{\hat{x}}}\) and \(T'_{{\hat{x}}}\), the vector \(\mathbf{x }\) can be split into two separate vectors:

$$\begin{aligned} \mathbf{x }=\mathbf{x }_{|\varOmega }=\mathbf{x }_{|T_{{\hat{x}}}}+\mathbf{x }_{|T'_{{\hat{x}}}} \end{aligned}$$
(6)

of the same length N. The vector \(\mathbf{x }_{|T_{{\hat{x}}}}\) relates to the part of the signal to be recovered. The second one, \(\mathbf{x }_{|T'_{{\hat{x}}}}\), which is left out of reconstruction (i.e., \(\hat{\mathbf{x }}_{|T'_{{\hat{x}}}}\equiv \mathbf{0 }\)), becomes the source of the approximation error. This error is defined as:

$$\begin{aligned} \epsilon _A={\mathbf{E }}\left[ \Vert {\mathbf{x }}_{|T'_{{\hat{x}}}} \Vert _2^2\right] =\Vert {\mathbf{x }}_{T'_{{\hat{x}}}} \Vert _2^2, \end{aligned}$$
(7)

and expresses how much the actual signal \(\mathbf{x }\) differs from its sparse representation \(\mathbf{x }_{|T_{{\hat{x}}}}\).

The estimation error \(\epsilon _E\) is defined as the variance of the LS signal estimate \(\hat{\mathbf{x }}_{T_{{\hat{x}}}}\) [10]. Due to the sparse approximation, the LS estimation can be affected not only by the measurement error \(\mathbf{e }\) but also by the unrecovered signal elements \(\mathbf{x }_{T'_{{\hat{x}}}}\). It is noticeable in the rewritten form of Eq. (2):

$$\begin{aligned} \mathbf{y }={\mathcal {A}}\mathbf{x }_{|T_{{\hat{x}}}}+{\mathcal {A}}\mathbf{x }_{|T'_{{\hat{x}}}}+\mathbf{e }={\mathcal {A}}_{T_{{\hat{x}}}}\mathbf{x }_{T_{{\hat{x}}}}+{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}+\mathbf{e }. \end{aligned}$$
(8)

Substituting (8) into (3), the least squares estimate \(\hat{\mathbf{x }}_{T_{{\hat{x}}}}\) can be expressed as:

$$\begin{aligned} \hat{\mathbf{x }}_{T_{{\hat{x}}}}=\mathbf{x }_{T_{{\hat{x}}}}+\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H\mathbf{e }+\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}, \end{aligned}$$
(9)

where the last element of (9) is the additional part of the estimation error independent of the distortion produced by noise \(\mathbf{e }\). In general case of the measurement matrix, when only noise \(\mathbf{e }\) is considered random the estimation error could be written as follows:

$$\begin{aligned} \begin{aligned} \epsilon _E&={\mathbf{E }}\left[ \Vert \hat{\mathbf{x }}(E)_{T_{{\hat{x}}}}-\mathbf{x }_{T_{{\hat{x}}}}\Vert _2^2\right] =\sigma _e^2 \text {trace}\{({\mathcal {A}}_{T_{{\hat{x}}}}^H {\mathcal {A}}_{T_{{\hat{x}}}})^{-1}\}+\\&\quad +\text {trace}\{({\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}})^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}({\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}})^{-1}\}. \end{aligned} \end{aligned}$$
(10)

Because, according to the assumption, the signal \({\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}\) can be considered as another noise process independent of \(\mathbf{e }\) then using the approximation error definition (7), the estimation error can be simplified to:

$$\begin{aligned} \epsilon _E=\sigma _e^2 \text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} +\sigma _a^2\epsilon _A\text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} \end{aligned}$$
(11)

Finally, using Eqs. (11) and (7) in (5), the MSE definition can be rewritten in closedform as follows:

$$\begin{aligned} \begin{aligned} \text {MSE}&=\sigma _e^2 \text {trace}\{\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\}+\\&\quad +(\sigma _a^2\text {trace}\{\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\}+1)\epsilon _A. \end{aligned} \end{aligned}$$
(12)

For a further analysis, the most important feature of the MSE error is its dependence on the approximation error \(\epsilon _{A}\). Figure 1 depicts an example of comparison between the MSE calculated according to Eq. (12) and a reference value obtained as average:

$$\begin{aligned} \text {MSE}_\mathrm{ref}=\frac{\sum _L \Vert \hat{\mathbf{x }}_{|T_{{\hat{x}}}}-\mathbf{x }\Vert _2^2}{L}, \end{aligned}$$
(13)

for only \(L=10\) independent trials with various measurement matrices \({\mathcal {A}}\), noise samples \(\mathbf{e }\), SNR ratios and a constant instance of signal \(\mathbf{x }\) in the single averaging. The small value of L proves a good convergence. An complex power-law decaying signal consistent with condition (1) was used as an example of the compressible signal (\(G=1\), \(r=0.85\)). The x-axis represents the number of elements estimated in the recovered signal. These elements are arranged in such a way that they always correspond to the highest amplitude elements in the original signal. This order of the MSE presentation is also the proper sequence of the support recovery by iterative compressed sensing methods.

The results presented in Fig. 1 show that it is possible to use (12) for MSE evaluation in case of the sparse approximation of compressible signal. Moreover, Eq. (12) is valid for all kind of signals.

Fig. 1
figure 1

MSE of a compressible signal estimation—the proposed closedform formula versus the statistical averaging

2.2 Residual Error Analysis

Energy of the residual error in LS procedure for a given support \({T_{{\hat{x}}}}\) is [10]:

$$\begin{aligned} \parallel \mathbf{r }_{T_{{\hat{x}}}}\parallel _2^2=\mathbf{y }^H\mathbf{y }-\mathbf{y }^H{\mathcal {A}}_{T_{{\hat{x}}}}\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H\mathbf{y }. \end{aligned}$$
(14)

Using (8), Eq. (14) can be rewritten after some mathematical manipulations as follows:

$$\begin{aligned} \begin{aligned} \parallel \mathbf{r }_{T_{{\hat{x}}}}\parallel _2^2&= \mathbf{x }_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}-\mathbf{x }_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}^HR{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}\\&\quad +\mathbf{e }^H\mathbf{e }+(\mathbf{x }_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}^H-\mathbf{x }_{T'_{{\hat{x}}}}^H{\mathcal {A}}_{T'_{{\hat{x}}}}^HR)\mathbf{e }\\&\quad +\mathbf{e }^H({\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}}-R{\mathcal {A}}_{T'_{{\hat{x}}}}\mathbf{x }_{T'_{{\hat{x}}}})-\mathbf{e }^HR\mathbf{e }, \end{aligned} \end{aligned}$$
(15)

where

$$\begin{aligned} {\mathcal {R}}={\mathcal {A}}_{T_{{\hat{x}}}}({\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}})^{-1}{\mathcal {A}}_{T_{{\hat{x}}}}^H, \end{aligned}$$
(16)

is the projection matrix with \({\hbox {tr}}\{{{\mathcal {R}}}\}=|T_{{\hat{x}}}|\) [9]. Using the definition of the approximation error (7), the expectation value of (15) over the stationary processes E (noise) and A (entries of the measurement matrix) is:

$$\begin{aligned} {\mathbf{E }}\left[ \parallel \mathbf{r }(E,A)_{T_{{\hat{x}}}}\parallel _2^2\right] =(M-|T_{{\hat{x}}}|)(\sigma _e^2+\sigma _a^2\epsilon _A). \end{aligned}$$
(17)

The result presented in (17) corresponds to the one obtained in [6] but here it is complemented by the influence of the approximation error.

Figure 2 depicts that the expectation value of the residual error (17) corresponds to the averaged value of (15) (named as \(r_\mathrm{ref}\)) obtained for the same settings as in the MSE analysis in the previous subsection. Hence, Eq. (17) is proposed to be used in the next mathematical analysis.

Fig. 2
figure 2

Expectation value of the residual error—the proposed closedform formula versus the statistical averaging

3 MSE-Based Decision Threshold

The residual error

$$\begin{aligned} \mathbf{r }^{(i)}=\mathbf{y }-{\mathcal {A}}\hat{\mathbf{x }}_{|T^{(i)}_{{\hat{x}}}}, \end{aligned}$$
(18)

is the basic measure that provides instant information about estimation quality in the ith iteration step of OGAs. It is commonly used to stop the iterative procedure if its energy drops below a predefined threshold. The paper proposes a threshold definition that is based on the assumed MSE level (\(\text {MSE}_\text {S}\)) of the finally estimated signal. In every step of the iterative procedure, the current energy of the residual error is tested against its expected value \(\varGamma _S\) for the given \(\text {MSE}_\text {S}\). The threshold \(\varGamma _S\) can be determined by combining Eqs. (12) and (17) and substituting the approximation error there. Finally the threshold level can be described as follows:

$$\begin{aligned} \begin{aligned} \varGamma _S&=(M-|T_{{\hat{x}}}|)\sigma _e^2\\&\quad +(M-|T_{{\hat{x}}}|)\sigma _e^2\frac{\text {SNR}\cdot \text {MSE}_\text {S}-\sigma _a^2\text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} }{\sigma _a^2\text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} +1}, \end{aligned} \end{aligned}$$
(19)

where \(\text {SNR}=\sigma _a^2\)/\(\sigma _e^2\).

If \(\text {MSE}_\text {S}\) value is selected too optimistically, the assumed quality may not be attainable during estimation process (due to specific properties of the signal and noise) and OGA’s iterations will not be interrupted. To prevent it, an additional stopping threshold is suggested. It is developed knowing, that two components of (12) are monotonic but in the opposite direction. If better and better support approximation of \({\mathbf {x}}\) is obtained, the approximation error decreases but the noise part of the estimation error increases. Thereby, there can be found a balance point:

$$\begin{aligned} \sigma _e^2 \text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} =\left( \sigma _a^2\text {trace}\left\{ \left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\right\} +1\right) \epsilon _A, \end{aligned}$$
(20)

that allows, when applied in (17), to define the so-called balanced threshold:

$$\begin{aligned} \begin{aligned} \varGamma _B&=(M-|T_{{\hat{x}}}|)\sigma _e^2+\\&\quad +(M-|T_{{\hat{x}}}|)\sigma _e^2\frac{\sigma _a^2\text {trace}\{\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\}}{\sigma _a^2\text {trace}\{\left( {\mathcal {A}}_{T_{{\hat{x}}}}^H{\mathcal {A}}_{T_{{\hat{x}}}}\right) ^{-1}\}+1}. \end{aligned} \end{aligned}$$
(21)

Consequently, using (20) and (21), the final stopping rule for any OGA can be determined as follows:

$$\begin{aligned} \text {STOP}~~\text {if}~~ \left\{ \begin{array} {l l} i=M, &{} ~ \\ \parallel \mathbf{r }^{(i)}\parallel _2^2< \max \{\varGamma _S, \varGamma _B\} \end{array} \right. . \end{aligned}$$
(22)
figure a

4 Numerical Experiments

The developed threshold definition was implemented in OMP algorithm (see Algorithm 1) and experimentally tested for power-law decaying signals with \(G=1\) and \(r=0.85\). The measurement matrix size was set to \(120\times 160\). The signals were recovered in noisy environment. The signal-to-noise ratio (SNR) varied in the range of 5[dB] to 25[dB] with 2[dB] step. MSE of the estimated signals was the main results of the performed simulations and the object of discussion thereafter. The simulations results were obtained averaging over \(K=50\) independent simulation runs for a single instance of the compressible signal and over \(P=50\) independent signals.

Fig. 3
figure 3

Estimation performance versus SNR for the assumed \(\text {MSE}_\text {S}\) condition

Figure 3 depicts the MSE performance of the signal estimation as a function of SNR ratio. The results obtained for the proposed stopping rule (22) were compared to the ones for the classical approach when the residual error energy is tested against the noise energy—\(\parallel \mathbf{r }^{(i)}\parallel _2^2<M\sigma _e^2\) [5]. There were considered two values of the quality factor: \( \text {MSE}_\text {S}=\{-15[\text {dB}]; -20[\text {dB}]\}\). Additionally, the reference results (labeled as “ref” in the figure) are provided. They are obtained according to the assumption that signal samples are restored in the descending order of their actual amplitudes. The depicted results show that it is possible to control the estimation process using the MSE value as the predetermined quality factor. Performance degradation for OMP algorithm is caused because the desirable recovery order valid for the “ref” case is violated by noise presence and the reconstructed signal structure. But it is the inherent problem of any CS methods and it is beyond the article’s scope. The OMP can recover the signal support with higher precision if SNR is relatively high. Therefore, the gap between MSEs for the OMP and the reference solution is reduced along with the noise reduction.

4.1 Implementation Issues

Contrary to the classical stopping rules, value of the proposed one varies across iterations according to Eqs. (19) and (21). Although, inverse of the matrix product \({\mathcal {A}}_{T_{{\hat{x}}}}^H {\mathcal {A}}_{T_{{\hat{x}}}}\) has the greatest computational complexity in these equations, it is worth noting that one is always determined earlier at every iteration step of OGA as the part of the LS estimation formula (see Eq. 3). Consequently, the complexity increase is minor in comparison with the main body of OGA (some new additions are required due to trace operation).

5 Conclusions

The paper presents the combined analysis of the residual error and the MSE of the estimated signal with respect to sparse approximation of the compressible signal using LS estimation procedure. The obtained results were then used to propose the practical rule that allows a reconstruction algorithm to control the signal recovery process using a predetermined value of MSE. The solution, which was implemented in OMP method, reduced an overall reconstruction complexity because the given recovery process was interrupted when the target quality level was reached. The threshold definition can be especially suggested for use in the recovery methods with proper support detection.

An interesting direction of a future work is to use the MSE-driven control mechanism for estimation of underlying sparse process using a structured (model-based) compressed sensing approach when quality metric is still evaluated in compressible domain. It could be useful in wireless transmission systems where channel estimation is often shifted from the channel impulse response domain to the propagation paths domain.