1 Introduction

Traditional compressed sensing theory is mainly used to process discrete and finite-dimensional digital signal. However, it is expected to develop a technology which can sample continuous and infinite dimensional analog signal at sub-Nyquist rate, so as to truly break the bandwidth limitation of existing ADC equipment and reconstruct original signal from the samples of baseband signal after low-pass filters and finally ease the pressure of hardware sampling. To address this problem, a variety of solutions have been proposed. Analog to information convertor (AIC) and Xampling for multiband analog signal [15] are two relatively mature technologies among them. Xampling uses modulated wideband converter [6] (MWC) to sample, whose results are infinite measurement vectors (IMV), and it cannot directly reconstruct signal using the traditional reconstruction algorithms. To solve this problem, we can reconstruct the original signal by turning IMV problem into multiple measurement vector (MMV) problem using continuous to finite (CTF) modular [7] under the premise of joint sparse [8].

The simultaneous orthogonal matching pursuit (SOMP) [9, 10], treated as a derivative algorithm of matching pursuit simultaneous (OMP) algorithm to solve MMV problem, is what used for reconstruction in CTF at present. Based on this method, many novel MMV algorithms have been proposed according to the derivative algorithms of OMP such as regularized orthogonal matching pursuit (ROMP) [11], stagewise orthogonal matching pursuit (StOMP) [12], compressive sampling matching pursuit (CoSaMP) [13], and subspace pursuit (SP) [14]. The main idea of these algorithms is firstly calculating the inner product between atoms in sensing matrix and columns in residual matrix. Then, several inner products corresponding to the same atom constitute an inner product vector. By calculating its 2-norm. We can find the maximum value whose corresponding atom is the matched atom. In fact, the inner product cannot measure the relevancy between atoms and residual matrix very well and sometimes will eventually lead to errors.

The idea of this paper is to change the inner product into the correlation coefficient, so that we can measure the matching degree between atoms and the residual better. In order to check the performance of the changement, the SOMP algorithm and its derivative algorithms such as MMV-ROMP, MMV-StOMP, MMV-CoSaMP, and MMV-SP, which can solve MMV problems, are chosen for comparison. The proposed algorithms are improved by changing the inner product into correlation coefficient as screening criterion of atoms, then compare the signal reconstruction probability of these methods with original algorithms. The results show that the improved algorithms have better reconstruction probability and there are less samples required.

2 Signal model of multiband

Assuming x(t) is a continuous signal in L 2 space which satisfies the square integration as follows.

$$ \int_{- \infty }^{+ \infty} {{\left| {x(t)} \right|}^{2}} dt < + \infty $$
((1))

And its Fourier transformation can be expressed as

$$ X(f) = \int_{- \infty }^{+ \infty} {x(t){e^{- j2\pi ft}}dt} $$
((2))

If X(f) is finite banded and its frequency spectrum is in [−1/2T,1/2T], then its Nyquist sampling frequency should be f NYQ=1/T. Besides, if X(f) satisfies the structure shown in Fig. 1, namely, it contains N disjoint sub-frequency bands whose bandwidths are all less than B, then x(t) is a multiband signal.

Fig. 1
figure 1

Model of multiband signal

3 Modulated wideband converter

The model of MWC for multiband signal is shown as Fig. 2.

Fig. 2
figure 2

Model of MWC

Multiband modulated wideband signal converter is shown in Fig. 2. There are m sampling channels, and the mixing function p i (t) is a pseudo-random sequence with T p -period. Its value is {+1,−1}, and it has M pulses in each period whose interval is T s . The values of p i (t) in the first k intervals are denoted as α ik , h(t) is a low-pass filter whose cutoff frequency is 1/2T s , and the sampling frequency is 1/T s . The signal x(t) is transmitted through m channels at the same time, and it is multiplied with different pseudo-random sequences in each branch whose cycle is abiding and obey the same distribution. The samples y i [n],n=1,2,…,m can be obtained after low-pass filters and low-speed sampling.

And the fourier transform can be expressed as

$$ {}{Y_{i}}({e^{j2\pi f{T_{s}}}}) = \sum\limits_{n = -\infty}^{\infty} {{y_{i}}[n]{e^{-j2\pi fn{T_{s}}}} = \sum\limits_{l = -{L_{0}}}^{{L_{0}}} {{c_{il}} X\left(\,f - l{f_{p}}\right)}} $$
((3))

where f p =1/T p , F s =[−f s /2,f s /2], \({L_{0}} = \left [ {\frac {{f_{\textit {NYQ}}} + {f_{s}}}{2{f_{p}}}} \right ]- 1\), and \({c_{\textit {il}}} = \frac {1}{{T_{p}}} \int _{0}^{T_{p}} {{p_{i}}(t){e^{-j \frac {2\pi }{T_{p}} lt}}} dt\).

For convenience, the formula can be expressed in matrix form

$$ \textbf{y}(f) = {\textbf{A}\alpha}(f)f \in {\textbf{F}_{s}} $$
((4))

From Eq. (4), it can be seen that y(f) is a m×1 dimension vector which is made of \({Y_{i}}\left ({e^{j2\pi f{T_{s}}}}\right)\). A is a M×L dimensional matrix which is composed of \({A_{\textit {il}}} = {c_{i, -l}} = c_{\textit {il}}^{*}, -{L_{0}} \leq l \leq {L_{0}}\), L=2L 0+1, and Lf NYQ/B in common. Moreover, α(f) is a L×1 dimensional vector which is composed of α i (f)=X(f+(iL 0−1)f p ),fF s . To recover a signal, we need to firstly solve the sparse spectrum α(f) from Eq. (4). Then, the estimated value \(\hat {\mathbf {x}}(t)\) of x(t) is obtained by using the inverse Fourier transform. However, f is defined over a continuous interval whose samples are infinite dimensional vector. Thus, in fact, this is a problem about how to gain the sparse solution from one more of the numerous less demand equations. Moreover, it cannot be calculated by the conventional compressed sensing reconstruction algorithms.Therefore, we should get the support set S of signal by using CTF module and MMV problem can be turned into SMV problem by joint sparsity; then, SOMP algorithm is adopted. In CTF, we first use y[n] to construct a matrix as follows

$$ \textbf{Q} = \int_{f \in {\textbf{F}_{s}}} {\textbf{y}(f)} {\textbf{y}^{H}}(f)df = \sum\limits_{n = -\infty}^{+\infty} {\textbf{y}[n]{\textbf{y}^{T}}[n]} $$
((5))

where y[n]=[y 1[n],y 2[n],…,y m [n]]T, then matrix Q is divided into Q=V V H, then we can get framework matrix V. Then, the sparsest support set \(\bar {\mathbf {U}}\) of V=AU is obtained. According to the support set \(\bar {\mathbf {U}}\) and y[n], we can solve the signal’s support set S, and finally, compose the matrix A S by the columns in the matrix A which correspond to the elements of S, then reconstruct sparse spectrum α(f) by using Eq. (6), so as to reconstruct the traditional signal.

$$ \begin{aligned} {\mathbf{\alpha}_{\textbf{s}}}(f) = \textbf{A}_{\textbf{s}}^{\mathbf{\dag}}\textbf{y}(f) \\ {\alpha_{i}}(f) = 0,i \notin \textbf{S} \end{aligned} $$
((6))

4 Improved simultaneous orthogonal matching pursuit algorithm

The reconstruction algorithm for the MWC system is usually simultaneous orthogonal matching pursuit algorithm, whose steps are as follows:

  • Input: M×L dimensional sensing matrix A, the number of sub frequency bands K, m×2K dimensional frame vector V, and residual threshold θ.

  • Output: 2K×1 dimensional support set S.

  • Initialization: support set S=, residual matrix R=V.

  • Iteration: Repeat the following steps until the residual is less than the threshold or the number of iteration reach K.

  • (1) P=A T R.

  • (2) d k =∥P k 2,k=1,2,…,L, P k is kth column of matrix P.

  • (3) \({z_{\text {k}}} = \frac {d_{k}}{{\left \| {\textbf {A}_{k}} \right \|}_{2}}\), k=1,2,…,L, z={z 1,z 2…,z L }, A k is the kth column of matrix A.

  • (4) Find the largest item in the vector z and add its corresponding index k into the support set S, and the symmetric index value L+1−k is also added to the support set V. It is generally considered that the traditional signal is real signal, and its support set is symmetric.

  • (5) Construct the matrix A S corresponding to the columns in matrix A and support set S.

  • (6) Evaluate \(\hat {\mathbf {U}}\), \(\hat {\mathbf {U}} = \mathbf {A}_{\textbf {S}}^{\mathbf {\dag }} \textbf {V} = {\left (\textbf {A}_{\textbf {S}}^{T}{\textbf {A}_{\textbf {S}}}\right)^{-1}} \textbf {A}_{\textbf {S}}^{T}\textbf {V}\).

  • (7) Update signal residuals \(\textbf {R} = \textbf {V} - {\textbf {A}_{\textbf {S}}}\hat {\mathbf {U}}\).

It can be found that the inner product between atoms in sensing matrix and columns in residual matrix can be obtained from steps (1) to (4) firstly. Then, several inner products corresponding to the same atom constitute an inner product vector. By calculating its 2-norm, we can find the maximum value, whose corresponding atom is matched atom. The relevancy between residual matrix and sensing matrix actually depends on the relevancy between residual matrix columns and atoms in the sensing matrix. However, because of the randomness of sensing matrix and sampling, the inner product cannot accurately reflect the relevancy between vectors. Wrong atoms can be selected when the number of sampling points is relatively small. However, if the correlation coefficient is used as the criterion, it can reflect the correlation between the random variables better. The reason for the difference can be traced to the calculation method of the inner product and correlation coefficient. For an n dimensional vector \(\vec {a}\) and \(\vec {b}\), their inner product formula is

$$ \vec{a} \bullet \vec{b} = \sum\limits_{i = 1}^{n} {{a_{i}}{b_{i}}} $$
((7))

In two dimensions, the inner product can be expressed as

$$ \vec{a} \bullet \vec{b} = \left| {\vec{a}} \right|\left| {\vec{b}} \right|\cos \theta $$
((8))

The inner product itself cannot reflect the relevancy between vectors because it is influenced by the norm of vectors. But if \(\vec {a}\) and \(\vec {b}\) are the unit vectors, the inner product can reflect the relevancy between \(\vec {a}\) and \(\vec {b}\). If \(\vec {b}\) is regarded as the column vector of the residual matrix, while the vector \(\vec {a}\) is regarded as the atom in the matrix A, then, the inner product can reflect the relevancy between the residuals and the atoms.With this relative, the correlation coefficient can also reflect the relevancy, and its formula is

$$ {r_{XY}} = \frac{\sum\limits_{i = 1}^{N} {\left({X_{i}} - \bar{X}\right)\left({Y_{i}} - \bar{Y}\right)}} {\sqrt{\sum\limits_{i = 1}^{N} {{\left({X_{i}} - \bar{X}\right)}^{2}}} \sqrt{\sum\limits_{i = 1}^{N} {{\left({Y_{i}} - \bar{Y}\right)}^{2}}}} $$
((9))

It also excludes the effect of the norm of atoms, and the mean value of atom is zero. Although both the inner product and the correlation coefficient can reflect the relevancy, but in some cases, for the same signal and sensing matrix, the atoms they selected are different, just as the following example: There are two atoms X=(1,2,3), Y=(2,2,3), signal Z=(4,5,6), after normalizing them, we can get X∗,Y∗, and Z∗. If we use the inner product as screening criterion of atoms, the inner product of X∗ and Z∗ is 0.975 and the inner product of Y∗ and Z∗ is 0.995.The most matched atoms should be Y∗. However, when the correlation coefficient is used as screening criterion of atoms, the correlation coefficient of X∗ and Z∗ is 1 and the correlation coefficient of Y∗ and Z∗ is 0.866. The most matched atoms should be X∗. The results are different using two criterions. Then which one is better? To solve this problem, we need to analyze the process of signal reconstruction.

In CTF, the signal support set S is calculated indirectly by calculating the most sparse solution \(\bar {\mathbf {U}}\) of V=AU, since the support set of \(\bar {\mathbf {U}}\) and the support set of signal y[n] are the same. Then, construct the matrix A S corresponding to the columns of matrix A and support set S. Finally, reconstruct the sparse spectrum α(f) and the original signal by using Eq. (6). If

$$ \textbf{A} = \left[ { \begin{array}{ccccccc} {a_{11}} & \cdots & {a_{1q}} & \cdots & {a_{1i}} & \cdots & {a_{1L}} \\ 0\vdots & & \vdots & & \vdots & & \vdots \\ {a_{r1}} & \cdots & {a_{rq}} & \cdots & {a_{ri}} & \cdots & {a_{rL}} \\ \vdots & & \vdots & & \vdots & & \vdots \\ {a_{m1}} & \cdots & {a_{mq}} & \cdots & {a_{mi}} & \cdots & {a_{mL}} \end{array}} \right] $$
((10))

In order to simplify the analysis, V=AU is expressed as

$$ \begin{aligned} &\left[ {\begin{array}{ccccccc} {a_{11}} & \cdots & {a_{1q}} & \cdots & {a_{1i}} & \cdots & {a_{1L}} \\ \vdots & & \vdots & & \vdots & & \vdots \\ {a_{r1}} & \cdots & {a_{rq}} & \cdots & {a_{ri}} & \cdots & {a_{rL}} \\ \vdots & & \vdots & & \vdots & & \vdots \\ {a_{m1}} & \cdots & {a_{mq}} & \cdots & {a_{mi}} & \cdots & {a_{mL}} \end{array}} \right] \times \\ &\left[ {\begin{array}{ccccc} 0 & \cdots & 0 & \cdots & 0 \\ \vdots & & \vdots & & \vdots \\ {u_{q1}} & \cdots & {u_{qt}} & \cdots & {u_{qp}} \\ \vdots & & \vdots & & \vdots \\ {u_{i1}} & \cdots & {u_{it}} & \cdots & {u_{ip}} \\ \vdots & & \vdots & & \vdots \\ 0 & \cdots & 0 & \cdots & 0 \end{array}} \right] = \left[ \begin{array}{c} {a_{1q}{u_{q1}} + {a_{1i}}{u_{i1}}} \\ \vdots \\ {{a_{rq}}{u_{q1}} + {a_{ri}}{u_{i1}}} \\ \vdots \\ {{a_{mq}}{u_{q1}} + {a_{mi}}{u_{i1}}} \end{array} \right. \\ &\left. {\begin{array}{ccc} {\cdots \ {a_{1q}}{u_{qt}} + {a_{1i}}{u_{it}}} & \cdots & {{a_{iq}}{u_{qp}} + {a_{1i}}{u_{ip}}} \\ \vdots & & \vdots \\ {\cdots \ {a_{rq}}{u_{qt}} + {a_{ri}}{u_{it}}} & \cdots & {{a_{rq}}{u_{qp}} + {a_{ri}}{u_{ip}}} \\ \vdots & & \vdots \\ {\cdots \ {a_{mq}}{u_{qt}} + {a_{mi}}{u_{it}}} & \cdots & {{a_{mq}}{u_{qp}} + {a_{mi}}{u_{ip}}} \end{array}} \right] = \\ &\left[{{A_{q}}{u_{q1}} + {A_{i}}{u_{i1}} \cdots {A_{q}}{u_{qt}} + {A_{i}}{u_{it}} \cdots {A_{q}}{u_{qp}} + {A_{i}}{u_{ip}}} \right] \end{aligned} $$
((11))

In which, A q and A i are the qth column and ith column of matrix A, respectively. According to SOMP algorithm, the method of using inner product as screening criterion of atoms can be expressed as

$$ \begin{aligned} &\lambda = {\left\| {{A_{s}}\textbf{R}} \right\|_{2}} = {\left\| {{A_{s}}\textbf{V}} \right\|_{2}} = \\ &{\left(\begin{array}{l} {\left({A_{s}^{T}}{A_{q}}{u_{q1}} + {A_{s}^{T}}{A_{i}}{u_{i1}}\right)^{2}} + \cdots \\ + {\left({A_{s}^{T}}{A_{q}}{u_{qp}} + {A_{s}^{T}}{A_{i}}{u_{ip}}\right)^{2}} \end{array} \right)^{1/2}} \end{aligned} $$
((12))

We hope that when s=q or s=i, λ has the maximum value. Since the column vectors of matrix A are approximately orthogonal, one of \({A_{s}^{T}}{A_{q}}{u_{\textit {qt}}}\) and \({A_{s}^{T}}{A_{i}}{u_{\textit {it}}}\) is relatively large, while the other one is quite small. Assuming

$$ \begin{aligned} \lambda &\approx \sqrt{{{\left({A_{s}^{T}}{A_{q}}{u_{q1}}\right)}^{2}} + \cdots + {{\left({A_{s}^{T}}{A_{q}}{u_{qp}}\right)}^{2}}} \\ &= \sqrt{{{\left({A_{s}^{T}}{A_{q}}\right)}^{2}}\left(u_{q1}^{2} + u_{q2}^{2} + \cdots u_{qp}^{2}\right)} \end{aligned} $$
((13))

where the value of λ is independent of the matrix U, only related to A s . Most of the time,

$$ {\left({A_{q}^{T}}{A_{q}}\right)^{2}} > {\left({A_{s}^{T}}{A_{q}}\right)^{2}},s \neq q $$
((14))

But sometimes, the result is opposite (even if the atoms are normalized). Then, the SOMP algorithm will select the wrong atoms. And if the correlation coefficient is used as screening criterion of atoms,

$$ {r_{{A_{s}}{A_{q}}}} \leq 1 $$
((15))

if and only if when s=q, the equality holds. Thus, using correlation coefficient as screening criterion of atoms is more accurate.

Since the sensing matrix A and the residual matrix R are multidimensional, we cannot calculate the correlation coefficient between them directly, so we need to calculate the correlation coefficient between atoms in matrix A and columns in residual matrix R, and then squaring the correlation coefficients and sum the correlation coefficients corresponding to the same atom; finally, extract the root of the sum and find the largest one, whose corresponding atom is matched atom. This improvement will only change the first step of the SOMP algorithm into \({P_{k}} = \sum \limits _{i = 1}^{m} {r_{{\textbf {A}_{k}}{\textbf {R}_{i}}}}\), other steps remain unchanged.

5 Simulation results

To test whether the improved algorithm can improve the reconstruction probability, we compare the performance of MMV-ROMP, MMV-StOMP, MMV-CoSaMP, MMV-SP, and SOMP and their improved algorithms. We perform 1000 Monte-Carlo simulation for each algorithm. The signal expression is

$$ x(t) = \sum\limits_{n = 1}^{N/2} {\sqrt{B{E_{i}}} {\text{sinc}}\left(B(t - {\tau_{i}})\cos (2\pi {f_{i}}(t - {\tau_{i}})\right)} $$
((16))

The signal contains N frequency bands which are origin symmetric. The maximum bandwidth of sub-frequency bands is 50 MHz. The carrier frequency f i is a uniform distribution random number among [−5 GHz,5 GHz]. The Nyquist sampling rate is f NYQ=10 GHz. The energy E i of each frequency band is random, time shift τ i is a uniform distribution random number among [0,1]μ s. L 0=97,L=195,M=195,SNR=0, the sampling period, and the period of the pseudo random sequence p i (t) are f p =f s =f NYQ/L=51.28 MHz, respectively. The value of p i (t) is either 1 or −1 with equal probability and obey Bernoulli distribution. The numbers of sampling channels in Figs. 3 and 7 are 100, 100, 50, 50, and 120. The simulation results are as follows.

Fig. 3
figure 3

Reconstruction probability comparison between MMV-CoSaMP and improved MMV-CoSaMP

As can be seen from the simulation, new algorithms are improved in reconstruction probability compared with the traditional algorithms, especially for MMV-ICoSaMP algorithm and MMV-ISP algorithm, the improvement is very significant. When traditional algorithms are completely unable to reconstruct the signal, as is shown in Figs. 3 and 4, the improved algorithm still can reconstruct the signal with 100 % probability. It also means that less samples were needed to reconstruct signals than traditional algorithms when the number of bands is unchanged. The other three algorithms can also improve the reconstruction probability to a limited extent, as is shown in Figs. 5, 6 and 7. Because of the improvement about correlation coefficient, the complexity of the algorithm is increased. In addition, relatively small SNR has an adverse effect on reconstruction probability of improved algorithm, as is shown in Figs. 8 and 9, where the number of sampling channels is 100 and the number of bands is 20.

Fig. 4
figure 4

Reconstruction probability comparison between MMV-SP and improved MMV-SP

Fig. 5
figure 5

Reconstruction probability comparison between MMV-ROMP and improved MMV-ROMP

Fig. 6
figure 6

Reconstruction probability comparison between SOMP and improved SOMP

Fig. 7
figure 7

Reconstruction probability comparison between MMV-StOMP and improved MMV-StOMP

Fig. 8
figure 8

Reconstruction probability comparison between MMV-CoSaMP and improved MMV-CoSaMP in different SNRs

Fig. 9
figure 9

Reconstruction probability comparison between MMV-SP and improved MMV-SP in different SNRs

Although changing the inner product into correlation coefficient as screening criterion of atoms can improve the performance of the algorithm, calculating correlation coefficient at each iteration will cost a lot of time, so we can consider the possibility of simplified this process. A significant difference between the inner product and correlation coefficient is that correlation coefficient excludes the impact of data mean, so we can consider whether it is feasible if the column vectors of sensing matrix are centered or normalized firstly and still use inner product as screening criterion of matching atoms. Simulation result shows that compared to traditional MMV-SP algorithms, normalization cannot improve the reconstruction probability, but centralization can. If the column vectors of sensing matrix are centered firstly and still use inner product as screening criterion of matching atoms, the reconstruction probability is silently worse than MMV-ISP algorithm, just as shown in Fig. 10, where MMV-SISP means simplified MMV-ISP. The advantage of this simplification is that it can reduce the complexity of the proposed algorithm so that the reconstruction time is about the same as MMV-SP algorithm.

Fig. 10
figure 10

Comparison of the reconstructed probability of the simplified and unsimplified algorithms

6 Conclusions

In order to improve the reconstruction probability of MWC sampling system, we improve the SOMP algorithm and its derivative algorithms in this paper by changing the inner product into correlation coefficient as screening criterion of atoms. The simulation experiments show that the improved algorithm can increase the reconstruction probability, especially for MMV-ICoSaMP algorithm and MMV-ISP algorithm, the improvement is very significant. For SOMP, MMV-ROMP, and MMV-StOMP algorithms, the improvement can also increase the reconstruction probability to a limited extent. It also means that less samples were needed to reconstruct signals than traditional algorithms when the number of bands is unchanged. Because of the improvement of correlation coefficient, the complexity of the algorithms increase, so the reconstruction time correspondingly increase and the algorithm is not suitable for the system with high real-time requirements. For systems that require high accuracy, the improved algorithms can be used. Since calculating correlation coefficient at each iteration will cost a lot of time, we also proposed a simplified algorithm, which can also improve reconstruction probability and reconstruction time is about the same as traditional algorithms. In addition, relatively small SNR has an adverse effect on reconstruction probability of improved algorithms, so the improved algorithms are more suitable for low noise channels.