Skip to main content

Sparse Signal Recovery with Exponential-Family Noise

  • Chapter
  • First Online:
Compressed Sensing & Sparse Filtering

Part of the book series: Signals and Communication Technology ((SCT))

Abstract

The problem of sparse signal recovery from a relatively small number of noisy measurements has been studied extensively in the recent compressed sensing literature. Typically, the signal reconstruction problem is formulated as \(l_1\)-regularized linear regression. From a statistical point of view, this problem is equivalent to maximum a posteriori probability (MAP) parameter estimation with Laplace prior on the vector of parameters (i.e., signal) and linear measurements disturbed by Gaussian noise. Classical results in compressed sensing (e.g., [7]) state sufficient conditions for accurate recovery of noisy signals in such linear-regression setting. A natural question to ask is whether one can accurately recover sparse signals under different noise assumptions. Herein, we extend the results of [7] to the general case of exponential-family noise that includes Gaussian noise as a particular case; the recovery problem is then formulated as \(l_1\)-regularized Generalized Linear Model (GLM) regression. We show that, under standard restricted isometry property (RIP) assumptions on the design matrix, \(l_1\)-minimization can provide stable recovery of a sparse signal in presence of exponential-family noise, and state some sufficient conditions on the noise distribution that guarantee such recovery.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Banerjee A, Merugu S, Dhillon IS, Ghosh J (2005) Clustering with Bregman divergences. J Mach Learn Res 6:1705–1749

    MathSciNet  MATH  Google Scholar 

  2. Banerjee A, Merugu S, Dhillon I, and Ghosh J (2004) Clustering with Bregman divergences. In: Proceedings of the fourth SIAM international conference on data mining, pp 234–245

    Google Scholar 

  3. Beygelzimer A, Kephart J, and Rish I (2007) Evaluation of optimization methods for network bottleneck diagnosis. In: Proceedings of ICAC-07

    Google Scholar 

  4. Candes E (2006) Compressive sampling. Int Cong Math 3:1433–1452

    MathSciNet  Google Scholar 

  5. Candes E, Romberg J (2006) Quantitative robust uncertainty principles and optimally sparse decompositions. Found Comput Math 6(2):227–254

    Article  MathSciNet  MATH  Google Scholar 

  6. Candes E, Romberg J, Tao T (2006) Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information. IEEE Trans Inf Theory 52(2):489–509

    Article  MathSciNet  MATH  Google Scholar 

  7. Candes E, Romberg J, Tao T (2006) Stable signal recovery from incomplete and inaccurate measurements. Commun Pure Appl Math 59(8):1207–1223

    Article  MathSciNet  MATH  Google Scholar 

  8. Candes E, Tao T (2005) Decoding by linear programming. IEEE Trans Inf Theory 51(12):4203–4215

    Google Scholar 

  9. Carroll MK, Cecchi GA, Rish I, Garg R, Rao AR (2009) Prediction and interpretation of distributed neural activity with sparse models. Neuroimage 44(1):112–122

    Google Scholar 

  10. Chandalia G, Rish I (2007) Blind source separation approach to performance diagnosis and dependency discovery. In: Proceedings of IMC-2007

    Google Scholar 

  11. Donoho D (2006) Compressed sensing. IEEE Trans Inf Theory 52(4):1289–1306

    Article  MathSciNet  Google Scholar 

  12. Donoho D (2006) For most large underdetermined systems of linear equations, the minimal ell-1 norm near-solution approximates the sparsest near-solution. Commun Pure Appl Math 59(7):907–934

    Article  MathSciNet  Google Scholar 

  13. Donoho D (2006) For most large underdetermined systems of linear equations, the minimal ell-1 norm solution is also the sparsest solution. Commun Pure Appl Math 59(6):797–829

    Google Scholar 

  14. Mitchell TM, Hutchinson R, Niculescu RS, Pereira F, Wang X, Just M, Newman S (2004) Learning to decode cognitive states from brain images. Mach Learn 57:145–175

    Article  MATH  Google Scholar 

  15. Negahban S, Ravikumar P, Wainwright MJ, Yu B (2009) A unified framework for the analysis of regularized \(M\)-estimators. In: Proceedings of neural information processing systems (NIPS)

    Google Scholar 

  16. Negahban S, Ravikumar P, Wainwright MJ, Yu B (2010) A unified framework for the analysis of regularized \(M\)-estimators. Technical Report 797, Department of Statistics, UC Berkeley

    Google Scholar 

  17. Park Mee-Young, Hastie Trevor (2007) An L1 regularization-path algorithm for generalized linear models. JRSSB 69(4):659–677

    Article  MathSciNet  Google Scholar 

  18. Rish I, Brodie M, Ma S, Odintsova N, Beygelzimer A, Grabarnik G, Hernandez K (2005) Adaptive diagnosis in distributed systems. IEEE Trans Neural Networks (special issue on Adaptive learning systems in communication networks) 16(5):1088–1109

    Google Scholar 

  19. Rish I, Grabarnik G, (2009) Sparse signal recovery with Exponential-family noise. In: Proceedings of the 47-th annual allerton conference on communication, control and, computing

    Google Scholar 

  20. Rockafeller RT, (1970) Convex analysis. Princeton university press. New Jersey

    Google Scholar 

  21. Zheng A, Rish I, Beygelzimer A (2005) Efficient test selection in active diagnosis via entropy approximation. In: Proceedings of UAI-05

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Irina Rish .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Rish, I., Grabarnik, G. (2014). Sparse Signal Recovery with Exponential-Family Noise. In: Carmi, A., Mihaylova, L., Godsill, S. (eds) Compressed Sensing & Sparse Filtering. Signals and Communication Technology. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-38398-4_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-38398-4_3

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-38397-7

  • Online ISBN: 978-3-642-38398-4

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics