Abstract
During the past decades, sparsity has been shown to be of significant importance in fields such as compression, signal sampling and analysis, machine learning, and optimization. In fact, most natural data can be sparsely represented, i.e., a small set of coefficients is sufficient to describe the data using an appropriate basis. Sparsity is also used to enhance interpretability in real-life applications, where the relevant information therein typically resides in a low dimensional space. However, the true underlying structure of many signal processing and machine learning problems is often more sophisticated than sparsity alone. In practice, what makes applications differ is the existence of sparsity patterns among coefficients. In order to better understand the impact of such structured sparsity patterns, in this chapter we review some realistic sparsity models and unify their convex and non-convex treatments. We start with the general group sparse model and then elaborate on two important special cases: the dispersive and hierarchical models. We also consider more general structures as defined by set functions and present their convex proxies. Further, we discuss efficient optimization solutions for structured sparsity problems and illustrate structured sparsity in action via three applications in image processing, neuronal signal processing, and confocal imaging.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Other convex structured models that can be described as the composition of a simple function over a linear transformation D can be found in [1].
- 2.
The proposed norm originates from the composite absolute penalties (CAP) convex norm, proposed in [133], according to which
$$\displaystyle\begin{array}{rcl} g(x) =\sum _{G_{i}}\left (\sum _{j\in G_{i}}\vert x_{j}\vert ^{\gamma }\right )^{p},& & {}\end{array}$$(12.19)for various values of γ and p. Observe that this model also includes the famous group sparse model where \(g(x) =\sum _{G_{i}}\|x_{G_{i}}\|_{2}\), described in Section 12.3, for p = 1∕2 and γ = 2.
- 3.
A regular quad-tree is a finite tree whose nodes have exactly four children, leaves excluded.
- 4.
A monotone function is a function that satisfies: ∀S ⊆ T ⊆ N, R(S) ≤ R(T).
- 5.
A symmetric function is a function that satisfies: \(\forall S \subseteq N,R(S) = R(N\setminus S)\).
- 6.
Actually, it is a norm iff \(N = \cup _{d_{i}>0}\ G_{i}\).
- 7.
Consider the following example: Let N = { 1, 2, 3, 4}, and \(\mathfrak{G} =\{ G_{1} =\{ 1\},G_{2} =\{ 2, 3\},G_{3} =\{ 1, 2, 4\}\}\), with weights defined as d i = | G i | . Then the inequality in Definition 7 is not satisfied for the sets S = { 1, 2} for which R sc (S) = 3, and U = { 1, 2, 4} for which R sc (U) = 4, with the addition of the element {3}.
- 8.
We acknowledge that there are other criteria that can be considered in practice; for completeness, in the simple sparsity case, we refer the reader to the ℓ 1-norm constrained linear regression (a.k.a. Lasso [115])—similarly, there are alternative optimization approaches for the discrete case [127]. However, our intention in this chapter is to use the most prevalent formulations used in practice.
- 9.
For example, ℓ 1-norm models well the ℓ 0-“norm.”
- 10.
- 11.
In [120], the authors consider a more general class of functions with no global Lipschitz constant L over their domain. The description of this material is out of the scope of this chapter and is left to the reader who is interested in deeper convex analysis and optimization.
References
Argyriou, A., Micchelli, C., Pontil, M., Shen, L., Xu, Y.: Efficient first order methods for linear composite regularizers (2000). arXiv preprint arXiv:1104.1436
Bach, F.: Structured sparsity-inducing norms through submodular functions. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 118–126 (2010)
Bach, F.: Learning with submodular functions: a convex optimization perspective (2011). arXiv preprint arXiv:1111.6453
Bah, B., Baldassarre, L., Cevher, V.: Model-based sketching and recovery with expanders. In: Proceedings of ACM-SIAM Symposium on Discrete Algorithms (SODA) (2014)
Baldassarre, L., Bhan, N., Cevher, V., Kyrillidis, A.: Group-sparse model selection: Hardness and relaxations (2013). arXiv preprint arXiv:1303.3207
Baraniuk, R.: Optimal tree approximation with wavelets. In: Proceedings of SPIE’s International Symposium on Optical Science, Engineering, and Instrumentation, pp. 196–207. International Society for Optics and Photonics (1999)
Baraniuk, R., DeVore, R., Kyriazis, G., Yu, X.: Near best tree approximation. Adv. Comput. Math. 16(4), 357–373 (2002)
Baraniuk, R., Cevher, V., Duarte, M., Hegde, C.: Model-based compressive sensing. IEEE Trans. Inf. Theory 56(4), 1982–2001 (2010)
Baraniuk, R., Cevher, V., Wakin, M.: Low-dimensional models for dimensionality reduction and signal recovery: a geometric perspective. Proc. IEEE 98(6), 959–971 (2010)
Beck, A., Teboulle, M.: A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2(1), 183–202 (2009)
Bengio, Y.: Learning deep architectures for AI. Found. Trends Mach. Learn. 2(1), 1–127 (2009)
Bertsekas, D.: Projected Newton methods for optimization problems with simple constraints. SIAM J. Control Optim. 20(2), 221–246 (1982)
Bhan, N., Baldassarre, L., Cevher, V.: Tractability of interpretability via selection of group-sparse models. In: Proceedings of IEEE International Symposium on Information Theory (ISIT) (2013)
Blumensath, T., Davies, M.: Iterative hard thresholding for compressed sensing. Appl. Comput. Harmon. Anal. 27(3), 265–274 (2009)
Blumensath, T., Davies, M.: Sampling theorems for signals from the union of finite-dimensional linear subspaces. IEEE Trans. Inf. Theory 55(4), 1872–1882 (2009)
Bonnans, J.: Local analysis of Newton-type methods for variational inequalities and nonlinear programming. Appl. Math. Optim. 29, 161–186 (1994)
Born, M., Wolf, E.: Principles of Optics: Electromagnetic Theory of Propagation, Interference and Diffraction of Light. 7th edn. Cambridge University Press, Cambridge, UK (1999)
Borwein, J., Lewis, A.: Convex Analysis and Nonlinear Optimization: Theory and Examples. Springer-Verlag, New York, US (2006)
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 3(1), 1–122 (2011)
Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge, UK (2004)
Buchbinder, N., Feldman, M., Naor, J., Schwartz, R.: A tight linear time 1∕2-approximation for unconstrained submodular maximization. In: IEEE 53rd Annual Symposium on Foundations of Computer Science (FOCS), pp. 649–658 (2012)
Buchbinder, N., Feldman, M., Naor, J., Schwartz, R.: Submodular maximization with cardinality constraints. In: Proceedings of ACM-SIAM Symposium on Discrete Algorithms (SODA) (2014)
Candes, E.: Compressive sampling. In: Proceedings of the International Congress of Mathematicians: Madrid, August 22–30, 2006: Invited Lectures, pp. 1433–1452 (2006)
Cartis, C., Thompson, A.: An exact tree projection algorithm for wavelets (2013). arXiv preprint arXiv:1304.4570
Cevher, V., Hegde, C., Duarte, M., Baraniuk, R.: Sparse signal recovery using Markov random fields. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation (2009)
Chambolle, A., De Vore, R., Lee, N., Lucier, B.: Nonlinear wavelet image processing: Variational problems, compression, and noise removal through wavelet shrinkage. IEEE Trans. Image Process. 7(3), 319–335 (1998)
Chambolle, A., Pock, T.: A first-order primal–dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vis. 40(1), 120–145 (2011)
Chandrasekaran, V., Recht, B., Parrilo, P., Willsky, A.: The convex geometry of linear inverse problems. Found. Comput. Math. 12, 805–849 (2012)
Chen, S., Donoho, D., Saunders, M.: Atomic decomposition by basis pursuit. SIAM J. Sci. Comput. 20(1), 33–61 (1998)
Combettes, P., Wajs, V.: Signal recovery by proximal forward–backward splitting. Multiscale Model. Simulat. 4(4), 1168–1200 (2005)
Crouse, M., Nowak, R., Baraniuk, R.: Wavelet-based statistical signal processing using hidden Markov models. IEEE Trans. Signal Process. 46(4), 886–902 (1998)
Dahl, J., Vandenberghe, L., Roychowdhury, V.: Covariance selection for nonchordal graphs via chordal embedding. Optim. Methods Softw. 23(4), 501–520 (2008)
Das, A., Dasgupta, A., Kumar, R.: Selecting diverse features via spectral regularization. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 1592–1600 (2012)
Das, A., Kempe, D.: Submodular meets spectral: Greedy algorithms for subset selection, sparse approximation and dictionary selection (2011). arXiv preprint arXiv:1102.3975
Daubechies, I., Defrise, M., De Mol, C.: An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. 57(11), 1413–1457 (2004)
Donoho, D.: Compressed sensing. IEEE Trans. Inf. Theory 52(4), 1289–1306 (2006)
Dughmi, S.: Submodular functions: extensions, distributions, and algorithms: a survey (2009). arXiv preprint arXiv:0912.0322
Efron, B., Hastie, T., Johnstone, I., Tibshirani, R.: Least angle regression. Ann. Stat. 32(2), 407–499 (2004)
El Halabi, M., Baldassarre, L., Cevher, V.: To convexify or not? Regression with clustering penalties on graphs. In: IEEE 5th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP), pp. 21–24 (20130
Eldar, Y., Mishali, M.: Robust recovery of signals from a structured union of subspaces. IEEE Trans. Inf. Theory 55(11), 5302–5316 (2009)
Foucart, S.: Hard thresholding pursuit: an algorithm for compressive sensing. SIAM J. Numer. Anal. 49(6), 2543–2563 (2011)
Friedman, J., Hastie, T., Tibshirani, R.: A note on the group lasso and a sparse group lasso (2010). arXiv preprint arXiv:1001.0736
Fujishige, S., Isotani, S.: A submodular function minimization algorithm based on the minimum-norm base. Pac. J. Optim. 7(1), 3–17 (2011)
Fujishige, S., Patkar, S.: Realization of set functions as cut functions of graphs and hypergraphs. Discret. Math. 226(1), 199–210 (2001)
Fukushima, M., Mine, H.: A generalized proximal point algorithm for certain non-convex minimization problems. Int. J. Syst. Sci. 12(8), 989–1000 (1981)
Gerstner, W., Kistler, W.: Spiking Neuron Models: Single Neurons, Populations, Plasticity. Cambridge University Press, Cambridge, UK (2002)
Gilbert, A., Indyk, P.: Sparse recovery using sparse matrices. Proc. IEEE 98(6), 937–947 (2010)
Girosi, F.: An equivalence between sparse approximation and support vector machines. Neural Comput. 10(6), 1455–1480 (1998)
Goldberg, A., Rao, S.: Beyond the flow decomposition barrier. J. ACM 45(5), 783–797 (1998)
Goldstein, T., Donoghue, B., Setzer, S.: Fast Alternating Direction Optimization Methods. CAM Report, pp. 12–35 (2012)
Goy, A., Psaltis, D.: Digital confocal microscope. Opt. Exp. 20(20), 22720 (2012)
Gramfort, A., Kowalski, M.: Improving M/EEG source localization with an inter-condition sparse prior. In: Proceedings of IEEE International Symposium on Biomedical Imaging (2009)
Guigue, V., Rakotomamonjy, A., Canu, S.: Kernel basis pursuit. In: Machine Learning, pp. 146–157. Springer-Verlag, Berlin, Heidelberg (2005)
He, B., Yuan, X.: On the O(1∕n) convergence rate of the Douglas–Rachford alternating direction method. SIAM J. Numer. Anal. 50, 700–709 (2012)
He, L., Carin, L.: Exploiting structure in wavelet-based Bayesian compressive sensing. IEEE Trans. Signal Process. 57(9), 3488–3497 (2009)
Heckerman, D., Geiger, D., Chickering, D.: Learning Bayesian networks: the combination of knowledge and statistical data. Mach. Learn. 20(3), 197–243 (1995)
Hegde, C., Duarte, M., Cevher, V.: Compressive sensing recovery of spike trains using a structured sparsity model. In: Signal Processing with Adaptive Sparse Structured Representations (SPARS) (2009)
Hsieh, C., Sustik, M., Dhillon, I., Ravikumar, P.: Sparse inverse covariance matrix estimation using quadratic approximation. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 2330–2338 (2011)
Hsieh, C., Sustik, M., Dhillon, I., Ravikumar, P., Poldrack, R.: BIG & QUIC: Sparse inverse covariance estimation for a million variables. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 3165–3173 (2013)
Huang, J., Zhang, T.: The benefit of group sparsity. Ann. Stat. 38(4), 1978–2004 (2010)
Huang, J., Zhang, T., Metaxas, D.: Learning with structured sparsity. J. Mach. Learn. Res. 12, 3371–3412 (2011)
Indyk, P., Razenshteyn, I.: On model-based RIP-1 matrices. In: Automata, Languages, and Programming, pp. 564–575. Springer-Verlag, Berlin, Heidelberg (2013)
International Neuroinformatics Coordinating Faculty.: Spike time prediction – challenge C (2009)
Jacob, L., Obozinski, G., Vert, J.P.: Group lasso with overlap and graph lasso. In: Proceedings of The 30th International Conference on Machine Learning (ICML) (2009)
Jalali, A., Ravikumar, P., Vasuki, V., Sanghavi, S.: On learning discrete graphical models using group-sparse regularization. In: Proceedings of International Conference on Artificial Intelligence and Statistics, pp. 378–387 (2011)
Jegelka, S., Lin, H., Bilmes, J.: On fast approximate submodular minimization. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 460–468 (2011)
Jenatton, R., Audibert, J.-Y., Bach, F.: Structured variable selection with sparsity-inducing norms. J. Mach. Learn. Res. 12, 2777–2824 (2011)
Jenatton, R., Gramfort, A., Michel, V., Obozinski, G., Bach, F., Thirion, B.: Multi-scale mining of fMRI data with hierarchical structured sparsity. In: Pattern Recognition in NeuroImaging (PRNI) (2011)
Jenatton, R., Mairal, J., Obozinski, G., Bach, F.: Proximal methods for hierarchical sparse coding. J. Mach. Learn. Res. 12, 2297–2334 (2011)
Johnstone, I.: On the distribution of the largest eigenvalue in principal components analysis. Ann. Stat. 29(2), 295–327 (2001)
Kim, S., Xing, E.: Tree-guided group lasso for multi-task regression with structured sparsity. In: Proceedings of The 30th International Conference on Machine Learning (ICML), pp. 543–550 (2010)
Kolmogorov, V., Zabin, R.: What energy functions can be minimized via graph cuts? IEEE Trans. Pattern Anal. Mach. Intell. 26(2), 147–159 (2004)
Krause, A., Cevher, V.: Submodular dictionary selection for sparse representation. In: Proceedings of The 30th International Conference on Machine Learning (ICML), pp. 567–574 (2010)
Kyrillidis, A., Cevher, V.: Recipes on hard thresholding methods. In: Proceedings of 4th IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP) (2011)
Kyrillidis, A., Cevher, V.: Combinatorial selection and least absolute shrinkage via the clash algorithm. In: Proceedings of International Symposium on Information Theory Proceedings (ISIT), pp. 2216–2220 (2012)
Kyrillidis, A., Cevher, V.: Fast proximal algorithms for self-concordant function minimization with application to sparse graph selection. In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6585–6589 (2013)
Kyrillidis, A., Puy, G., Cevher, V.: Hard thresholding with norm constraints. In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3645–3648 (2012)
Lee, J., Hastie, T.: Structure learning of mixed graphical models. In: Proceedings of the Sixteenth International Conference on Artificial Intelligence and Statistics, pp. 388–396 (2013)
Loh, P., Wainwright, M.: Structure estimation for discrete graphical models: generalized covariance matrices and their inverses. Ann. Stat. 41(6), 3022–3049 (2013)
Lovász, L.: Submodular functions and convexity. In: Mathematical Programming The State of the Art, pp. 235–257. Springer-Verlag, Berlin, Heidelberg (1983)
Lustig, M., Donoho, D., Pauly, J.: Sparse MRI: the application of compressed sensing for rapid MR imaging. Magn. Reson. Med. 58(6), 1182–1195 (2007)
Mallat, S.: A Wavelet Tour of Signal Processing. Academic Press, Burlington, MA, US (1999)
Mallat, S., Zhang, Z.: Matching pursuits with time–frequency dictionaries. IEEE Trans. Signal Process. 41(12), 3397–3415 (1993)
Martins, A., Smith, N., Aguiar, P., Figueiredo, M.: Structured sparsity in structured prediction. In: proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 1500–1511 (2011)
McCoy, B., Wu, T.: The Two-Dimensional Ising Model. Harvard University Press, Cambridge, MA, US (1973)
Meier, L., Van De Geer, S., Bühlmann, P.: The group lasso for logistic regression. J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 70(1), 53–71 (2008)
Minsky, M.: Microscopy Apparatus. US Patent 3,013,467 (1961)
Mosci, S., Villa, S., Verri, A., Rosasco, L.: A primal–dual algorithm for group ℓ 1 regularization with overlapping groups. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation (2010)
Narasimhan, M., Jojic, N., Bilmes, J.: Q-Clustering. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation (2005)
Natarajan, B.: Sparse approximate solutions to linear systems. SIAM J. Comput. 24(2), 227–234 (1995)
Needell, D., Tropp, J.: COSAMP: iterative signal recovery from incomplete and inaccurate samples. Appl. Comput. Harmon. Anal. 26(3), 301–321 (2009)
Nemhauser, G., Wolsey, L.: Integer and Combinatorial Optimization, vol. 18. Wiley, New York (1988)
Nemhauser, G., Wolsey, L., Fisher, M.: An analysis of approximations for maximizing submodular set functions. Math. Program. 14(1), 265–294 (1978)
Nemirovskii, A.: Proximal-method with rate of convergence \(\mathcal{O}(1/t)\) for variational inequalities with Lipschitz continuous monotone operators and smooth convex–concave saddle point problems. SIAM J. Optim. 15(1), 229–251 (2004)
Nesterov, Y.: A method of solving a convex programming problem with convergence rate O(1∕k 2). Sov. Math. Dokl. 27, 372–376 (1983)
Nesterov, Y.: Excessive gap technique in nonsmooth convex minimization. SIAM J. Optim. 16(1), 235–249 (2005)
Nesterov, Y.: Smooth minimization of nonsmooth functions. Math. Program. 103(1), 127–152 (2005)
Nesterov, Y.: Primal–dual subgradient methods for convex problems. Math. Program. 120(1, Ser. B), 221–259 (2009)
Obozinski, G., Bach, F.: Convex relaxation for combinatorial penalties (2012). arXiv preprint arXiv:1205.1240
Obozinski, G., Jacob, L., Vert, J.: Group lasso with overlaps: The latent group lasso approach (2011). arXiv preprint arXiv:1110.0413
Orlin, J.: A faster strongly polynomial time algorithm for submodular function minimization. Math. Program. 118(2), 237–251 (2009)
Puig, A., Wiesel, A., Zaas, A., Woods, C., Ginsburg, G., Fleury, G., Hero, A.: Order-preserving factor analysis—application to longitudinal gene expression. IEEE Trans. Signal Process. 59, 4447–4458 (2011)
Rao, N., Nowak, R., Wright, S., Kingsbury, N.: Convex approaches to model wavelet sparsity patterns. In: Proceedings of 18th IEEE International Conference on Image Processing (ICIP), pp. 1917–1920 (2011)
Rao, N., Recht, B., Nowak, R.: Signal recovery in unions of subspaces with applications to compressive imaging (2012). arXiv preprint arXiv:1209.3079
Rapaport, F., Barillot, E., Vert, J.: Classification of arrayCGH data using fused SVM. Bioinformatics 24(13), 375–i382 (2008)
Rebafka, T., Lvy-Leduc, C., Charbit, M.: OMP-type algorithm with structured sparsity patterns for multipath radar signals (2011). arXiv preprint arXiv:1103.5158
Robinson, S.: Strongly regular generalized equations. Math. Oper. Res. 5, 43–62 (1980)
Schmidt, M., Roux, N.L., Bach, F.: Convergence rates of inexact proximal-gradient methods for convex optimization. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation (2011)
Seeger, M.: On the Submodularity of Linear Experimental Design. Technical Report (2009)
Shapiro, J.: Embedded image coding using zero trees of wavelet coefficients. IEEE Trans. Signal Process. 41(12), 3445–3462 (1993)
Sheppard, C., Shotton, D.: Confocal Laser Scanning Microscopy. BIOS Scientific Publishers, Garland Science, New York, US (1997)
Simon, N., Friedman, J., Hastie, T., Tibshirani, R.: A sparse-group lasso. J. Comput. Graph. Stat. 22(2), 231–245 (2013)
Stojnic, M., Parvaresh, F., Hassibi, B.: On the reconstruction of block-sparse signals with an optimal number of measurements. IEEE Trans. Signal Process. 57(8) 3075–3085 (2009)
Subramanian, A., Tamayo, P., Mootha, V.K., Mukherjee, S., Ebert, B.L., Gillette, M.A., Paulovich, A., Pomeroy, S.L., Golub, T.R., Lander, E.S., et al. Gene set enrichment analysis: a knowledge-based approach for interpreting genome-wide expression profiles. Proc. Natl. Acad. Sci. U. S. A. 102(43), 15545–15550 (2005)
Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Methodol.) 28(1), 267–288 (1996)
Tibshirani, R., Saunders, M., Rosset, S., Zhu, J., Knight, K.: Sparsity and smoothness via the fused lasso. J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 67(1), 91–108 (2005)
Tran-Dinh, Q., Cevher, V.: An Optimal Primal–Dual Decomposition Framework. Technical Report, LIONS – EPFL (2014)
Tran-Dinh, Q., Cevher, V.: A Unified Optimal Primal–Dual Framework for Constrained Convex Minimization. Technical Report, LIONS, pp. 1–32 (2014)
Tran-Dinh, Q., Cevher, V.: Constrained convex minimization via model-based excessive gap. In: Proceedings of the Neural Information Processing Systems Foundation Conference (NIPS) (2014)
Tran Dinh, Q., Kyrillidis, A., Cevher, V.: Composite self-concordant minimization (2013). arXiv preprint arXiv:1308.2867
Tran Dinh, Q., Kyrillidis, A., Cevher, V.: A proximal Newton framework for composite minimization: graph learning without Cholesky decompositions and matrix inversions. In: Proceedings of The 30th International Conference on Machine Learning (ICML), pp. 271–279 (2013)
Tropp, J., Gilbert, A.: Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans. Inf. Theory 53(12), 4655–4666 (2007)
Tseng, P.: Applications of splitting algorithm to decomposition in convex programming and variational inequalities. SIAM J. Optim. 29, 119–138 (1991)
Villa, S., Rosasco, L., Mosci, S., Verri, A.: Proximal methods for the latent group lasso penalty. Comput. Optim. Appl. 58(2), 1–27 (2012)
Villa, S., Salzo, S., Baldassarre, L., Verri, A.: Accelerated and inexact forward–backward algorithms. SIAM J. Optim. 23(3), 1607–1633 (2013)
Vincent, M., Hansen, N.: Sparse group lasso and high dimensional multinomial classification. Comput. Stat. Data Anal. 71, 771–786 (2014)
Wright, S., Nowak, R., Figueiredo, M.: Sparse reconstruction by separable approximation. IEEE Trans. Signal Process. 57(7), 2479–2493 (2009)
Wright, S., Nocedal, J.: Numerical Optimization. Springer, New York (1999)
Yuan, L., Liu, J., Ye, J.: Efficient methods for overlapping group lasso. In: Proceedings of Neural Information Processing Systems (NIPS) Foundation, pp. 352–360 (2011)
Yuan, M., Lin, Y.: Model selection and estimation in regression with grouped variables. J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 68(1), 49–67 (2006)
Zeng, X., Figueiredo, M.: A novel sparsity and clustering regularization (2013). arXiv preprint arXiv:1310.4945
Zhang, Z., Shi, Y., Yin, B.: MR images reconstruction based on TV-group sparse model. In: Proceedings of IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6 (2013)
Zhao, P., Rocha, G., Yu, B.: The composite absolute penalties family for grouped and hierarchical variable selection. Ann. Stat. 37(6A), 3468–3497 (2009)
Zhou, H., Sehl, M.E., Sinsheimer, J.S., Lange, K.: Association screening of common and rare genetic variants by penalized regression. Bioinformatics 26(19), 2375 (2010)
Zhou, Y., Jin, R., Hoi, S.: Exclusive lasso for multi-task feature selection. In: Proceedings of International Conference on Artificial Intelligence and Statistics, pp. 988–995 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Kyrillidis, A., Baldassarre, L., Halabi, M.E., Tran-Dinh, Q., Cevher, V. (2015). Structured Sparsity: Discrete and Convex Approaches. In: Boche, H., Calderbank, R., Kutyniok, G., Vybíral, J. (eds) Compressed Sensing and its Applications. Applied and Numerical Harmonic Analysis. Birkhäuser, Cham. https://doi.org/10.1007/978-3-319-16042-9_12
Download citation
DOI: https://doi.org/10.1007/978-3-319-16042-9_12
Publisher Name: Birkhäuser, Cham
Print ISBN: 978-3-319-16041-2
Online ISBN: 978-3-319-16042-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)