Abstract
The paper deals with the problem of finding sparse solutions to systems of polynomial equations possibly perturbed by noise. In particular, we show how these solutions can be recovered from group-sparse solutions of a derived system of linear equations. Then, two approaches are considered to find these group-sparse solutions. The first one is based on a convex relaxation resulting in a second-order cone programming formulation which can benefit from efficient reweighting techniques for sparsity enhancement. For this approach, sufficient conditions for the exact recovery of the sparsest solution to the polynomial system are derived in the noiseless setting, while stable recovery results are obtained for the noisy case. Though lacking a similar analysis, the second approach provides a more computationally efficient algorithm based on a greedy strategy adding the groups one-by-one. With respect to previous work, the proposed methods recover the sparsest solution in a very short computing time while remaining at least as accurate in terms of the probability of success. This probability is empirically analyzed to emphasize the relationship between the ability of the methods to solve the polynomial system and the sparsity of the solution.
Similar content being viewed by others
Notes
For example, consider the monomials \(u_1=x_1,\,u_2=x_2,\,u_3=x_1x_2,\,u_4=x_1^2 x_2\), then the structure of \(u_3\) and \(u_4\) is enforced by \(u_3=u_1u_2\) and \(u_4 = u_1u_3\). But note that since these constraints are relaxed in the final formulation, the estimation can yield \(u_3\ne u_1u_2\), which then recursively implies that all monomial constraints involving \(u_3\) are meaningless.
Note that since all the groups have the same number of variables, they need not be weighted by a function of the number of variables in each group.
The maximal number of iterations is the number of groups \(n\), but if the correct sparsity pattern is recovered then the algorithm stops earlier.
The code for the proposed methods is available at http://www.loria.fr/~lauer/software/.
Assume \(d> n\), then, the assumption of the Proposition leads to \(n> n(\Vert \varvec{x}\Vert _0 + n)\) and \(\Vert \varvec{x}\Vert _0 + n< 1\) which is impossible since \(n\ge 1\).
References
Andersen, E.D., Andersen, K.D.: The MOSEK interior point optimizer for linear programming: an implementation of the homogeneous algorithm. High Perform. Optim. 33, 197–232 (2000)
Balan, R., Casazza, P., Edidin, D.: On signal reconstruction without phase. Appl. Comput. Harmon. Anal. 20, 345–356 (2006)
Bandeira, A.S., Cahill, J., Mixon, D.G., Nelson, A.A.: Saving phase: injectivity and stability for phase retrieval. Appl. Comput. Harmon. Anal. 37(1), 106–125 (2014)
Beck, A., Eldar, Y.C.: Sparsity constrained nonlinear optimization: optimality conditions and algorithms. SIAM J. Optim. 23(3), 1480–1509 (2013)
Blumensath, T.: Compressed sensing with nonlinear observations and related nonlinear optimisation problems. IEEE Trans. Inf. Theory 59(6), 3466–3474 (2013)
Blumensath, T., Davies, M.E.: Gradient pursuit for non-linear sparse signal modelling. In: European Signal Processing Conference (EUSIPCO), pp. 25–29 (2008)
Bruckstein, A.M., Donoho, D.L., Elad, M.: From sparse solutions of systems of equations to sparse modeling of signals and images. SIAM Rev. 51(1), 34–81 (2009)
Candès, E.J.: Compressive sampling. In: Proceedings of the International Congress of Mathematicians: Invited Lectures, pp. 1433–1452 (2006)
Candès, E.J., Eldar, Y.C., Strohmer, T., Voroninski, V.: Phase retrieval via matrix completion. SIAM J. Imaging Sci. 6(1), 199–225 (2013)
Candès, E.J., Strohmer, T., Voroninski, V.: Phaselift: exact and stable signal recovery from magnitude measurements via convex programming. Commun. Pure Appl. Math. 66(8), 1241–1274 (2013)
Candès, E.J., Wakin, M.B., Boyd, S.P.: Enhancing sparsity by reweighted \(\ell _1\) minimization. J. Fourier Anal. Appl. 14(5), 877–905 (2008)
Deng, W., Yin, W., Zhang, Y.: Group sparse optimization by alternating direction method. Technical report TR11-06, Department of Computational and Applied Mathematics, Rice University (2011)
Donoho, D.L.: Compressed sensing. IEEE Trans. Inf. Theory 52(4), 1289–1306 (2006)
Donoho, D.L., Elad, M., Temlyakov, V.N.: Stable recovery of sparse overcomplete representations in the presence of noise. IEEE Trans. Inf. Theory 52(1), 6–18 (2006)
Donoho, D.L., Huo, X.: Uncertainty principles and ideal atomic decomposition. IEEE Trans. Inf. Theory 47(7), 2845–2862 (2001)
Ehler, M., Fornasier, M., Sigl, J.: Quasi-linear compressed sensing. Multiscale Model. Simul. 12(2), 725–754 (2014)
Eldar, Y.C., Kutyniok, G. (eds.): Compressed Sensing: Theory and Applications. Cambridge University Press, Cambridge (2012)
Fienup, J.: Phase retrieval algorithms: a comparison. Appl. Opt. 21(15), 2758–2769 (1982)
Foucart, S., Rauhut, H.: A Mathematical Introduction to Compressive Sensing. Springer, Berlin (2013)
Gerchberg, R., Saxton, W.: A practical algorithm for the determination of phase from image and diffraction plane pictures. Optik 35, 237–246 (1972)
Gonsalves, R.: Phase retrieval from modulus data. J. Opt. Soc. Am. 66(9), 961–964 (1976)
Grant, M., Boyd, S.: Graph implementations for nonsmooth convex programs. In: Blondel, V., Boyd, S., Kimura, H. (eds.) Recent Advances in Learning and Control, Lecture Notes in Control and Information Sciences, pp. 95–110. Springer. http://stanford.edu/~boyd/graph_dcp.html (2008)
Grant, M., Boyd, S.: CVX: Matlab software for disciplined convex programming, version 2.0 beta. http://cvxr.com/cvx (2013)
Kohler, D., Mandel, L.: Source reconstruction from the modulus of the correlation function: a practical approach to the phase problem of optical coherence theory. J. Opt. Soc. Am. 63(2), 126–134 (1973)
Le, V.L., Lauer, F., Bloch, G.: Selective \(\ell _1\) minimization for sparse recovery. IEEE Trans. Autom. Control 59(11) (to appear) (2014)
Marchesini, S.: Phase retrieval and saddle-point optimization. J. Opt. Soc. Am. A 24(10), 3289–3296 (2007)
Ohlsson, H., Eldar, Y.C.: On conditions for uniqueness in sparse phase retrieval. CoRR arXiv:abs/1308.5447 (2013)
Ohlsson, H., Yang, A.Y., Dong, R., Sastry, S.: Nonlinear basis pursuit. In: Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, pp. 315–319 (2013)
Ohlsson, H., Yang, A.Y., Dong, R., Verhaegen, M., Sastry, S.: Quadratic basis pursuit. arXiv:1301.7002 (2013)
Ranieri, J., Chebira, A., Lu, Y.M., Vetterli, M.: Phase retrieval for sparse signals: uniqueness conditions. CoRR arXiv:abs/1308.3058 (2013)
Vidal, R., Ma, Y., Sastry, S.: Generalized principal component analysis (GPCA). IEEE Trans. Pattern Anal. Mach. Intell. 27(12), 1945–1959 (2005)
Acknowledgments
Henrik Ohlsson gratefully acknowledges support from the NSF project FORCES (Foundations Of Resilient CybEr-physical Systems), the Swedish Research Council in the Linnaeus center CADICS, the European Research Council under the advanced grant LEARN, contract 267381, a postdoctoral grant from the Sweden–America Foundation, donated by ASEA’s Fellowship Fund, and a postdoctoral grant from the Swedish Research Council.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix 1: Bound on the sparsity level of \(\phi \)
Lemma 3
For all \((a,b,d) \in (\mathbb {N}^*)^3\) such that \(a\ge d(b+d)\), the inequality
holds.
Proof
For \(q\le d\), we can bound the terms in the sum as
where we used \(a\ge d(b+d)\) in the second inequality. Then
\(\square \)
Proposition 2
Let the mapping \(\phi : \mathbb {R}^n \rightarrow \mathbb {R}^M\) be defined as above. Then, with \(d\ge 3\) and \(n\ge d(\Vert \varvec{x}\Vert _0 + d)\), the vector \(\phi (\varvec{x})\) is sparser than the vector \(\varvec{x}\) in the sense that the inequality
holds for all \(\varvec{x}\in \mathbb {R}^n\).
Proof
By construction, the number of nonzeros in \(\phi (\varvec{x})\) is equal to the sum over \(q\), \(1\le q\le d\), of the number of monomials of degree \(q\) in \(\Vert \varvec{x}\Vert _0\) variables:
The assumption \(n\ge d(\Vert \varvec{x}\Vert _0 + d)\) implies thatFootnote 6 \(d\le n\). With \(d\le n\), we have
which yields
Now, on the one hand we have
and on the other hand, Lemma 3 yields
Thus,
\(\square \)
Appendix 2: Other conditions for sparse recovery
The following uses the exact value of \(\Vert \varvec{\phi }_0\Vert _0\).
Theorem 7
Let \(\varvec{x}_0\) denote the unique solution to (1)–(2). If the inequality
holds, then the solution \(\hat{\varvec{\phi }}\) to (5) is unique and equal to \(\phi (\varvec{x}_0)\), thus providing \(\hat{\varvec{x}} = \varvec{x}_0\).
Another more compact but slightly less tight result is as follows.
Theorem 8
Let \(\varvec{x}_0\) denote the unique solution to (1)–(2). If the inequality
holds, then the solution \(\hat{\varvec{\phi }}\) to (5) is unique and equal to \(\phi (\varvec{x}_0)\), thus providing \(\hat{\varvec{x}} = \varvec{x}_0\).
Proof
Since the terms in the sum of Theorem 7 form an increasing sequence, we have
which yields the sought statement by application of Theorem 7. \(\square \)
Appendix 3: Value of \(m\)
Let us define \(m\) as the number of monomials involving a base variable \(x\) with a degree \(\ge 1\). It can be computed as the sum over \(q,\, 0\le q\le d-1\) of the number of monomials of degree \(q\) in \(n-1\) variables times the remaining degree \(d-q\) (since each monomial in \(n-1\) variables can be multiplied by \(x\) or \(x^2 \ldots \) or \(x^{d-q}\) to produce a monomial of degree at most \(d\) in \(n\) variables):
Another technique computes \(m\) as the total number of all monomials minus the number of monomials not involving \(x\) which is the number of monomials in \(n-1\) variables:
Rights and permissions
About this article
Cite this article
Lauer, F., Ohlsson, H. Finding sparse solutions of systems of polynomial equations via group-sparsity optimization. J Glob Optim 62, 319–349 (2015). https://doi.org/10.1007/s10898-014-0225-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10898-014-0225-8