Skip to main content
Log in

An unexpected connection between Bayes A-optimal designs and the group lasso

  • Regular Article
  • Published:
Statistical Papers Aims and scope Submit manuscript

A Publisher Correction to this article was published on 12 September 2019

This article has been updated

Abstract

We show that the A-optimal design optimization problem over m design points in \({\mathbb {R}}^n\) is equivalent to minimizing a quadratic function plus a group lasso sparsity inducing term over \(n\times m\) real matrices. This observation allows to describe several new algorithms for A-optimal design based on splitting and block coordinate decomposition. These techniques are well known and proved powerful to treat large scale problems in machine learning and signal processing communities. The proposed algorithms come with rigorous convergence guarantees and convergence rate estimate stemming from the optimization literature. Performances are illustrated on synthetic benchmarks and compared to existing methods for solving the optimal design problem.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

Change history

  • 12 September 2019

    Unfortunately, due to a technical error, the articles published in issues 60:2 and 60:3 received incorrect pagination. Please find here the corrected Tables of Contents. We apologize to the authors of the articles and the readers.

Notes

  1. When \(\mu \) is the uniform measure over the design space, we point out that the IMSE criterion is sometimes called I-optimality, or IV-optimality (for integrated variance); see Pukelsheim (1993).

References

  • Atkinson A, Donev A (1992) Optimum experimental designs. Oxford Statistical Science Series, vol 8. Oxford University Press, Oxford

    Google Scholar 

  • Bach F (2008) Consistency of the group lasso and multiple kernel learning. J Mach Learn Res 9:1179–1225 (Jun)

    MathSciNet  MATH  Google Scholar 

  • Bach F, Jenatton R, Mairal J, Obozinski G (2012) Optimization with sparsity-inducing penalties. Found Trends Mach Learn 4(1):1–106

    Article  MATH  Google Scholar 

  • Beck A, Teboulle M (2009) A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J Imaging Sci 2(1):183–202

    Article  MathSciNet  MATH  Google Scholar 

  • Beck A, Pauwels E, Sabach S (2015) The cyclic block conditional gradient method for convex optimization problems. SIAM J Optim 25(4):2024–2049

    Article  MathSciNet  MATH  Google Scholar 

  • Bertsekas DP (1999) Nonlinear programming, 2nd edn. Athena Scientific, Belmont

    MATH  Google Scholar 

  • Böhning D (1986) A vertex-exchange-method in D-optimal design theory. Metrika 33(1):337–347

    Article  MathSciNet  MATH  Google Scholar 

  • Borwein J, Lewis AS (2010) Convex analysis and nonlinear optimization: theory and examples. Springer, New York

    Google Scholar 

  • Boyd S, Vandenberghe L (2004) Convex optimization. Cambridge University Press, Cambridge

    Book  MATH  Google Scholar 

  • Braun G, Pokutta S, Zink D (2017) Lazifying conditional gradient algorithms. In: International conference on machine learning, pp 566–575

  • Černỳ M, Hladík M (2012) Two complexity results on c-optimality in experimental design. Comput Optim Appl 51(3):1397–1408

    Article  MathSciNet  MATH  Google Scholar 

  • Chaloner K (1984) Optimal Bayesian experimental design for linear models. Ann Stat 12:283–300

    Article  MathSciNet  MATH  Google Scholar 

  • Chambolle A, Dossal C (2015) On the convergence of the iterates of “FISTA”. J Optim Theory Appl 166(3):25

    Article  MATH  Google Scholar 

  • Combettes PL, Pesquet JC (2011) Proximal splitting methods in signal processing. In: Fixed-point algorithms for inverse problems in science and engineering, pp 185–212. Springer, New York

  • Duncan G, DeGroot M (1976) A mean squared error approach to optimal design theory. In: Proceedings of the 1976 conference on Information: science and systems, pp 217–221. The John Hopkins University, Baltimore

  • Fedorov V (1972) Theory of optimal experiments. Academic Press, New York. Translated and edited by W. J. Studden and E.M. Klimko

  • Fedorov V (1996) Design of spatial experiments: model fitting and prediction. In: Gosh S, Rao C (eds) Handbook of statistics, Chap. 16, vol 13. Elsevier, New York, pp 515–553

    Google Scholar 

  • Fedorov V, Lee J (2000) Design of experiments in statistics. In: Wolkowicz H, Saigal R, Vandenberghe L (eds) Handbook of semidefinite programming. Kluwer, Boston

    Google Scholar 

  • Frank M, Wolfe P (1956) An algorithm for quadratic programming. Naval Res Logist Q 3(1–2):95–110

    Article  MathSciNet  Google Scholar 

  • Gauthier B, Pronzato L (2016) Optimal design for prediction in random field models via covariance kernel expansions. In: mODa 11-advances in model-oriented design and analysis, pp 103–111. Springer

  • Gauthier B, Pronzato L (2017) Convex relaxation for imse optimal design in random-field models. Comput Stat Data Anal 113:375–394

    Article  MathSciNet  MATH  Google Scholar 

  • Harman R, Filová L, Richtárik P (2018) A randomized exchange algorithm for computing optimal approximate designs of experiments . arXiv:1801.05661

  • Kerdreux T, Pedregosa F, d’Aspremont A (2018) Frank-Wolfe with subsampling oracle. arXiv:1803.07348

  • Lukemire J, Mandal A, Wong W (2018) D-QPSO: a quantum-behaved particle swarm technique for finding d-optimal designs with discrete and continuous factors and a binary response. In: Technometrics, pp 1–27

  • Nesterov Y (1983) A method of solving a convex programming problem with convergence rate O (1/k2). Sov Math Dokl 27:372–376

    MATH  Google Scholar 

  • Pilz J (1983) Bayesian estimation and experimental design in linear regression models, vol 55. Teubner-Texte zur Mathematik, Leipzig

    MATH  Google Scholar 

  • Pronzato L (2013) A delimitation of the support of optimal designs for kiefer’s \(\phi \)p-class of criteria. Stat Probab Lett 83(12):2721–2728

    Article  MathSciNet  MATH  Google Scholar 

  • Pukelsheim F (1993) Optimal design of experiments. Wiley, New York

    MATH  Google Scholar 

  • Pukelsheim F, Rieder S (1992) Efficient rounding of approximate designs. Biometrika 79:763–770

    Article  MathSciNet  Google Scholar 

  • Rockafellar RT (1970) Convex analysis. Princeton Mathematical Series, vol 28. Princeton University Press, Princeton

    Google Scholar 

  • Rockafellar RT, Wets RJB (2009) Variational analysis, vol 317. Springer, New York

    MATH  Google Scholar 

  • Sagnol G (2011) Computing optimal designs of multiresponse experiments reduces to second-order cone programming. J Stat Plan Inference 141(5):1684–1708

    Article  MathSciNet  MATH  Google Scholar 

  • Sagnol G, Harman R (2015) Computing exact D-optimal designs by mixed integer second-order cone programming. Ann Stat 43(5):2198–2224

    Article  MathSciNet  MATH  Google Scholar 

  • Sagnol G, Hege H.C, Weiser M (2016) Using sparse kernels to design computer experiments with tunable precision. In: Proceedings of the 22nd international conference on computational statistics, pp 397–408

  • Silvey S, Titterington D, Torsney B (1978) An algorithm for optimal designs on a finite design space. Commun Stat Theory Methods 7(14):1379–1389

    Article  MATH  Google Scholar 

  • Spöck G, Pilz J (2010) Spatial sampling design and covariance-robust minimax prediction based on convex design ideas. Stoch Environ Res Risk Assess 24(3):463–482

    Article  MATH  Google Scholar 

  • Tanaka K, Miyakawa M (2013) The group lasso for design of experiments. arXiv:1308.1196

  • Wright SJ (2015) Coordinate descent algorithms. Math Progr 151(1):3–34

    Article  MathSciNet  MATH  Google Scholar 

  • Wynn H (1970) The sequential generation of \(D\)-optimum experimental designs. Ann Math Stat 41:1655–1664

    Article  MathSciNet  MATH  Google Scholar 

  • Yu Y (2010) Monotonic convergence of a general algorithm for computing optimal designs. Ann Stat 38(3):1593–1606

    Article  MathSciNet  MATH  Google Scholar 

  • Yu Y (2011) D-optimal designs via a cocktail algorithm. Stat Comput 21(4):475–481

    Article  MathSciNet  MATH  Google Scholar 

  • Yuan M, Lin Y (2006) Model selection and estimation in regression with grouped variables. J R Stat Soc B 68(1):49–67

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

Both authors would like to thank two anonymous reviewers for their careful reading and detailed comments which helped improve the quality of this manuscript.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guillaume Sagnol.

Additional information

This work was initiated when the first author was invited in Toulouse by the Chair in Applied Mathematics OQUAIDO, gathering partners in technological research (BRGM, CEA, IFPEN, IRSN, Safran, Storengy) and academia (CNRS, Ecole Centrale de Lyon, Mines Saint-Etienne, University of Grenoble, University of Nice, University of Toulouse) around advanced methods for Computer Experiments. The research of the first author is carried out in the framework of MATHEON supported by Einstein Foundation Berlin.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Sagnol, G., Pauwels, E. An unexpected connection between Bayes A-optimal designs and the group lasso. Stat Papers 60, 565–584 (2019). https://doi.org/10.1007/s00362-018-01062-y

Download citation

  • Received:

  • Revised:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00362-018-01062-y

Keywords

Mathematics Subject Classification

Navigation