Skip to main content
Log in

The predictive Lasso

  • Published:
Statistics and Computing Aims and scope Submit manuscript

Abstract

We propose a shrinkage procedure for simultaneous variable selection and estimation in generalized linear models (GLMs) with an explicit predictive motivation. The procedure estimates the coefficients by minimizing the Kullback-Leibler divergence of a set of predictive distributions to the corresponding predictive distributions for the full model, subject to an l 1 constraint on the coefficient vector. This results in selection of a parsimonious model with similar predictive performance to the full model. Thanks to its similar form to the original Lasso problem for GLMs, our procedure can benefit from available l 1-regularization path algorithms. Simulation studies and real data examples confirm the efficiency of our method in terms of predictive performance on future observations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  • Aitchison, J.: Goodness of prediction fit. Biometrika 62, 547–554 (1975)

    Article  MathSciNet  MATH  Google Scholar 

  • Bailey, C.: Smart Exercise: Burning Fat, Getting Fit. Houghton-Mifflin, Boston (1994)

    Google Scholar 

  • Brown, P.J., Vannucci, M., Fearn, T.: Bayes model averaging with selection of regressors. J. R. Stat. Soc. B 64, 519–536 (2002)

    Article  MathSciNet  MATH  Google Scholar 

  • Burnham, K.P., Anderson, D.: Model Selection and Multimodel Inference: A Practical Information-Theoretic Approach. Springer, New York (2002)

    MATH  Google Scholar 

  • Chen, M.H., Ibrahim, J.G.: Conjugate priors for generalized linear models. Stat. Sin. 13, 461–476 (2003)

    MathSciNet  MATH  Google Scholar 

  • Dupuis, J.A., Robert, C.P.: Variable selection in qualitative models via an entropic explanatory power. J. Stat. Plan. Inference 111, 77–94 (2003)

    Article  MathSciNet  MATH  Google Scholar 

  • Geisser, S.: Discussion of “Sampling and Bayes’ inference in scientific modelling and robustness'' by G.E.P. Box. J. R. Stat. Soc., Ser. A 143, 416–417 (1980)

    Google Scholar 

  • Geisser, S.: Predictive Inference: An Introduction. Chapman & Hall, New York (1993)

    MATH  Google Scholar 

  • Gelman, A., Jakulin, A., Grazia, P., Su, Y.-S.: A weakly informative default prior distribution for logistic and other regression models. Ann. Appl. Stat. 2, 1360–1383 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  • Gneiting, T., Raftery, A.: Strictly proper scoring rules, prediction, and estimation. J. Am. Stat. Assoc. 102, 359–378 (2007)

    Article  MathSciNet  MATH  Google Scholar 

  • Good, I.J.: Rational decisions. J. R. Stat. Soc. B 14, 107–114 (1952)

    MathSciNet  Google Scholar 

  • Hersbach, H.: Decomposition of the continuous ranked probability score for ensemble prediction systems. Weather Forecast. 15, 559–570 (2000)

    Article  Google Scholar 

  • Hoeting, J.A., Madigan, D., Raftery, A.E., Volinsky, C.T.: Bayesian model averaging: a tutorial. Stat. Sci. 14, 382–417 (1999)

    Article  MathSciNet  MATH  Google Scholar 

  • Johnson, R.W.: Fitting percentage of body fat to simple body measurements. J. Stat. Educ. 4, 1 (1996)

    Google Scholar 

  • Leng, C., Tran, M.-N., Nott, D.J.: Bayesian adaptive Lasso. Technical Report (2010). arXiv:1009.2300v1

  • Lindley, D.V.: The choice of variables in multiple regression (with discussion). J. R. Stat. Soc. B 30, 31–66 (1968)

    MathSciNet  Google Scholar 

  • Nott, D.J., Leng, C.: Bayesian projection approaches to variable selection in generalized linear models. Comput. Stat. Data Anal. 54, 3227–3241 (2010)

    Article  MathSciNet  Google Scholar 

  • O’Hagan, A., Forster, J.: The Advanced Theory of Statistics, Bayesian Inference, vol. 2B. Edward Arnold, London (2004)

    Google Scholar 

  • Park, T., Casella, G.: The Bayesian Lasso. J. Am. Stat. Assoc. 103, 681–686 (2008)

    Article  MathSciNet  MATH  Google Scholar 

  • Raftery, A.E., Madigan, D., Hoeting, J.A.: Bayesian model averaging for linear regression models. J. Am. Stat. Assoc. 92, 179–191 (1997)

    MathSciNet  MATH  Google Scholar 

  • Tibshirani, R.: Regression shrinkage and selection via the Lasso. J. R. Stat. Soc. B 58, 267–288 (1996)

    MathSciNet  MATH  Google Scholar 

  • Tran, M.N.: A criterion for optimal predictive model selection. Commun. Stat., Theory Methods 40, 893–906 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  • Vehtari, A., Lampinen, J.: Model selection via predictive explanatory power. Report B38, Laboratory of Computational Engineering, Helsinki University of Technology (2004)

  • Zellner, A.: On assessing prior distributions and Bayesian regression analysis with g-prior distributions. In: Bayesian Inference and Decision Techniques: Essays in Honour of Bruno De Finetti, pp. 233–243. North-Holland, Amsterdam (1986)

    Google Scholar 

  • Zhao, P., Yu, B.: On model selection consistency of Lasso. J. Mach. Learn. Res. 7, 2541–2563 (2006)

    MathSciNet  MATH  Google Scholar 

  • Zou, H.: The adaptive Lasso and its oracle properties. J. Am. Stat. Assoc. 101, 1418–1429 (2006)

    Article  MATH  Google Scholar 

  • Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. R. Stat. Soc. B 67, 301–320 (2005)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Minh-Ngoc Tran.

Additional information

The authors would like to thank the Editor, Associate Editor and referees for insightful comments which helped to improve the manuscript.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Tran, MN., Nott, D.J. & Leng, C. The predictive Lasso. Stat Comput 22, 1069–1084 (2012). https://doi.org/10.1007/s11222-011-9279-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11222-011-9279-3

Keywords

Navigation