Abstract
We address the development of a robust variable selection procedure using the density power divergence with the least absolute shrinkage and selection operator (LASSO). It produces robust estimates of the regression parameters and simultaneously selects the important explanatory variables. The asymptotic distribution of the regression coefficients is derived. The widely used model selection procedures based on the classical information criteria often show very poor performance in the presence of heavy-tailed error or outliers. For this purpose, we introduce a robust version of the Mallows’s \(C_p\) statistic based on our proposed method. The simulation studies show that the robust variable selection technique outperforms the classical likelihood-based techniques in the presence of outliers. The performance of the proposed method is explored through a healthcare data analysis.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Akaike, H.: Information theory and an extension of the maximum likelihood principle. In: Proceedings 2nd International Symposium on Information Theory, pp. 267–281. Akadémiai Kiadó, Budapest (1973)
Bassett, G., Jr., Koenker, R.: Asymptotic theory of least absolute error regression. J. Am. Statist Assoc. 73(363), 618–622 (1978)
Basu, A., Harris, I.R., Hjort, N.L., Jones, M.C.: Robust and efficient estimation by minimising a density power divergence. Biometrika 85(3), 549–559 (1998)
Fan, J., Li, R.: Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Statist Assoc. 96(456), 1348–1360 (2001)
Frank, L.E., Friedman, J.H.: A statistical view of some chemometrics regression tools. Technometrics 35(2), 109–135 (1993)
Ghosh, A., Basu, A.: Robust estimation for independent non-homogeneous observations using density power divergence with applications to linear regression. Electron. J. Statist 7, 2420–2456 (2013)
Ghosh, A., Majumdar, S.: Ultrahigh-dimensional robust and efficient sparse regression using non-concave penalized density power divergence. IEEE Trans. Inform. Theor. 66(12), 7812–7827 (2020)
Huber, P.J.: Robust Statistics. Wiley, New York (1981)
Kawashima, T., Fujisawa, H.: Robust and sparse regression via \(\gamma \)-divergence. Entropy 19(11), 608:e19110608 (2017)
Koenker, R., Hallock, K.F.: Quantile regression. J. Econ. Perspect. 15(4), 143–156 (2001)
Li, G., Peng, H., Zhu, L.: Nonconcave penalized \(M\)-estimation with a diverging number of parameters. Statist. Sinica 21(1), 391–419 (2011)
Mallows, C.L.: Some comments on \({C}_p\). Technometrics 15(4), 661–675 (1973)
Ronchetti, E.: Robust model selection in regression. Statist. Probab. Lett. 3(1), 21–23 (1985)
Ronchetti, E., Staudte, R.G.: A robust version of Mallows’ \(C_P\). J. Am. Statist. Assoc. 89(426), 550–559 (1994)
Schwarz, G.: Estimating the dimension of a model. Ann. Statist. 6(2), 461–464 (1978)
Tibshirani, R.: Regression shrinkage and selection via the lasso. J. Roy. Statist Soc. Ser. B 58(1), 267–288 (1996)
Wang, H., Li, G., Jiang, G.: Robust regression shrinkage and consistent variable selection through the LAD-Lasso. J. Bus. Econ. Stat. 25(3), 347–355 (2007)
Zhang, C.H.: Nearly unbiased variable selection under minimax concave penalty. Ann. Statist 38(2), 894–942 (2010)
Zou, H.: The adaptive lasso and its oracle properties. J. Am. Statist Assoc. 101(476), 1418–1429 (2006)
Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. Roy. Statist Soc. Ser. B 67(2), 301–320 (2005)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Mandal, A., Ghosh, S. (2023). Robust LASSO and Its Applications in Healthcare Data. In: Balakrishnan, N., Gil, M.Á., Martín, N., Morales, D., Pardo, M.d.C. (eds) Trends in Mathematical, Information and Data Sciences. Studies in Systems, Decision and Control, vol 445. Springer, Cham. https://doi.org/10.1007/978-3-031-04137-2_33
Download citation
DOI: https://doi.org/10.1007/978-3-031-04137-2_33
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-04136-5
Online ISBN: 978-3-031-04137-2
eBook Packages: EngineeringEngineering (R0)