Deep Ensemble Sparse Regression Network for Alzheimer’s Disease Diagnosis
For neuroimaging-based brain disease diagnosis, sparse regression models have proved their effectiveness in handling high-dimensional data but with a small number of samples. In this paper, we propose a novel framework that utilizes sparse regression models as target-level representation learner and builds a deep convolutional neural network for clinical decision making. Specifically, we first train multiple sparse regression models, each of which has different values of a regularization control parameter, and use the outputs of the trained regression models as target-level representations. Note that sparse regression models trained with different values of a regularization control parameter potentially select different sets of features from the original ones, thereby they have different powers to predict the response values, i.e., a clinical label and clinical scores in our work. We then construct a deep convolutional neural network by taking the target-level representations as input. Our deep network learns to optimally fuse the predicted response variables, i.e., target-level representations, from the same sparse response model(s) and also those from the neighboring sparse response models. To our best knowledge, this is the first work that systematically integrates sparse regression models with deep neural network. In our experiments with ADNI cohort, we validated the effectiveness of the proposed method by achieving the highest classification accuracies in three different tasks of Alzheimer’s disease and mild cognitive impairment identification.
This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2015R1C1A1A01052216) and also partially supported by Institute for Information & communications Technology Promotion (IITP) grant funded by the Korea government (MSIP) (No. B0101-15-0307, Basic Software Research in Human-level Lifelong Machine Learning (Machine Learning Center)).
- 1.Cotter, A., Shamir, O., Srebro, N., Sridharan, K.: Better mini-batch algorithms via accelerated gradient methods. In: Advances in Neural Information Processing Systems, vol. 24, pp. 1647–1655 (2011)Google Scholar
- 3.Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd International Conference on Machine Learning, pp. 448–456 (2015)Google Scholar
- 4.Kabani, N., MacDonald, D., Holmes, C., Evans, A.: A 3D atlas of the human brain. NeuroImage 7(4), S717 (1998)Google Scholar
- 5.Liu, J., Ji, S., Ye, J.: SLEP: Sparse Learning with Efficient Projections. Arizona State University (2009)Google Scholar
- 6.Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning, pp. 807–814 (2010)Google Scholar
- 12.Wang, H., Nie, F., Huang, H., Risacher, S., Saykin, A.J., Shen, L.: Identifying AD-sensitive and cognition-relevant imaging biomarkers via joint classification and regression. In: Fichtinger, G., Martel, A., Peters, T. (eds.) MICCAI 2011. LNCS, vol. 6893, pp. 115–123. Springer, Heidelberg (2011). doi: 10.1007/978-3-642-23626-6_15 CrossRefGoogle Scholar
- 15.Zhou, J., Liu, J., Narayan, V.A., Ye, J.: Modeling disease progression via fused sparse group lasso. In: Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1095–1103 (2012)Google Scholar