Skip to main content
Log in

Selective Cascade of Residual ExtraTrees

  • Original Research
  • Published:
SN Computer Science Aims and scope Submit manuscript

Abstract

We propose a novel tree-based ensemble method named Selective Cascade of Residual ExtraTrees (SCORE). SCORE draws inspiration from representation learning, incorporates regularized regression with variable selection features, and utilizes boosting to improve prediction and reduce generalization errors. We also develop a variable importance measure to increase the explainability of SCORE. Our computer experiments show that SCORE provides comparable or superior performance in prediction against ExtraTrees, random forest, gradient boosting machine, and neural networks; and the proposed variable importance measure for SCORE is comparable to studied benchmark methods. Finally, the predictive performance of SCORE remains stable across hyper-parameter values, suggesting potential robustness to hyper-parameter specification.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

References

  1. Dietterich TG. An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Mach Learn. 2000;40:139–57.

    Article  Google Scholar 

  2. Breiman L. Bagging predictors. Mach Learn. 1996;24(2):123–40.

    MATH  Google Scholar 

  3. Ho TK. The random subspace method for constructing decision forests. IEEE Trans Pattern Anal Mach Intell. 1998;20(8):832–44.

    Article  Google Scholar 

  4. Ho TK. Random decision forest. In: Proceedings of the 3rd international conference on document analysis and recognition; 1995. pp. 278–282.

  5. Geurts P, Ernst D, Wehenkel L. Extremely randomized trees. Mach Learn. 2006;63(1):3–42.

    Article  Google Scholar 

  6. Fan W, Wang H, Yu P, Ma S. Is random model better? On its accuracy and efficiency. In: Third IEEE international conference on data mining, no. October 2015. IEEE Comput Soc; 2003. pp. 51–58.

  7. Liu FT, Ting KM, Yu Y, Zhou Z-H. Spectrum of variable-random trees. J Artif Intell Res. 2008;32:355–84.

    Article  Google Scholar 

  8. Friedman JH. Greedy function approximation: a gradient boosting machine. Ann Stat. 2001;29(5):1189–232.

    Article  MathSciNet  Google Scholar 

  9. Mason L, Baxter J, Bartlett P, Frean M. Boosting algorithms as gradient descent. Adv Neural Inf Process Syst. 2000;2000:512–8.

    Google Scholar 

  10. Hinton GE, Osindero S, Teh Y-W. A fast learning algorithm for deep belief nets. Neural Comput. 2006;18(7):1527–54.

    Article  MathSciNet  Google Scholar 

  11. Kong Y, Yu T. A deep neural network model using random forest to extract feature representation for gene expression data classification. Sci Rep. 2018;8:1–9.

    Article  MathSciNet  Google Scholar 

  12. Zhou ZH, Feng J. Deep forest: towards an alternative to deep neural networks. In: IJCAI international joint conference on artificial intelligence; 2017. pp. 3553–3559.

  13. Feng J, Yu Y, Zhou ZH. Multi-layered gradient boosting decision trees. Adv Neural Inf Process Syst. 2018;2018:3551–61.

    Google Scholar 

  14. Bengio Y, Courville A, Vincent P. Representation learning: a review and new perspectives. IEEE Trans Pattern Anal Mach Intell. 2013;35(8):1798–828.

    Article  Google Scholar 

  15. Freund Y, Schapire RE. A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci. 1997;55(1):119–39.

    Article  MathSciNet  Google Scholar 

  16. Schapire RE. The strength of weak learnability. Mach Learn. 1990;5(2):197–227.

    Google Scholar 

  17. Li N, Zhou ZH. Selective ensemble under regularization framework. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics; 2009.

  18. Bengio Y, Courville A, Vincent P. Representation learning: a review and new perspectives. IEEE Trans Pattern Anal Mach Intell. 2013;35:1798–828.

    Article  Google Scholar 

  19. Brown G, Wyatt JL, Tino P. Managing diversity in regression ensembles. J Mach Learn Res. 2005;6:1621–2650.

    MathSciNet  MATH  Google Scholar 

  20. Tibshirani R. Regression shrinkage and selection via the Lasso. J R Stat Soc Ser B (Methodol). 1996;58(1):267–88.

    MathSciNet  MATH  Google Scholar 

  21. Helliwell J, Layard R, Sachs J. World Happiness Report 2019. In: Sustainable development solutions network, New York, Tech. Rep., 2019. https://worldhappiness.report/ed/2019/.

  22. Olden JD, Joy MK, Death RG. An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data. Ecol Model. 2004;178(3–4):389–97.

    Article  Google Scholar 

Download references

Acknowledgements

We thank Dr. Gitta Lubke and two anonymous referees for their useful and constructive comments on the project and the manuscript.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Qimin Liu.

Ethics declarations

Conflicts of interest

The authors declare that they have no conflicts of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

See Tables 5, 6.

Table 5 Input attributes in Boston Housing Data
Table 6 Input attributes in World Happiness Report Data

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, Q., Liu, F. Selective Cascade of Residual ExtraTrees. SN COMPUT. SCI. 1, 354 (2020). https://doi.org/10.1007/s42979-020-00358-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s42979-020-00358-x

Keywords

Navigation