Skip to main content
Log in

A boosting approach for corporate failure prediction

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Predicting corporate failure is an important management science problem. This is a typical classification question where the objective is to determine which indicators are involved in the failure/success of a corporation. Despite the importance of this problem, until now only classical machine learning tools have been considered to tackle this classification task. The objective of this paper is twofold. On the one hand, we introduce novel discerning measures to rank independent variables in a generic classification task. On the other hand, we apply boosting techniques to improve the accuracy of a classification tree. We apply this methodology to a set of European firms, considering the usual predicting variables such as financial ratios, as well as including novel variables rarely used before in corporate failure prediction, such as firm size, activity and legal structure. We show that our approach decreases the generalization error about thirty percent with respect to the error produced with a classification tree. In addition, the most important ratios deal with profitability and indebtedness, as is usual in failure prediction studies.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Proceedings of the 13th international conference on machine learning. Morgan Kaufmann, pp 148–146.

  2. Freund Y, Schapire RE (1997) A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci 55(1):119–139

    Article  MATH  MathSciNet  Google Scholar 

  3. Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140

    MATH  MathSciNet  Google Scholar 

  4. Frydman H, Altman E, Kao D (1985) Introducing recursive partitioning for financial classification: the case of financial distress. J Finance 269–291

  5. Friedman J, Hastie T, Tibshirani R (2000) Additive logistic regression: a statistical view of boosting. Ann Stat 38(2):391–293

    MathSciNet  Google Scholar 

  6. Banfield RE, Hall LO, Bowyer KW, Bhadoria D, Kegelmeyer WP, Eschrich S (2004) A comparison of ensemble creation techniques. In: Roli F, Kittler J, Windeatt T (eds) Multiple classifier systems, vol. 3077 of Lecture notes in computer science. Springer, Cagliari, Italy, pp 223–232

  7. Dietterich TG (2000) Ensemble methods in machine learning. In: Kittler J, Roli F (eds) Multiple classifier systems, vol 1857 of Lecture notes in computer science. Springer, Cagliari, Italy, pp 1–15

  8. Kuncheva LI, Bezdek JC, Duin RPW (2001) Decision templates for multiple classifier fusion: and experimental comparison. Pattern Recog 34:299–314

    Article  MATH  Google Scholar 

  9. Lam L (2000) Classifier combinations: implementations and theoretical issues. In: Kittler J, Roli F (eds) Multiple classifier systems, vol 1857 of Lecture notes in computer science. Springer, Cagliari, Italy, pp 78–86

  10. Valentini G, Masulli F (2002) Ensembles of learning machines. In: Marinaro M, Tagliaferri R (eds) Neural nets WIRN Vietri-02, vol. 2486 of Lecture notes in computer science. Springer-Verlag, Heidelberg, Germany, pp 3–19

  11. Breiman L (1998) Arcing classifiers. Ann Statist 26(3):801–849

    Article  MATH  MathSciNet  Google Scholar 

  12. Bauer E, Kohavi R (1999) An empirical comparison of voting classification algorithm: bagging, boosting and variants. Mach Learn 36:105–142

    Article  Google Scholar 

  13. Drucker H, Cortes C (1996) Boosting decision trees. In: Touretzky DS, Mozer MC, Hasselmo ME (eds) Advances in neural information processing systems 8: Proc. of the NIPS’95, vol. 8. The MIT Press, pp 479–485

  14. Quinlan JR (1996) Bagging, boosting, and C4.5. In: Proceedings of the thirteenth national conference on artificial intelligence and the eighth innovative applications of artificial intelligence conference. AAAI Press/MIT Press, Menlo Park, pp 725–730

  15. Schapire RE (2002) The boosting approach to machine learning: an overview. In: Workshop on nonlinear estimation and classification. MSRI

  16. Beaver WH (1966) Financial ratios as predictors of failure. Empirical research in accounting. Selected studies. Supplement to vol. 4 of Journal of Accounting Research, pp 71–111

  17. Altman EI (1968) Financial ratios, discriminant analysis and the prediction of corporate bankruptcy. J Finance 23(4):589–609

    Article  Google Scholar 

  18. Olshon JA (1980) Financial ratios and the probabilistic prediction of bankruptcy. J Account Res 18(1):5–12

    Google Scholar 

  19. Zmijewski M (1984) Methodological issues related to the estimation of financial distress prediction models. J Account Res 22:59–86

    Article  Google Scholar 

  20. Wilson RL, Sharda R (1994) Bankruptcy prediction using neural network. Decis Support Syst 11:545–557

    Article  Google Scholar 

  21. Laitinen T, Kankaanpaa M (1999) Comparative analysis of failure prediction methods: the Finish case. Eur Account Rev 8(1):67–92

    Article  Google Scholar 

  22. R Development Core Team (2004) R: a language and environment for statistical computing. R foundation for statistical computing, Viena. http://www.R-project.org

  23. Breiman L, Friedman JH, Olshen R, Stone CJ (1984) Clasification and regresion trees. Belmont, Wadsworth International Group

    Google Scholar 

  24. Opitz D, Maclin R (1999) Popular ensemble methods: an empirical study. J Artif Intell Res 11:169–198

    MATH  Google Scholar 

  25. Freund Y, Schapire RE, Bartlett P, Lee WS (1998) Boosting the margin: a new explanation for the effectiveness of voting methods. Ann Stat 26(5):1651–1686

    Article  MATH  MathSciNet  Google Scholar 

  26. Kuncheva LI (2004) Combining pattern classifiers. Methods and algorithms. Wiley.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Esteban Alfaro Cortés.

Additional information

E. A. Cortés · M. G. Martínez · N. G. Rubio. The authors teach Statistics at the Faculty of Economic and Business Sciences in the University of Castilla-La Mancha. Esteban Alfaro completed his degree in Business in 1999 and got his Ph.D. in Economics in 2005, both in the University of Castilla-La Mancha. His thesis dealt with the application of ensemble classifiers to corporate failure prediction. Matías Gámez got his degree in Mathematics at the University of Granada in 1991 and finished a Master in Applied Statistics a year after. He completed his Ph.D. in Economics at the University of Castilla-La Mancha in 1998 on the application of geo-statistical techniques to the estimation of housing prices. Noelia García got her degree in Economics at the University of Madrid (UAM) in 1996 and completed her Ph.D. in Economics in 2004 on the construction of an intelligent and automated system for property valuation through the combination of neural nets and a geographic information system (GIS). Current research deals with spatial statistics and the combination of classifiers (decision trees and neural nets) for solving heated topics in the Economics.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Cortés, E.A., Martínez, M.G. & Rubio, N.G. A boosting approach for corporate failure prediction. Appl Intell 27, 29–37 (2007). https://doi.org/10.1007/s10489-006-0028-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-006-0028-9

Keywords

Navigation