Skip to main content
Log in

Corporate bankruptcy prediction: a high dimensional analysis

  • Published:
Review of Accounting Studies Aims and scope Submit manuscript

Abstract

Much bankruptcy research has relied on parametric models, such as multiple discriminant analysis and logit, which can only handle a finite number of predictors (Altman in The Journal of Finance 23 (4), 589–609, 1968; Ohlson in Journal of Accounting Research 18 (1), 109–131, 1980). The gradient boosting model is a statistical learning method that overcomes this limitation. The model accommodates very large numbers of predictors which can be rank ordered, from best to worst, based on their overall predictive power (Friedman in The Annals of Statistics 29 (5), 1189–1232, 2001; Hastie et al. 2009). Using a sample of 1115 US bankruptcy filings and 91 predictor variables, the study finds that non-traditional variables, such as ownership structure/concentration and CEO compensation are among the strongest predictors overall. The next best predictors are unscaled market and accounting variables that proxy for size effects. This is followed by market-price measures and financial ratios. The weakest predictors overall included macro-economic variables, analyst recommendations/forecasts and industry variables.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Notes

  1. A weak classifier is one that predicts a little better than a random guess.

  2. Gradient boosting has conceptual similarities with forward stepwise regression. This method starts with an intercept term and sequentially adds variables based on statistical criteria such as model-fit improvement. However, stepwise has many limitations not shared by gradient boosting. For instance, parameter estimates, R-squares and t-values can be severely over-stated and lack interpretability (see Harrell, 2001).

  3. In contrast to conventional models, gradient boosting estimation is less affected by outliers, variable scaling/transformations, inclusion of irrelevant inputs, missing observations, database errors, and even data snooping—this type of researcher intervention can introduce significant bias to the modeling exercise (Freedman 2010).

  4. The averages discussed above ignore lower impacting variables with RVIs less than 10.

  5. Hillegeist et al. (2004) argue that asset volatility is important because it indicates the likelihood that the value of the firm’s assets will decline to such a degree that the firm will be unable to repay its debts. In addition to providing direct measures of volatility, Beaver et al. (2005) argue that market price variables are appealing because they reflect a rich and comprehensive mix of information, which includes financial statement data as a subset. Furthermore, market prices can also be measured with a finer partition of time. Whereas most bankruptcy studies use annual data, market prices can exploit the availability of daily prices (Frino et al. 2007).

  6. According to Beaver et al. (2005), this slight loss in predictive power appears to be due to increased discretion in financial reporting or the increase in intangible assets not being offset by improvements due to additional FASB standards.

  7. However, the international evidence is somewhat mixed on this issue (e.g., Frino et al. 2014).

  8. Insider ownership is also likely to play a similar role. If the directors and CEO are also stockholders, their interests are expected to be more closely aligned with the longer-term financial performance of the firm. Insiders with higher levels of stock ownership will have stronger incentives to limit the organization’s exposure to bankruptcy risks, which inevitably leads to significant destruction in stockholder value.

  9. Macroeconomic variables are extracted from a commercial source TradingEconomics.com. Real GDP and real GDP growth are key measures of overall economic health and prosperity. A strong and growing economy is expected to lead to lower default risk and lower overall probability of corporate bankruptcy. Higher interest rates are often associated with general tightness in the economy and increased likelihood of liquidity and debt-servicing pressures on firms. Inflation is a widely cited economic indicator, and the general perception is that high inflation is unhealthy for the economy, although its economic impacts can be ambiguous (Figlewski et al. 2012; Jones et al. 2015). There is also a common perception that a high ratio of public debt to GDP and high unemployment is a sign of broad economic weakness and vulnerability. Hence I expect these indicators to be positively associated with corporate failure. The NBER recession indicator, the Michigan sentiment index, and the leading index provide more direct measures of current economic sentiment and conditions. The Moody’s AAA and BBB seasoned bond yields are broad measures of credit risk. For instance, rising bond yields indicate more risk in the economy including default risk.

  10. This is based on their review of more than 150 empirical studies in the field.

  11. Boosting is a general method for improving the performance of learning algorithms (the idea of combining and weighting many weak classifiers into a powerful ‘voting’ committee of classifiers). While most applications of boosting use decision trees as the base classifier, some studies have examined how well the boosting technique works when neural networks or logit are used as the base classifier (examples of such studies are included in Appendix Table 12).

  12. This study offers a number of improvements over this literature which is reflected in the generally stronger empirical results. For instance, this study uses a much larger sample size; a much wider range of predictive inputs (and related interaction effects); and applies a consistent definition of corporate failure (Chapter 11 filings). I also use a prize winning commercial gradient boosting package known as TreeNet®, which is well known for its predictive power. I discuss these issues in more detail in Section 6.

  13. An RVI above zero indicates that the variable adds something to the predictive success of the overall model.

  14. The antecedents of modern boosting models come from the classification and regression trees (CART™) technique. However, despite the early popularity of CART™ (particularly in health diagnostics), the technique became associated with a number of limitations, most notably high variance. (It does not generalize well.) More sophisticated techniques, starting with bagging (Breiman 1996), began to develop, which significantly improved on the performance of CART™. Bagging can dramatically reduce the variance of unstable procedures (like trees), leading to improved prediction outcomes. While bagging is a major improvement on CART™, more sophisticated boosting methodologies, such as random forests, adaptive boosting (AdaBoost), and gradient boosting, began to develop. Random forests are essentially a refined form of bagging. The technique improves on bagging by “de-correlating” the trees, which maximizes the reduction in variance. (See Hastie et al. 2009 for details.) There is now an extensive literature devoted to the gradient boosting framework (including related approaches such as AdaBoost), which highlight the many advantages of this approach and, more particularly, the superior forecasting accuracy of the model. Recent applications include biological sciences (such as DNA research), text and speech recognition and processing, satellite imaging analysis (such as oil spill detection), cyber security, geological mapping, and credit risk. Specific applications of boosting to bankruptcy research are provided in Appendix Table 12.

  15. Both AdaBoost and gradient boosting are conceptually similar techniques. Both approaches boost the performance of a base classifier by iteratively focusing attention on observations that are difficult to predict. AdaBoost achieves this by increasing the weight on observations that were incorrectly classified in the previous round. With gradient boosting, difficult observations are identified by large residuals computed in the previous iterations. The idea behind gradient boosting is to build the new base classifiers to be maximally correlated with the negative gradient of the loss function, across the whole tree ensemble (Friedman 2001).

  16. The squared relative importance of variable X is the sum of such squared improvements over all internal nodes for which it was chosen as the splitting variable (Hastie et al. 2009).

  17. As shown by Hastie et al. (2009, p.368), it is straightforward to adapt these expressions to classification problems.

  18. At the end of the process, the entire family of dependence curves is averaged and centered to produce the final dependence plot on X 1 . (More details are provided in “Introduction to TreeNet” Salford Systems, San Diego, 2015.)

  19. If there are two perfectly correlated predictors in the dataset, the RVI would be 100% for the first one found by gradient boosting and zero for the other (i.e., the second variable is redundant). If two variables are strongly but not perfectly correlated, the gradient boosting algorithm simply partitions the correlated variables, looking for the greatest space between them that can enhance predictive power. A predictive variable will gain a higher RVI, while the other high correlated predictor gets a low RVI, because of redundancy in signal. (It may still have some signal because the correlation is not perfect.)

  20. This approach does involve look-ahead bias. The impact of look-ahead bias on empirical results is examined further in Section 6. Studies by Shumway (2001), Hillegeist et al., (2004) and Beaver et al., (2005) use hazard/duration models to predict corporate bankruptcy. For this type of modeling, it is clearly more appropriate to code a firm as bankrupt only in the year of bankruptcy as hazard models are designed to predict time-to-event using a survival function.

  21. A random guess describes a horizontal curve through the unit interval and has an AUC of exactly 0.5. As a minimum, classifiers are expected to perform >.5 (ie., better than random guessing), whereas an AUC score of 1 represents perfect classification accuracy (zero Type I and Type II errors).

  22. Note that the RVIs reported in Table 3 incorporate all important interaction effects.

  23. I re-estimated a gradient boosting model using all the variables in Table 3 but removing all financial variables with RVIs less than 10 (around 30+ variables were dropped as a result). However, I found that the RVIs of the remaining variables did not change significantly.

  24. Based on the same 70/30 random allocation to learn and test samples used for the gradient boosting model in Table 3.

  25. The parameter estimate is an “average” effect and has the same value across all observations.

  26. Other market based variables evidence some interesting nonlinear effects. Consistent with previous literature, my analysis of the marginal effects indicates that stock price volatility also exhibits a strong positive relationship with the failure outcome (high stock price volatility increases the probability of failure). Higher levels of short interest are also increase with the failure outcome, but again the impacts are nonlinear and vary over different levels of short interest.

  27. Several other accounting based variables exhibit similar nonlinear relationships with the bankruptcy outcome. For instance, the EBIT to total assets variable displays a similar pattern to cash flow returns.

  28. The commercial version of gradient boosting, TreeNet®, used for this study, allows for automatic detection of all interaction effects. According to Salford Systems, interaction effects are based on comparisons with a genuine bivariate plot (where variables are allowed to interact) and an additive combination of the two corresponding univariate plots. By determining the differences between the two response surfaces, the strength of interaction effect can be measured for a given pair of variables. The core idea behind interaction testing in TreeNet® is that variables can only interact within individual trees (see “Introduction to TreeNet,” Salford Systems, San Diego, 2015).

  29. Measure 1 also shows the effect of the interacting pair normalized to the overall response surface (total variation of predicted response).

  30. The two-way interaction effects suggest that higher levels of ownership concentration and institutional ownership reduce the impact of excess returns on the failure outcome.

  31. According to Salford Systems, TreeNet®‘s high predictive accuracy comes from the algorithm’s capacity to generate thousands of small decision trees built in a “sequential error–correcting process that converges to a highly accurate model.” Other reasons include the power of TreeNet®‘s interaction detection engine. TreeNet® establishes whether higher order interactions are required in a predictive model. The interaction detection feature not only helps improve model performance (sometimes dramatically) but also assists in the discovery of valuable new variables and patterns. This software has automatic features for detecting higher order interaction effects across any number of variables (not available in most packages or older methods). TreeNet® also has optimization features to help find the analyst find the best tree depth which maximizes predictive performance.

  32. The predictive power of interaction effects can readily be isolated by TreeNet® by estimating a model with and without interaction effects and comparing the AUCs accordingly.

  33. A basic issue in any machine learning model is the learn curve—this shows how well the model predicts as the training sample is increased. The smaller the sample size, the higher the expected bias in the model. Model variance is also expected to diminish as the sample size is monotonically increased. A model also needs to be tested out of sample, and the more sample available, the more confidence we can have in the generalization error of the model. It goes without saying that the efficient estimation of more complex models involving many feature vectors and higher order interaction effects requires larger sample sizes, particularly when comparing predicting performance across alternative models.

  34. Jones et al. (2015) reported that gradient boosting and AdaBoost strongly outperformed when decision trees are used as the base classifier. For instance, they examined boosting models where both logit and probit were used as the base classifier, but they did not perform as strongly.

  35. This is double TreeNet®‘s battery petition default setting of 10-repeat models.

  36. Bankruptcy researchers often want to predict the success of their models a number of years prior to failure, not just in the year of failure. For instance, testing the predictive power of a model at t = −3 (three years from failure) requires knowledge of which companies failed in t = 0 (year of failure). Hence look-ahead bias is unavoidable in this situation.

References

  • Alfaro, E., García, N., Gámez, M., & Elizondo, D. (2008). Bankruptcy forecasting: An empirical comparison of AdaBoost and neural networks. Decision Support Systems, 45, 110–122.

    Article  Google Scholar 

  • Altman, E. I. (1968). Financial ratios, discriminant analysis and the prediction of corporate bankruptcy. The Journal of Finance, 23(4), 589–609.

    Article  Google Scholar 

  • Altman, E. I. (2002). Bankruptcy, credit risk, and high yield junk bonds. Malden: Blackwell Publishers.

    Google Scholar 

  • Altman, E. I., Haldeman, R. G., & Narayanan, P. (1977). ZETA analysis: A new model to identify bankruptcy risk of corporations. Journal of Banking & Finance, 1(1), 29–54.

    Article  Google Scholar 

  • Altman, E. I., Marco, G., & Varetto, F. (1994). Corporate distress diagnosis: Comparisons using linear discriminant analysis and neural networks (the Italian experience). Journal of Banking & Finance, 18(3), 505–529.

    Article  Google Scholar 

  • Ashbaugh, H., Collins, D. W., & LaFond, R. (2006). The effects of corporate governance on firms’ credit ratings. Journal of Accounting and Economics, 52(2), 737–783.

    Google Scholar 

  • Asquith, P., Gertner, R., & Scharfstein, D. (1994). Anatomy of financial distress: An examination of junk-bond issuers. The Quarterly Journal of Economics, 109(3), 625–658.

    Article  Google Scholar 

  • Bangia, A., Diebold, F., Kronimus, F. X., Schagen, A., Schuermann, T., & T. (2002). Ratings migration and the business cycle, with application to credit portfolio stress testing. Journal of Banking & Finance, 26, 445–474.

    Article  Google Scholar 

  • Barth, M., Beaver, W., & Landsman, W. (1998). Relative valuation roles of equity book value and net income as a function of financial health. Journal of Accounting and Economics, 25, 1–34.

    Article  Google Scholar 

  • Beaver, W. H. (1966). Financial ratios as predictors of failure. Journal of Accounting Research, 4, 71–111.

    Article  Google Scholar 

  • Beaver, W. H., McNichols, M. F., & Rhie, J. W. (2005). Have financial statements become less informative? Evidence from the ability of financial ratios to predict bankruptcy. Review of Accounting Studies, 10, 93–122.

    Article  Google Scholar 

  • Bharath, S., & Shumway, T. (2008). Forecasting default with the Merton distance to default model. Review of Financial Studies, 21(3), 1339–1369.

    Article  Google Scholar 

  • Billings, B. (1999). Revisiting the relation between the default risk of debt and the earnings response coefficient. Accounting Review, 74(4), 509–522.

    Article  Google Scholar 

  • Black, N. S. (1998). In P. Newman (Ed.), “shareholder activism and corporate governance in the United States” in The New Palgrave Dictionary of Economics and the Law. New York: Stockton Press.

    Google Scholar 

  • Breiman, L. (1996). Bagging predictors. Machine Learning, 26, 123–140.

    Google Scholar 

  • Breiman, L., Friedman, J. H., Olshen, R. A., & Stone, C. J. (1984). Classification and regression trees. Wadsworth statistics/probability series: CRC Press.

    Google Scholar 

  • Chandra, D. K., Ravi, R., & Bose, L. (2009). Failure prediction of dotcom companies using hybrid intelligent techniques. Expert Systems with Applications, 36, 4830–4837.

    Article  Google Scholar 

  • Chava, S., & Jarrow, R. A. (2004). Bankruptcy prediction with industry effects. Review of Finance, 8, 537–569.

    Article  Google Scholar 

  • Clarke, J., Ferris, S. P., Jayaraman, N., & Lee, J. (2006). Are analyst recommendations biased? Evidence from corporate bankruptcies. Journal of Financial and Quantitative Analysis, 41, 169–196.

    Article  Google Scholar 

  • Cortés, E. A., Martínez, M. G., & Rubio, N. G. (2007). A boosting approach for corporate failure prediction. Applied Intelligence, 27(1), 29–37.

    Article  Google Scholar 

  • Daily, C. M., & Dalton, D. R. (1994). Bankruptcy and corporate governance: The impact of Board composition and structure. The Academy of Management Journal, 37(6), 1603–1617.

    Article  Google Scholar 

  • DeAngelo, H., & DeAngelo, L. (1990). Dividend policy and financial distress: An empirical investigation of troubled NYSE firms. The Journal of Finance, 45(5), 1415–1431.

    Article  Google Scholar 

  • Dichev, I. D. (1998). Is the risk of bankruptcy a systematic risk? The Journal of Finance, 53(3), 1131–1147.

    Article  Google Scholar 

  • Duffie, D., & Singleton, K. J. (2003). Credit risk, pricing, measurements, and management. Princeton: Princeton University Press.

    Google Scholar 

  • Ederington, L. H., & Goh, J. C. (1998). The bond agency ratings and stock analysts: Who knows what when? Journal of Financial and Quantitative Analysis, 33, 569–585.

    Article  Google Scholar 

  • Fedorova, E., Gilenko, E., & Dovzhenkon, S. (2013). Bankruptcy prediction for Russian companies: Application of combined classifiers. Expert Systems with Applications, 40, 7285–7293.

    Article  Google Scholar 

  • Figlewski, S., Frydman, H., & Liang, W. (2012). Modeling the effect of macroeconomic factors on corporate default and credit rating transitions. International Review of Economics and Finance, 21(1), 87–105.

    Article  Google Scholar 

  • Foster, G. (1986). Financial statement analysis (2nd ed.). Englewood Cliffs: Prentice Hall.

    Google Scholar 

  • Freedman, D.A. (2010). Statistical Methods and Causal Inference: A Dialogue with the Social Sciences, edited by collier, D., J.S., Sekhon, and P.B. Stark. Cambridge University Press, New York.

  • Friedman, J. (2001). Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29(5), 1189–1232.

    Article  Google Scholar 

  • Friedman, J., & Meulman, J. (2003). Multiple additive regression trees with application in epidemiology. Statistics in Medicine, 22(9), 1365–1381.

    Article  Google Scholar 

  • Frino, A., Jones, S., Wong, J., & J. (2007). Market behaviour around bankruptcy announcements: Evidence from the Australian stock Exchange. Accounting and Finance, 47(4), 713–730.

    Article  Google Scholar 

  • Frino, A., Jones, S., Lepone, A., & Wong, J. (2014). Market behavior of institutional investors around bankruptcy announcements. Journal of Business Finance and Accounting, 41, 270–295.

    Article  Google Scholar 

  • Harrell, F. (2001). Regression modeling strategies with applications to linear models, logistic and ordinal regression and survival analysis. New York: Springer.

  • Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements of statistical learning: Data mining, inference and prediction (2nd ed.). New York: Springer.

    Book  Google Scholar 

  • Hensher, D. A., Jones, S., & Greene, W. H. (2007). An error component logit analysis of corporate bankruptcy and insolvency. The Economic Record, 83(260), 86–103.

    Article  Google Scholar 

  • Heo, J., & Yang, J. W. (2014). AdaBoost based bankruptcy forecasting of Korean construction companies. Applied Soft Computing, 24, 494–499.

    Article  Google Scholar 

  • Hillegeist, S. A., Keating, E. K., Cram, D. P., & Lundstedt, K. G. (2004). Assessing the probability of bankruptcy. Review of Accounting Studies, 9(1), 5–34.

    Article  Google Scholar 

  • Hung, C., & Chen, J. H. (2009). A selective ensemble based on expected probabilities for bankruptcy prediction. Expert Systems with Applications, 36, 5297–5303.

    Article  Google Scholar 

  • Jones, S. (2011). Does the capitalization of intangible assets increase the predictability of corporate failure. Accounting Horizons, 1(25), 41–70.

    Article  Google Scholar 

  • Jones, S., & Hensher, D. A. (2004). Predicting firm financial distress: A mixed logit model. Accounting Review, 79(4), 1011–1038.

    Article  Google Scholar 

  • Jones, S., & Hensher, D. A. (2008). Advances in credit risk modelling and corporate bankruptcy prediction. Cambridge: Cambridge University Press.

    Book  Google Scholar 

  • Jones, S., & Johnstone, D. (2012). Analyst recommendations, earnings forecasts and corporate bankruptcy: Recent evidence. Journal of Behavioral Finance, 13(4), 281–298.

    Article  Google Scholar 

  • Jones, S., Johnstone, D., & Wilson, R. (2015). An empirical evaluation of the performance of binary classifiers in the prediction of credit ratings changes. Journal of Banking & Finance, 56(7), 72–85.

    Article  Google Scholar 

  • Jones, S., Johnstone, D., & Wilson, R. (2017). Predicting corporate bankruptcy: An evaluation of alternative statistical models. Journal of Business Finance and Accounting, 44(1–2), 3–34.

    Article  Google Scholar 

  • Karas, M., & Režňáková, R. (2014). A parametric or nonparametric approach for creating a new bankruptcy prediction model: The evidence from the Czech Republic. International Journal of Mathematical Models, 8, 214-223.

  • Kealhofer, S., and M. Kurbat. The default prediction power of the Merton approach, relative to debt ratings and accounting variables. Moody's KMV, 2001.

  • Keenan, S. C., Sobehart, J., & Hamilton, D. T. (1999). Predicting default rates: A forecasting model for moody's issuer-based default rates. Special Comment: Moody’s Investors Service.

    Google Scholar 

  • Kim, M. J., & Kang, D. K. (2010). Ensemble with neural networks for bankruptcy prediction. Expert Systems with Applications, 37, 3373–3379.

    Article  Google Scholar 

  • Kim, M. J., & Kang, D. K. (2015). Geometric mean based boosting algorithm with over-sampling to resolve data imbalance problem for bankruptcy prediction. Expert Systems with Applications, 42, 1074–1082.

    Article  Google Scholar 

  • Kim, S. Y., & Upneja, A. (2014). Predicting restaurant financial distress using decision tree and AdaBoosted decision tree models. Economic Modelling, 36, 354–362.

    Article  Google Scholar 

  • Koopman, S. J., Kräussl, R., Lucas, A., & Monteiro, A. B. (2009). Credit cycles and macro fundamentals. Journal of Empirical Finance, 16(1), 42–54.

    Article  Google Scholar 

  • Koopman, S. J., Lucas, A., & Schwaab, B. (2011). Modeling frailty-correlated defaults using many macroeconomic covariates. Journal of Econometrics, 162(2), 312–325.

    Article  Google Scholar 

  • Ohlson, J. A. (1980). Financial ratios and the probabilistic prediction of bankruptcy. Journal of Accounting Research, 18(1), 109–131.

    Article  Google Scholar 

  • Olson, D. L., Delen, D., & Mengm, Y. (2012). Comparative analysis of data mining methods for bankruptcy prediction. Decision Support Systems, 52, 464–473.

    Article  Google Scholar 

  • Precourt, E., and H.R. Oppenheimer (2013). “What do institutional investors know and act on before almost everyone Else: Evidence from corporate bankruptcies.” Seminar Paper at University of Rhode Island, Fall 2013.

  • Ramalingegowda, S. (2014). Evidence from impending bankrupt firms that long horizon institutional investors are informed about future firm value. Review of Accounting Studies, 19(2), 1009–1045.

    Article  Google Scholar 

  • Salford Systems. (2015). Introduction to TreeNet. San Diego.

    Google Scholar 

  • Schapire, R., & Freund, Y. (2012). Boosting: Foundations and algorithms. USA: MIT Press.

    Google Scholar 

  • Shleifer, A., & Vishny, R. W. (1997). A survey of corporate governance. The Journal of Finance, 52, 737–783.

    Article  Google Scholar 

  • Shumway, T. (2001). Forecasting bankruptcy more accurately: A simple hazard model. Journal of Business, 74(1), 101–124.

    Article  Google Scholar 

  • Sloan, R. G. (1996). Do stock prices fully reflect information in accruals and cash flows about future earnings? Accounting Review, 71(3), 289–315.

    Google Scholar 

  • Sun, J., Jia, M., & Hui, L. (2011). AdaBoost ensemble for financial distress prediction: An empirical comparison with data from Chinese listed companies. Expert Systems with Applications, 38(8), 9305–9312.

    Article  Google Scholar 

  • Swets, J. A., Dawes, R. M., & Monahan, J. (2000). Better decisions through science. Scientific American, 283, 82–87.

    Article  Google Scholar 

  • Tsai, F., Hsu, Y. F., & Yen, D. C. (2014). A comparative study of classifier ensembles for bankruptcy prediction. Applied Soft Computing, 24, 977–984.

    Article  Google Scholar 

  • Virág, M., & Nyitrai, T. (2014). The application of ensemble methods in forecasting bankruptcy. Financial & Economic Review, 13(4), 178–193.

    Google Scholar 

  • Wang, G., Ma, J., & Yang, S. (2014). An improved boosting based on feature selection for corporate bankruptcy prediction. Expert Systems with Applications, 41, 2353–2361.

    Article  Google Scholar 

  • West, D., Dellan, S., & Qia, J. (2005). Neural network ensemble strategies for financial decision applications. Computers & Operations Research, 32, 2543–2559.

    Article  Google Scholar 

  • White, H. (2000). A reality check for data snooping. Econometrica, 68(5), 1097–1126.

    Article  Google Scholar 

  • Zmijewski, M. E. (1984). Methodological issues related to the estimation of financial distress prediction models. Journal of Accounting Research, 22, 59–82.

    Article  Google Scholar 

Download references

Acknowledgements

Financial support for this project was provided by the Australian Research Council: ARC A7769.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Stewart Jones.

Additional information

Data availability

From the public sources identified in the paper.

Appendices

Appendix 1

Table 11 Variable Definitions

Appendix 2

Table 12 Applications of boosting models to bankruptcy prediction

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jones, S. Corporate bankruptcy prediction: a high dimensional analysis. Rev Account Stud 22, 1366–1422 (2017). https://doi.org/10.1007/s11142-017-9407-1

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11142-017-9407-1

Keywords

JEL classification

Navigation