Skip to main content

Stability of Random-Projection Based Classifiers. The Bayes Error Perspective

  • Conference paper
  • First Online:
Stochastic Models, Statistics and Their Applications (SMSA 2019)

Abstract

In this paper we investigate the Bayes error and the stability of Bayes’ error when the dimension of the classification problem is reduced using random projections. We restrict our attention to the two-class problem. Furthermore, we assume that distributions in classes come from multivariate normal distributions with the same covariance matrices, i.e., differing only in the means. This is one of the few situations when the Bayes error expression can be written in a simple form of a compact final formula. The bias and the variance of the classification error introduced by random projections are determined. Both full-dimensional normal distributions and singular distributions were considered with a real dimension smaller than the ambient dimension. These results allow for the separation of the impact of random dimension reduction from the impact of the learning sample and provide lower bounds on classification errors. Relatively low variance of the Bayes error introduced by random projections confirms the stability of the random-projection based classifiers, at least under the proposed assumptions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Abramowitz, M., Stegun, I.A. (eds.): Handbook of Mathematical Functions with Formulas, Graphs, and Mathematical Tables, 9th printing. Dover, New York (1972)

    Google Scholar 

  2. Askey, R.A., Olde Daalhuis, A.B.: Generalized hypergeometric functions and Meijer G-function. In: NIST Handbook of Mathematical Functions, U.S. Department of Commerce, Washington, DC, pp. 403–418. https://dlmf.nist.gov/16 (2010)

  3. Breiman, L.: Arcing clasifiers. Ann. Stat. 26(3), 801–849 (1998)

    Article  Google Scholar 

  4. Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001)

    Article  Google Scholar 

  5. Dasgupta, S.: Experiments with random projections. In: Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence, pp. 143–151 (2000)

    Google Scholar 

  6. Dasgupta, S., Gupta, A.: An elementary proof of a theorem of Johnson and Lindenstrauss. Random Struct. Algorithms 22(1), 60–65 (2003)

    Article  MathSciNet  Google Scholar 

  7. Devroye, L., Gyrfi, L., Lugosi, G.: Probabilistic Theory of Pattern Recognition. Springer, New York (1996)

    Book  Google Scholar 

  8. Duda, R., Hart, P.: Pattern Classification and Scene Analysis. Wiley, New York (1973)

    MATH  Google Scholar 

  9. Frankl, P., Maehara, H.: Some geometric applications of the beta distribution. Ann. Inst. Stat. Math. 42(3), 463–474 (1990)

    Article  MathSciNet  Google Scholar 

  10. Golub, G., Van Loan, C.F.: Matrix Computations. The Johns Hopkins University Press, Baltimore (1996)

    MATH  Google Scholar 

  11. Johnson, W.B., Lindenstrauss, J.: Extensions of Lipshitz mapping into Hilbert space. Contemp. Math. 26, 189–206 (1984)

    Article  Google Scholar 

  12. Kaski, S.: Dimensionality reduction by random mapping: fast similarity computation for clustering. In: Proceedings of the IEEE International Joint Conference on Neural Networks, vol. 1, pp. 413–418 (1998)

    Google Scholar 

  13. Lugosi, G., Pawlak, M.: On the posterior-probability estimate of the error rate of nonparametric classification rules. EEE Trans. Inf. Theory 40(2), 475–481 (1994)

    Article  MathSciNet  Google Scholar 

  14. Mathai, A.M., Saxena, R.K.: Generalized Hypergeometric Functions with Applications in Statistics and Physical Sciences. Springer, New York (1973)

    Book  Google Scholar 

  15. Meckes, E.: Approximation of projections of random vectors. J. Theor. Probab. 25(2), 333–352 (2012)

    Article  MathSciNet  Google Scholar 

  16. Rao, C.R.: Linear Statistical Inference and Its Applications, wyd II. Wiley, New York (1973)

    Book  Google Scholar 

  17. Skubalska-Rafajłowicz, E.: Relative stability of random projection-based image classification. In: 17th International Conference, ICAISC 2018 Zakopane, Poland, June 37, 2018. Proceedings, Part I, LNCS, vol. 10841, pp. 702–713 (2018)

    Google Scholar 

  18. Slater, Lucy Joan: Generalized Hypergeometric Functions. Cambridge University Press, Cambridge (1966)

    MATH  Google Scholar 

  19. Vempala, S.: The Random Projection Method. American Mathematical Society, Providence (2004)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ewa Skubalska-Rafajłowicz .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Skubalska-Rafajłowicz, E. (2019). Stability of Random-Projection Based Classifiers. The Bayes Error Perspective. In: Steland, A., Rafajłowicz, E., Okhrin, O. (eds) Stochastic Models, Statistics and Their Applications. SMSA 2019. Springer Proceedings in Mathematics & Statistics, vol 294. Springer, Cham. https://doi.org/10.1007/978-3-030-28665-1_9

Download citation

Publish with us

Policies and ethics