Skip to main content

Generalization Error Bounds for Aggregate Classifiers

  • Chapter
  • 2086 Accesses

Part of the book series: Lecture Notes in Statistics ((LNS,volume 171))

Abstract

The use of multiple classifiers has raised much interest in the statistical learning community in the past few years. The basic principle of multiple classifiers algorithms, also called aggregation or ensemble or voting methods, is to construct, according to some algorithm, several (generally a few dozens) different classifiers belonging to a certain family (e.g. support vector machines, classification trees, neural nets...). The “aggregate” classifier is then obtained by majority vote among the outputs of the single constructed classifiers once they are presented a new instance. For some algorithms the majority vote is replaced by a weighted vote, with weights prescribed by the aggregation algorithm. Classical references about this kind of methods include [2, 3, 9, 8, 12, 14, 19].

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Y. Amit and G. Blanchard. Multiple randomized classifiers: MRCL. Technical report, University of Chicago, 2000.

    Google Scholar 

  2. Y. Amit and D. Geman. Shape quantization and recognition with randomized trees. Neural Computation, 9:1545–1588, 1997.

    Article  Google Scholar 

  3. Y. Amit, D. Geman, and K. Wilder. Joint induction of shape features and tree classifiers. IEEE Trans. PAMI, 19(11):1300–1306, 1997.

    Article  Google Scholar 

  4. G. Blanchard. The “progressive mixture” estimator for regression trees. Annales de l’LH.R, 35(6):793–820, 1999.

    MathSciNet  MATH  Google Scholar 

  5. G. Blanchard. Mixture and aggregation of estimators for pattern recognition. Application to decision trees. PhD disseration, Université Paris-Nord, 2001. (In English, with an introductory part in French). Available at http://www.math.u-psud.fr/~blanchard/publi/these.ps.gz.

  6. L. Breiman. Bagging predictors. Machine Learning, 26:123–140, 1996.

    Google Scholar 

  7. L. Breiman. Prediction games and arcing algorithms. Technical report, Statistics department, University of California at Berkeley, December 1997.

    Google Scholar 

  8. L. Breiman. Arcing classifiers. The annals of statistics, 26(3):801–849, 1998.

    Article  MathSciNet  MATH  Google Scholar 

  9. L. Breiman. Random forests — random features. Technical report, University of California, Berkeley, 1999.

    Google Scholar 

  10. H. Chipman, E. I. George, and E. McCulloch. Bayesian CART model search. JASA, 93:935–947, September 1998.

    Article  Google Scholar 

  11. N. Christianini and J. Shawe-Taylor. An introduction to Support Vector Machines. Cambridge University Press, 2000.

    Google Scholar 

  12. T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output codes. J. Artificial Intell. Res., 2:263–286, 1995.

    MATH  Google Scholar 

  13. Y. Freund and R. E. Schapire. Game theory, on-line prediciton and Boosting. In Proceedings of the 9th annual conference on computational learning theory, 1996.

    Google Scholar 

  14. J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. The Annals of Statistics, 28:337–374, 2000.

    Article  MathSciNet  MATH  Google Scholar 

  15. V. Kolchinkskii and D. Panchenko. Empirical margin distributions and bounding the generalization error of combined classifiers. To appear in Ann. Statist.

    Google Scholar 

  16. J. Langford, M. Seeger, and N. Megiddo. An improved predictive accuracy bound for averaging classifiers. NIPS 2001.

    Google Scholar 

  17. G. Lugosi. Lectures on statistical learning theory. Presented at the Garchy Seminar on Mathematical Statistics and Applications, available at http://www.econ.upf.es/~lugosi, 2000.

  18. G. Ratsch, T. Onoda, and K.-R. Müller. Soft margins for AdaBoost. Machine Learning, 2000.

    Google Scholar 

  19. R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margins: a new explanation for the effectiveness of voting methods. Annals of Statistics, 26(5):1651–1686, 1998.

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Springer Science+Business Media New York

About this chapter

Cite this chapter

Blanchard, G. (2003). Generalization Error Bounds for Aggregate Classifiers. In: Denison, D.D., Hansen, M.H., Holmes, C.C., Mallick, B., Yu, B. (eds) Nonlinear Estimation and Classification. Lecture Notes in Statistics, vol 171. Springer, New York, NY. https://doi.org/10.1007/978-0-387-21579-2_23

Download citation

  • DOI: https://doi.org/10.1007/978-0-387-21579-2_23

  • Publisher Name: Springer, New York, NY

  • Print ISBN: 978-0-387-95471-4

  • Online ISBN: 978-0-387-21579-2

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics