Skip to main content

An AdaBoost for Efficient Use of Confidences of Weak Hypotheses on Text Categorization

  • Conference paper
PRICAI 2014: Trends in Artificial Intelligence (PRICAI 2014)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 8862))

Included in the following conference series:

Abstract

We propose a boosting algorithm based on AdaBoost for using real-valued weak hypotheses that return confidences of their classifications as real numbers with an approximated upper bound of the training error. The approximated upper bound is induced with Bernoulli’s inequality and the upper bound enables us to analytically calculate a confidence-value that satisfies a reduction in the original upper bound. The experimental results on the Reuters-21578 data set and an Amazon review data show that our boosting algorithm with the perceptron attains better accuracy than Support Vector Machines, decision stumps-based boosting algorithms and a perceptron.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Freund, Y., Schapire, R.E.: A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences 55(1) (1997)

    Google Scholar 

  2. Schapire, R.E., Singer, Y.: Improved boosting algorithms using confidence-rated predictions. Machine Learning 37(3), 297–336 (1999)

    Article  MATH  Google Scholar 

  3. Cohen, W.W., Singer, Y.: A simple, fast, and effective rule learner. In: AAAI 1999/IAAI 1999, pp. 335–342 (1999)

    Google Scholar 

  4. Schapire, R.E., Singer, Y.: Boostexter: A boosting-based system for text categorization. Machine Learning 39(2/3), 135–168 (2000)

    Article  MATH  Google Scholar 

  5. Rätsch, G., Onoda, T., Müller, K.R.: Soft margins for adaboost. Machine Learning 42(3), 287–320 (2001)

    Article  MATH  Google Scholar 

  6. Pfahringer, B., Holmes, G., Kirkby, R.: Optimizing the induction of alternating decision trees. In: Cheung, D., Williams, G.J., Li, Q. (eds.) PAKDD 2001. LNCS (LNAI), vol. 2035, pp. 477–487. Springer, Heidelberg (2001)

    Chapter  Google Scholar 

  7. Jin, R., Liu, Y., Si, L., Carbonell, J., Hauptmann, A.G.: A new boosting algorithm using input-dependent regularizer. In: Proc. of ICML 2003 (2003)

    Google Scholar 

  8. Busa-Fekete, R., Kégl, B.: Fast boosting using adversarial bandits. In: Proc. of ICML 2010, pp. 143–150 (2010)

    Google Scholar 

  9. Eaton, E., desJardins, M.: Selective transfer between learning tasks using task-based boosting. In: AAAI (2011)

    Google Scholar 

  10. Quinlan, J.R.: Bagging, boosting, and c4.5. In: AAAI/IAAI, vol. 1, pp. 725–730 (1996)

    Google Scholar 

  11. Collins, M., Schapire, R.E., Singer, Y.: Logistic regression, adaboost and bregman distances. Machine Learning 48(1-3), 253–285 (2002)

    Article  MATH  Google Scholar 

  12. Nock, R., Nielsen, F.: A real generalization of discrete adaboost. Artif. Intell. 171(1), 25–41 (2007)

    Article  MATH  MathSciNet  Google Scholar 

  13. Rosenblatt, F.: The perceptron: A probabilistic model for information storage and organization in the brain 65(6), 386–408 (1958)

    Google Scholar 

  14. Collins, M.: Discriminative training methods for Hidden Markov Models: theory and experiments with perceptron algorithms. In: Proc. of EMNLP 2002, pp. 1–8 (2002)

    Google Scholar 

  15. Crammer, K., Dekel, O., Keshet, J., Shalev-Shwartz, S., Singer, Y.: Online passive-aggressive algorithms. Journal of Machine Learning Research 7, 551–585 (2006)

    MATH  MathSciNet  Google Scholar 

  16. Escudero, G., Màrquez, L., Rigau, G.: Boosting applied to word sense disambiguation. In: Lopez de Mantaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 129–141. Springer, Heidelberg (2000)

    Chapter  Google Scholar 

  17. Carreras, X., Màrques, L., Padró, L.: Named entity extraction using adaboost. In: Proc. of CoNLL 2002, pp. 167–170 (2002)

    Google Scholar 

  18. Carreras, X., Màrquez, L., Punyakanok, V., Roth, D.: Learning and inference for clause identification. In: Elomaa, T., Mannila, H., Toivonen, H. (eds.) ECML 2002. LNCS (LNAI), vol. 2430, pp. 35–47. Springer, Heidelberg (2002)

    Chapter  Google Scholar 

  19. Schapire, R.E.: Theoretical views of boosting and applications. In: Watanabe, O., Yokomori, T. (eds.) ALT 1999. LNCS (LNAI), vol. 1720, pp. 13–25. Springer, Heidelberg (1999)

    Chapter  Google Scholar 

  20. Kégl, B., Busa-Fekete, R.: Boosting products of base classifiers. In: Proc. of ICML 2009, pp. 497–504 (2009)

    Google Scholar 

  21. Vapnik, V.: Statistical Learning Theory. John Wiley & Sons (1998)

    Google Scholar 

  22. Blitzer, J., Dredze, M., Pereira, F.: Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In: ACL, pp. 440–447 (2007)

    Google Scholar 

  23. Friedman, J., Hastie, T., Tibshirani, R.: Additive logistic regression: A statistical view of boosting. The Annals of Statistics 28(2), 337–407 (2000)

    Article  MATH  MathSciNet  Google Scholar 

  24. Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann (1993)

    Google Scholar 

  25. Shalev-Shwartz, S., Singer, Y., Srebro, N., Cotter, A.: Pegasos: primal estimated sub-gradient solver for svm. Math. Program. 127(1), 3–30 (2011)

    Article  MATH  MathSciNet  Google Scholar 

  26. Friedman, J.: Greedy function approximation: A gradient boosting machine. Annals of Statistics 29, 1189–1232 (2000)

    Article  Google Scholar 

  27. Geurts, P., Wehenkel, L., d’Alché Buc, F.: Gradient boosting for kernelized output spaces. In: Proc. of ICML 2007, pp. 289–296 (2007)

    Google Scholar 

  28. Li, L.: Perceptron learning with random coordinate descent (2005)

    Google Scholar 

  29. Abrich, R., Bovbel, P.: CSC2515 fall 2011 boosting with perceptrons and decision stumps (2011)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Iwakura, T., Saitou, T., Okamoto, S. (2014). An AdaBoost for Efficient Use of Confidences of Weak Hypotheses on Text Categorization. In: Pham, DN., Park, SB. (eds) PRICAI 2014: Trends in Artificial Intelligence. PRICAI 2014. Lecture Notes in Computer Science(), vol 8862. Springer, Cham. https://doi.org/10.1007/978-3-319-13560-1_62

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-13560-1_62

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-13559-5

  • Online ISBN: 978-3-319-13560-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics