Skip to main content

Multi Criteria Wrapper Improvements to Naive Bayes Learning

  • Conference paper
Intelligent Data Engineering and Automated Learning – IDEAL 2006 (IDEAL 2006)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 4224))

Abstract

Feature subset selection using a wrapper means to perform a search for an optimal set of attributes using the Machine Learning Algorithm as a black box. The Naive Bayes Classifier is based on the assumption of independence among the values of the attributes given the class value. Consequently, its effectiveness may decrease when the attributes are interdependent. We present FBL, a wrapper that uses information about dependencies to guide the search for the optimal subset of features and we use the Naive Bayes Classifier as the black-box Machine Learning algorithm. Experimental results show that FBL allows the Naive Bayes Classifier to achieve greater accuracies, and that FBL performs better than other classical filters and wrappers.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Almuallim, H., Dietterich, T.G.: Learning with Many Irrelevant Features. In: 9th National Conference on Artificial Intelligence, pp. 547–552. Mit Press, Cambridge (1991)

    Google Scholar 

  2. Blake, C.L., Merz, C.J.: UCI Repository of Machine Learning Databases. Irvine, University of California. Department of Information and Computer Science, http://www.ics.uci.edu/mlearn/MLRepository.html

  3. Cortizo, J.C., Giraldez, J.I.: Discovering Data Dependencies in Web Content Mining. In: IADIS International Conference WWW/Internet, pp. 881–884 (2004)

    Google Scholar 

  4. Exposito, D., Giraldez, J.I.: Control MultiAgente del trfico rodado mediante Redes WIFI Conferencia Iberoamericana. IADIS WWW/Internet, pp. 473–476 (2004)

    Google Scholar 

  5. Fisher, R.: Statistical Methods for Research Workers. Macmillan Pub. Co., Basingstoke (1925)

    Google Scholar 

  6. Gomez, J.M., Buenaga, M., Cortizo, J.C.: The Role of Word Sense Disambiguation in Automated Text Categorization. In: Montoyo, A., Muńoz, R., Métais, E. (eds.) NLDB 2005. LNCS, vol. 3513, pp. 298–309. Springer, Heidelberg (2005)

    Chapter  Google Scholar 

  7. Jakulin, A., Bratko, I.: Analyzing Attribute Dependencies. In: Proceedings of Knowledge Discovery in Data (PKDD), LNAI, pp. 229–240. Springer, Heidelberg (2003)

    Google Scholar 

  8. John, G.H., Kohavi, R., Pfleger, K.: Irrelevant Features and the Subset Selection Problem. In: Proceedings of the International Conference on Machine Learning, pp. 121–129 (1994)

    Google Scholar 

  9. Kira, K., Rendell, L.A.: A Practical Approach to Feature Subset Selection. In: 9th International Conference on Machine Learning. Morgan Kaufmann, San Francisco (1992)

    Google Scholar 

  10. Kononenko, I.: Comparison of inductive and naive Bayesian learning approaches to automatic knowledge adquisition. In: Wielinga, B. (ed.) Current trends in Knowledge Adquisition, pp. 190–197. IOS Press, Amsterdam (1990)

    Google Scholar 

  11. Kohavi, R., Sommerfield, D.: Feature Subset Selection Using the Wrapper Method: Overfitting and Dynamic Search Space Topology. In: 1st International Conference on Knowledge Discovery and Data Mining, pp. 192–197 (1995)

    Google Scholar 

  12. Kullback, S., Leibler, R.A.: On Information and Sufficiency. Annals of Mathematical Statistics 22, 79–86 (1951)

    Article  MATH  MathSciNet  Google Scholar 

  13. Langley, P.: Induction of recursive Bayesian Classifiers. In: Proceedings of the 1993 European Conference on Machine Learning, pp. 153–164 (1993)

    Google Scholar 

  14. Montes, C.: Metodo de Induccion Total. PhD Thesis. Universidad Politcnica de Madrid, Boadilla del Monte, Spain

    Google Scholar 

  15. Neter, J., Kutner, M.H., Wasserman, W., Nachtsheim, C.J.: Applied Linear Statistical Models. Irwin Editors (1996)

    Google Scholar 

  16. Pazzani, M.: Searching for dependencies in Bayesian Classifiers. Artificial Intelligence and Statistics IV. Springer, Heidelberg (1997)

    Google Scholar 

  17. Shlens, J.: A tutorial on Principal Component Analysis. Systems Neurobiology Laboratory, Salk Institute for Biological Studies (2005)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Cortizo, J.C., Giraldez, I. (2006). Multi Criteria Wrapper Improvements to Naive Bayes Learning. In: Corchado, E., Yin, H., Botti, V., Fyfe, C. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2006. IDEAL 2006. Lecture Notes in Computer Science, vol 4224. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11875581_51

Download citation

  • DOI: https://doi.org/10.1007/11875581_51

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-45485-4

  • Online ISBN: 978-3-540-45487-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics