Abstract
Feature subset selection using a wrapper means to perform a search for an optimal set of attributes using the Machine Learning Algorithm as a black box. The Naive Bayes Classifier is based on the assumption of independence among the values of the attributes given the class value. Consequently, its effectiveness may decrease when the attributes are interdependent. We present FBL, a wrapper that uses information about dependencies to guide the search for the optimal subset of features and we use the Naive Bayes Classifier as the black-box Machine Learning algorithm. Experimental results show that FBL allows the Naive Bayes Classifier to achieve greater accuracies, and that FBL performs better than other classical filters and wrappers.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Almuallim, H., Dietterich, T.G.: Learning with Many Irrelevant Features. In: 9th National Conference on Artificial Intelligence, pp. 547–552. Mit Press, Cambridge (1991)
Blake, C.L., Merz, C.J.: UCI Repository of Machine Learning Databases. Irvine, University of California. Department of Information and Computer Science, http://www.ics.uci.edu/mlearn/MLRepository.html
Cortizo, J.C., Giraldez, J.I.: Discovering Data Dependencies in Web Content Mining. In: IADIS International Conference WWW/Internet, pp. 881–884 (2004)
Exposito, D., Giraldez, J.I.: Control MultiAgente del trfico rodado mediante Redes WIFI Conferencia Iberoamericana. IADIS WWW/Internet, pp. 473–476 (2004)
Fisher, R.: Statistical Methods for Research Workers. Macmillan Pub. Co., Basingstoke (1925)
Gomez, J.M., Buenaga, M., Cortizo, J.C.: The Role of Word Sense Disambiguation in Automated Text Categorization. In: Montoyo, A., Muńoz, R., Métais, E. (eds.) NLDB 2005. LNCS, vol. 3513, pp. 298–309. Springer, Heidelberg (2005)
Jakulin, A., Bratko, I.: Analyzing Attribute Dependencies. In: Proceedings of Knowledge Discovery in Data (PKDD), LNAI, pp. 229–240. Springer, Heidelberg (2003)
John, G.H., Kohavi, R., Pfleger, K.: Irrelevant Features and the Subset Selection Problem. In: Proceedings of the International Conference on Machine Learning, pp. 121–129 (1994)
Kira, K., Rendell, L.A.: A Practical Approach to Feature Subset Selection. In: 9th International Conference on Machine Learning. Morgan Kaufmann, San Francisco (1992)
Kononenko, I.: Comparison of inductive and naive Bayesian learning approaches to automatic knowledge adquisition. In: Wielinga, B. (ed.) Current trends in Knowledge Adquisition, pp. 190–197. IOS Press, Amsterdam (1990)
Kohavi, R., Sommerfield, D.: Feature Subset Selection Using the Wrapper Method: Overfitting and Dynamic Search Space Topology. In: 1st International Conference on Knowledge Discovery and Data Mining, pp. 192–197 (1995)
Kullback, S., Leibler, R.A.: On Information and Sufficiency. Annals of Mathematical Statistics 22, 79–86 (1951)
Langley, P.: Induction of recursive Bayesian Classifiers. In: Proceedings of the 1993 European Conference on Machine Learning, pp. 153–164 (1993)
Montes, C.: Metodo de Induccion Total. PhD Thesis. Universidad Politcnica de Madrid, Boadilla del Monte, Spain
Neter, J., Kutner, M.H., Wasserman, W., Nachtsheim, C.J.: Applied Linear Statistical Models. Irwin Editors (1996)
Pazzani, M.: Searching for dependencies in Bayesian Classifiers. Artificial Intelligence and Statistics IV. Springer, Heidelberg (1997)
Shlens, J.: A tutorial on Principal Component Analysis. Systems Neurobiology Laboratory, Salk Institute for Biological Studies (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Cortizo, J.C., Giraldez, I. (2006). Multi Criteria Wrapper Improvements to Naive Bayes Learning. In: Corchado, E., Yin, H., Botti, V., Fyfe, C. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2006. IDEAL 2006. Lecture Notes in Computer Science, vol 4224. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11875581_51
Download citation
DOI: https://doi.org/10.1007/11875581_51
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-45485-4
Online ISBN: 978-3-540-45487-8
eBook Packages: Computer ScienceComputer Science (R0)