Abstract
Real world problems often contain noise features which can decrease effectiveness of classification models. This article proposes a filter-based technique to select a minimal set of features for classification problems. The proposed method employs fuzzification of original features based on irregular-shaped membership functions created by genetic algorithm and particle swarm optimization, and a feature selection process using two criterion functions to evaluate feature subsets. The first function is applied to eliminate features with redundant effects, and the second function is applied to select a feature subset that maximizes inter-class distances and minimize intra-class distances. Standard machine learning data sets in various sizes and complexities are used in experiments. The results show that the proposed technique is effective and performs well in comparisons with other research.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and Regression Trees. Wadsworht, Pacific Grove (1984)
Kennedy, J., Eberhart, B.: Particle swarm optimization. In: Proceedings of IEEE International Conference on Neural Network, Pert, Australia, pp. 1942–1948 (1995)
Engelbrecht, A.P.: Computational Intelligence: An Introduction, 2nd edn. Wiley, New York (2007)
Fleuret, F.: Fast binary feature selection with conditional mutual information. J. Mach. Learn. Res. 5(11), 1531–1555 (2004)
Gan, J.Q., Awwad Shiekh Hasan, B., Tsui, C.S.L.: A hybrid approach to feature subset selection for brain-computer interface design. In: Yin, H., Wang, W., Rayward-Smith, V. (eds.) IDEAL 2011. LNCS, vol. 6936, pp. 279–286. Springer, Heidelberg (2011). doi:10.1007/978-3-642-23878-9_34
Haindl, M., Somol, P., Ververidis, D., Kotropoulos, C.: Feature selection based on mutual correlation. In: Martínez-Trinidad, J.F., Carrasco Ochoa, J.A., Kittler, J. (eds.) CIARP 2006. LNCS, vol. 4225, pp. 569–577. Springer, Heidelberg (2006). doi:10.1007/11892755_59
Huang, H., Pasquier, M., Quek, C.: HiCEFS – A Hierarchical Coevolutionary Approach for the Dynamic Generation of Fuzzy System, pp. 3426–3443. IEEE Congress on Evolutionary Computation, CEC (2007)
Jabeen, H., Jalil, Z., Baig, A.: Opposition based initialization in particle swarm optimization (O-PSO). In: Proceedings of Genetic and Evolutionary Computation Conference, Montreal, Canada, pp. 2047–2052 (2009)
Jalali, L., Nasiri, M., Minaei, B.: A hybrid feature selection method based on fuzzy feature selection and consistency measures. In: Intelligent Computing and Intelligent System (ICIS), pp. 718–722 (2009)
Li, Y., Wu, Z.F.: Fuzzy feature selection based on min-max learning rule and extension matrix. Pattern Recogn. 41, 217–226 (2008)
Maroño, N.S., Betanzos, A.A., Castillo, E.: A new wrapper method for feature subset selection. In: Proceedings-European Symposium on Artificial Neural Networks, pp. 515–520 (2005)
Pudil, P., Novovičová, J., Kittler, J.: Floating search methods in feature selection. Pattern Recogn. Lett. 15, 1119–1125 (1994)
Somol, P., Novovičová, J., Pudil, P.: Flexible-hybrid sequential floating search in statistical feature selection. In: Yeung, D.-Y., Kwok, J.T., Fred, A., Roli, F., Ridder, D. (eds.) SSPR/SPR 2006. LNCS, vol. 4109, pp. 632–639. Springer, Heidelberg (2006). doi:10.1007/11815921_69
Vieira, S.M., Sousa, J.M.C., Kaymak, U.: Fuzzy criteria for feature selection. Fuzzy Sets Syst. 189, 1–18 (2012)
Yu, L., Liu, H.: Feature selection for high-dimensional data: a fast correlation-based filter solution. In: Proceedings of the Twentieth International Conference on Machine Learning (ICML-2003) (2003)
Zhang, W., Ma, D., Wei, J., Liang, H.: A parameter selection strategy for particle swarm optimization based on particle positions. Expert Syst. Appl. 41, 3576–3584 (2014)
Zhang, L.X., Wang, J.X., Zhao, Y.N., Yang, Z.H.: A novel hybrid feature selection algorithm: using relief estimation for GA-wrapper search. In: Proceedings of the Second International Conference on Machine Learning and Cybernetics, pp. 380–384 (2003)
Zhou, Y., Weng, F., Wu, L., Schmidt, H.: A fast algorithm for feature selection in conditional maximum entropy modeling. In: Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing, pp. 153–159 (2003)
Zhuo, L., Zheng, J., Wang, F., Li, X., Ai, B., Qian, J.: A genetic algorithm based wrapper feature selection method for classification of hyperspectral images using support vector machine. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. XXXVII Par B7, 397–402 (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Sornil, O. (2016). Filter-Based Feature Selection Using Two Criterion Functions and Evolutionary Fuzzification. In: Sombattheera, C., Stolzenburg, F., Lin, F., Nayak, A. (eds) Multi-disciplinary Trends in Artificial Intelligence. MIWAI 2016. Lecture Notes in Computer Science(), vol 10053. Springer, Cham. https://doi.org/10.1007/978-3-319-49397-8_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-49397-8_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-49396-1
Online ISBN: 978-3-319-49397-8
eBook Packages: Computer ScienceComputer Science (R0)