Abstract
Incremental Attribute Learning (IAL) is a novel machine learning strategy, where features are gradually trained in one or more according to some orderings. In IAL, feature ordering is a special preprocessing. Apart from time-consuming contribution-based feature ordering methods, feature ordering also can be derived by filter criteria. In this paper, a novel criterion based on Discriminability, a distribution-based metric, and Entropy is presented to give ranks for feature ordering, which has been validated in both two-category and multivariable classification problems by neural networks. Final experimental results show that the new metric is not only applicable for IAL, but also able to obtain better performance in lower error rates.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Liu, H.: Evolving feature selection. IEEE Intelligent Systems 20(6), 64–76 (2005)
Weiss, S.H., Indurkhya, N.: Predictive data mining: a practical guide. Morgan Kaufmann Publishers, San Francisco (1998)
Chao, S., Wong, F.: An incremental decision tree learning methodology regarding attributes in medical data mining. In: Proc. of the 8th Int’l Conf. on Machine Learning and Cybernetics, Baoding, pp. 1694–1699 (2009)
Agrawal, R.K., Bala, R.: Incremental Bayesian classification for multivariate normal distribution data. Pattern Recognition Letters 29(13), 1873–1876 (2008)
Guan, S.U., Liu, J.: Feature selection for modular networks based on incremental training. Journal of Intelligent Systems 14(4), 353–383 (2005)
Zhu, F., Guan, S.U.: Ordered incremental training for GA-based classifiers. Pattern Recognition Letters 26(14), 2135–2151 (2005)
Guan, S.U., Li, S.: Incremental learning with respect to new incoming input attributes. Neural Processing Letters 14(3), 241–260 (2001)
Guan, S.U., Li, S.: Parallel growing and training of neural networks using output parallelism. IEEE Trans. on Neural Networks 13(3), 542–550 (2002)
Bermejo, P., de la Ossa, L., Gámez, J.A., Puerta, J.M.: Fast wrapper feature subset selection in high-dimensional datasets by means of filter re-ranking. Knowledge-Based Systems 25(1), 35–44 (2012)
Peng, H., Long, F., Ding, C.: Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(8), 1226–1238 (2005)
Wang, T., Guan, S.U., Liu, F.: Ordered Incremental Attribute Learning based on mRMR and Neural Networks. International Journal of Design, Analysis and Tools for Integrated Circuits and Systems 2(2), 86–90 (2011)
Wang, T., Guan, S.-U., Liu, F.: Feature Discriminability for Pattern Classification Based on Neural Incremental Attribute Learning. In: Wang, Y.L., Li, T.R. (eds.) ISKE 2011. AISC, vol. 122, pp. 275–280. Springer, Heidelberg (2011)
Guan, S.U., Liu, J.: Incremental neural network training with an increasing input dimension. Journal of Intelligent Systems 13(1), 43–69 (2004)
Ripley, B.D.: Pattern Recognition and Neural Networks. Cambridge University Press, Cambridge (1996)
Guan, S.U., Liu, J.: Incremental Ordered Neural Network Training. Journal of Intelligent Systems 12(3), 137–172 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Wang, T., Guan, SU., Liu, F. (2012). Entropic Feature Discrimination Ability for Pattern Classification Based on Neural IAL. In: Wang, J., Yen, G.G., Polycarpou, M.M. (eds) Advances in Neural Networks – ISNN 2012. ISNN 2012. Lecture Notes in Computer Science, vol 7368. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-31362-2_4
Download citation
DOI: https://doi.org/10.1007/978-3-642-31362-2_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-31361-5
Online ISBN: 978-3-642-31362-2
eBook Packages: Computer ScienceComputer Science (R0)