Abstract
Artificial neural network (ANN) is a very useful tool in solving learning problems. Boosting the performances of ANN can be mainly concluded from two aspects: optimizing the architecture of ANN and normalizing the raw data for ANN. In this paper, a novel method which improves the effects of ANN by preprocessing the raw data is proposed. It totally leverages the fact that different features should play different roles. The raw data set is firstly preprocessed by principle component analysis (PCA), and then its principle components are weighted by their corresponding eigenvalues. Several aspects of analysis are carried out to analyze its theory and the applicable occasions. Three classification problems are launched by an active learning algorithm to verify the proposed method. From the empirical results, conclusion comes to the fact that the proposed method can significantly improve the performance of ANN.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Jin, F., Sun, S.: Neural Network Multitask Learning for Traffic Flow Forecasting. In: Proceedings of the International Joint Conference on Neural Networks, pp. 1898–1902 (2008)
Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. John Wiley and Sons, New York (2001)
Breiman, L.: Bagging Predictors. Machine Learning 24(2), 123–140 (1996)
Freund, Y., Schapire, R.E.: A Short Introduction to Boosting. Journal of Japanese Society for Artificial Intelligence 14(5), 771–780 (1999)
Sun, S.: Ensemble Learning Methods for Classifying EEG Signals. In: Haindl, M., Kittler, J., Roli, F. (eds.) MCS 2007. LNCS, vol. 4472, pp. 113–120. Springer, Heidelberg (2007)
Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, Oxford (1995)
Anderson, H., Black, T.: Multivariate Data Analysis. Prentice-Hall, London (1998)
Song, Q., Kasabov, N.: WDN-RBF: Weighted Data Normalization for Radial Basic Function Type Neural Networks. In: Proceedings of IEEE International Joint Conference on Neural Networks, pp. 2095–2098 (2004)
Furlanello, C., Giuliani, D.: Combining Local PCA and Radial Basis Function Networks for Speaker Normalization. In: Proceedings of IEEE Workshop on Neural Networks for Signal Processing, pp. 233–242 (1995)
Howley, T., Madden, M.G., O’Connell, M., Ryder, A.G.: The Effect of Principle Domponent Analysis on Machine Learning Accuracy with High Dimensional Spectral Data. In: Proceedings the 25th SGAI International Conference on Innovative Techniques and Applications of Artificial Intelligence, pp. 209–222 (2005)
Jolliffe, I.T.: Principal Component Analysis, 2nd edn. Springer, New York (2002)
Tong, S.: Active Learning: Theory and Applications. PhD thesis, Stanford University, Stanford (2001)
Yang, L.: Distance Metric Learning: A Comprehensive Survey. Michigan State University, Michigan (2006)
Cristianini, N., Taylor, J.S.: An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University Press, Cambridge (2000)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, Q., Sun, S. (2009). Weighted Data Normalization Based on Eigenvalues for Artificial Neural Network Classification. In: Leung, C.S., Lee, M., Chan, J.H. (eds) Neural Information Processing. ICONIP 2009. Lecture Notes in Computer Science, vol 5863. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-10677-4_39
Download citation
DOI: https://doi.org/10.1007/978-3-642-10677-4_39
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-10676-7
Online ISBN: 978-3-642-10677-4
eBook Packages: Computer ScienceComputer Science (R0)