Abstract
The level of generalization of neural networks is heavily dependent on the quality of the training data. That is, some of the training patterns can be redundant or irrelevant. It has been shown that with careful dynamic selection of training patterns, better generalization performance may be obtained. Nevertheless, generalization is carried out independently of the novel patterns to be approximated. In this paper, we present a learning method that automatically selects the most appropriate training patterns to the new sample to be predicted. The proposed method has been applied to Radial Basis Neural Networks, whose generalization capability is usually very poor. The learning strategy slows down the response of the network in the generalisation phase. However, this does not introduces a significance limitation in the application of the method because of the fast training of Radial Basis Neural Networks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Moody J.E. and Darken C.J.: Fast Learning in Networks of Locally-Tuned Processing Units. Neural Computation 1, (1989), 281–294.
Poggio T. and Girosi F.: Networks for approximation and learning. Proceedings of the IEEE, 78, (1990), 1481–1497.
Park J. and Sandberg I.W.: Approximation and Radial-Basis-Function Networks. Neural Computation, 5, (1993), 305–316.
Abu-Mostafa Y. S.: The Vapnik-Chervonenkis dimension: information versus complexity in learning. Neural Conputation 1, (1989), 312–317.
Cohn D., L. Atlas and R. Ladner: Improving Generalisation with Active Learning, Machine Learning, Vol 15, (1994), 201–221.
Vijayakumar S. and H. Ogawa: Improving Generalization Abolity through Active Learning. IEICE Transactions on Information and Sytems. Vol E82-D,2, (1999), 480–487.
Atkeson C. G., A. W. Moore and S. Schaal. Locally Weighted Learning. Artificial Intelligence Review 11, (1997), 11–73.
Wettschereck D., D.W. Aha and T. Mohri: A review and Empirical Evaluation of feature weighting methods for a class of lazy learning algorithms. Artificial Intelligence Review 11, (1997), 273–314.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Valls, J.M., Isasi, P., Galván, I.M. (2001). Deferring the Learning for Better Generalization in Radial Basis Neural Networks. In: Dorffner, G., Bischof, H., Hornik, K. (eds) Artificial Neural Networks — ICANN 2001. ICANN 2001. Lecture Notes in Computer Science, vol 2130. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44668-0_27
Download citation
DOI: https://doi.org/10.1007/3-540-44668-0_27
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42486-4
Online ISBN: 978-3-540-44668-2
eBook Packages: Springer Book Archive