Abstract
In this paper, we propose a new Expectation-Maximization (EM) algorithm which speeds up the training of feedforward networks with local activation functions such as the Radial Basis Function (RBF) network. The core of the conventional EM algorithm for supervised learning of feedforward networks consists of decomposing the observations into their individual output units and then estimating the parameters of each unit separately. In previously proposed approaches, at eac h E-step the residual is decomposed equally among the units or proportionally to the weights of the output layer. However, this approach tends to slow down the training of networks with local activation units. To overcome this drawback in this paper we use a new E-step which applies a soft decomposition of the residual among the units. Inparticular, the residual is decomposed according to the probability of each RBF unit given each input-output pattern. It is shown that this variant not only speeds up the training in comparison with other EM-type algorithms, but also provides better results than a global gradient-descent technique since it has the capability of avoiding some unwanted minima of the cost function.
This work has been supported by the European Community and the Spanish Government through FEDER project 1FD97-1863-C02-01.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
S. Haykin, Neural Networks: A Comprehensive Foundation, Macmillan, New York, 1994.
C. Bishop, Neural Networks for Pattern Recognition, Clarendon Press, Oxford, 1997.
D. S. Broomhead, D. Lowe, “ Multivariable functional interpolation and adaptive networks”, Complex Systems, vol. 2, pp. 321–355, 1988.
J. E. Moody, C. J. Darken, “Fast learning in networks of locally-tuned processing units”, Neural Computation, vol. 1, pp. 281–294, 1989.
N. B. Karayanis, “Gradient descent learning of radial basis neural network”, Proc. 1997 IEEE Int. Joint Conf. Neural Networks, Houston, TX, pp. 1825–1820, 1997.
D. Lowe, “Adaptive radial basis function function nonlinearities and the problem of generalization, 1st IEE Conf. on Artificial Neural Networks, London, UK, pp. 171–175, 1989.
I. Santamaría, et al “A nonlinear MESFET model for intermodulation analysis using a GRBF Network”, Neurocomputing, vol. 25, pp. 1–18, 1999.
A. P. Dempster, N. M. Laird, D. B. Rubin, “Maximum likelihood from incomplete data via the EM algorithm”, J. Roy. Statisc. Soc B, vol. 39, pp. 1–38, 1977.
M. I. Jordan, R. A. Jacobs, “Hierarchical mixtures of experts and the EM algorithm”, Neural Computation, vol. 6, pp. 181–214, 1994.
Z. Ghahramani, M. I. Jordan, “Supervised learning from incomplete data via an EM approach”, in Advances in NIPS VI, J. D. Cowan, G. Tesauro and J. Alspector, Eds., San Mateo, CA: Morgan Kaufmann, 1994.
S. Ma, C. Ji, J. Farmer “An efficient EM-based training algorithm for feedforward neural networks”, Neural Networks, vol. 10, pp. 243–256, 1997.
S. Ma, C. Ji, “Fast training of recurrent networks based on the EM algorithm”, IEEE Trans. on Neural Networks, vol. 9, pp. 11–26, 1998.
M. Feder, E. Weinstein, “Parameter estimation of superimposed signals using the EM algorithm”, IEEE Trans. Acoust., Speech, Signal Processing, vol. 36, pp. 477–489, 1988.
S. Chen, C. F. N. Cowan, P. M. Grant, “Orthogonal least squares learning algorithm for radial basis functions networks”, IEEE Trans. on Neural Networks, vol. 2, pp.302–309, 1991.
V. Cherkassky, D. Gehring, F. Mulier, “Comparison of adaptive methods for function estimation from samples”, IEEE Trans. on Neural Networks, vol. 7, no. 4, pp.969–984, 1996.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Lázaro, M., Santamaría, I., Pantaleón, C. (2001). Accelerating the Convergence of EM-Based Training Algorithms for RBF Networks. In: Mira, J., Prieto, A. (eds) Connectionist Models of Neurons, Learning Processes, and Artificial Intelligence. IWANN 2001. Lecture Notes in Computer Science, vol 2084. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45720-8_40
Download citation
DOI: https://doi.org/10.1007/3-540-45720-8_40
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42235-8
Online ISBN: 978-3-540-45720-6
eBook Packages: Springer Book Archive