Skip to main content

Accelerating the Convergence of EM-Based Training Algorithms for RBF Networks

  • Conference paper
  • First Online:
Connectionist Models of Neurons, Learning Processes, and Artificial Intelligence (IWANN 2001)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2084))

Included in the following conference series:

Abstract

In this paper, we propose a new Expectation-Maximization (EM) algorithm which speeds up the training of feedforward networks with local activation functions such as the Radial Basis Function (RBF) network. The core of the conventional EM algorithm for supervised learning of feedforward networks consists of decomposing the observations into their individual output units and then estimating the parameters of each unit separately. In previously proposed approaches, at eac h E-step the residual is decomposed equally among the units or proportionally to the weights of the output layer. However, this approach tends to slow down the training of networks with local activation units. To overcome this drawback in this paper we use a new E-step which applies a soft decomposition of the residual among the units. Inparticular, the residual is decomposed according to the probability of each RBF unit given each input-output pattern. It is shown that this variant not only speeds up the training in comparison with other EM-type algorithms, but also provides better results than a global gradient-descent technique since it has the capability of avoiding some unwanted minima of the cost function.

This work has been supported by the European Community and the Spanish Government through FEDER project 1FD97-1863-C02-01.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. S. Haykin, Neural Networks: A Comprehensive Foundation, Macmillan, New York, 1994.

    MATH  Google Scholar 

  2. C. Bishop, Neural Networks for Pattern Recognition, Clarendon Press, Oxford, 1997.

    Google Scholar 

  3. D. S. Broomhead, D. Lowe, “ Multivariable functional interpolation and adaptive networks”, Complex Systems, vol. 2, pp. 321–355, 1988.

    MATH  MathSciNet  Google Scholar 

  4. J. E. Moody, C. J. Darken, “Fast learning in networks of locally-tuned processing units”, Neural Computation, vol. 1, pp. 281–294, 1989.

    Article  Google Scholar 

  5. N. B. Karayanis, “Gradient descent learning of radial basis neural network”, Proc. 1997 IEEE Int. Joint Conf. Neural Networks, Houston, TX, pp. 1825–1820, 1997.

    Google Scholar 

  6. D. Lowe, “Adaptive radial basis function function nonlinearities and the problem of generalization, 1st IEE Conf. on Artificial Neural Networks, London, UK, pp. 171–175, 1989.

    Google Scholar 

  7. I. Santamaría, et al “A nonlinear MESFET model for intermodulation analysis using a GRBF Network”, Neurocomputing, vol. 25, pp. 1–18, 1999.

    Article  Google Scholar 

  8. A. P. Dempster, N. M. Laird, D. B. Rubin, “Maximum likelihood from incomplete data via the EM algorithm”, J. Roy. Statisc. Soc B, vol. 39, pp. 1–38, 1977.

    MATH  MathSciNet  Google Scholar 

  9. M. I. Jordan, R. A. Jacobs, “Hierarchical mixtures of experts and the EM algorithm”, Neural Computation, vol. 6, pp. 181–214, 1994.

    Article  Google Scholar 

  10. Z. Ghahramani, M. I. Jordan, “Supervised learning from incomplete data via an EM approach”, in Advances in NIPS VI, J. D. Cowan, G. Tesauro and J. Alspector, Eds., San Mateo, CA: Morgan Kaufmann, 1994.

    Google Scholar 

  11. S. Ma, C. Ji, J. Farmer “An efficient EM-based training algorithm for feedforward neural networks”, Neural Networks, vol. 10, pp. 243–256, 1997.

    Article  Google Scholar 

  12. S. Ma, C. Ji, “Fast training of recurrent networks based on the EM algorithm”, IEEE Trans. on Neural Networks, vol. 9, pp. 11–26, 1998.

    Article  Google Scholar 

  13. M. Feder, E. Weinstein, “Parameter estimation of superimposed signals using the EM algorithm”, IEEE Trans. Acoust., Speech, Signal Processing, vol. 36, pp. 477–489, 1988.

    Article  MATH  Google Scholar 

  14. S. Chen, C. F. N. Cowan, P. M. Grant, “Orthogonal least squares learning algorithm for radial basis functions networks”, IEEE Trans. on Neural Networks, vol. 2, pp.302–309, 1991.

    Article  Google Scholar 

  15. V. Cherkassky, D. Gehring, F. Mulier, “Comparison of adaptive methods for function estimation from samples”, IEEE Trans. on Neural Networks, vol. 7, no. 4, pp.969–984, 1996.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Lázaro, M., Santamaría, I., Pantaleón, C. (2001). Accelerating the Convergence of EM-Based Training Algorithms for RBF Networks. In: Mira, J., Prieto, A. (eds) Connectionist Models of Neurons, Learning Processes, and Artificial Intelligence. IWANN 2001. Lecture Notes in Computer Science, vol 2084. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45720-8_40

Download citation

  • DOI: https://doi.org/10.1007/3-540-45720-8_40

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-42235-8

  • Online ISBN: 978-3-540-45720-6

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics