Abstract
This chapter reviews the Expectation Maximisation (EM) algorithm, and points out its principled advantage over a conventional gradient descent scheme. It is shown, however, that its application to the GM network is not directly feasible since the M-step of this algorithm is intractable with respect to one class of network parameters. A simple simulation demonstrates that as a consequence of this bottleneck effect, a gain in training speed can only be achieved at the price of a concomitant deterioration in the generalisation performance. It is therefore suggested to combine the EM algorithm with the RVFL concept discussed in the previous chapter. The parameter adaptation rules for the resulting GM-RVFL network are derived, and questions of numerical stability of the algorithm discussed. The superiority of this scheme over training a GM model with standard gradient descent will be demonstrated in the next chapter.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this chapter
Cite this chapter
Husmeier, D. (1999). Improved Training Scheme Combining the Expectation Maximisation (EM) Algorithm with the RVFL Approach. In: Neural Networks for Conditional Probability Estimation. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0847-4_7
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0847-4_7
Publisher Name: Springer, London
Print ISBN: 978-1-85233-095-8
Online ISBN: 978-1-4471-0847-4
eBook Packages: Springer Book Archive