Abstract
An error function E for a mixture model is derived from a maximum likelihood approach. The derivation of a gradient descent scheme is performed for both the DSM and the GM networks, and leads to a modified form of the backpropagation algorithm. However, a straightforward application of this method is shown to suffer from considerable inherent convergence problems due to large curvature variations of the error surface. A simple rectification scheme based on a curvature-based shape modification of E is presented.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this chapter
Cite this chapter
Husmeier, D. (1999). A Maximum Likelihood Training Scheme. In: Neural Networks for Conditional Probability Estimation. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0847-4_3
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0847-4_3
Publisher Name: Springer, London
Print ISBN: 978-1-85233-095-8
Online ISBN: 978-1-4471-0847-4
eBook Packages: Springer Book Archive