Abstract
The structure of a universal approximator network for predicting conditional probability densities is derived, and it is shown that the resulting architecture can deal with both stochastic and determinstic processes. Two variants, the derivative-of-sigmoid mixture (DSM) and the Gaussian mixture (GM) networks are presented, and their relation to a stochastic kernel expansion is noted. The chapter concludes with a comparison between these models and several relevant alternative approaches which have recently been introduced to the neural network community.
Keywords
- Hide Layer
- Gaussian Mixture Model
- Expectation Maximisation Algorithm
- Conditional Density
- Moment Generate Function
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this chapter
Cite this chapter
Husmeier, D. (1999). A Universal Approximator Network for Predicting Conditional Probability Densities. In: Neural Networks for Conditional Probability Estimation. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0847-4_2
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0847-4_2
Publisher Name: Springer, London
Print ISBN: 978-1-85233-095-8
Online ISBN: 978-1-4471-0847-4
eBook Packages: Springer Book Archive