Abstract
This chapter generalises the Bayesian evidence scheme, introduced to the neural network community by David MacKay for the regularisation of networks under the assumption of Gaussian homoscedastic noise, to the prediction of arbitrary conditional probability densities. The idea is to optimise parameters and hyperparameters seperately, and to find the mode with respect to the hyperparameters only after the parameters have been integrated out. This integration is carried out by Gaussian approximation, which requires the calculation of the Hessian of the error function at the mode. The derivation of this matrix can be accomplished with a generalised version of the EM algorithm, as exposed in detail in the appendix.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London Limited
About this chapter
Cite this chapter
Husmeier, D. (1999). The Bayesian Evidence Scheme for Regularisation. In: Neural Networks for Conditional Probability Estimation. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0847-4_10
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0847-4_10
Publisher Name: Springer, London
Print ISBN: 978-1-85233-095-8
Online ISBN: 978-1-4471-0847-4
eBook Packages: Springer Book Archive