Abstract
The learning of large neural networks is an ill-posed problem and there is generally a continuum of possible set of admissible weights. In this case, we cannot rely anymore on asymptotic properties of maximum likelihood estimators to approximate confidence intervals. Applying the Bayesian learning paradigm to neural networks or to generalized linear models results in a powerful framework that can be used for estimating the density of predictors. Within this approach, the uncertainty about parameters is expressed and measured by probabilities. This formulation allows for a probabilistic treatment of our a priori knowledge about parameters based on Markov Chain Monte Carlo methods. In order to explain those methods that are based on simulations, we need to review the main features of Markov chains.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
\(\mathcal {B}(\mathcal {X})\) is the sigma algebra defined on \(\mathcal {X}\).
- 2.
g.c.d. : greatest common divisor.
References
Buntine WL, Weigend AS (1991) Bayesian back-propagation. Complex Systems 5:603–643
de Freitas JFG (1999) Bayesian methods for neural networks. PhD Thesis, Department of Engineering, Cambridge University, Cambridge
Guyon I, Gunn S, Nikravesh M, Zadeh L (2006) Feature extraction, foundations and applications. Springer, New York
Hastie T, Tibshirani R, Friedman JH (2009) The elements of statistical learning: data mining, inference, and prediction. Springer series in statistics, 2nd edn. Springer, New York
Hastings W (1970) Monte Carlo sampling methods using Markov chains and their applications. Biometrika 57:97–109
Johnson N (2009) A study of the NIPS feature selection challenge. Standford working paper
Korb KB, Nicholson AE (2010) Bayesian artificial intelligence, 2nd edn. Chapman & Hall/CRC Computer Science & Data Analysis
Mackay DJC (1992) A practical Bayesian framework for backpropagation networks. Neural Comput 4(3):448–472
Metropolis N, Rosenbluth AW, Rosenbluth MN, Teller AH, Teller E (1953). Equations of state calculations by fast computing machines. J Chem Phys 21:1087–1092
Neal RM (1996) Bayesian learning for neural networks. Lecture notes in statistics no. 118. Springer, New York
Neal R, Zhang J (2006) High dimensional classification with Bayesian neural networks and Dirichlet diffusion trees. In: Feature extraction, foundations and applications. Springer, New-York, pp 265–296
Robert CP, Casella G (2004) Monte Carlo statistical methods. Springer, New York
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Denuit, M., Hainaut, D., Trufin, J. (2019). Bayesian Neural Networks and GLM. In: Effective Statistical Learning Methods for Actuaries III. Springer Actuarial(). Springer, Cham. https://doi.org/10.1007/978-3-030-25827-6_2
Download citation
DOI: https://doi.org/10.1007/978-3-030-25827-6_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-25826-9
Online ISBN: 978-3-030-25827-6
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)