Abstract
Centuries ago, the Boltzmann distribution, also called the Gibbs distribution, was proposed. This energy-based distribution was found to be useful for statistically modelling physical systems. One of these systems was the Ising model, which modelled interacting particles with binary spins. Later, it was discovered that the Ising model could be a neural network. Therefore, the Hopfield network was proposed, which implemented an Ising model in a network for modelling memory.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
A greedy algorithm makes every decision based on the most benefit at the current step and does not consider the final outcome at the final step. This greedy approach hopes that the final step will obtain a good result by small best steps based on their current benefits.
- 2.
Simulated annealing is a metaheuristic optimization algorithm in which a temperature parameter controls the amount of global search versus local search. This gradually reduces the temperature to decrease the exploration and increase the exploitation of the search space.
- 3.
In hashing, a hash function is used to map data of arbitrary size to fixed-size values.
References
David H Ackley, Geoffrey E Hinton, and Terrence J Sejnowski. “A learning algorithm for Boltzmann machines”. In: Cognitive science 9.1 (1985), pp. 147–169.
Diego Alberici, Pierluigi Contucci, and Emanuele Mingione. “Deep Boltzmann machines: rigorous results at arbitrary depth”. In: Annales Henri Poincaré. Springer. 2021, pp. 1–24.
Diego Alberici et al. “Annealing and replica-symmetry in deep Boltzmann machines”. In: Journal of Statistical Physics 180.1 (2020), pp. 665–677.
Yoshua Bengio et al. “Greedy layer-wise training of deep networks”. In: Advances in neural information processing systems. 2007, pp. 153–160.
Christopher M Bishop. “Pattern recognition”. In: Machine learning 128.9 (2006).
Ludwig Boltzmann. “Studien uber das Gleichgewicht der lebenden Kraft”. In: Wissenschafiliche Abhandlungen 1 (1868), pp. 49–96.
Bernhard E Boser, Isabelle M Guyon, and Vladimir N Vapnik. “A training algorithm for optimal margin classifiers”. In: Proceedings of the fifth annual workshop on Computational learning theory. 1992, pp. 144–152.
Stephen G Brush. “History of the Lenz-Ising model”. In: Reviews of modern physics 39.4 (1967), p. 883.
Sean Carroll. From eternity to here: the quest for the ultimate theory of time. Penguin, 2010.
Peter Dayan et al. “The Helmholtz machine”. In: Neural computation 7.5 (1995), pp. 889–904.
Raaz Dwivedi et al. “Log-concave sampling: Metropolis-Hastings algorithms are fast!” In: Conference on learning theory. PMLR. 2018, pp. 793–797.
Carol Bates Edwards. Multivariate and multiple Poisson distributions. Iowa State University, 1962.
Stuart Geman and Donald Geman. “Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images”. In: IEEE Transactions on pattern analysis and machine intelligence. PAMI-6.6 (1984), pp. 721–741.
Benyamin Ghojogh and Mark Crowley. “The theory behind overfitting, cross validation, regularization, bagging, and boosting: tutorial”. In: arXiv preprint arXiv:1905.12787 (2019).
Benyamin Ghojogh et al. “Sampling algorithms, from survey sampling to Monte Carlo methods: Tutorial and literature review”. In: arXiv preprint arXiv:2011.00901 (2020).
J Willard Gibbs. Elementary principles in statistical mechanics. Courier Corporation, 1902.
Xavier Glorot, Antoine Bordes, and Yoshua Bengio. “Deep sparse rectifier neural networks”. In: Proceedings of the fourteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings. 2011, pp. 315–323.
Ian Goodfellow et al. “Multi-prediction deep Boltzmann machines”. In: Advances in Neural Information Processing Systems 26 (2013), pp. 548–556.
Donald Hebb. The Organization of Behavior. Wiley & Sons, New York, 1949.
Geoffrey E Hinton. “A practical guide to training restricted Boltzmann machines”. In: Neural networks: Tricks of the trade. Springer, 2012, pp. 599–619.
Geoffrey E Hinton. “Boltzmann machine”. In: Scholarpedia 2.5 (2007), p. 1668.
Geoffrey E Hinton. “Deep belief networks”. In: Scholarpedia 4.5 (2009), p. 5947.
Geoffrey E Hinton. “Training products of experts by minimizing contrastive divergence”. In: Neural computation 14.8 (2002), pp. 1771–1800.
Geoffrey E Hinton, Simon Osindero, and Yee-Whye Teh. “A fast learning algorithm for deep belief nets”. In: Neural computation 18.7 (2006), pp. 1527–1554.
Geoffrey E Hinton and Ruslan R Salakhutdinov. “Reducing the dimensionality of data with neural networks”. In: Science 313.5786 (2006), pp. 504–507.
Geoffrey E Hinton and Russ R Salakhutdinov. “A better way to pretrain deep Boltzmann machines”. In: Advances in Neural Information Processing Systems 25 (2012), pp. 2447–2455.
Geoffrey E Hinton and Terrence J Sejnowski. “Optimal perceptual inference”. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition. Vol. 448. IEEE, 1983.
John J Hopfield. “Neural networks and physical systems with emergent collective computational abilities”. In: Proceedings of the national academy of sciences 79.8 (1982), pp. 2554–2558.
John J Hopfield. “Neurons with graded response have collective computational properties like those of two-state neurons”. In: Proceedings of the national academy of sciences 81.10 (1984), pp. 3088–3092.
Kerson Huang. Statistical Mechanics. John Wiley & Sons, 1987.
Ernst Ising. “Beitrag zur theorie des ferromagnetismus”. In: Zeitschrift für Physik 31.1 (1925), pp. 253–258.
Andrej Karpathy and Li Fei-Fei. “Deep visual-semantic alignments for generating image descriptions”. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2015, pp. 3128–3137.
Scott Kirkpatrick, C Daniel Gelatt, and Mario P Vecchi. “Optimization by simulated annealing”. In: science 220.4598 (1983), pp. 671–680.
Daphne Koller and Nir Friedman. Probabilistic graphical models: principles and techniques. MIT press, 2009.
Alex Krizhevsky and Geoff Hinton. “Convolutional deep belief networks on CIFAR-10”. In: Unpublished manuscript 40.7 (2010), pp. 1–9.
Dmitry Krotov. “Hierarchical Associative Memory”. In: arXiv preprint arXiv:2107.06446 (2021).
Dmitry Krotov and John Hopfield. “Large associative memory problem in neurobiology and machine learning”. In: International Conference on Learning Representations (ICLR). 2021.
Dmitry Krotov and John J Hopfield. “Dense associative memory for pattern recognition”. In: Advances in neural information processing systems 29 (2016), pp. 1172–1180.
Yann LeCun, Yoshua Bengio, and Geoffrey Hinton. “Deep learning”. In: nature 521.7553 (2015), pp. 436–444.
Yann LeCun et al. “A tutorial on energy-based learning”. In: Predicting structured data 1 (2006).
Wilhelm Lenz. “Beitršge zum verstšndnis der magnetischen eigenschaften in festen kšrpern”. In: Physikalische Z 21 (1920), pp. 613–615.
William A Little. “The existence of persistent states in the brain”. In: Mathematical biosciences 19.1–2 (1974), pp. 101–120.
Jan Melchior, Asja Fischer, and Laurenz Wiskott. “How to center deep Boltzmann machines”. In: The Journal of Machine Learning Research 17.1 (2016), pp. 3387–3447.
Abdel-rahman Mohamed, George Dahl, Geoffrey Hinton, et al. “Deep belief networks for phone recognition”. In: Nips workshop on deep learning for speech recognition and related applications. Vol. 1. 9. Vancouver, Canada. 2009, p. 39.
Abdel-rahman Mohamed, George E Dahl, and Geoffrey Hinton. “Acoustic modeling using deep belief networks”. In: IEEE transactions on audio, speech, and language processing 20.1 (2011), pp. 14–22.
Abdel-rahman Mohamed and Geoffrey Hinton. “Phone recognition using restricted Boltzmann machines”. In: 2010 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE. 2010, pp. 4354–4357.
Mehdi Molkaraie. “Marginal Densities, Factor Graph Duality and High-Temperature Series Expansions”. In: International Conference on Artificial Intelligence and Statistics. 2020, pp. 256–265.
Mehdi Molkaraie. “The primal versus the dual Ising model”. In: 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton). IEEE. 2017, pp. 53–60.
Grégoire Montavon and Klaus-Robert Müller. “Deep Boltzmann machines and the centering trick”. In: Neural networks: tricks of the trade. Springer, 2012, pp. 621–637.
Chi Nhan Duong et al. “Beyond principal components: Deep Boltzmann machines for face modeling”. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015, pp. 4786–4794.
Leandro Aparecido Passos and Joao Paulo Papa. “Temperature-based deep Boltzmann machines”. In: Neural Processing Letters 48.1 (2018), pp. 95–107.
Hubert Ramsauer et al. “Hopfield networks is all you need”. In: arXiv preprint arXiv:2008.02217 (2020).
David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. “Learning representations by back-propagating errors”. In: Nature 323.6088 (1986), pp. 533–536.
Ruslan Salakhutdinov. “Learning deep Boltzmann machines using adaptive MCMC”. In: Proceedings of the 27th International Conference on Machine Learning. 2010, pp. 943–950.
Ruslan Salakhutdinov and Geoffrey Hinton. “An efficient learning procedure for deep Boltzmann machines”. In: Neural computation 24.8 (2012), pp. 1967–2006.
Ruslan Salakhutdinov and Geoffrey Hinton. “Deep Boltzmann machines”. In: Artificial intelligence and statistics. PMLR. 2009, pp. 448–455.
Ruslan Salakhutdinov and Geoffrey Hinton. “Semantic hashing”. In: International Journal of Approximate Reasoning 50.7 (2009), pp. 969–978.
Ruslan Salakhutdinov and Hugo Larochelle. “Efficient learning of deep Boltzmann machines”. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings. 2010, pp. 693–700.
Nitish Srivastava and Ruslan Salakhutdinov. “Multimodal Learning with Deep Boltzmann Machines”. In: Advances in neural information processing systems. Vol. 1. 2012, p. 2.
Nitish Srivastava and Ruslan Salakhutdinov. “Multimodal learning with deep Boltzmann machines”. In: Journal of Machine Learning Research 15.1 (2014), pp. 2949–2980.
Nitish Srivastava, Ruslan R Salakhutdinov, and Geoffrey E Hinton. “Modeling documents with deep Boltzmann machines”. In: arXiv preprint arXiv:1309.6865 (2013).
Nitish Srivastava et al. “Dropout: a simple way to prevent neural networks from overfitting”. In: The journal of machine learning research 15.1 (2014), pp. 1929–1958.
Ilya Sutskever, Geoffrey E Hinton, and Graham W Taylor. “The recurrent temporal restricted Boltzmann machine”. In: Advances in neural information processing systems. 2009, pp. 1601–1608.
Graham W Taylor, Geoffrey E Hinton, and Sam T Roweis. “Modeling human motion using binary latent variables”. In: Advances in neural information processing systems. 2007, pp. 1345–1352.
Laurens Van Der Maaten. “Learning a parametric embedding by preserving local structure”. In: Artificial Intelligence and Statistics. 2009, pp. 384–391.
Max Welling, Michal Rosen-Zvi, and Geoffrey E Hinton. “Exponential Family Harmoniums with an Application to Information Retrieval.” In: Advances in neural information processing systems. Vol. 4. 2004, pp. 1481–1488.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Ghojogh, B., Crowley, M., Karray, F., Ghodsi, A. (2023). Restricted Boltzmann Machine and Deep Belief Network. In: Elements of Dimensionality Reduction and Manifold Learning. Springer, Cham. https://doi.org/10.1007/978-3-031-10602-6_18
Download citation
DOI: https://doi.org/10.1007/978-3-031-10602-6_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-10601-9
Online ISBN: 978-3-031-10602-6
eBook Packages: Computer ScienceComputer Science (R0)