Abstract
The goal of this work is to present an efficient implementation of the Backpropagation (BP) algorithm to train Artificial Neural Networks with general feedforward topology. This will lead us to the “consecutive retrieval problem” that studies how to arrange efficiently sets into a sequence so that every set appears contiguously in the sequence. The BP implementation is analyzed, comparing efficiency results with another similar tool. Together with the BP implementation, the data description and manipulation features of our toolkit facilitates the development of experiments in numerous fields.
This work has been partially supported by the Spanish Government under contract TIN2006-12767 and by the Generalitat Valenciana under contract GVA06/302.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bishop, C.M.: Neural networks for pattern recognition. Oxford University Press, Oxford (1996)
Bridle, J.: Probabilistic interpretation of feedforward classification network outputs, with relationships to statistical pattern recognition. In: Neuro-computing: Algorithms, Architectures, and Applications, pp. 227–236. Springer, Heidelberg (1989)
Garey, M.R., Johnson, D.S.: Computers and Intractability. A Guide to the Theory of NP-Completeness. W.H. Freeman and Company, New York (1979)
Ghosh, S.P.: File organization: the consecutive retrieval property. Commun. ACM 15(9), 802–808 (1972)
Hassibi, B., Stork, D.G.: Second order derivatives for network pruning: Optimal brain surgeon. In: Advances in NIPS, vol. 5, pp. 164–171. Morgan Kaufmann, San Francisco (1993)
Ierusalimschy, R.: Programming in Lua. Published by Lua.org (December 2003)
Kou, L.T.: Polynomial complete consecutive information retrieval problems. SIAM J. Comput. 6, 67–75 (1977)
LeCun, Y., Denker, J., Solla, S., Howard, R.E., Jackel, L.D.: Optimal brain damage. In: Advances in NIPS II, Morgan Kaufmann, San Francisco (1990)
Nethercote, N., Fitzhardinge, J.: Valgrind: A Program Supervision Framework. Electronic Notes in Theoretical Computer Science 89(2) (2003)
Plaut, D., Nowlan, S., Hinton, G.: Experiment on learning by back propagation. Technical Report CMU-CS-86-126, Department of Computer Science, Carnegie Mellon University (1986)
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning internal representations by error propagation. In: PDP: Computational models of cognition and perception, I, pp. 319–362. MIT Press, Cambridge (1986)
Werbos, P.: Backpropagation: Past and future. In: Proceedings of the IEEE International Conference on Neural Networks, pp. 343–353. IEEE Computer Society Press, Los Alamitos (1988)
Zamora Martínez, F.: Implementación eficiente del algoritmo de retropropagación del error con momentum para redes hacia delante generales. Proyecto Final de Carrera (2005)
Zell, A., Mache, N., Huebner, R., Schmalzl, M., Sommer, T., Korb, T.: SNNS: Stuttgart Neural Network Simulator. User manual Version 4.1. Technical report, Stuttgart (1995)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer Berlin Heidelberg
About this paper
Cite this paper
España-Boquera, S., Zamora-Martínez, F., Castro-Bleda, M.J., Gorbe-Moya, J. (2007). Efficient BP Algorithms for General Feedforward Neural Networks. In: Mira, J., Álvarez, J.R. (eds) Bio-inspired Modeling of Cognitive Tasks. IWINAC 2007. Lecture Notes in Computer Science, vol 4527. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-73053-8_33
Download citation
DOI: https://doi.org/10.1007/978-3-540-73053-8_33
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-73052-1
Online ISBN: 978-3-540-73053-8
eBook Packages: Computer ScienceComputer Science (R0)