Abstract
We outline the main models and developments in the broad field of artificial neural networks (GlossaryTerm
ANN
). A brief introduction to biological neurons motivates the initial formal neuron model – the perceptron. We then study how such formal neurons can be generalized and connected in network structures. Starting with the biologically motivated layered structure of GlossaryTermANN
(feed-forward GlossaryTermANN
), the networks are then generalized to include feedback loops (recurrent GlossaryTermANN
) and even more abstract generalized forms of feedback connections (recursive neuronal networks) enabling processing of structured data, such as sequences, trees, and graphs. We also introduce GlossaryTermANN
models capable of forming topographic lower-dimensional maps of data (self-organizing maps). For each GlossaryTermANN
type we outline the basic principles of training the corresponding GlossaryTermANN
models on an appropriate data collection.Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Abbreviations
- ANN:
-
artificial neural network
- BPTT:
-
back-propagation through time
- DAG:
-
directed acyclic graph
- ESN:
-
echo state network
- FPM:
-
fractal prediction machine
- LSM:
-
liquid state machine
- LSTM:
-
long short term memory
- RBF:
-
radial basis function
- RecNN:
-
recursive neural network
- RNN:
-
recurrent neural network
- RTRL:
-
real-time recurrent learning
- SD:
-
structured data
- SOM:
-
self-organizing map
- SRN:
-
simple recurrent network
- TDNN:
-
time delay neural network
References
F. Rosenblatt: The perceptron, a probabilistic model for information storage and organization in the brain, Psychol. Rev. 62, 386–408 (1958)
D.E. Rumelhart, G.E. Hinton, R.J. Williams: Learning internal representations by error propagation. In: Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Vol. 1 Foundations, ed. by D.E. Rumelhart, J.L. McClelland (MIT Press/Bradford Books, Cambridge 1986) pp. 318–363
J. Zurada: Introduction to Artificial Neural Systems (West Publ., St. Paul 1992)
K. Hornik, M. Stinchocombe, H. White: Multilayer feedforward networks are universal approximators, Neural Netw. 2, 359–366 (1989)
D.J.C. MacKay: Bayesian interpolation, Neural Comput. 4(3), 415–447 (1992)
S. Haykin: Neural Networks and Learning Machines (Prentice Hall, Upper Saddle River 2009)
C. Bishop: Neural Networks for Pattern Recognition (Oxford Univ. Press, Oxford 1995)
T. Sejnowski, C. Rosenberg: Parallel networks that learn to pronounce English text, Complex Syst. 1, 145–168 (1987)
A. Weibel: Modular construction of time-delay neural networks for speech recognition, Neural Comput. 1, 39–46 (1989)
J.L. Elman: Finding structure in time, Cogn. Sci. 14, 179–211 (1990)
M.I. Jordan: Serial order: A parallel distributed processing approach. In: Advances in Connectionist Theory, ed. by J.L. Elman, D.E. Rumelhart (Erlbaum, Hillsdale 1989)
Y. Bengio, R. Cardin, R. DeMori: Speaker independent speech recognition with neural networks and speech knowledge. In: Advances in Neural Information Processing Systems II, ed. by D.S. Touretzky (Morgan Kaufmann, San Mateo 1990) pp. 218–225
P.J. Werbos: Generalization of backpropagation with application to a recurrent gas market model, Neural Netw. 1(4), 339–356 (1988)
R.J. Williams, D. Zipser: A learning algorithm for continually running fully recurrent neural networks, Neural Comput. 1(2), 270–280 (1989)
Y. Bengio, P. Simard, P. Frasconi: Learning long-term dependencies with gradient descent is difficult, IEEE Trans. Neural Netw. 5(2), 157–166 (1994)
T. Lin, B.G. Horne, P. Tino, C.L. Giles: Learning long-temr dependencies with NARX recurrent neural networks, IEEE Trans. Neural Netw. 7(6), 1329–1338 (1996)
S. Hochreiter, J. Schmidhuber: Long short-term memory, Neural Comput. 9(8), 1735–1780 (1997)
M. Lukosevicius, H. Jaeger: Overview of Reservoir Recipes, Technical Report, Vol. 11 (School of Engineering and Science, Jacobs University, Bremen 2007)
A. Graves, M. Liwicki, S. Fernandez, R. Bertolami, H. Bunke, J. Schmidhuber: A novel connectionist system for improved unconstrained handwriting recognition, IEEE Trans. Pattern Anal. Mach. Intell. 31, 5 (2009)
S. Hochreiter, M. Heusel, K. Obermayer: Fast model-based protein homology detection without alignment, Bioinformatics 23(14), 1728–1736 (2007)
H. Jaeger, H. Hass: Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless telecommunication, Science 304, 78–80 (2004)
W. Maass, T. Natschlager, H. Markram: Real-time computing without stable states: A new framework for neural computation based on perturbations, Neural Comput. 14(11), 2531–2560 (2002)
P. Tino, G. Dorffner: Predicting the future of discrete sequences from fractal representations of the past, Mach. Learn. 45(2), 187–218 (2001)
M.H. Tong, A. Bicket, E. Christiansen, G. Cottrell: Learning grammatical structure with echo state network, Neural Netw. 20, 424–432 (2007)
K. Ishii, T. van der Zant, V. Becanovic, P. Ploger: Identification of motion with echo state network, Proc. OCEANS 2004 MTS/IEEE-TECHNO-OCEAN Conf., Vol. 3 (2004) pp. 1205–1210
L. Medsker, L.C. Jain: Recurrent Neural Networks: Design and Applications (CRC, Boca Raton 1999)
J. Kolen, S.C. Kremer: A Field Guide to Dynamical Recurrent Networks (IEEE, New York 2001)
D. Mandic, J. Chambers: Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability (Wiley, New York 2001)
J.B. MacQueen: Some models for classification and analysis if multivariate observations, Proc. 5th Berkeley Symp. Math. Stat. Probab. (Univ. California Press, Oakland 1967) pp. 281–297
M.D. Buhmann: Radial Basis Functions: Theory and Implementations (Cambridge Univ. Press, Cambridge 2003)
G.-B. Huang, Q.-Y. Zhu, C.-K. Siew: Extreme learning machine: theory and applications, Neurocomputing 70, 489–501 (2006)
T. Kohonen: Self-Organizing Maps, Springer Series in Information Sciences, Vol. 30 (Springer, Berlin, Heidelberg 2001)
T. Kohonen, E. Oja, O. Simula, A. Visa, J. Kangas: Engineering applications of the self-organizing map, Proc. IEEE 84(10), 1358–1384 (1996)
T. Koskela, M. Varsta, J. Heikkonen, K. Kaski: Recurrent SOM with local linear models in time series prediction, 6th Eur. Symp. Artif. Neural Netw. (1998) pp. 167–172
T. Voegtlin: Recursive self-organizing maps, Neural Netw. 15(8/9), 979–992 (2002)
M. Strickert, B. Hammer: Merge som for temporal data, Neurocomputing 64, 39–72 (2005)
M. Hagenbuchner, A. Sperduti, A. Tsoi: Self-organizing map for adaptive processing of structured data, IEEE Trans. Neural Netw. 14(3), 491–505 (2003)
A. Sperduti, A. Starita: Supervised neural networks for the classification of structures, IEEE Trans. Neural Netw. 8(3), 714–735 (1997)
P. Frasconi, M. Gori, A. Sperduti: A general framework for adaptive processing of data structures, IEEE Trans. Neural Netw. 9(5), 768–786 (1998)
B. Hammer, A. Micheli, A. Sperduti: Universal approximation capability of cascade correlation for structures, Neural Comput. 17(5), 1109–1159 (2005)
A. Micheli: Neural network for graphs: A contextual constructive approach, IEEE Trans. Neural Netw. 20(3), 498–511 (2009)
B. Hammer, A. Micheli, A. Sperduti, M. Strickert: A general framework for unsupervised processing of structured data, Neurocomputing 57, 3–35 (2004)
M. Hagenbuchner, A. Sperduti, A.-C. Tsoi: Graph self-organizing maps for cyclic and unbounded graphs, Neurocomputing 72(7–9), 1419–1430 (2009)
Y. Bengio, Y. LeCun: Greedy Layer-Wise Training of Deep Network. In: Advances in Neural Information Processing Systems 19, ed. by B. Schölkopf, J. Platt, T. Hofmann (MIT Press, Cambridge 2006) pp. 153–160
D.C. Ciresan, U. Meier, L.M. Gambardella, J. Schmidhuber: Deep big simple neural nets for handwritten digit recognition, Neural Comput. 22(12), 3207–3220 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Tino, P., Benuskova, L., Sperduti, A. (2015). Artificial Neural Network Models. In: Kacprzyk, J., Pedrycz, W. (eds) Springer Handbook of Computational Intelligence. Springer Handbooks. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-662-43505-2_27
Download citation
DOI: https://doi.org/10.1007/978-3-662-43505-2_27
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-662-43504-5
Online ISBN: 978-3-662-43505-2
eBook Packages: EngineeringEngineering (R0)