Abstract
In this chapter we introduce a general model of (artificial) neural networks that captures (more or less) all special forms, which we consider in the following chapters. We start by defining the structure of an (artificial) neural network and then describe generally the operation and finally the training of an (artificial) neural network.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
A loop is an edge/connection from a vertex to this vertex itself, that is, an edge e=(v,v) with a vertex v∈V.
- 2.
A topological ordering is a numbering of the vertices of a directed graph, such that all edges are directed from a vertex with a lower number to a vertex with a higher number. A topological ordering exists only for acyclic graphs, that is, for feed forward networks. For feed forward networks, an update following a topological ordering ensures that all inputs of a neuron are already available (have already been computed), before it (re-)computes its own activation and output.
- 3.
The second formula is based on the maximum likelihood estimator for the variance of a normal distribution. In statistics often the unbiased estimator is preferred, which differs from the one used above only by using |L|−1 instead of |L|. For the normalization this difference is negligible.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag London
About this chapter
Cite this chapter
Kruse, R., Borgelt, C., Klawonn, F., Moewes, C., Steinbrecher, M., Held, P. (2013). General Neural Networks. In: Computational Intelligence. Texts in Computer Science. Springer, London. https://doi.org/10.1007/978-1-4471-5013-8_4
Download citation
DOI: https://doi.org/10.1007/978-1-4471-5013-8_4
Publisher Name: Springer, London
Print ISBN: 978-1-4471-5012-1
Online ISBN: 978-1-4471-5013-8
eBook Packages: Computer ScienceComputer Science (R0)