Abstract
Recent interest in neural networks is largely triggered by the idea that knowledge from this field may be advantageously applied to problems arising in massively parallel computing. Besides increasing processing velocity, parallelism provides possibilities for improved fault tolerance and graceful degradation. In technical devices, this has been achieved by back-up hardware added in parallel to the main system. Biological systems take advantage of parallelism in many other respects including natural implementations, minimization of the number of computation steps, exploitation of signal redundancy, and a balanced distribution of processing tasks between all subsystems. As a result, reliability and accuracy of computation become exchangable. We present examples for these principles of biological information processing and discuss how parallelism is used for their implementation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
see “Collection of References from all Contributions”, pp. 513–540
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1989 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
von Seelen, W., Mallot, H.A. (1989). Parallelism and Redundancy in Neural Networks. In: Eckmiller, R., v.d. Malsburg, C. (eds) Neural Computers. Springer Study Edition, vol 41. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-83740-1_7
Download citation
DOI: https://doi.org/10.1007/978-3-642-83740-1_7
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-50892-2
Online ISBN: 978-3-642-83740-1
eBook Packages: Springer Book Archive