Abstract
Before we enter into the discussion of how one can derive a learning rule for multilayered perceptrons, it is useful to consider a simple example. We choose the exclusive-or (xor) function, with which we are already familiar. In order to circumvent the “no-go” theorem derived for simple perceptrons in the previous section, we add a hidden layer containing two neurons which receive signals from the input neurons and feed the output neuron (see Fig. 6.1). We denote the states of the hidden neurons by the variables s j , (j = 1,..., N h ).1 The synaptic connections between the hidden neurons and the output neurons are denoted by w ij ; those between the input layer and the hidden layer by \( {\bar w_{jk}} \). The threshold potentials of the output neurons are called ϑ i ; those of the hidden neurons are called \( {\bar \vartheta _j} \).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1990 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Müller, B., Reinhardt, J. (1990). Multilayered Perceptrons. In: Neural Networks. Physics of Neural Networks. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-97239-3_6
Download citation
DOI: https://doi.org/10.1007/978-3-642-97239-3_6
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-97241-6
Online ISBN: 978-3-642-97239-3
eBook Packages: Springer Book Archive