Abstract
Various tradeoffs between time and space efficiency are possible in connectionistic models. One obvious effect is that while an ability to generalize effectively permits improved space efficiency, attempted generalization leads to increased learning time when generalization is not appropriate. For example, a maximal generalization over observed positive instances includes, by definition, as many unobserved (possibly negative) instances as possible. This is a significant drawback when specific instance learning is known to be appropriate. For example, in order to recognize a particular stimulus (e.g., a picture of a person) as having been seen before, the representation should be as specific as possible to avoid incorrectly responding to similar stimuli. A second problem with generalization is that, even when generalization is appropriate, an incremental learning system that stores only a single generalization hypothesis can make repeated mistakes on the same input patterns, a situation which need not occur with specific instance learning. Perceptron training is good at learning generalizations, but poor at learning specific instances.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1990 Birkhäuser Boston
About this chapter
Cite this chapter
Hampson, S.E. (1990). Learning and Using Specific Instances. In: Connectionistic Problem Solving. Birkhäuser Boston. https://doi.org/10.1007/978-1-4684-6770-3_4
Download citation
DOI: https://doi.org/10.1007/978-1-4684-6770-3_4
Publisher Name: Birkhäuser Boston
Print ISBN: 978-0-8176-3450-6
Online ISBN: 978-1-4684-6770-3
eBook Packages: Springer Book Archive