Skip to main content
Log in

Strategies for improving neural net generalisation

  • Articles
  • Published:
Neural Computing & Applications Aims and scope Submit manuscript

Abstract

We address the problem of training multilayer perceptrons to instantiate a target function. In particular, we explore the accuracy of the trained network on a test set of previously unseen patterns — the generalisation ability of the trained network. We systematically evaluate alternative strategies designed to improve the generalisation performance. The basic idea is to generate a diverse set of networks, each of which is designed to be an implementation of the target function. We then have a set of trained, alternative versions — a version set. The goal is to achieve ‘useful diversity’ within this set, and thus generate potential for improved generalisation performance of the set as a wholewhen compared to the performance of any individual version. We define this notion of ‘useful diversity’, we define a metric for it, we explore a number of ways of generating it, and we present the results of an empirical study of a number of strategies for exploiting it to achieve maximum generalisation performance. The strategies encompass statistical measures as well as a ‘selectornet’ approach which proves to be particularly promising. The selector net is a form of ‘metanet’ that operates in conjunction with a version set.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Littlewood B, Miller DR. Conceptual modeling of coincident failures in multiversion software. IEEE Trans Software Eng 1989; 15(12)

  2. Partridge D. Network generalization differences quantified. Technical Report 291, Department of Computer Science, University of Exeter, 1994

  3. Partridge D, Sharkey NE. Use of neural computing in multiversion software reliability. In Redmill F, Anderson T. (eds). Technology and Assessment of Safety-Critical Systems. Springer-Verlag, London, 1994, pp 224–235

    Google Scholar 

  4. Partridge D, Sharkey NE. Neural computing for software reliability. Expert Systems, 1994

  5. Denker J, Schwartz D, Wittner B, Solla S, Howard R, Jackel L, Hopfield J. Large automatic learning rule extraction and generalisation. Complex Systems, 1987; 1

  6. Holden SB. Neural networks and the VC dimension. Technical Report CUED/F-INFENG/TR.119, Department of Engineering, University of Cambridge, 1992

  7. Holden SB, Rayner PJW. Generalization and PAC learning: some new results for the class of generalized single layer networks. IEEE Trans Neural Networks 1994 (in press)

  8. Partridge D, Collins T. Neural net training: random versus systematic. Technical Report 302, Department of Computer Science, University of Exeter, 1994. (To appear in: JG Taylor (ed). Adaptive Computing and Information Processing: Neural Networks)

  9. Sharkey NE, Sharkey AJC. Adaptive generalisation. Technical Report 256, Department of Computer Science, University of Exeter, 1993

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Derek Partridge.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Partridge, D., Griffith, N. Strategies for improving neural net generalisation. Neural Comput & Applic 3, 27–37 (1995). https://doi.org/10.1007/BF01414174

Download citation

  • Received:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF01414174

Keywords

Navigation