Skip to main content
Log in

Optimizing neural network architectures using generalization error estimators

  • Published:
Radiophysics and Quantum Electronics Aims and scope

Abstract

This paper addresses the optimization of neural network architectures. It is suggested to optimize the architecture by selecting the model with minimal estimated averaged generalization error. We consider a least-squares (LS) criterion for estimating neural network models, i.e., the associated model weights are estimated by minimizing the LS criterion. The quality of a particular estimated model is measured by the average generalization error. This is defined as the expected squared prediction error on a novel input-output sample averaged over all possible training sets. An essential part of the suggested architecture optimization scheme is to calculate an estimate of the average generalization error. We suggest using the GEN-estimator [9, 10] which allows for dealing with nonlinear, incomplete models, i.e., models which are not capable of modeling the underlying nonlinear relationship perfectly. In most neural network applications, it is impossible to suggest a perfect model, and consequently the ability to handle incomplete models is urgent. A concise derivation of the GEN-estimator is provided, and its qualities are demonstrated by comparative numerical studies.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. H. Akaike, “Fitting autoregressive models for prediction,”Ann. Inst. Statistical Math.,21, 243 (1969).

    Google Scholar 

  2. N. R. Draper and H. Smith,Applied Regression Analysis, John Wiley and Sons, New York (1981).

    Google Scholar 

  3. D. B. Fogel,IEEE Trans. Neural Networks,2, No. 5, 490 (1991)

    Google Scholar 

  4. L. K. Hausen,Neural Networks,6, 393 (1993).

    Google Scholar 

  5. L. K. Hansen and P. Salamon,IEEE Trans. Pattern Analysis Machine Intelligence,12, No. 10, 993 (1990).

    Google Scholar 

  6. J. Hertz, A. Krogh, and R. G. Palmer,Introduction to the Theory of Neural Computation, Addison-Wesley, Redwood City, California (1991).

    Google Scholar 

  7. K. Hornik, M. Stinchcombe, and H. White,Neural Networks,3, No. 5, 551 (1990).

    Google Scholar 

  8. R. Kannurpatti and G. W. Hart,IEEE Trans. Information Theory,37, No. 5 1441 (1991)

    Google Scholar 

  9. J. Larsen, in:Neural Networks for Signals, S. Y. Kung, F. Fallside, J. A. Sorensen, and C. A. Kamm (eds.), IEEE, Piscataway, New Jersey (1992), p. 29.

    Google Scholar 

  10. J. Larsen, “Design of neural network filters,”Ph.D. Thesis, The Technical University of Denmark, Electronics Institute, March, 1993.

  11. J. Moody, in:Proceedings of the First IEEE Workshop on Neural Networks for Signal Processing, B. H. Juang, S. Y. Kung, and C. A. Kamm (eds.), IEEE, Piscataway, New Jersey (1991), p. 1.

    Google Scholar 

  12. J. Moody, in:Advances in Neural Information Processing Systems 4, Proceedings of the 1991 Conference, J. E. Moody, S. J. Hanson, and R. P. Lippmann (eds.), Morgan Kaufmann Publishers, San Mateo, California (1992), p. 847.

    Google Scholar 

  13. M. Rosenblatt,Stationary Sequences and Random Fields, Birkhäuser, Boston, Massachusetts (1985).

    Google Scholar 

  14. G. A. F. Seber and C. J. Wild,Nonlinear Regression, John Wiley and Sons, New York (1989).

    Google Scholar 

Download references

Authors

Additional information

The Computational Neural Network Center, Electronics Institute, The Technical University of Denmark, Building 349. Translated from Izvestiya Vysshikh Uchebnykh Zavedenii, Radiofizika, Vol. 37, No. 9, pp. 1131–1147, September, 1994.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Larsen, J. Optimizing neural network architectures using generalization error estimators. Radiophys Quantum Electron 37, 729–740 (1994). https://doi.org/10.1007/BF01039612

Download citation

  • Received:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF01039612

Keywords

Navigation