Skip to main content

Improving the generalization performance of multi-layer-perceptrons with population-based incremental learning

  • Applications of Evolutionary Computation Evolutionary Computation in Machine Learning, Neural Networks, and Fuzzy Systems
  • Conference paper
  • First Online:
Parallel Problem Solving from Nature — PPSN IV (PPSN 1996)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1141))

Included in the following conference series:

Abstract

Based on Population-Based Incremental Learning (PBIL) we present a new approach for the evolution of neural network architectures and their corresponding weights. The main idea is to use a probability vector rather than bit strings to represent a population of networks in each generation. We show that crucial issues of neural network training can effectively be integrated into the PBIL framework. First, a Quasi-Newton method for local weight optimization is integrated and the moving average update rule of the PBIL is extended to continuous parameters in order to transmit the best network to the next generation. Second, and more important, we incorporate cross-validation to focus the evolution towards networks with optimal generalization performance. A comparison with standard pruning and stopped-training algorithms shows that our approach effectively finds small networks with increased generalization ability.

This author gratefully acknowledges support by the German BMBF (project EVOALG, a cooperation of Informatik Centrum Dortmund, Siemens AG München, and Humboldt-Universität zu Berlin), grant 01 IB 403 A.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Baluja, C., Caruana R.: Removing the genetics from the standard genetic algorithm, Proc. of the Twelfth Int. Conference on Machine Learning (1995)

    Google Scholar 

  2. Braun, H., Zagorski, P.: ENZO-M, A Hybrid Approach for Optimizing Neural Networks by Evolution and Learning, Parallel Problem Solving from Nature, Springer (1994)

    Google Scholar 

  3. Fletcher, R. Practical methods for optimization, John Wiley and Sons, Chichester (1995)

    Google Scholar 

  4. Harp, S., Samad, T., Guha, A.: Designing application-specific neural networks using the genetic algorithm, Advances in Neural Information Processing Systems 2, Morgan Kaufmann, San Mateo, CA (1990)

    Google Scholar 

  5. Hergert, F.,Finnoff, W. and Zimmermann H.: A comparison of weight elimination methods for reducing complexity in neural networks, Int. Joint Conf. on Neural Networks, Baltimore (1992)

    Google Scholar 

  6. Liu, Y.: Neural Network Model Selection using Asymptotic Jackknife Estimator and Cross-Validation, Advances in Neural Information Processing Systems 4, Morgan Kaufmann, San Mateo, CA (1992)

    Google Scholar 

  7. Svarer, C., Hansen, L., Larsen, J.: On design and evaluation of tapped-delay neural network architectures, IEEE International Conference on Neural Networks, San Francisco (1993)

    Google Scholar 

  8. Tong, H., Lim, K,: Threshold autoregression, limit cycles and cyclical data, Journ. Roy. Stat. Soc. B, 42 (1980) 245

    Google Scholar 

  9. Weigend, A., Rummelhart, D., Huberman, B.: Predicting the future: A connectionist approach, Int. Jour. of Neural Systems (1990)

    Google Scholar 

  10. Goldberg, D.: Gentic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley, Redwood City (1989)

    Google Scholar 

  11. Schwefel, H.-P.: Evolution and Optimium Seeking, John Wiley and Suns, Chichester (1995)

    Google Scholar 

  12. Hertz, J., Krogh, A. and Palmer, R. Introduction to the theory of neural computation, Addison-Wesley, Redwood City (1991)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Hans-Michael Voigt Werner Ebeling Ingo Rechenberg Hans-Paul Schwefel

Rights and permissions

Reprints and permissions

Copyright information

© 1996 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Galić, E., Höhfeld, M. (1996). Improving the generalization performance of multi-layer-perceptrons with population-based incremental learning. In: Voigt, HM., Ebeling, W., Rechenberg, I., Schwefel, HP. (eds) Parallel Problem Solving from Nature — PPSN IV. PPSN 1996. Lecture Notes in Computer Science, vol 1141. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-61723-X_1037

Download citation

  • DOI: https://doi.org/10.1007/3-540-61723-X_1037

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-61723-5

  • Online ISBN: 978-3-540-70668-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics