Skip to main content

Optimizing Weights by Genetic Algorithm for Neural Network Ensemble

  • Conference paper
Advances in Neural Networks – ISNN 2004 (ISNN 2004)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 3173))

Included in the following conference series:

Abstract

Combining the outputs of several neural networks into an aggregate output often gives improved accuracy over any individual output. The set of networks is known as an ensemble. Neural network ensembles are effective techniques to improve the generalization of a neural network system. This paper presents an ensemble method for regression that has advantages over simple weighted or weighted average combining techniques. After the training of component neural networks, genetic algorithm is used to optimize the combining weights of component networks. Compared with ordinary weighted methods, the method proposed in this paper achieved high predicting accuracy on five test datasets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hansen, L.K., Salamon, P.: Neural Network Ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence 12(10), 993–1001 (1990)

    Article  Google Scholar 

  2. Zhou, Z.H., Wu, J.X., Tang, W.: Ensembling Neural Networks: Many Could be Better than All. Artificial Intelligence 137, 239–263 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  3. Borra, S., Ciaccio, A.D.: Improving Nonparametric Regression Methods by Bagging and Boosting. Computational Statistics & Data Analysis 38, 407–420 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  4. Schapire, R.E.: A Brief Introduction to Boosting. In: Thomas, D. (ed.) Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence, Stockholm, Sweden, pp. 1–6 (1999)

    Google Scholar 

  5. Houck, C.R., Joines, J.A., Kay, M.G.: A Genetic Algorithm for Function Optimization: A Matlab Implementation. Technical Report NCSU-IE-TR-95-09, North Carolina State University, Raleigh, NC (1995)

    Google Scholar 

  6. The Mathworks (ed.): Neural Network Toolbox User’s Guide (version 4). The Mathworks, Inc., Natick, Massachussets (2001)

    Google Scholar 

  7. Granitto, P.M., Verdes, P.F., Navone, H.D., Ceccatto, H.A.: Aggregation Algorithms for Neural Network Ensemble Construction. In: Werner, B. (ed.) Proceedings of the VII Brazilian Symposium on Neural Networks, pp. 178–183. IEEE Computer Society, Pernambuco (2002)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2004 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Shen, ZQ., Kong, FS. (2004). Optimizing Weights by Genetic Algorithm for Neural Network Ensemble. In: Yin, FL., Wang, J., Guo, C. (eds) Advances in Neural Networks – ISNN 2004. ISNN 2004. Lecture Notes in Computer Science, vol 3173. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-28647-9_55

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-28647-9_55

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-22841-7

  • Online ISBN: 978-3-540-28647-9

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics