Skip to main content

Selecting Variables for Neural Network Committees

  • Conference paper
Advances in Neural Networks - ISNN 2006 (ISNN 2006)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 3971))

Included in the following conference series:

Abstract

The aim of the variable selection is threefold: to reduce model complexity, to promote diversity of committee networks, and to find a trade-off between the accuracy and diversity of the networks. To achieve the goal, the steps of neural network training, aggregation, and elimination of irrelevant input variables are integrated based on the negative correlation learning [1] error function. Experimental tests performed on three real world problems have shown that statistically significant improvements in classification performance can be achieved from neural network committees trained according to the technique proposed.

This work was supported in part by the Lithuanian State Science and Studies Foundation.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Liu, Y., Yao, X.: Ensemble Learning via Negatine Correlation. Neural Networks 12, 1399–1404 (1999)

    Article  Google Scholar 

  2. Bacauskiene, M., Verikas, A.: Selecting Salient Features for Classification Based on Neural Network Committees. Pattern Recognition Letters 25, 1879–1891 (2004)

    Article  Google Scholar 

  3. Verikas, A., Lipnickas, A., Malmqvist, K.: Selecting Neural Networks for Making a Committee Decision. In: Dorronsoro, J.R. (ed.) ICANN 2002. LNCS, vol. 2415, pp. 420–425. Springer, Heidelberg (2002)

    Chapter  Google Scholar 

  4. Monari, G., Dreyfus, G.: Local Overfitting Control via Leverages. Neural Computation 14, 1481–1506 (2002)

    Article  MATH  Google Scholar 

  5. Kuncheva, L.I., Bezdek, J.C., Duin, R.P.W.: Decision Templates for Multiple Classifier Fusion. Pattern Recognition 34, 299–314 (2001)

    Article  MATH  Google Scholar 

  6. Verikas, A., Lipnickas, A., Malmqvist, K., Bacauskiene, M., Gelzinis, A.: Soft Combination of Neural Classifiers: A Comparative Study. Pattern Recognition Letters 20, 429–444 (1999)

    Article  Google Scholar 

  7. Setiono, R., Liu, H.: Neural-Network Feature Selector. IEEE Transactions on Neural Networks 8, 654–662 (1997)

    Article  Google Scholar 

  8. Verikas, A., Bacauskiene, M.: Feature Selection with Neural Networks. Pattern Recognition Letters 23, 1323–1335 (2002)

    Article  MATH  Google Scholar 

  9. Cohen, P.R.: Empirical Methods for Artificial Intelligence. MIT Press, Cambridge (1995)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Bacauskiene, M., Cibulskis, V., Verikas, A. (2006). Selecting Variables for Neural Network Committees. In: Wang, J., Yi, Z., Zurada, J.M., Lu, BL., Yin, H. (eds) Advances in Neural Networks - ISNN 2006. ISNN 2006. Lecture Notes in Computer Science, vol 3971. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11759966_123

Download citation

  • DOI: https://doi.org/10.1007/11759966_123

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-34439-1

  • Online ISBN: 978-3-540-34440-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics