A Neural Adaptive Algorithm for Feature Selection and Classification of High Dimensionality Data

  • Elisabetta Binaghi
  • Ignazio Gallo
  • Mirco Boschetti
  • P. Alessandro Brivio
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3617)


In this paper, we propose a novel method which involves neural adaptive techniques for identifying salient features and for classifying high dimensionality data. In particular a network pruning algorithm acting on MultiLayer Perceptron topology is the foundation of the feature selection strategy. Feature selection is implemented within the back-propagation learning process and based on a measure of saliency derived from bell functions positioned between input and hidden layers and adaptively varied in shape and position during learning. Performances were evaluated experimentally within a Remote Sensing study, aimed to classify hyperspectral data. A comparison analysis was conducted with Support Vector Machine and conventional statistical and neural techniques. As seen in the experimental context, the adaptive neural classifier showed a competitive behavior with respect to the other classifiers considered; it performed a selection of the most relevant features and showed a robust behavior operating under minimal training and noisy situations.


Feature Selection Hide Neuron High Dimensionality Data Adaptive Model Spectral Angle Mapper 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


  1. 1.
    Fukunaga, K., Hayes, R.R.: Effects of Sample Size Classifier Design. IEEE Trans. Pattern Analysis and Machine Intelligence 11(8), 873–885 (1989)CrossRefGoogle Scholar
  2. 2.
    Jain, A.K., Duin, R.P., Mao, J.: Statistical Pattern Recognition: a review. IEEE Trans. on Pattern Analysis and Machine Intelligence 22, 4–37 (2000)CrossRefGoogle Scholar
  3. 3.
    Jain, A., Zongker, D.: Feature Selection: Evaluation, Application, and Small Sample Performance. IEEE Trans. on Pattern Analysis and Machine Intelligence 19(2), 153–158 (1997)CrossRefGoogle Scholar
  4. 4.
    Pao, Y.H.: Adaptive Pattern Recognition and Neural Networks. Addison-Wesley, MA (1989)zbMATHGoogle Scholar
  5. 5.
    Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, Oxford (1995)Google Scholar
  6. 6.
    Reed, R.: Pruning Algorithms - a survey. IEEE Trans. Neural Networks 5, 740–747 (1993)CrossRefGoogle Scholar
  7. 7.
    Rumelhart, H., Hinton, G.E., Williams, R.J.: Learning Internal Representation by Error Propagation. In: Rumelhart, H., Mc Lelland, J.L. (eds.) Parallel Distributed Processing, pp. 318–362. MIT Press, Cambridge (1986)Google Scholar
  8. 8.
    Van Genderen, J.L., Lock, B.F., Vass, P.A.: Remote Sensing: Statistical testing of thematic map accuracy. In: Remote Sensing of Environment, vol. 7, pp. 3–14 (1978)Google Scholar
  9. 9.
    Kruse, F.A., Lefkoff, A.B., Boardman, J.B., Heidebrecht, K.B., Shapiro, A.T., Barloon, P.J., Goetz, A.F.H.: The Spectral Image Processing System (SIPS) - Interactive Visualization and Analysis of Imaging spectrometer Data. In: Remote Sensing of Environment, vol. 44, pp. 145–163 (1993)Google Scholar
  10. 10.
    ENVI, The Environment for Visualizing Images. Research Systems Inc.,
  11. 11.
    Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)zbMATHGoogle Scholar
  12. 12.
    Chang, C.-C., Lin, C.-J.: LIBSVM: a library for support vector machines (2001), Software available at

Copyright information

© Springer-Verlag Berlin Heidelberg 2005

Authors and Affiliations

  • Elisabetta Binaghi
    • 1
  • Ignazio Gallo
    • 1
  • Mirco Boschetti
    • 2
  • P. Alessandro Brivio
    • 2
  1. 1.Dipartimento di Informatica e ComunicazioneUniversita’ degli Studi, dell’InsubriaVareseItaly
  2. 2.Institute for Electromagnetic Sensing of the EnvironmentCNR-IREAMilanItaly

Personalised recommendations