Abstract
A neural network architecture is introduced which implements a supervised clustering algorithm for the classification of feature vectors. The network is selforganising, and is able to adapt to the shape of the underlying pattern distribution as well as detect novel input vectors during training. It is also capable of determining the relative importance of the feature components for classification. The architecture is a hybrid of supervised and unsupervised networks, and combines the strengths of three wellknown architectures: learning vector quantisation, backpro-pagation and adaptive resonance theory. Network performance is compared to that of learning vector quantisation, back-propagation and cascade-correlation. It is found that performance is generally as good as or better than the performance of these other architectures, while training time is considerably shorter. However, the main advantage of the hybrid architecture is its ability to gain insight into the feature pattern space.
Similar content being viewed by others
Abbreviations
- O j :
-
The output value of thejth unit
- I i :
-
Theith component of the input pattern
- W ij :
-
The weight of the cluster connection between theith input and thejth unit
- B ij :
-
The weight of the shape connection between theith input and thejth unit
- N :
-
The dimension of the input patterns
- v j :
-
The vigilance parameter of thejth unit
- v init :
-
The initial vigilance parameter value
- v rate :
-
The change in the vigilance parameter value
- X i :
-
Theith direction in anN-dimensional coordinate system
- T k :
-
The classification tag of thekth unit
- C :
-
The classification tag of the current input vector
- α(p) :
-
The learning rate at thepth epoch for the cluster weights
- p :
-
The current epoch
- P :
-
The total number of epochs
- E k :
-
The error associated with thekth unit
- β :
-
The constant learning rate for the shape weights
- a j :
-
The age in epochs of thejth unit
References
Jain AK. Cluster analysis. In Handbook of Pattern Recognition and Image Processing, Young TY, Fu KS (eds.), Academic Press, 1986, 33–57
Kaufman L, Rousseeuw PJ. Finding Groups in Data, an Introduction to Cluster Analysis, John Wiley & Sons, 1990
Lillesand TM, Kiefer RW. Remote Sensing and Image Interpretation, 3rd ed., John Wiley & Sons, 1994
Kohonen T. Improved versions of learning vector quantization. Proc Int Joint Conf Neural Networks 1990; 1: 545–550
Fausett LV. Fundamentals of Neural Networks, Architectures, Algorithms and Applications, Prentice-Hall, 1994
Fritzke B. Growing cell structures — a self-organizing network for unsupervised and supervised learning. Neural Networks 1994; 7:1441–1460
Moody J, Darken C. Learning with localized receptive fields. Proc Connectionist Summer School, Morgan Kaufmann, 1988; 133–143
Rouspini EH. Numerical methods for fuzzy clustering. Infor Sci 1970; 2: 319–350
Musavi M, Ahmed W, Chan K, Fairs K, Hummels D. On the training of radial basis functions classifiers. Neural Networks 1992; 5: 595–603
Nowlan S, Hinton G. Adaptive soft weight tying using gaussian mixtures. In: Advances in Neural Information Processing Systems 4, Moody JE, Hanson SJ, Lippman RP (eds.), Morgan Kaufmann, 1992; 993–1000
Bors A, Gabboujh M. Minimal topology for a radial basis functions neural network for pattern classification. Digital Signal Processing 1994; 4: 173–188
Nowlan S. Maximum likelihood competitive learning. In: Advances in Neural Information Processing Systems 2, Touretzky D (ed.), Morgan Kaufmann, 1990: 574–582
Reilly DL, Cooper LN, Elbaum C. A neural model for category learning. Biological Cybernetics 1982; 45: 35–41
Platt J. A resource-allocating network for function interpolation. Neural Computation 1991; 3: 213–225
Carpenter GA, Grossberg S. Pattern Recognition by Self-Organizing Neural Networks, MIT Press, 1991
Rumelhart DE, Hinton GE, Williams RJ. Learning internal representations by error propagation. In: Parallel Distributed Processing: Explorations in the Microstructure of Cognition, Vol. 1: Foundations, Rumelhart DE, McClelland JL (eds.), MIT Press, 1986; 318–362
Fahlman SE, Lebiere C. The Cascade-Correlation learning architecture. In: Advances in Neural Information Processing Systems 2, Touretzky D (ed.), Morgan Kaufmann, 1990; 524–532
Lang KJ, Witbrock MJ. Learning to tell two spirals apart. In: Proc Connectionist Summer School, Touretzky D, Hinton G, Sejnowski T (eds.), Morgan Kaufmann, 1988; 52–59.
Brodatz P. Textures, A Photographic Album for Artists and Designers. Dover, 1966
Haralick RN, Shanmugam K, Dinstein I. Textural features for image classification, IEEE Trans Systems Man Cybernetics 1973; 3: 610–621
Augusteijn MF, Clemens LE, Shaw KA. Performance evaluation of texture measures for ground cover identification in satellite images by means of a neural network classifier. IEEE Trans Geosci Remote Sensing 1995; 33: 616–626
Weszka JS, Dyer CR, Rosenfeld A. A comparative study of algorithms. IEEE Trans Systems Man Cybernetics 1976; 6: 269–285
Conners RW, Harlow CA. A theoretical comparison of texture algorithms. IEEE Trans Pattern Analysis Machine Intelligence, 1980; 2: 204–222
du Buf JMH, Kardan M, Spann M. Texture feature performance for image segmentation. Pattern Recogn. 1990; 23: 291–309
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Augusteijn, M.F., Steck, U.J. Supervised adaptive clustering: A hybrid neural network clustering algorithm. Neural Comput & Applic 7, 78–89 (1998). https://doi.org/10.1007/BF01413712
Issue Date:
DOI: https://doi.org/10.1007/BF01413712