Reducing the Training Times of Neural Classifiers with Dataset Condensing
In this paper we apply a k-nearest-neighbour-based data condensing algorithm to the training sets of multi-layer perceptron neural networks. By removing the overlapping data and retaining only training exemplars adjacent to the decision boundary we are able to significantly speed the network training time while achieving an undegraded misclassification rate compared to a network trained on the unedited training set. We report results on a range of synthetic and real datasets which indicate that a speed-up of an order of magnitude in the network training time is typical.
KeywordsNeural networks data editing pattern classifiers
- Kraaijveld M A, Duin R P W, “An Optimal Stopping Criterion for Backpropagation Learning”, Neural Network World, Vol 1, No 6, pp 365–370, June 1991.Google Scholar
- Hara K, Nakayama K, “Training Data Selection Method for Generalization by Multilayer Neural Networks”, IEICE Trans Fundamentals, vol E81–A, No 3, pp 374–381, 1998.Google Scholar
- Downloadable from ftp://ftp.ics.uci.edu/pub/learning-databases/