Abstract
We report on our approach, CBAmethod3E, which was submitted to the NIPS 2003 Feature Selection Challenge on Dec. 8, 2003. Our approach consists of combining filtering techniques for variable selection, information gain and feature correlation, with Support Vector Machines for induction. We ranked 13th overall and ranked 6th as a group. It is worth pointing out that our feature selection method was very successful in selecting the second smallest set of features among the top-20 submissions, and in identifying almost all probes in the datasets, resulting in the challenge’s best performance on the latter benchmark.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
B.E. Boser, I. Guyon, and V. Vapnik. A training algorithm for optimal margin classifiers. In Fifth Annual Workshop on Computational Learning Theory, pages 144–152. ACM, 1992.
P.S. Bradley and O.L. Mangasarian. Feature selection via concave minimization and support vector machines. In Proc15th IntĊonfṀachine Learning, pages 82–90, 1998.
C.C. Chang and C.J. Lin. LIBSVM: a library for support vector machines, 2001. Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvm.
C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20(3):273–297, 1995.
U. Fayyad and K. Irani. Multi-interval discretization of continuous-valued attributes for classification learning. In Proc10th IntĊonfṀachine Learning, pages 194–201, 1993.
I. Guyon, J. Weston, S. Barnhill, and V. Vapnik. Gene selection for cancer classification using support vector machines. Machine Learning, 46:389–422, 2002.
G. Karakoulas. Cost-effective classification for credit scoring. In Proc3rd IntĊonfȦI Applications on Wall Street, 1995.
T. Mitchell. Machine Learning. McGraw-Hill, New York, 1997.
M. Momma and K.P. Bennett. A pattern search method for model selection of support vector regression. In R. Grossman, J. Han, V. Kumar, H. Mannila, and R. Motwani, editors, Proceedings of the Second SIAM International Conference on Data Mining, pages 261–274. SIAM, 2002.
J. Platt. Fast Training of Support Vector Machines using Sequential Minimal Optimization, chapter 12, pages 185–208. MIT Press, 1999.
P. Turney. Types of cost in inductive concept learning. In Workshop cost-sensitive learning, Proc. 17th Int. Conf. Machine Learning, pages 15–21, 2000.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Roobaert, D., Karakoulas, G., Chawla, N.V. (2006). Information Gain, Correlation and Support Vector Machines. In: Guyon, I., Nikravesh, M., Gunn, S., Zadeh, L.A. (eds) Feature Extraction. Studies in Fuzziness and Soft Computing, vol 207. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-35488-8_23
Download citation
DOI: https://doi.org/10.1007/978-3-540-35488-8_23
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-35487-1
Online ISBN: 978-3-540-35488-8
eBook Packages: EngineeringEngineering (R0)