Abstract
The computational complexity of a classification algorithm should be reduced to a sufficient minimum by reducing the number of features considered. We can either select the most informative features or extract a new, smaller set of features using a (linear) combination of the original features. Principal component analysis (PCA) implements the second option by diagonalizing the covariance matrix to find directions in feature space corresponding to the directions of greatest variance. Linear discriminant analysis (LDA) also reduces the dimensionality of a problem, but specifically finds the most useful directions for separating classes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Fisher, R.A.: The use of multiple measurements in taxonomic problems. Ann. Eugen. 7, 179–188 (1936)
Kruskal, J.B., Wish, M.: Multidimensional Scaling. Sage, Beverly Hills, CA (1977)
Van der Heijden, F., Duin, R.P.W., de Ridder, D., Tax, D.M.J.: Classification, Parameter Estimation and State Estimation. Wiley, Chichester (2004)
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media New York
About this chapter
Cite this chapter
Dougherty, G. (2013). Feature Extraction and Selection. In: Pattern Recognition and Classification. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-5323-9_7
Download citation
DOI: https://doi.org/10.1007/978-1-4614-5323-9_7
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-5322-2
Online ISBN: 978-1-4614-5323-9
eBook Packages: Computer ScienceComputer Science (R0)