Abstract
Large amount of data have been collected routinely in the course of day-to-day work in different fields. Typically, the datasets constantly grow accumulating a large number of features, which are not equally important in decision-making. Moreover, the information often lacks completeness and has relatively low information density. Dimensionality reduction is a fundamental area of research in data mining domain. Rough Set Theory (RST), based on a mathematical concept, has become very popular in dimensionality reduction of large datasets. The method is used to determine a subset of attributes called reduct which can predict the decision concepts. In the paper, the concepts of discernibility relation and attribute dependency are integrated for the formation of a compact reduct set which not only reduces the complexity but also helps to achieve higher accuracy of the system. Performance of the proposed method has been evaluated by comparing classification accuracy with some existing dimension reduction algorithms, demonstrating superior result.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Raymer, M.L., et al.: Dimensionality Reduction using Genetic Algorithms. IEEE Transactions on Evolutionary Computation 4(2), 164–171 (2000)
Carreira-Perpinan, M.A.: A Review of Dimension Reduction Techniques. Technical Report CS-96-09, Department of Computer Science, University of Sheffield (1997)
Gordon, A.D.: Classification, 2nd edn. Monographs on Statistics and Applied Probability. Chapman and Hall/CRC, London (1999) ISBN: 9781584880134
Pal, S.K., Mitra, S.: Multi-Layer Perceptron, fuzzy Sets and Classification. IEEE Trans. Neural Networks 3, 683–697 (1992)
Komorowski, J., Pawalk, Z., Polkowski, S.A.: Rough Sets: A Tutorial. In: Pal, S.K., Skowron, A. (eds.) Rough Fuzzy Hybridization: A New Trend in Decision-Making, pp. 3–98. Springer, Berlin (1999)
Pawlak, Z.: Rough Set Theory and Its Applications to Data Analysis. Cybernetics and Systems 29, 661–688 (1998)
Garey, M., Johnson, D.: Computers and Intractability- A Guide to the Theory of NP-Completeness. Freeman, New York (1979)
Swiniarski, R.W.: Rough Sets Methods in Feature Reduction and Classification. International Journal of Applied Mathematics and Computer Science 11(3), 565–582 (2001)
Qu, G., Hariri, S., Yousif, M.: A New Dependency and Correlation Analysis for Features. IEEE Transactions on Knowledge and Data Engineering 17(9), 1199–1207 (2005)
Novovičová, J., Somol, P., Haindl, M., Pudil, P.: Conditional Mutual Information Based Feature Selection for Classification Task. In: Rueda, L., Mery, D., Kittler, J. (eds.) CIARP 2007. LNCS, vol. 4756, pp. 417–426. Springer, Heidelberg (2007)
Freitas, A.: A Genetic Programming Framework for Two Data Mining Tasks- Classification and Generalized Rule Induction. In: Conf. on Genetic Programming, USA, pp. 96–101 (1997)
Deng, D., Huang, H.: Dynamic Reduction Based on Rough Sets in Incomplete Decision Systems. In: Yao, J., Lingras, P., Wu, W.-Z., Szczuka, M.S., Cercone, N.J., Ślęzak, D. (eds.) RSKT 2007. LNCS (LNAI), vol. 4481, pp. 76–83. Springer, Heidelberg (2007)
Quinlan, J., Rivest, R.: Inferring Decision Trees using the Minimum Description Length Principle. Inf. Comput. 80, 227–248 (1989)
Hansen, M., Yu, B.: Model Selection and the Principle of Minimum Description Length. J. Am. Stat. Assoc. 96, 746–774 (2001)
Quinlan, J.R.: The Minimum Description Length and Categorical Theories. In: Proceedings 11th International Conference on Machine Learning, New Brunswick, pp. 233–241. Morgan Kaufmann, San Francisco (1994)
Roman, W.S., Larry, H.: Rough sets as a Frontend as Neural-Networks Texture Classifiers. Neuro-Computing 36, 85–102 (2001)
Kuncheva, L.I.: Combining Pattern Classifiers, Methods and Algorithms. Wiley Interscience, New York (2005)
Fumera, G., Roli, F.: Analysis of Error-Reject Trade off in Linearly Combined Multiple Classifiers. Pattern Recognition 37, 1245–1265 (2004)
Kittler, J., Hatef, M.: On Combining Classifiers. IEEE Transactions on Pattern Analysis and Machine Intelligence 20(3), 226–239 (1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Das, A.K., Chakrabarty, S., Sengupta, S. (2012). Formation of a Compact Reduct Set Based on Discernibility Relation and Attribute Dependency of Rough Set Theory. In: Venugopal, K.R., Patnaik, L.M. (eds) Wireless Networks and Computational Intelligence. ICIP 2012. Communications in Computer and Information Science, vol 292. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-31686-9_30
Download citation
DOI: https://doi.org/10.1007/978-3-642-31686-9_30
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-31685-2
Online ISBN: 978-3-642-31686-9
eBook Packages: Computer ScienceComputer Science (R0)