Summary
This paper aimed at two following objectives. One was the introduction of a new measure (R-measure) of dependency between groups of attributes in a data set, inspired by the notion of dependency of attribute in the rough set theory. The second was the application of this measure to the problem of attribute selection in decision tree induction, and an experimental comparative evaluation of decision tree systems using R-measure and other different attribute selection measures most of them are widely used in machine learning: gain-ratio, gini-index, d N distance, relevance, x 2.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bairn, P.W. (1988): A method for attribute selection in inductive learning systems. IEEE Trans. on PAMI, 10, 888–896.
Breiuran, L., Friedman, J., Olshen, R., Stone, C. (1984): Classification and Regression Trees, Belmont, CA: Wadsworth.
Buntine, W., Niblett, T. (1991): A further comparison of splitting rules for decision-tree induction. Machine Learning, 8, 75–85
Dougherty, J., Kohavi, R. and Sahami, M. (1995): Supervised and Unsupervised Discretization of Continuous Features. Proceedings 12th International Conference on Machine Learning, Morgan Kaufmann, 194–202.
Ho, T.B., Nguyen, T.D. (1997): An interactive-graphic system for decision tree induction (under review).
Kononenko, I. (1995): On biases in estimating multi-valued attributes. Proc. 14th Inter. Joint. Conf. on Artificial Intelligence, Montreal, Morgan Kaufmann, 1034–1040.
Kohavi, R (1995): A study of cross-validation and bootstrap for accuracy estimation and model selection. Proc. Int. Joint Conf. on Artificial Intelligence IJCAI’95, 1137–1143.
Liu, W.Z., White, A.P. (1994): The importance of attribute selection measures in decision tree induction. Machine Learning, 15, 25–41.
López de Mantaras, R. (1991): A distance-based attribute selection measure for decision tree induction. Machine Learning, 6, 81–92.
Mingers, J. (1989): An empirical comparison of selection measures for decision-tree induction. Machine Learning, 3, 319–342.
Pawlak, Z. (1991): Rough Sets: Theoretical Aspects of Reasoning About Data,Kluwer Academic Publishers.
Pawlak, Z., Grzymala-Busse, J., Slowinski, R., Ziarko, W. (1995): Rough sets. Communications of the ACM, 38, 89–95.
Quinlan, J. R. (1993): C4.5: Programs for Machine Learning,Morgan Kaufmann.
Wille, R. (1992): Concept lattice and conceptual knowledge systems. Computers and Mathematics with Applications, 23, 493–515.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1998 Springer Japan
About this paper
Cite this paper
Ho, T.B., Nguyen, T.D., Kimura, M. (1998). Induction of Decision Trees Based on the Rough Set Theory. In: Hayashi, C., Yajima, K., Bock, HH., Ohsumi, N., Tanaka, Y., Baba, Y. (eds) Data Science, Classification, and Related Methods. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Tokyo. https://doi.org/10.1007/978-4-431-65950-1_22
Download citation
DOI: https://doi.org/10.1007/978-4-431-65950-1_22
Publisher Name: Springer, Tokyo
Print ISBN: 978-4-431-70208-5
Online ISBN: 978-4-431-65950-1
eBook Packages: Springer Book Archive