Abstract
Induction of decision trees belongs to the most popular algorithms used in machine learning and data mining. This process will result in a single tree that can be use both for classification of new examples and for description the partitioning of the training set. In the paper we propose an alternative approach that is related to the idea of finding all interesting relations (usually association rules, but in our case all interesting trees) in given data. When building the so called exploration trees, we consider not a single best attribute for branching but more ”good” attributes for each split. The proposed method will be compared with the ”standard” C4.5 algorithm on several data sets from the loan application domain.
We propose this algorithm in the framework of the GUHA method, a genuine exploratory analysis method that aims at finding all patterns, that are true in the analyzed data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Agrawal, R., Imielinski, T., Swami, A.: Mining Association Rules Between Sets of Items in Large Databases. In: SIGMOD Conference, pp. 207–216 (1993)
Biggs, D., deVille, B., Suen, E.: A method of choosing multiway partitions for classification and decision trees. Journal of Applied Statistics 18(1), 49–62 (1991)
Breiman, L.: Random Forrests. Machine Learning 45, 5–32 (2001)
Breiman, L., Friedman, J.H., Ohlsen, R.A., Stone, P.J.: Classification and Regression Trees. Wadsworth, Belmont (1984)
Freund, Y., Mason, L.: The Alternating Decision Tree Learning Algorithm. In: Procedings of the 16th Int. Conference on Machine Learning, vol. 1, pp. 124–133. Morgan Kaufman, San Francisco (1999)
Freund, Y., Schapirem, R.E.: Experiments with a new boosting algorithm. In: Procedings of the 13th Int. Conference on Machine Learning, pp. 148–156. Morgan Kaufmann, San Francisco (1996)
Hájek, P., Havránek, T.: Mechanising Hypothesis Formation Mathematical Foundations for a General Theory. Springer, Heidelberg (1978)
Kohavi, R., Kunz, C.: Option Decision Trees with Majority Notes. In: Procedings of the 14th Int. Conference on Machine Learning, pp. 161–169. Morgan Kaufmann, San Francisco (1997)
Quinlan, J.R.: Induction of decision trees. Machine Learning 1(1), 81–106 (1986)
Quinlan, J.R.: C4.5: Programs for machine learning. Morgan Kaufmann, San Francisco (1993)
Ralbovský, M.: History and Future Development of the Ferda system. Mundus Symbolicus 15, 143–147 (2007)
Rauch, J., Šimůnek, M.: An Alternative Approach to Mining Association Rules. In: Proc. Foundations of Data Mining and Knowledge Discovery. Springer, Heidelberg (2005)
Šimůnek, M.: Academic KDD Project LISp-Miner. In: Advances in Soft Computing Intelligent Systems Design and Applications, vol. 272, pp. 263–272. Springer, Heidelberg (2003)
UCI Machine Learning Repository, http://www.ics.uci.edu/mlearn/MLRepository.html
Weka - Data Mining with Open Source Machine Learning Software, http://www.cs.waikato.ac.nz/ml/weka/
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Berka, P. (2011). ETree Miner: A New GUHA Procedure for Building Exploration Trees. In: Kryszkiewicz, M., Rybinski, H., Skowron, A., Raś, Z.W. (eds) Foundations of Intelligent Systems. ISMIS 2011. Lecture Notes in Computer Science(), vol 6804. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-21916-0_11
Download citation
DOI: https://doi.org/10.1007/978-3-642-21916-0_11
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-21915-3
Online ISBN: 978-3-642-21916-0
eBook Packages: Computer ScienceComputer Science (R0)