Abstract
Object oriented design divides complex algorithms and data structures into smaller and simpler components, specializing in solving extracted subproblems. As a result, also in the approach to a general framework for DT induction, the algorithms can be composed by a number of compatible components. In the framework described in Chap. 3, even the simplest DT induction algorithms are composed of several components responsible for such tasks as performing search process, estimating split quality, pruning and so on.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
The example has been prepared especially for this illustration. It has not been published in any article as an approach claiming to be right.
- 2.
Since the counts in Table 5.4 are summed over 5 algorithms, the maximum possible value is \(5*21=105\). The highest score of 56 means wins in more than half of the tests.
References
Almuallim H (1996) An efficient algorithm for optimal pruning of decision trees. Artif Intell 83(2):347–362. http://dx.doi.org/10.1016/0004-3702(95)00060-7
Bohanec M, Bratko I (1994) Trading accuracy for simplicity in decision trees. Mach Learn 15:223–250. http://dx.doi.org/10.1007/BF00993345, 10.1007/BF00993345
Breiman L, Friedman JH, Olshen A, Stone CJ (1984) Classification and regression trees. Wadsworth, Belmont
Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140
Cestnik B, Bratko I (1991) On estimating probabilities in tree pruning. In: Kodratoff Y (ed) Machine learning—EWSL-91, Lecture notes in computer science, vol 482. Springer, Berlin, pp 138–150. http://dx.doi.org/10.1007/BFb0017010, 10.1007/BFb0017010
Dietterich TG (2000) An experimental comparison of three methods for constructing ensembles of decision trees: bagging boosting and randomization. Mach Learn 40(2):139–157. doi:10.1023/A:1007607513941, http://dx.doi.org/10.1023/A:1007607513941
Esposito F, Malerba D, Semeraro G (1997) A comparative analysis of methods for pruning decision trees. IEEE Trans Pattern Anal Mach Intell 19(5):476–491
Fournier D, Crémilleux B (2002) A quality index for decision tree pruning. Knowl-Based Syst 15(1–2):37–43
Frank A, Asuncion A (2010) UCI machine learning repository. http://archive.ics.uci.edu/ml
Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Machine learning: proceedings of the thirteenth international conference
Gehrke J, Ramakrishnan R, Ganti V (2000) Rainforest—a framework for fast decision tree construction of large datasets. Data Min Knowl Discov 4:127–162. http://dx.doi.org/10.1023/A:1009839829793
Grąbczewski K (2011a) Separability of split value criterion with weighted separation gains. In: Perner P (ed) Machine learning and data mining in pattern recognition, Lecture notes in computer science, vol 6871. Springer, Berlin, pp 88–98. http://dx.doi.org/10.1007/978-3-642-23199-5_7
Grąbczewski K (2011b) Validated decision trees versus collective decisions. In: Jedrzejowicz P, Nguyen N, Hoang K (eds) Computational collective intelligence. Technologies and applications, Lecture notes in computer science, vol 6923. Springer, Berlin, pp 342–351. http://dx.doi.org/10.1007/978-3-642-23938-0_35
Grąbczewski K, Jankowski N (2006) Mining for complex models comprising feature selection and classification. In: Guyon I, Gunn S, Nikravesh M, Zadeh L (eds) Feature extraction, foundations and applications. Studies in fuzziness and soft computing. Springer, Heidelberg, pp 473–489
Grąbczewski K, Duch W (1999) A general purpose separability criterion for classification systems. In: Proceedings of the 4th conference on neural networks and their applications, Zakopane, Poland, pp 203–208
Grąbczewski K, Duch W (2000) The separability of split value criterion. In: Proceedings of the 5th conference on neural networks and their applications, Zakopane, Poland, pp 201–208
Grąbczewski K, Jankowski N (2011) Saving time and memory in computational intelligence system with machine unification and task spooling. Knowl-Based Syst 24:570–588. http://dx.doi.org/10.1016/j.knosys.2011.01.003
Jankowski N, Grąbczewski K (2007) Handwritten digit recognition—road to contest victory. In: IEEE symposium series on computational intelligence. IEEE Press, USA, pp 491–498
Kononenko I (1998) The minimum description length based decision tree pruning. In: Lee HY, Motoda H (eds) PRICAI’98: topics in artificial intelligence, Lecture notes in computer science, vol 1531. Springer, Berlin, pp 228–237
Lim TS, Loh WY, Shih YS (2000) A comparison of prediction accuracy, complexity, and training time of thirty-three old and new classification algorithms. Mach Learn 40:203–228
Loh WY, Shih YS (1997) Split selection methods for classification trees. Stat. Sinica 7:815–840
Mingers J (1989) An empirical comparison of pruning methods for decision tree induction. Mach Learn 4(2):227–243
Niblett T, Bratko I (1986) Learning decision rules in noisy domains. In: Proceedings of expert systems’86, the 6th annual technical conference on research and development in expert systems III. Cambridge University Press, New York, pp 25–34
Quinlan JR (1987) Simplifying decision trees. Int J Man-Mach Stud 27(3):221–234. http://dx.doi.org/10.1016/S0020-7373(87)80053-6
Quinlan JR (1993) C 4.5: programs for machine learning. Morgan Kaufmann, San Mateo
Quinlan JR (1996) Bagging, boosting, and C4.5. In: Proceedings of the thirteenth national conference on artificial intelligence and eighth innovative applications of artificial intelligence conference, AAAI 96, IAAI 96, vol 1. AAAI Press/MIT Press, Portland, Oregon, pp 725–730
Torres-Sospedra J, Hernández-Espinosa C, Fernández-Redondo M (2007) Averaged conservative boosting: introducing a new method to build ensembles of neural networks. In: de Sá J, Alexandre L, Duch W, Mandic D (eds) Artificial neural networks—ICANN 2007, Lecture notes in computer science, vol 4668. Springer, Berlin, pp 309–318
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Grąbczewski, K. (2014). Meta-Level Analysis of Decision Tree Induction. In: Meta-Learning in Decision Tree Induction. Studies in Computational Intelligence, vol 498. Springer, Cham. https://doi.org/10.1007/978-3-319-00960-5_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-00960-5_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-00959-9
Online ISBN: 978-3-319-00960-5
eBook Packages: EngineeringEngineering (R0)