Skip to main content

Meta-Level Analysis of Decision Tree Induction

  • Chapter
  • First Online:
Meta-Learning in Decision Tree Induction

Part of the book series: Studies in Computational Intelligence ((SCI,volume 498))

  • 1825 Accesses

Abstract

Object oriented design divides complex algorithms and data structures into smaller and simpler components, specializing in solving extracted subproblems. As a result, also in the approach to a general framework for DT induction, the algorithms can be composed by a number of compatible components. In the framework described in Chap. 3, even the simplest DT induction algorithms are composed of several components responsible for such tasks as performing search process, estimating split quality, pruning and so on.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The example has been prepared especially for this illustration. It has not been published in any article as an approach claiming to be right.

  2. 2.

    Since the counts in Table 5.4 are summed over 5 algorithms, the maximum possible value is \(5*21=105\). The highest score of 56 means wins in more than half of the tests.

References

  • Almuallim H (1996) An efficient algorithm for optimal pruning of decision trees. Artif Intell 83(2):347–362. http://dx.doi.org/10.1016/0004-3702(95)00060-7

  • Bohanec M, Bratko I (1994) Trading accuracy for simplicity in decision trees. Mach Learn 15:223–250. http://dx.doi.org/10.1007/BF00993345, 10.1007/BF00993345

    Google Scholar 

  • Breiman L, Friedman JH, Olshen A, Stone CJ (1984) Classification and regression trees. Wadsworth, Belmont

    MATH  Google Scholar 

  • Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140

    MATH  MathSciNet  Google Scholar 

  • Cestnik B, Bratko I (1991) On estimating probabilities in tree pruning. In: Kodratoff Y (ed) Machine learning—EWSL-91, Lecture notes in computer science, vol 482. Springer, Berlin, pp 138–150. http://dx.doi.org/10.1007/BFb0017010, 10.1007/BFb0017010

  • Dietterich TG (2000) An experimental comparison of three methods for constructing ensembles of decision trees: bagging boosting and randomization. Mach Learn 40(2):139–157. doi:10.1023/A:1007607513941, http://dx.doi.org/10.1023/A:1007607513941

  • Esposito F, Malerba D, Semeraro G (1997) A comparative analysis of methods for pruning decision trees. IEEE Trans Pattern Anal Mach Intell 19(5):476–491

    Article  Google Scholar 

  • Fournier D, Crémilleux B (2002) A quality index for decision tree pruning. Knowl-Based Syst 15(1–2):37–43

    Article  Google Scholar 

  • Frank A, Asuncion A (2010) UCI machine learning repository. http://archive.ics.uci.edu/ml

  • Freund Y, Schapire RE (1996) Experiments with a new boosting algorithm. In: Machine learning: proceedings of the thirteenth international conference

    Google Scholar 

  • Gehrke J, Ramakrishnan R, Ganti V (2000) Rainforest—a framework for fast decision tree construction of large datasets. Data Min Knowl Discov 4:127–162. http://dx.doi.org/10.1023/A:1009839829793

  • Grąbczewski K (2011a) Separability of split value criterion with weighted separation gains. In: Perner P (ed) Machine learning and data mining in pattern recognition, Lecture notes in computer science, vol 6871. Springer, Berlin, pp 88–98. http://dx.doi.org/10.1007/978-3-642-23199-5_7

  • Grąbczewski K (2011b) Validated decision trees versus collective decisions. In: Jedrzejowicz P, Nguyen N, Hoang K (eds) Computational collective intelligence. Technologies and applications, Lecture notes in computer science, vol 6923. Springer, Berlin, pp 342–351. http://dx.doi.org/10.1007/978-3-642-23938-0_35

  • Grąbczewski K, Jankowski N (2006) Mining for complex models comprising feature selection and classification. In: Guyon I, Gunn S, Nikravesh M, Zadeh L (eds) Feature extraction, foundations and applications. Studies in fuzziness and soft computing. Springer, Heidelberg, pp 473–489

    Google Scholar 

  • Grąbczewski K, Duch W (1999) A general purpose separability criterion for classification systems. In: Proceedings of the 4th conference on neural networks and their applications, Zakopane, Poland, pp 203–208

    Google Scholar 

  • Grąbczewski K, Duch W (2000) The separability of split value criterion. In: Proceedings of the 5th conference on neural networks and their applications, Zakopane, Poland, pp 201–208

    Google Scholar 

  • Grąbczewski K, Jankowski N (2011) Saving time and memory in computational intelligence system with machine unification and task spooling. Knowl-Based Syst 24:570–588. http://dx.doi.org/10.1016/j.knosys.2011.01.003

  • Jankowski N, Grąbczewski K (2007) Handwritten digit recognition—road to contest victory. In: IEEE symposium series on computational intelligence. IEEE Press, USA, pp 491–498

    Google Scholar 

  • Kononenko I (1998) The minimum description length based decision tree pruning. In: Lee HY, Motoda H (eds) PRICAI’98: topics in artificial intelligence, Lecture notes in computer science, vol 1531. Springer, Berlin, pp 228–237

    Google Scholar 

  • Lim TS, Loh WY, Shih YS (2000) A comparison of prediction accuracy, complexity, and training time of thirty-three old and new classification algorithms. Mach Learn 40:203–228

    Article  MATH  Google Scholar 

  • Loh WY, Shih YS (1997) Split selection methods for classification trees. Stat. Sinica 7:815–840

    MATH  MathSciNet  Google Scholar 

  • Mingers J (1989) An empirical comparison of pruning methods for decision tree induction. Mach Learn 4(2):227–243

    Google Scholar 

  • Niblett T, Bratko I (1986) Learning decision rules in noisy domains. In: Proceedings of expert systems’86, the 6th annual technical conference on research and development in expert systems III. Cambridge University Press, New York, pp 25–34

    Google Scholar 

  • Quinlan JR (1987) Simplifying decision trees. Int J Man-Mach Stud 27(3):221–234. http://dx.doi.org/10.1016/S0020-7373(87)80053-6

  • Quinlan JR (1993) C 4.5: programs for machine learning. Morgan Kaufmann, San Mateo

    Google Scholar 

  • Quinlan JR (1996) Bagging, boosting, and C4.5. In: Proceedings of the thirteenth national conference on artificial intelligence and eighth innovative applications of artificial intelligence conference, AAAI 96, IAAI 96, vol 1. AAAI Press/MIT Press, Portland, Oregon, pp 725–730

    Google Scholar 

  • Torres-Sospedra J, Hernández-Espinosa C, Fernández-Redondo M (2007) Averaged conservative boosting: introducing a new method to build ensembles of neural networks. In: de Sá J, Alexandre L, Duch W, Mandic D (eds) Artificial neural networks—ICANN 2007, Lecture notes in computer science, vol 4668. Springer, Berlin, pp 309–318

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Krzysztof Grąbczewski .

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Grąbczewski, K. (2014). Meta-Level Analysis of Decision Tree Induction. In: Meta-Learning in Decision Tree Induction. Studies in Computational Intelligence, vol 498. Springer, Cham. https://doi.org/10.1007/978-3-319-00960-5_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-00960-5_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-00959-9

  • Online ISBN: 978-3-319-00960-5

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics