Skip to main content

An Appropriate Abstraction for Constructing a Compact Decision Tree

  • Conference paper
  • First Online:
Discovery Science (DS 2000)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1967))

Included in the following conference series:

Abstract

In general, it is considered that pre-processings for data mining are necessary techniques to remove irrelevant and meaningless aspects of data before applying data mining algorithms. From this viewpoint, we have considered pre-processing for detecting a decision tree, and already proposed a notion of Information Theoretical Abstraction, and implemented a system ITA. Given a relational database and a family of possible abstractions for its attribute values, called an abstraction hierarchy, our system ITA selects the best abstraction among the possible ones so that class distributions needed to perform our classification task are preserved, and generalizes database according to the best abstraction. According to our previous experiment, just one application of abstraction for the whole database has shown its effectiveness in reducing the size of the detected decision tree, without making the classification accuracy worse. However, since such classification systems as C4.5 perform serial attribute-selection repeatedly, ITA does not generally guarantee the preservingness of class distributions, given a sequence of attribute-selections. For this reason, in this paper, we propose a new version of ITA, called iterative ITA, so that it tries to keep the class distributions in each attribute selection step as possibly as we can.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Fayyad, U.N., Piatetsky-Shapiro, G., Smyth, P. and Uthurusamy, R.(eds.): Advances in Knowledge Discovery and Data Mining. AAAI/MIT Press, 1996.

    Google Scholar 

  2. Han, J. and Fu, Y.: Attribute-Oriented Induction in Data Mining. In:[1], pp.399–421, 1996.

    Google Scholar 

  3. Kudoh, Y. and Haraguchi, M.: An Appropriate Abstraction for an Attribute-Oriented Induction Proceeding of The Second International Conference on Discovery Science, LNAI 1721, pp.43–55, 1999.

    Google Scholar 

  4. Quinlan, J.R.: C4.5-Programs for Machine Learning, Morgan Kaufmann, 1993.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2000 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kudoh, Y., Haraguchi, M. (2000). An Appropriate Abstraction for Constructing a Compact Decision Tree. In: Arikawa, S., Morishita, S. (eds) Discovery Science. DS 2000. Lecture Notes in Computer Science(), vol 1967. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44418-1_33

Download citation

  • DOI: https://doi.org/10.1007/3-540-44418-1_33

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-41352-3

  • Online ISBN: 978-3-540-44418-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics