Skip to main content

Evaluating Decision Trees Grown with Asymmetric Entropies

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 4994))

Abstract

We propose to evaluate the quality of decision trees grown on imbalanced datasets with a splitting criterion based on an asymmetric entropy measure. To deal with the class imbalance problem in machine learning, especially with decision trees, different authors proposed such asymmetric splitting criteria. After the tree is grown a decision rule has to be assigned to each leaf. The classical Bayesian rule that selects the more frequent class is irrelevant when the dataset is strongly imbalanced. A best suited assignment rule taking asymmetry into account must be adopted. But how can we then evaluate the resulting prediction model? Indeed the usual error rate is irrelevant when the classes are strongly imbalanced. Appropriate evaluation measures are required in such cases. We consider ROC curves and recall/precision graphs for evaluating the performance of decision trees grown from imbalanced datasets. These evaluation criteria are used for comparing trees obtained with an asymmetric splitting criterion with those grown with a symmetric one. In this paper we only consider the cases involving 2 classes.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Provost, F.: Learning with imbalanced data sets. In: AAAI 2000 Workshop on Imbalanced Data Sets (2000)

    Google Scholar 

  2. Barandela, R., Sánchez, J.S., García, V., Rangel, E.: Strategies for learning in class imbalance problems. Pattern Recognition 36(3), 849–851 (2003)

    Article  Google Scholar 

  3. Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification And Regression Trees. Chapman and Hall, New York (1984)

    MATH  Google Scholar 

  4. Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)

    Google Scholar 

  5. Zighed, D.A., Marcellin, S., Ritschard, G.: Mesure d’entropie asymétrique et consistante. In: EGC, pp. 81–86 (2007)

    Google Scholar 

  6. Rényi, A.: On measures of entropy and information. In: 4th Berkely Symp. Math. Statist. Probability, vol. 1, pp. 547–561 (1960)

    Google Scholar 

  7. Aczel, J., Daroczy, Z.: On measures of information and their characterizations (1975)

    Google Scholar 

  8. Zighed, D., Rakotomalala, R.: Graphe d’induction Apprentissage et Data Mining. Hermés, Paris (2000)

    Google Scholar 

  9. Lallich, S., Lenca, P., Vaillant, B.: Construction d’une entropie décentrée pour l’apprentissage supervisé. In: 3éme Atelier Qualité des Connaissances á partir des Données (QDC-EGC 2007), Namur, Belgique, pp. 45–54 (2007)

    Google Scholar 

  10. Thomas, J., Jouve, P.E., Nicoloyannis, N.: Mesure non symétrique pour l’évaluation de modéles, utilisation pour les jeux de données déséquilibrés. In: 3éme Atelier Qualité des Connaissances á partir des Données (QDC-EGC 2007), Namur, Belgique (2007)

    Google Scholar 

  11. Marcellin, S., Zighed, D., Ritschard, G.: An asymmetric entropy measure for decision trees. In: 11th Information Processing and Management of Uncertainty in knowledge-based systems (IPMU 2006), Paris, France, pp. 1292–1299 (2006)

    Google Scholar 

  12. Ritschard, G., Zighed, D., Marcellin, S.: Données déséquilibrées, entropie décentrée et indice d’implication. In: Gras, R., Orús, P., Pinaud, B., Gregori, P. (eds.) Nouveaux apports théoriques á l’analyse statistique implicative et applications (actes des 4émes rencontres ASI4, Castellón de la Plana (Espan̈a), Departament de Matemàtiques, Universitat Jaume, October 18-21, 2007, vol. I, pp. 315–327 (2007)

    Google Scholar 

  13. Egan, J.: Signal detection theory and roc analysis. Series in Cognition and Perception (1975)

    Google Scholar 

  14. Fawcett, T.: An introduction to roc analysis. Pattern Recognition Letter 27(8), 861–874 (2006)

    Article  MathSciNet  Google Scholar 

  15. Hettich, S., Bay, S.D.: The uci kdd archive (1999)

    Google Scholar 

  16. Chen, C., Liaw, A., Breiman, L.: Using random forest to learn imbalanced data (2004)

    Google Scholar 

  17. Chai, X., Deng, L., Yang, Q.: Ling: Test-cost sensitive naive bayes classification. In: ICDM (2005)

    Google Scholar 

  18. Sebastiani, F.: Machine learning in automated text categorization. ACM Comput. Surv. 34(1), 1–47 (2002)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Aijun An Stan Matwin Zbigniew W. Raś Dominik Ślęzak

Rights and permissions

Reprints and permissions

Copyright information

© 2008 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Marcellin, S., Zighed, D.A., Ritschard, G. (2008). Evaluating Decision Trees Grown with Asymmetric Entropies. In: An, A., Matwin, S., Raś, Z.W., Ślęzak, D. (eds) Foundations of Intelligent Systems. ISMIS 2008. Lecture Notes in Computer Science(), vol 4994. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-68123-6_6

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-68123-6_6

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-68122-9

  • Online ISBN: 978-3-540-68123-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics