Comparative Analysis of Decision Tree Algorithms: ID3, C4.5 and Random Forest

  • Shiju Sathyadevan
  • Remya R. Nair
Conference paper
Part of the Smart Innovation, Systems and Technologies book series (SIST, volume 31)


To analyze the raw data manually and find the correct information from it is a tough process. But Data mining technique automatically detect the relevant patterns or information from the raw data, using the data mining algorithms. In Data mining algorithms, Decision trees are the best and commonly used approach for representing the data. Using these Decision trees, data can be represented as a most visualizing form. Many different decision tree algorithms are used for the data mining technique. Each algorithm gives a unique decision tree from the input data. This paper focus on the comparison of different decision tree algorithms for data analysis.


Iterative dichotomiser 3 (ID3) C4.5 Randomforest 



We thank our college Amrita School of Engineering, Amritapuri and Amrita Center of Cyber Security, Amritapuri for giving us an opportunity to be a part of the internship program that leads to the development of this work. Many thanks to Shiju Sathyadevan for countless discussions and feedback that help me to complete the work successfully.


  1. 1.
    Rokach, L., Maimon, O.: Decision trees. In Maimon, O.Z., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook vol. 6, pp. 165–192. Springer, Heidelberg (2005)Google Scholar
  2. 2.
    Peng, X., Guo, H., Pang, J.: Performance analysis between different decision trees for uncertain data. In: Proceedings of the 2012 International Conference on Computer Science and Service System, CSSS ’12, pp. 574–577. Washington, DC (2012)Google Scholar
  3. 3.
    Jin, C., De-lin, L., Fen-xiang, M.: An improved ID3 decision tree algorithm. In: Proceedings of 2009 4th International Conference on Computer Science and Education, vol. 1 (2009)Google Scholar
  4. 4.
    Wang, M., Chai, R.: Improved classification attribute selection scheme for decision for decision tree. Comp. Eng. Appl. 3, 127–129 (2010)Google Scholar
  5. 5.
    Guan, C., Zeng, X.: An improved ID3 based on weighted modified information gain. In: Seventh International Conference on Computational Intelligence and Security, vol. 1, pp. 1283–1285. IEEE Computer Society, Washington, DC (2011)Google Scholar
  6. 6.
    Quinlan, J.R.: C4.5 program for machine learning, vol. 16, pp. 21–30. Morgan Kaufmann Publishers Inc., San Francisco (1993)Google Scholar
  7. 7.
    Quinlan, J.R.: Improved use of continuous attributes inc 4.5. J. Artif. Intell. Res. 4, 77–90 (1996)MATHGoogle Scholar
  8. 8.
    de Vries, A.P., Roelleke, T.: Relevance information: a loss of entropy but a gain for idf? In: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 282–289 (2005)Google Scholar
  9. 9.
    Ruggieri, S.: Efficient C4.5. IEEE Trans. Knowl. Data Eng. 2, 438–444 (2002)Google Scholar
  10. 10.
    Kotsiantis, S.: A hybrid decision tree classifier. J. Intell. Fuzzy Syst. 1, 327–336 (2014)Google Scholar
  11. 11.
    Fern, A., Givan, R.: Online ensemble learning: an empirical study. Mach. Learn 53, 279–286 (2000) (Morgan Kaufmann Publishers Inc., San Francisco)Google Scholar
  12. 12.
    Quinlan, J.R.: C4.5: Programs for machine learning, vol. 16, pp. 235–240. Morgan Kaufmann Publishers Inc., San Francisco, CA (1993)Google Scholar
  13. 13.
    Bache, K., Lichman, M.: UCI machine learning repository. Digit. Libr. 1 (2013)Google Scholar
  14. 14.
    Abele, J., Mantas, C.J.: Improving experimental studies about ensembles of classifiers for bankruptcy prediction and credit scoring. Expert Syst. Appl. Int. J. 41, 3825–3830. Pergamon Press Inc., USA (2014)Google Scholar
  15. 15.
    Olson, D.L., Delen, D., Meng, Y.: Comparative analysis of data mining methods for bankruptcy prediction. Decis. Support Syst. 2, 464–473 (2012)Google Scholar
  16. 16.
    Grzymala-Busse, J.W.: Rule induction. In: Maimon, O.Z., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook, vol. 6, pp. 277–294. Springer, Heidelberg (2005)Google Scholar
  17. 17.
    Maimon, O., Rokach, L.: The Data Mining and Knowledge Discovery Handbook, vol. 2, pp. 1–17. Springer, Heidelerg (2005)Google Scholar
  18. 18.
    Podgorelec, V., Kokol, P., Stiglic, B., Rozman, I.: Decision trees: an overview and their use in medicine. J. Med. Syst. 5, 445–463 (2002)Google Scholar
  19. 19.
    Povalej, P., Kokol, P.: End user friendly data mining with decision trees: a reality or a wish? In: Proceedings of the 2007 Annual Conference on International Conference on Computer Engineering and Applications, vol. 3, pp. 35–40. Stevens Point, Wisconsin (2007)Google Scholar

Copyright information

© Springer India 2015

Authors and Affiliations

  1. 1.Amrita Center for Cyber Security Systems and NetworksAmrita UniversityKollamIndia

Personalised recommendations