Advertisement

Applying Attribute Selection Algorithms in Academic Performance Prediction

  • Mu Lin WongEmail author
  • S. Senthil
Conference paper
Part of the Lecture Notes on Data Engineering and Communications Technologies book series (LNDECT, volume 26)

Abstract

There is an essential need to identify effective and efficient attribute selection algorithms for developing predictive and descriptive classifiers to predict students’ academic performance. Personal, Academic and Socio-economic data were collected from 100 MCA students through questionnaire and cleansed. Eight attribute selection algorithms available in WEKA were applied to identify the best attributes. These attribute sets were then employed by four rule-based classifiers and four tree-based classifiers to examine the rate of accuracy, recall and AUC. The results proved that dimension reduction improves classifier performance by a maximum margin of 9% in accuracy, 25% in recall and 45% in AUC, when attributes were selected by subset-based algorithms. Both academic and socio-economic data are required to develop effective classifiers.

Keywords

Educational data mining Academic performance prediction Attribute selection algorithm Rule-based classifiers Tree-based classifiers Historical data 

References

  1. 1.
    Romero, C., Ventura, S.: Educational data mining: a survey from 1995 to 2005. Expert Syst. Appl. 33, 135–146 (2007)CrossRefGoogle Scholar
  2. 2.
    Romero, C., Ventura, S.: Educational data mining: a review of the state of the art. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 40(6), 601–618 (2010)CrossRefGoogle Scholar
  3. 3.
    Pena-Ayala, A.: Educational data mining: a survey and a data mining-based analysis of recent works. Expert Syst. Appl. 41, 1432–1462 (2014)CrossRefGoogle Scholar
  4. 4.
    Ganesh, S.H., Christy, A.J.: Applications of educational data mining: a survey. In: 2nd IEEE International Conference on Innovations in Information Embedded and Communication Systems, pp. 1–6. IEEE Xplore (2015)Google Scholar
  5. 5.
    Sukhija, K., Jindal, M., Aggarwal, N.: The recent state of educational data mining: a survey and future visions. In: 3rd IEEE International Conference on MOOCs, Innovation and Technology in Education, pp. 354–359. IEEE Xplore (2015)Google Scholar
  6. 6.
    Anoopkumar, M., Rahman, A.M.J.M.Z.: A review on data mining techniques and factors used in educational data mining to predict student amelioration. In: International Conference on Data Mining and Advanced Computing, pp. 122–133. IEEE Xplore (2016)Google Scholar
  7. 7.
    Lei, C., Li, K.F.: Academic performance predictors. In 29th International Conference on Advanced Information Networking and Applications Workshops, pp. 577–581. IEEE Xplore (2015)Google Scholar
  8. 8.
    Vuttipittayamongkol, P.: Predicting factors of academic performance. In: 2nd Asian Conference on Defence Technology, pp. 161–166. IEEE Xplore (2016)Google Scholar
  9. 9.
    Hasbun, T., Araya, A., Villalon, J.: Extracurricular activities as dropout prediction factors in higher education using decision trees. In: 16th IEEE International Conference on Advanced Learning Technologies, pp. 242–244. IEEE Xplore (2016)Google Scholar
  10. 10.
    Conijn, R., Snijders, C., Kleingeld, A., Matzat, U.: Predicting student performance from LMS data: a comparison of 17 blended courses using moodle LMS. IEEE Trans. Learn. Technol. 10(1), 17–29 (2017)CrossRefGoogle Scholar
  11. 11.
    Pardo, A., Han, F., Ellis, R.A.: Combining university student self-regulated learning indicators and engagement with online learning events to predict academic performance. IEEE Trans. Learn. Technol. 10(1), 82–92 (2017)CrossRefGoogle Scholar
  12. 12.
    Cano, A., Zafra, A., Ventura, S.: An ınterpretable classification rule mining algorithm. Inf. Sci. 240, 1–20 (2013)CrossRefGoogle Scholar
  13. 13.
    Huysmans, J., Dejaeger, K., Mues, C., Vanthienen, J., Baesens, B.: An empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. Decis. Support Syst. 51, 141–154 (2011)CrossRefGoogle Scholar
  14. 14.
    Cohen, W..: Fast effective rule induction. In: 12th International Conference on Machine Learning, pp. 115–123 (1995)CrossRefGoogle Scholar
  15. 15.
    Martin, B.: Instance-Based learning: Nearest Neighbor with Generalization. Hamilton, New Zealand (1995)Google Scholar
  16. 16.
    Roy, S.: Nearest Neighbor with Generalization. Christchurch, New Zealand (2002)Google Scholar
  17. 17.
    Frank, E., Witten, I.H.: Generating accurate rule sets without global optimization. In: 15th International Conference on Machine Learning, pp. 144–151 (1998)Google Scholar
  18. 18.
    Gaines, B.R., Compton, P.: Induction of ripple-down rules applied to modeling large databases. J. Intell. Inf. Syst. 5(3), 211–228 (1995)CrossRefGoogle Scholar
  19. 19.
    Freund, Y., Mason, L.: The alternating decision tree learning algorithm. In: 16th International Conference on Machine Learning, pp. 124–133. Bled, Slovenia (1999)Google Scholar
  20. 20.
    Witten, I.H., Frank, E., Hall, M.: Data Mining: Practical Machine Learning Tools and Techniques, 3rd edn. Morgan Kaufmann, Burlington (2011)Google Scholar
  21. 21.
    Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Francisco (1993)Google Scholar
  22. 22.
    Holmes, G., Pfahringer, B., Kirkby, R., Frank, E., Hall, M.: Multiclass alternating decision trees. In: ECML, pp. 161–172 (2001)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.School of Computer Science and ApplicationsREVA UniversityBengaluruIndia

Personalised recommendations