Skip to main content

Comparative Analysis of Traditional and Optimization Algorithms for Feature Selection

  • Conference paper
  • First Online:
Artificial Intelligence and Speech Technology (AIST 2021)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1546))

  • 972 Accesses

Abstract

Machine learning enables the automation of the system to generate results without direct assistance from the environment once the machine is trained for all possible scenarios. This is achieved by a series of processes such as collecting relevant data in raw format, exploratory data analysis, selection and implementation of required models, evaluation of those models, and so forth. The initial stage of the entire pipeline involves the necessary task of feature selection. The feature selection process includes extracting more informative features from the pool of input attributes to enhance the predictions made by machine learning models. The proposed approach focuses on the traditional feature selection algorithms and bio-inspired modified Ant Colony Optimization (ACO) algorithm to remove redundant and irrelevant features. In addition, the proposed methodology provides a comparative analysis of their performances. The results show that the modified ACO computed fewer error percentages in the Linear Regression Model of the dataset. In contrast, the traditional methods used outperformed the modified ACO in the SVR model.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Dorigo, M., Birattari, M.: Ant colony optimization. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning. Springer, Boston (2011). https://doi.org/10.1007/978-0-387-30164-8_22

  2. Xue, B., Zhang, M., Browne, W.N.: Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans. Cybern. 43(6), 1656–1671 (2013). https://doi.org/10.1109/TSMCB.2012.2227469

    Article  Google Scholar 

  3. Liu, Y., Wang, G., Chen, H., Dong, H., Zhu, X., Wang, S.: An improved particle swarm optimization for feature selection. J. Bionic Eng. 8(2), 191–200 (2011)

    Article  Google Scholar 

  4. Ghaemi, M., Feizi-Derakhshi, M.-R.: Feature selection using Forest Optimization Algorithm. Pattern Recogn. 60, 121–129 (2016)

    Article  Google Scholar 

  5. Moradi, P., Rostami, M.: Integration of graph clustering with ant colony optimization for feature selection. Knowl.-Based Syst. 84, 144–161 (2015)

    Google Scholar 

  6. Daelemans, W., Hoste, Véronique., De Meulder, F., Naudts, B.: Combined optimization of feature selection and algorithm parameters in machine learning of language. In: Lavrač, N., Gamberger, D., Blockeel, H., Todorovski, L. (eds.) ECML 2003. LNCS (LNAI), vol. 2837, pp. 84–95. Springer, Heidelberg (2003). https://doi.org/10.1007/978-3-540-39857-8_10

    Chapter  MATH  Google Scholar 

  7. Emary, E., Zawbaa, H.M., Ghany, K.K.A., Hassanien, A.E., Parv, B.: Firefly optimization algorithm for feature selection. In: Proceedings of the 7th Balkan Conference on Informatics Conference (BCI 2015). Association for Computing Machinery, New York (2015). Article 26, 1–7. https://doi.org/10.1145/2801081.2801091

  8. Zhang, L., Mistry, K., Lim, C.P., Neoh, S.C.: Feature selection using firefly optimization for classification and regression models. Decis. Support Syst. 106 (2018)

    Google Scholar 

  9. Yan, Z., Yuan, C.: Ant colony optimization for feature selection in face recognition. In: Zhang, D., Jain, A.K. (eds.) ICBA 2004. LNCS, vol. 3072, pp. 221–226. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-25948-0_31

    Chapter  Google Scholar 

  10. https://www.kaggle.com/uciml/student-alcohol-consumption

  11. https://www.google.com/url?q=https://www.kaggle.com/kumarajarshi/life-expectancywho&sa=D&source=hangouts&usg=1623059761384000&usg=AFQjCNFet3_t0XG6MXcGx2-gPY7mXgX9wQ

    Google Scholar 

  12. Pearson’s correlation coefficient. In: Kirch, W. (eds.) Encyclopedia of Public Health. Springer, Dordrecht (2008). https://doi.org/10.1007/978-1-4020-5614-7_2569

  13. Ross, B.C.: Mutual information between discrete and continuous data sets. PLoS ONE 9(2), e87357 (2014). https://doi.org/10.1371/journal.pone.0087357

    Article  Google Scholar 

  14. McHugh, M.L.: The chi-square test of independence. Biochem. Med. (Zagreb) 23(2), 143–149 (2013). https://doi.org/10.11613/bm.2013.018

    Article  Google Scholar 

  15. Dorigo, M.: Ant colony optimization. Scholarpedia 2(3), 1461 (2007). revision #90969

    Google Scholar 

  16. Obson, J.D.: Multiple linear regression. In: Applied Multivariate Data Analysis. STS. Springer, New York (1991). https://doi.org/10.1007/978-1-4612-0955-3_4

  17. Basak, D., Pal, S., Patranabis, D.: Support vector regression. Neural Inf. Process. – Lett. Rev. 11 (2007)

    Google Scholar 

  18. Mean absolute error. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning and Data Mining. Springer, Boston (2017). https://doi.org/10.1007/978-1-4899-7687-1_953

  19. Mean squared error. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning. Springer, Boston (2011). https://doi.org/10.1007/978-0-387-30164-8_528

  20. Shekhar, S., Xiong, H.: Root-mean-square error. In: Shekhar, S., Xiong, H. (eds.) Encyclopedia of GIS. Springer, Boston (2008). https://doi.org/10.1007/978-0-387-35973-1_1142

  21. Refaeilzadeh, P., Tang, L., Liu, H.: Cross-validation. In: Liu, L., Özsu, M.T. (eds.) Encyclopedia of Database Systems. Springer, Boston (2009). https://doi.org/10.1007/978-0-387-39940-9_565

  22. Rathee, N., Joshi, N., Kaur, J.: Sentiment analysis using machine learning techniques on Python. In: 2018 Second International conference on Intelligent Computing and Control Systems (ICICCS), pp. 779–785 (2018). https://doi.org/10.1109/ICCONS.2018.8663224

  23. Kumar, V., Rathee, N.: Knowledge discovery from database using an integration of clustering and classification. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 2(3) (2011)

    Google Scholar 

  24. Ankita, N.S.: Improved link prediction using PCA. Int. J. Anal. Appl. 17(4), 578–585 (2019)

    MATH  Google Scholar 

  25. Rathee, N., Chhillar, R.S.: Generation and optimization of test paths using modified ACO. Int. J. Control Theory Appl. (2017)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sakshi Singhal .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Singhal, S., Sharma, R., Malhotra, N., Rathee, N. (2022). Comparative Analysis of Traditional and Optimization Algorithms for Feature Selection. In: Dev, A., Agrawal, S.S., Sharma, A. (eds) Artificial Intelligence and Speech Technology. AIST 2021. Communications in Computer and Information Science, vol 1546. Springer, Cham. https://doi.org/10.1007/978-3-030-95711-7_35

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-95711-7_35

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-95710-0

  • Online ISBN: 978-3-030-95711-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics