Abstract
Machine learning enables the automation of the system to generate results without direct assistance from the environment once the machine is trained for all possible scenarios. This is achieved by a series of processes such as collecting relevant data in raw format, exploratory data analysis, selection and implementation of required models, evaluation of those models, and so forth. The initial stage of the entire pipeline involves the necessary task of feature selection. The feature selection process includes extracting more informative features from the pool of input attributes to enhance the predictions made by machine learning models. The proposed approach focuses on the traditional feature selection algorithms and bio-inspired modified Ant Colony Optimization (ACO) algorithm to remove redundant and irrelevant features. In addition, the proposed methodology provides a comparative analysis of their performances. The results show that the modified ACO computed fewer error percentages in the Linear Regression Model of the dataset. In contrast, the traditional methods used outperformed the modified ACO in the SVR model.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Dorigo, M., Birattari, M.: Ant colony optimization. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning. Springer, Boston (2011). https://doi.org/10.1007/978-0-387-30164-8_22
Xue, B., Zhang, M., Browne, W.N.: Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans. Cybern. 43(6), 1656–1671 (2013). https://doi.org/10.1109/TSMCB.2012.2227469
Liu, Y., Wang, G., Chen, H., Dong, H., Zhu, X., Wang, S.: An improved particle swarm optimization for feature selection. J. Bionic Eng. 8(2), 191–200 (2011)
Ghaemi, M., Feizi-Derakhshi, M.-R.: Feature selection using Forest Optimization Algorithm. Pattern Recogn. 60, 121–129 (2016)
Moradi, P., Rostami, M.: Integration of graph clustering with ant colony optimization for feature selection. Knowl.-Based Syst. 84, 144–161 (2015)
Daelemans, W., Hoste, Véronique., De Meulder, F., Naudts, B.: Combined optimization of feature selection and algorithm parameters in machine learning of language. In: Lavrač, N., Gamberger, D., Blockeel, H., Todorovski, L. (eds.) ECML 2003. LNCS (LNAI), vol. 2837, pp. 84–95. Springer, Heidelberg (2003). https://doi.org/10.1007/978-3-540-39857-8_10
Emary, E., Zawbaa, H.M., Ghany, K.K.A., Hassanien, A.E., Parv, B.: Firefly optimization algorithm for feature selection. In: Proceedings of the 7th Balkan Conference on Informatics Conference (BCI 2015). Association for Computing Machinery, New York (2015). Article 26, 1–7. https://doi.org/10.1145/2801081.2801091
Zhang, L., Mistry, K., Lim, C.P., Neoh, S.C.: Feature selection using firefly optimization for classification and regression models. Decis. Support Syst. 106 (2018)
Yan, Z., Yuan, C.: Ant colony optimization for feature selection in face recognition. In: Zhang, D., Jain, A.K. (eds.) ICBA 2004. LNCS, vol. 3072, pp. 221–226. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-25948-0_31
https://www.google.com/url?q=https://www.kaggle.com/kumarajarshi/life-expectancywho&sa=D&source=hangouts&usg=1623059761384000&usg=AFQjCNFet3_t0XG6MXcGx2-gPY7mXgX9wQ
Pearson’s correlation coefficient. In: Kirch, W. (eds.) Encyclopedia of Public Health. Springer, Dordrecht (2008). https://doi.org/10.1007/978-1-4020-5614-7_2569
Ross, B.C.: Mutual information between discrete and continuous data sets. PLoS ONE 9(2), e87357 (2014). https://doi.org/10.1371/journal.pone.0087357
McHugh, M.L.: The chi-square test of independence. Biochem. Med. (Zagreb) 23(2), 143–149 (2013). https://doi.org/10.11613/bm.2013.018
Dorigo, M.: Ant colony optimization. Scholarpedia 2(3), 1461 (2007). revision #90969
Obson, J.D.: Multiple linear regression. In: Applied Multivariate Data Analysis. STS. Springer, New York (1991). https://doi.org/10.1007/978-1-4612-0955-3_4
Basak, D., Pal, S., Patranabis, D.: Support vector regression. Neural Inf. Process. – Lett. Rev. 11 (2007)
Mean absolute error. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning and Data Mining. Springer, Boston (2017). https://doi.org/10.1007/978-1-4899-7687-1_953
Mean squared error. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning. Springer, Boston (2011). https://doi.org/10.1007/978-0-387-30164-8_528
Shekhar, S., Xiong, H.: Root-mean-square error. In: Shekhar, S., Xiong, H. (eds.) Encyclopedia of GIS. Springer, Boston (2008). https://doi.org/10.1007/978-0-387-35973-1_1142
Refaeilzadeh, P., Tang, L., Liu, H.: Cross-validation. In: Liu, L., Özsu, M.T. (eds.) Encyclopedia of Database Systems. Springer, Boston (2009). https://doi.org/10.1007/978-0-387-39940-9_565
Rathee, N., Joshi, N., Kaur, J.: Sentiment analysis using machine learning techniques on Python. In: 2018 Second International conference on Intelligent Computing and Control Systems (ICICCS), pp. 779–785 (2018). https://doi.org/10.1109/ICCONS.2018.8663224
Kumar, V., Rathee, N.: Knowledge discovery from database using an integration of clustering and classification. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 2(3) (2011)
Ankita, N.S.: Improved link prediction using PCA. Int. J. Anal. Appl. 17(4), 578–585 (2019)
Rathee, N., Chhillar, R.S.: Generation and optimization of test paths using modified ACO. Int. J. Control Theory Appl. (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Singhal, S., Sharma, R., Malhotra, N., Rathee, N. (2022). Comparative Analysis of Traditional and Optimization Algorithms for Feature Selection. In: Dev, A., Agrawal, S.S., Sharma, A. (eds) Artificial Intelligence and Speech Technology. AIST 2021. Communications in Computer and Information Science, vol 1546. Springer, Cham. https://doi.org/10.1007/978-3-030-95711-7_35
Download citation
DOI: https://doi.org/10.1007/978-3-030-95711-7_35
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-95710-0
Online ISBN: 978-3-030-95711-7
eBook Packages: Computer ScienceComputer Science (R0)