Skip to main content
Log in

Learning the transfer function in binary metaheuristic algorithm for feature selection in classification problems

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

One of the most challenging issues in pattern recognition is the data attribution selection process. Feature selection plays a key role in solving problems with high-dimensional data and is a fundamental step in pre-processing many classifications and machine learning problems. The feature selection method reduces the amount of data and increases the category precision. Unrelated data, which can lead to inappropriate classification, is thus removed to obtain fewer features. In this paper, the Binary Gray Wolf Optimization algorithm uses the Wrapper method for feature selection. The transfer function is an essential part of BGWO to map a continuous value to a binary value. In this study, eight transfer functions are divided into two families: S-shaped and V-shaped. Previous research has used only one transfer function for the whole algorithm, and all wolves in the whole algorithm deal with this transfer function. In this paper, each wolf has its own transfer function. Because algorithms are evolutionary meta-innovations and can optimize themselves, each wolf can play a role in the whole algorithm at any stage while optimizing itself and adapting to its community, and not just depend on one transfer function. In the proposed method, eight transfer functions are divided into two families, S-shaped and V-shaped. This article proposes two approaches for learning the transfer function, by selecting the transfer function and the slope of these functions. In the first approach, we add three or two binary bits to the initial population. If two bits are added, four modes of the transfer function are available, and if three bits are added, eight transfer functions are achievable. These bits are used as a criterion for selecting a predefined transfer function for each wolf. So, in the proposed method, each wolf has its transfer function. During the implementation of the algorithm, the wolves update their position according to the evaluation function and learning. In the second approach, ten or twenty-one binary bits are added to the initial population. If ten binary bits are used, we will have a transfer function, and \(2^{10}\) coefficient modes are available for the slope of the transfer function. If twenty-one binary bits are used, we have two transfer functions available. So, there are \(2^{10}\) modes for the gradient of the transfer function. These bits are used as a criterion for selecting the transfer function and the coefficient affecting the slope of these functions. In both ideas, after each iteration of the algorithm, the position of the wolves is updated and based on the evaluation function, the alpha wolf is identified and the transfer function is selected. With subsequent iterations, the algorithm learns and optimizes the transfer function to achieve the best feature selection with the smallest error. Experimental results on ten UCI datasets show that selecting the obtained feature subset with high classification accuracy is efficient.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Ziemba P, Piwowarski M (2011) Feature selection methods in data mining techniques, Prace Naukowe Uniwersytetu Ekonomicznego we Wrocławiu, 213–223

  2. Chen R-C, Dewi C, Huang S-W, Caraka RE (2020) Selecting critical features for data classification based on machine learning methods. J Big Data 7:1–6

    Article  Google Scholar 

  3. Zawbaa HM, Emary E, Grosan C, Snasel V (2018) Large-dimensionality small-instance set feature selection: a hybrid bio-inspired heuristic approach. Swarm Evolution Comput 42:29–42

    Article  Google Scholar 

  4. Zamani H, Nadimi-Shahraki M-H (2016) Feature selection based on whale optimization algorithm for diseases diagnosis. Int J Comput Sci Inform Security 14(9):1243

    Google Scholar 

  5. Hsu H-H, Hsieh C-W, Lu M-D (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38:8144–8150

    Article  Google Scholar 

  6. Gangavarapu T, Patil N (2019) A novel filter–wrapper hybrid greedy ensemble approach optimized using the genetic algorithm to reduce the dimensionality of high-dimensional biomedical datasets. Applied Soft Computing 81:105538

    Article  Google Scholar 

  7. Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: a mutual information-based feature selection method. Expert Syst Appl 41(14):6371–6385

    Article  Google Scholar 

  8. Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43(6):1656–1671

    Article  Google Scholar 

  9. Zhang R, Nie F, Li X, Wei X (2019) Feature selection with multiview data: a survey. Inform Fusion 50:158–167

    Article  Google Scholar 

  10. Sánchez-Maroño N, Alonso-Betanzos A, Tombilla-Sanromán M (2007) Filter methods for feature selection–a comparative study. In: Intelligent data engineering and automated learning–ideal, pp 178–187

  11. Nnamoko NA, Arshad FN, England D, Vora J, Norman J (2014) Evaluation of filter and wrapper methods for feature selection in supervised machine learning. In: Conference: the 15th annual postgraduate symposium on the convergence of telecommunication, networking and broadcasting at: liverpool

  12. Cai J, Luo J, Wang S, Yang S (2018) Feature selection in machine learning: a new perspective. Neurocomputing 300:70–79

    Article  Google Scholar 

  13. Venkatesh B, Anuradha J (2019) A review of feature selection and its methods. Cybern Inform Technol 19(1):3–26

    MathSciNet  Google Scholar 

  14. Wu Y, Liu Y, Wang Y, Shi Y, Zhao X (2018) JCDSA: a joint covariate detection tool for survival analysis on tumor expression profiles. BMC Bioinform 19(1):1–8

    Article  Google Scholar 

  15. Yang R, Zhang C, Zhang L, Gao R (2018) A two-step feature selection method to predict Cancerlectins by Multiview features and synthetic minority oversampling technique. BioMed Res Int. https://doi.org/10.1155/2018/9364182

    Article  Google Scholar 

  16. Masoudi Sobhanzadeh Y, Motieghader H, Masoudi-Nejad A (2019) Feature Select: a software for feature selection based on machine learning approaches. BMC Bioinform 20(170):1–17

    Google Scholar 

  17. Metin SK (2018) Feature selection in multiword expression recognition. Expert Syst Appl 92(C):106–123

    Article  Google Scholar 

  18. Saxena AK, Dubey VK, Wang J (2017) Hybrid feature selection methods for high-dimensional multi-class datasets. Int J Data Min Modell Manag 9(4):315

    Google Scholar 

  19. Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A (2012) An ensemble of filters and classifiers for microarray data classification. Pattern Recogn 45(1):531–539

    Article  Google Scholar 

  20. Hoque N, Singh M, Bhattacharyya DK (2018) EFS-MI: an ensemble feature selection method for classification. Complex Intell Syst 4:105–118

    Article  Google Scholar 

  21. Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A, Benitez JM, Herrera F (2014) A review of microarray datasets and applied feature selection methods. Inf Sci 282:111–135

    Article  Google Scholar 

  22. Agrawal P, Abutarboush HF, Ganesh T, Mohamed AW (2021) Metaheuristic Algorithms on feature selection: A survey of one decade of research (2009–2019). IEEE Access 9:26766–26791

    Article  Google Scholar 

  23. Jiang Y, Luo Q, Wei Y, Abualigah L, Zhou Y (2021) An efficient binary Gradient-based optimizer for feature selection. Math Biosci Eng 18(4):3813–3854

    Article  MATH  Google Scholar 

  24. Emary E, Zawbaa HM, Hassanien AE (2016) Binary grey wolf optimization approaches for feature selection. Neuro Comput 172:371–381

    Google Scholar 

  25. Mirjalili S, Lewis A (2013) S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol Comput 9:114

    Article  Google Scholar 

  26. Mirjalili S, Yang X-S, Mirjalili SM (2014) Binary bat algorithm. Neurl Comput Appl 25(3):663–681

    Article  Google Scholar 

  27. Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61

    Article  Google Scholar 

  28. Zhang J, Hong L, Liu Q (2021) An improved whale optimization algorithm for the traveling salesman problem. Symmetry 13(1):48

    Article  Google Scholar 

  29. Hussien AG, Houssein EH, Hassanien AE (2017) A binary whale optimization algorithm with hyperbolic tangent fitness function for feature selection. In: Eighth international conference on intelligent computing and information systems (ICICIS)

  30. Hu P, Pan JS, Chu SC (2020) Improved Binary Grey Wolf Optimizer and Its application for a feature selection. Knowledg Based Syst 195:105746

    Article  Google Scholar 

  31. Altman NS (1992) An introduction to kernel and nearest-neighbor nonpara-metric regression. Am Statist 46(3):175–185

    Google Scholar 

  32. Ghosh M et al (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176

    Article  Google Scholar 

  33. Chellammal S, Sharmila R (2019) Recommendation of attributes for heart disease prediction using correlation measure. Int J Recent Technol Eng (IJRTE) 8(23):870–875

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Contributions

Zahra Nassiri: Programmer Validation, Visualization, Investigation, Writing—original draft. Hesam Omranpour: Supervision, Project administration, Conceptualization, Methodology, Writing—reviewing and editing.

Corresponding author

Correspondence to Hesam Omranpour.

Ethics declarations

Conflicts of interest

The authors declare that they have no known competing reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Nassiri, Z., Omranpour, H. Learning the transfer function in binary metaheuristic algorithm for feature selection in classification problems. Neural Comput & Applic 35, 1915–1929 (2023). https://doi.org/10.1007/s00521-022-07869-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07869-z

Keywords

Navigation