Skip to main content
Log in

MLFS-CCDE: multi-objective large-scale feature selection by cooperative coevolutionary differential evolution

  • Regular Research Paper
  • Published:
Memetic Computing Aims and scope Submit manuscript

Abstract

Feature selection is a pre-processing procedure of choosing the optimal feature subsets for constructing model, yet it is difficult to satisfy the requirements of reducing number of features and maintaining classification accuracy. Towards this problem, we propose novel multi-objectives large-scale cooperative coevolutionary algorithm for three-objectives feature selection, termed MLFS-CCDE. Firstly, a cooperative searching framework is designed for efficiently and effectively seeking for the optimal feature subset. Secondly, in the framework, three objectives, feature’s number, classification accuracy and total information gain are established for guiding the evolution of features’ combination. Thirdly, in framework’s decomposition process, cluster-based decomposition strategy is elaborated for reducing the computation; in framework’s coevolution process, dual indicator-based representatives are elaborated for balancing the representative solution’ convergence and diversity. Finally, to verify framework’s practicability, a heart disease diagnosis system based on MLFS-CCDE framework is constructed in cardiology. Numerical experiments demonstrate that the proposed MLFS-CCDE outperforms its competitors in terms of both classification accuracy and metrics of features’ number.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

References

  1. Benteng M, Xia Y (2017) A tribe competition-based genetic algorithm for feature selection in pattern classification. Appl Soft Comput 58:328–338

    Google Scholar 

  2. Otero FEB, Freitas AA, Johnson CG (2010) A hierarchical multi-label classification ant colony algorithm for protein function prediction. Memetic Comput 2(3):165–181

    Google Scholar 

  3. Yong J, He F, Li H, Zhou W (2019) A novel bat algorithm based on cross boundary learning and uniform explosion strategy. Appl Math J Chin Univ 34(4):482–504

    MathSciNet  MATH  Google Scholar 

  4. Kale A, Sonavane S (2017) Hybrid feature subset selection approach for fuzzy-extreme learning machine. Data-Enabled Discov Appl 1(1):10

    Google Scholar 

  5. Hoai NB, Bing X, Peter A, Hisao I, Mengjie Z (2020) Multiple reference points-based decomposition for multi-objective feature selection in classification: static and dynamic mechanisms. IEEE Trans Evol Comput 24(1):170–184

    Google Scholar 

  6. Bach NH, Xue B, Mengjie Z (2016) A subset similarity guided method for multi-objective feature selection. Lect Notes Artif Intell 9592:298–310

    Google Scholar 

  7. Dong H, Sun J, Li T, Ding R, Sun X (2020) A multi-objective algorithm for multi-label filter feature selection problem. Appl Intell (7)

  8. Monika R, Yun TG, Jonathon C (2020) Multi-objective-based feature selection for ddos attack detection in iot networks. Iet Netw 9(3):120–127

    Google Scholar 

  9. Majid K, Wael L, Narges A, Dimitrios H (2018) Feature selection for nonstationary data: application to human recognition using medical biometrics. IEEE Trans Cybern 48(5):1446–1459

    Google Scholar 

  10. Quqn Q, He F, Li H (2020) A multi-phase blending method with incremental intensity for training detection networks. Vis Comput 34(1):1–26

    Google Scholar 

  11. Komeili M, Louis W, Armanfard N, Hatzinakos D (2017) Feature selection for nonstationary data: Application to human recognition using medical biometrics. IEEE Trans Cybern 1–14

  12. Luo J, He F, Li H, Liang Y (2020) A novel whale optimization algorithm with filtering disturbance and non-linear step. Int J Bio-Inspir Comput 16:137–148

    Google Scholar 

  13. He CL, Zhang Y, Gong DW, Wu B (2020) Multi-objective feature selection based on artificial bee colony for hyperspectral images

  14. Tian D (2016) A multi-objective genetic local search algorithm for optimal feature subset selection. In: 2016 International conference on computational science and computational intelligence (CSCI) (pp 1089–1094). IEEE

  15. Gu S, ChengR, Jin Y (2016) Feature selection for high-dimensional classification using a competitive swarm optimizer. Soft Comput 1–12

  16. Ishibuchi H, Tsukamoto N, Nojima T (2008) Evolutionary many-objective optimization: a short review. In: 2008 IEEE congress on evolutionary computation (IEEE world congress on computational intelligence) (pp 2419–2426). IEEE

  17. Huang B, Buckley B, Kechadi TM (2010) Multi-objective feature selection by using nsga-ii for customer churn prediction in telecommunications. Expert Syst Appl 37(5):3638–3646

    Google Scholar 

  18. Zawbaa Hossam M, Eid E, Crina G, Vaclav S (2018) Large-dimensionality small-instance set feature selection: a hybrid bio-inspired heuristic approach. Swarm Evolut Comput 42:29–42

    Google Scholar 

  19. Li AD, Bing XB, Zhang M (2020) Multi-objective feature selection using hybridization of a genetic algorithm and direct multisearch for key quality characteristic selection. Inf Sci 523:245–265

    MathSciNet  Google Scholar 

  20. Guo X, Wang X, Wang M, Wang Y (2012) A new objective reduction algorithm for many-objective problems: employing mutual information and clustering algorithm. In: 2012 Eighth international conference on computational intelligence and security (pp 11–16). IEEE

  21. Takfarinas S, David B, Goetz B, Anthony V (2018) Is seeding a good strategy in multi-objective feature selection when feature models evolve? Inf Softw Technol 95(3):266–280

    Google Scholar 

  22. Shenkai G, Ran C, Yaochu J (2018) Feature selection for high-dimensional classification using a competitive swarm optimizer. Soft Comput 22(3):811–822

    Google Scholar 

  23. Binder M, MoosbauerJ, Thomas J, Bischl B (2020) Multi-objective hyperparameter tuning and feature selection using filter ensembles. In: GECCO ’20: genetic and evolutionary computation conference

  24. Konrad D (2011) Performance evaluation of artificial bee colony optimization and new selection schemes. Memetic Comput 3(3):149–162

    Google Scholar 

  25. Huang CL, Wang CJ (2006) A ga-based feature selection and parameters optimization for support vector machines. Expert Syst Appl 31(2):231–240

    Google Scholar 

  26. Lin SW, Tseng TY, Chen SC, Huang JF (2006) A sa-based feature selection and parameter optimization approach for support vector machine. In: 2006 IEEE international conference on systems, man and cybernetics (vol 4, pp 3144–3145). IEEE

  27. Cheng Lung Huang (2009) Aco-based hybrid classification system with feature subset selection and model parameters optimization. Neurocomputing 73(1):438–448

    Google Scholar 

  28. Linqiang P, Cheng H, Ye T, Handing W, Xingyi Z, Yaochu J (2019) A classification-based surrogate-assisted evolutionary algorithm for expensive many-objective optimization. IEEE Trans Evol Comput 23(1):74–88

    Google Scholar 

  29. Ferrante N, Ville T (2009) Scale factor local search in differential evolution. Memetic Comput 1(2):153–171

    Google Scholar 

  30. Li K, He F, Yu H, Chen X (2019) A parallel and robust object tracking approach synthesizing adaptive Bayesian learning and improved incremental subspace learning. Front Comput Sci 13(5):1116–1135

    Google Scholar 

  31. Bach NH, Xue B, Peter A (2018) Pso with surrogate models for feature selection: static and dynamic clustering-based methods. Memetic Comput 5:1–10

    Google Scholar 

  32. Yudong Z, Shuihua W, Preetha P, Genlin J (2014) Binary pso with mutation operator for feature selection using decision tree applied to spam detection. Knowl-Based Syst 64(1):22–31

    Google Scholar 

  33. Yeh CL, Wei CH, Chung Jui T, Hong YC (2008) Improved binary pso for feature selection using gene expression data. Comput Biol Chem 32(1):29–38

    MATH  Google Scholar 

  34. Potter MA, De Jong KA (2014) Cooperative coevolution: an architecture for evolving coadapted subcomponents. Evol Comput 8(1):1–29

    Google Scholar 

  35. Antonio LM, Coello Coello CA (2016) Indicator-based cooperative coevolution for multi-objective optimization. In: Evolutionary computation

  36. Antonio LM, Coello Coello CA (2018) Coevolutionary multi-objective evolutionary algorithms: a survey of the state-of-the-art. IEEE Trans Evolut Comput 22(6):851–865

    Google Scholar 

  37. Hammami M, Bechikh S, Hung CC, Said LB (2018) A multi-objective hybrid filter-wrapper evolutionary approach for feature selection. Memetic Comput 1–16

  38. Huimin Z, Sinha Atish P, Ge W (2009) Effects of feature construction on classification performance: an empirical study in bank failure prediction. Expert Syst Appl 36(2):2633–2644

    Google Scholar 

  39. Lu D, Weng Q (2007) A survey of image classification methods and techniques for improving classification performance. Int J Remote Sens 28(5):823–870

    Google Scholar 

  40. Adler J, Parmryd I (2010) Quantifying colocalization by correlation: the Pearson correlation coefficient is superior to the Mander’s overlap coefficient. Cytometry Part A, 77a(8):733–742

  41. Hanchuan P, Fuhui L, Chris D (2005) Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238

    Google Scholar 

  42. Karakaya G, Galelli S, Ahipasaoglu SD, Taormina R (2016) Identifying (quasi) equally informative subsets in feature selection problems for classification: a max-relevance min-redundancy approach. IEEE Trans Cybern 46(6):1424–1437

    Google Scholar 

  43. Dua D, Graff C (2017) UCI machine learning repository

  44. Wang Z, Qu L, Xin J, Yang H, Gao X (2018) A unified distributed elm framework with supervised, semi-supervised and unsupervised big data learning. Memetic Comput 1–11

  45. Bin HG, Zhu Qin Yu, Kheong SC (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1):489–501

    Google Scholar 

  46. Hamdani TM, Won J-M, Alimi AM, Karray F (2007) Multi-objective feature selection with nsga ii. In: International conference on adaptive and natural computing algorithms (pp 240–247). Springer

  47. Das SK, Mohanty R, Mohanty M, Mahamaya M (2020) Multi-objective feature selection (mofs) algorithms for prediction of liquefaction susceptibility of soil based on in situ test methods. Nat Hazards (11)

  48. Usman AM, Yusof UK, Naim S (2020) Filter-based multi-objective feature selection using nsga iii and cuckoo optimisation algorithm. IEEE Access (99):1

  49. Yu h, He F, Pan Y (2019) A novel segmentation model for medical images with intensity inhomogeneity based on adaptive perturbation. Multimedia Tools Appl 78(9):11779–11798

    Google Scholar 

  50. Hou N, He F, Chen Y (2020) An efficient gpu-based parallel tabu search algorithm for hardware/software co-design. Front Comput Sci 14(5):145316

    Google Scholar 

  51. Luo J, He F, Yong J (2020) An efficient and robust bat algorithm with fusion of opposition-based learning and whale optimization algorithm. Intell Data Anal 24(3):581–606

    Google Scholar 

  52. Zhang J, He F, Chen Y (2020) A new haze removal approach for sky/river alike scenes based on external and internal clues. Multimedia Tools Appl 79:2085–2107

    Google Scholar 

  53. António CC (2014) A memetic algorithm based on multiple learning procedures for global optimal design of composite structures. Memetic Comput 6(2):113–131

    Google Scholar 

  54. Pan Y, He F, Yu H (2020) Learning social representations with deep autoencoder for recommender system. World Wide Web 23(4):2259–2279

    Google Scholar 

  55. Li H, Jingjing M, Maoguo G, Qiongzhi J, Licheng J (2015) Change detection in synthetic aperture radar images based on evolutionary multi-objective optimization with ensemble learning. Memetic Comput 7(4):275–289

    Google Scholar 

  56. Liang Y, He F, Zeng X (2020) 3d mesh simplification with feature preservation based on whale optimization algorithm and differential evolution. Integrated Comput-Aided Eng 27(4):417–435

    Google Scholar 

  57. Zhang S, He F (2020) Drcdn: learning deep residual convolutional dehazing networks. Vis Comput 36(9):1797–1808

    Google Scholar 

  58. Zhang S, He F, Ren W, Yao J (2020) Joint learning of image detail and transmission map for single image dehazing. Vis Comput 36(2):305–316

    Google Scholar 

  59. Pan Y, He F, Yu H (2019) A novel enhanced collaborative autoencoder with knowledge distillation for top-n recommender systems. Neurocomputing 332:137–148

    Google Scholar 

  60. Chen X, He F, Yu H (2019) A matting method based on full feature coverage. Multimedia Tools Appl 78(9):11173–11201

  61. Chen Y, He F, Li H, Zhang D, Wu Y (2020) A full migration bbo algorithm with enhanced population quality bounds for multimodal biomedical image registration. Appl Soft Comput 93:106335

    Google Scholar 

  62. Cai W, He F, Lv X, Cheng Y (2020) A semi-transparent selective undo algorithm for multi-user collaborative editors. Front Comput Sci. https://doi.org/10.1007/s11704-020-9518-x

    Article  Google Scholar 

  63. Pan YT, He FZ, Yu HP (2020) A correlative denoising autoencoder to model social influence for top-n recommender system. Front Comput Sci 14(3):143301

    Google Scholar 

Download references

Funding

This work is supported by the National Natural Science Foundation of China under Grant No. 62072348, the Science and Technology Major Project of Hubei Province (Next-Generation AI Technologies) under Grant No. 2019AEA170 and Translational Medicine and Interdisciplinary Research Joint Fund of Zhongnan Hospital of Wuhan University under Grant No. ZNJC201917.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Fazhi He.

Ethics declarations

Conflict of interest

Haoran Li declares that he has no conflict of interest. Fazhi He declares that he has no conflict of interest. Yiling Chen declares that she has no conflict of interest. Yiteng Pan declares that he has no conflict of interest.

Ethical approval

This article does not contain any studies with human participants performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, H., He, F., Chen, Y. et al. MLFS-CCDE: multi-objective large-scale feature selection by cooperative coevolutionary differential evolution. Memetic Comp. 13, 1–18 (2021). https://doi.org/10.1007/s12293-021-00328-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12293-021-00328-7

Keywords

Navigation