Skip to main content
Log in

A filter-based bare-bone particle swarm optimization algorithm for unsupervised feature selection

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Due to good exploration capability, particle swarm optimization (PSO) has shown advantages on solving supervised feature selection problems. Compared with supervised and semi-supervised cases, unsupervised feature selection becomes very difficult as a result of no label information. This paper studies a novel PSO-based unsupervised feature selection method, called filter-based bare-bone particle swarm optimization algorithm (FBPSO). Two filter-based strategies are proposed to speed up the convergence of the algorithm. One is a space reduction strategy based on average mutual information, which is used to remove irrelevant and weakly relevant features fast; another is a local filter search strategy based on feature redundancy, which is used to improve the exploitation capability of the swarm. And, a feature similarity-based evaluation function and a parameter-free update strategy of particle are introduced to enhance the performance of FBPSO. Experimental results on some typical datasets confirm superiority and effectiveness of the proposed FBPSO.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  1. Abualigah LM, Khader AT (2017) Unsupervised text feature selection technique based on hybrid particle swarm optimization algorithm with genetic operators for the text clustering. J Supercomput 73(11):4773–4795

    Article  Google Scholar 

  2. Abualigah LM, Khader AT, Al-Betar MA (2016) Unsupervised feature selection technique based on harmony search algorithm for improving the text clustering. In: the 7th International Conference on Computer Science and Information Technology (pp.1–6)

  3. Adeli A, Broumandnia A (2018) Image steganalysis using improved particle swarm optimization based feature selection. Appl Intell 48(6):1609–1622

    Article  Google Scholar 

  4. Arun V, Krishna M, Arunkumar BV, Padma SK, Shyam V (2018) Exploratory boosted feature selection and neural network framework for depression classification. International Journal of Interactive Multimedia and Artificial Intelligence 5(3):61–71

    Article  Google Scholar 

  5. Bagwari P, Saxena B, Balodhi M, Bijalwan V (2017) Comparison of feedforward network and radial basis function to detect leukemia. International Journal of Interactive Multimedia and Artificial Intelligence 4(5):55–57

    Article  Google Scholar 

  6. Barani F, Mirhosseini M, Nezamabadi-pour H (2017) Application of binary quantum-inspired gravitational search algorithm in feature subset selection. Appl Intell 47(2):304–318

    Article  Google Scholar 

  7. Bhadra T, Bandyopadhyay S (2015) Unsupervised feature selection using an improved version of Differential Evolution. Expert Syst Appl 42(8):4042–4053

    Article  Google Scholar 

  8. Bi N, Tan J, Lai JH, Suen CY (2018) High-dimensional supervised feature selection via optimized kernel mutual information. Expert Syst Appl 108:81–95

    Article  Google Scholar 

  9. Cai D, Zhang C, He X (2010) Unsupervised feature selection for multi-cluster data. In: ACM International Conference on Knowledge Discovery and Data Mining (pp.333-342)

  10. Dua D, Karra Taniskidou E (2017) UCI Machine Learning Repository. Irvine: University of California, School of Information and Computer Science. http://archive.ics.uci.edu/ml

  11. Gao WF, Hu L, Zhang P, Wang F (2018) Feature selection by integrating two groups of feature evaluation criteria. Expert Syst Appl 110:11–19

    Article  Google Scholar 

  12. García S, Molina D, Lozano M, Herrera F (2009) A study on the use of non-parametric tests for analyzing the evolutionary algorithms' behaviour: a case study on the CEC'2005 special session on real parameter optimization. J Heuristics 15(6):617–644

    Article  MATH  Google Scholar 

  13. Gong DW, Liu K (2018) A multi-objective optimization model and its evolution-based solutions for the fingertip localization problem. Pattern Recogn 74:385–405

    Article  Google Scholar 

  14. Hancer E, Xue B, Zhang MJ, Karaboga D, Akay B (2018) Pareto front feature selection based on artificial bee colony optimization. Inf Sci 422:462–479

    Article  Google Scholar 

  15. He JR, Bi YZ, Ding LX, Li ZK, Wang SW (2017) Unsupervised feature selection based on decision graph. Neural Comput & Applic 28(10):3047–3059

    Article  Google Scholar 

  16. He X, Cai D, Niyogi P (2005) Laplacian score for feature selection. In: International Conference on Neural Information Processing Systems (pp. 507–514)

  17. Hou C, Nie F, Li X, Yi D, Wu Y (2014) Joint embedding learning and sparse regression: a framework for unsupervised feature selection. IEEE Transactions on Cybernetics 44(6):793–804

    Article  Google Scholar 

  18. Iranmehr E, Shouraki SB, Faraji MM (2017) Unsupervised feature selection for phoneme sound classification using particle swarm optimization. In: Iranian Joint Congress on Fuzzy and Intelligent Systems (pp. 86–90)

  19. Kennedy J (2003) Bare-bones particle swarms. In: Proceedings of the Swarm Intelligence Symposium (pp. 80–87)

  20. Kennedy J, Eberhart R (1995) Particle swarm optimization. In: IEEE International Conference on Neural Networks (pp.1–7)

  21. Kennedy J, Eberhart RC (1997) A discrete binary version of the particle swarm algorithm. IEEE International Conference on Systems 5:4104–4108

    Google Scholar 

  22. Kimovski D, Ortega J, Ortiz A, Banos R (2015) Parallel alternatives for evolutionary multi-objective optimization in unsupervised feature selection. Expert Syst Appl 42(9):4239–4252

    Article  Google Scholar 

  23. Li Z, Yang Y, Liu J, Zhou X, Lu H (2012) Unsupervised feature selection using nonnegative spectral analysis. In: Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence (pp.1026–1032)

  24. Meza J, Espitia H, Montenegro C, Giménez E, González R (2017) MOVPSO: vortex multi-objective particle swarm optimization. Appl Soft Comput 52:1042–1057

    Article  Google Scholar 

  25. Meza J, Espitia H, Montenegro C, Crespo RG (2016) Statistical analysis of a multi-objective optimization algorithm based on a model of particles with vorticity behavior. Soft Comput 20(9):3521–3536

    Article  Google Scholar 

  26. Mitra P, Murthy CA, Pal SK (2002) Unsupervised feature selection using feature similarity. IEEE Transactions on Pattern Analysis & Machine Intelligence 24(3):301–312

    Article  Google Scholar 

  27. Sheikhpour R, Sarram MA, Sheikhpour E (2018) Semi-supervised sparse feature selection via graph Laplacian based scatter matrix for regression problems. Inf Sci 468:14–28

    Article  Google Scholar 

  28. Tabakhi S, Moradi P, Akhlaghian F (2014) An unsupervised feature selection algorithm based on ant colony optimization. Eng Appl Artif Intell 32(6):112–123

    Article  Google Scholar 

  29. Talukdar U, Hazarika SM, Gan JQ (2018) A Kernel Partial least square based feature selection method. Pattern Recogn 83:91–106

    Article  Google Scholar 

  30. Tang C, Zhu XZ, Chen JJ, Wang PC, Liu XW, Tian J (2018) Robust graph regularized unsupervised feature selection. Expert Syst Appl 96:64–76

    Article  Google Scholar 

  31. Wang D, Nie F, Huang H (2014) Unsupervised feature selection via unified trace ratio formulation and k-means clustering. In: Proceeding of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases (pp. 306–321)

  32. Wang Y, Feng LZ, Zhu JM (2018) Novel artificial bee colony based feature selection method for filtering redundant information. Appl Intell 48(4):868–885

    Article  Google Scholar 

  33. Wang Y, Wang J, Liao H (2017) Unsupervised feature selection based on Markov blanket and particle swarm optimization. J Syst Eng Electron 28(1):151–161

    Article  Google Scholar 

  34. Zhai Y, Ong YS, Tsang IW (2014) The emerging "Big Dimensionality". IEEE Comput Intell Mag 9(3):14–26

    Article  Google Scholar 

  35. Zhang Y, Song XF, Gong DW (2017) A return-cost-based binary firefly algorithm for feature selection. Inf Sci 418:561–574

    Article  Google Scholar 

  36. Zhang Y, Gong DW, Cheng J (2017) Multi-objective particle swarm optimization approach for cost-based feature selection in classification. IEEE/ACM Transactions on Computational Biology and Bioinformatics 14(1):64–75

    Article  Google Scholar 

  37. Zhang Y, Gong DW, Hu Y (2015) Feature selection algorithm based on bare-bones particle swarm optimization. Neurocomputing 148(1):150–157

    Article  Google Scholar 

  38. Zhang Y, Gong DW, Zhang JH (2013) Robotic path planning in uncertain environment using multi-objective particle swarm optimization. Neurocomputing 103:172–185

    Article  Google Scholar 

  39. Zhao Z, Morstatter F, Sharma S, Anand A, Liu H (2010) Advancing feature selection research. Arizona State University, Phoehix

    Google Scholar 

  40. Zhu PF, Xu Q, Hu QH, Zhang CQ (2018) Co-regularized unsupervised feature selection. Neurocomputing 275:2855–2863

    Article  Google Scholar 

Download references

Acknowledgements

This work was jointly supported by the National Natural Science Foundation of China (No. 61876185), and Six Talents Peaks Project of Jiangsu Province (No. DZXX-053).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Yong Zhang or Hai-Gang Li.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, Y., Li, HG., Wang, Q. et al. A filter-based bare-bone particle swarm optimization algorithm for unsupervised feature selection. Appl Intell 49, 2889–2898 (2019). https://doi.org/10.1007/s10489-019-01420-9

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-019-01420-9

Keywords

Navigation