An Improved BPSO Algorithm for Feature Selection

  • Lalit KumarEmail author
  • Kusum Kumari Bharti
Conference paper
Part of the Lecture Notes in Electrical Engineering book series (LNEE, volume 524)


In machine learning and data mining tasks, feature selection has been used to select the relevant subset of features. Traditionally, high-dimensional datasets have so many redundant and irrelevant features, which degrade the performance of clustering. Therefore, feature selection is necessary to improve the clustering performance. In this paper, we select the optimal subset of features and perform cluster analysis simultaneously using modified-BPSO (Binary Particle Swarm Optimization) and K-means. Optimality of clusters is measured by various cluster validation indices. By comparing the overall performance of the modified-BPSO with the BPSO and BMFOA (Binary Moth Flame Optimization Algorithm) on six real datasets drawn from the UC Irvine Machine Learning Repository, the results show that the performance of the proposed method is better than other methods involved in the paper.


BPSO BMFO Cluster validation index Data clustering Feature selection Swarm intelligence 


  1. 1.
    Guyon, I., & Elisseeff, A. (2003). An introduction to variable and feature selection. The Journal of Machine Learning Research, 3, 1157–1182.zbMATHGoogle Scholar
  2. 2.
    Kabir, M., Shahjahan, M., & Murase, K. (2012). A new hybrid ant colony optimization algorithm for feature selection. Expert Systems with Applications, 39(3), 3747–3763.CrossRefGoogle Scholar
  3. 3.
    Prakash, J., & Singh, P. (2015). Particle swarm optimization with k-means for simultaneous feature selection and data clustering. In IEEE 2015 Second International Conference on Soft Computing and Machine Intelligence (ISCMI), Nov 23 (pp. 74–78).Google Scholar
  4. 4.
    Gwak, J., Jeon, M., & Pedrycz, W. (2016). Bolstering efficient SSGAs based on an ensemble of probabilistic variable-wise crossover strategies. Soft Computing, 20(6), 2149–2176.CrossRefGoogle Scholar
  5. 5.
    Yang, S.: Nature-inspired optimization algorithms. Elsevier (2014).Google Scholar
  6. 6.
    Ali, M., Awad, N., Suganthan, P., & Reynolds, R. (2017). An adaptive multipopulation differential evolution with dynamic population reduction. IEEE Transactions on Cybernetics, 47(9), 2768–2779.CrossRefGoogle Scholar
  7. 7.
    Moradi, P., & Gholampour, M. (2016). A hybrid particle swarm optimization for featuresubset selection by integrating a novel local search strategy. Applied Soft Computing, 43, 117–130.CrossRefGoogle Scholar
  8. 8.
    Nanda, S., & Panda, G. (2014). A survey on nature inspired metaheuristic algorithms for partitional clustering. Swarm and Evolutionary Computation, 16, 1–8.CrossRefGoogle Scholar
  9. 9.
    Kennedy, J. (2011). Particle swarm optimization. In Encyclopedia of machine learning (pp. 760–766). US: Springer.Google Scholar
  10. 10.
    Marini, F., & Walczak, B. (2015). Particle swarm optimization (PSO). A tutorial. Chemometrics and Intelligent Laboratory Systems, 149, 153–165.CrossRefGoogle Scholar
  11. 11.
    Kennedy, J., & Eberhart, R. (1997). A discrete binary version of the particle swarm algorithm. In 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation (Vol. 5). IEEE.Google Scholar
  12. 12.
    Liu, J., Mei, Y., & Li, X. (2016). An analysis of the inertia weight parameter for binary particle swarm optimization. IEEE Transactions on Evolutionary Computation, 20(5), 666–681.CrossRefGoogle Scholar
  13. 13.
    Mirjalili, S. (2015). Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowledge-Based Systems, 89, 228–249.CrossRefGoogle Scholar
  14. 14.
    Reddy, S., Panwar, L., Panigrahi, B., & Kumar, R. (2017). Solution to unit commitment in power system operation planning using modified moth flame optimization algorithm (MMFOA): A flame selection based computational technique. Journal of Computational Science, 1–18.Google Scholar
  15. 15.
    Kaufman, L., & Rousseeuw, P. (2009). Finding groups in data: an introduction to cluster analysis. Probability and statistics. Wiley.Google Scholar
  16. 16.
    Bezdek, J., & Pal, N. (1995). Cluster validation with generalized Dunn’s indices. In Proceedings of the Second New Zealand International Two-Stream Conference on Artificial Neural Networks and Expert Systems, Nov 20 (pp. 190–193). IEEE.Google Scholar
  17. 17.
    Davies, D., & Bouldin, D. (1979). A cluster separation measure. In IEEE Transactions On Pattern Analysis And Machine Intelligence, Apr 2 (pp. 224–227).CrossRefGoogle Scholar
  18. 18.
    Jain, A., & Dubes, R. (1988). Algorithms for clustering data (Vol. 6). Prentice Hall Englewood Cliffs.Google Scholar
  19. 19.
    Prakash, J., & Singh, P. (2012). An effective hybrid method based on de, ga, and k-means for data clustering. In Proceedings of the Second International Conference on Soft Computing for Problem Solving (SocProS 2012), Dec 28–30 (pp. 1561–1572). Springer.Google Scholar
  20. 20.
    Jain, I., Jain, V., & Jain, R. (2018). Correlation feature selection based improved-binary Particle Swarm Optimization for gene selection and cancer classification. Applied Soft Computing, 62, 203–215.CrossRefGoogle Scholar
  21. 21.
    Frank, A. UCI Machine Learning Repository.

Copyright information

© Springer Nature Singapore Pte Ltd. 2019

Authors and Affiliations

  1. 1.Design and ManufacturingPDPM-Indian Institute of Information TechnologyJabalpurIndia

Personalised recommendations