Skip to main content

Multi-label Feature Selection Method Based on Multivariate Mutual Information and Particle Swarm Optimization

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2018)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 11304))

Included in the following conference series:

Abstract

Multi-label feature selection has become an indispensable pre-processing step to deal with possible irrelevant and redundant features, to decrease computational burdens, improve classification performance and enhance model interpretability, in multi-label learning. Mutual information (MI) between two random variables is widely used to describe feature-label relevance and feature-feature redundancy. Furthermore, multivariate mutual information (MMI) is approximated via limiting three-degree interactions to speed up its computation, and then is used to characterize relevance between selected feature subset and label subset. In this paper, we combine MMI-based relevance with MI-based redundancy to define a new max-relevance and min-redundancy feature selection criterion (simply MMI). To search for a globally optimal solution, we add an auxiliary mutation operation to existing binary particle swarm optimization with mutation to control the number of selected features strictly to form a new PSO variant: M2BPSO. Integrating MMI with M2BPSO builds a novel multi-label feature selection method: MMI-PSO. The experiments on four benchmark data sets demonstrate the effectiveness of our proposed algorithm, according to four instance-based classification evaluation metrics, compared with three state-of-the-art feature selection approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://computer.njnu.edu.cn/Lab/LABIC/LABIC_Software.html.

References

  1. Tsoumakas, G., Katakis, I.: Multi-label classification: an overview. Int. J. Data Warehouse Min. 3(3), 1–13 (2007)

    Article  Google Scholar 

  2. Zhang, M., Zhou, Z.: A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng. 26(8), 1338–1351 (2014)

    Article  Google Scholar 

  3. Gibaja, E., Ventura, S.: A tutorial on multilabel learning. ACM Comput. Surv. 47(3), 1–38 (2015). Article 52

    Article  Google Scholar 

  4. Herrera, F., Charte, F., Rivera, A.J., del Jesus, M.J.: Multilabel Classification: Problem Analysis, Metrics and Techniques. Springer, Switzerland (2016). https://doi.org/10.1007/978-3-319-41111-8

    Book  Google Scholar 

  5. Kashef, S., Nezamabadi-pour, H., Nipour, B.: Multilabel feature selection: a comprehensiove review and guide experiments. WIREs Data Min. Knowl. Discov. 8(2), e1240 (2018)

    Article  Google Scholar 

  6. Pereira, R., Plastino, A., Zadrozny, B., Merschmann, L.H.C.: Categorizing feature selection methods for multi-label classification. Artif. Intell. Rev. 49(1), 57–78 (2018)

    Article  Google Scholar 

  7. Vergara, J.R., Estevez, P.A.: A review of feature selection methods based on mutual information. Neural Comput. Appl. 24(1), 175–186 (2014)

    Article  Google Scholar 

  8. Cai, J., Luo, J., Wang, S., Yang, S.: Feature selection in machine learning: a new perspective. Neurocomputing 300, 70–79 (2018)

    Article  Google Scholar 

  9. McGill, W.J.: Multivariate information transmission. Trans. IRE Prof. Group Inf. Theor. 4(4), 93–111 (1954)

    Article  MathSciNet  Google Scholar 

  10. Lee, J., Kim, D.W.: Feature selection for multi-label classification using multivariate mutual information. Pattern Recognit. Lett. 34(3), 349–357 (2013)

    Article  Google Scholar 

  11. Lee, J., Kim, D.W.: Fast multi-label feature selection based on information-theoretic feature ranking. Pattern Recognit. 48(9), 2761–2771 (2015)

    Article  Google Scholar 

  12. Lin, Y., Hu, Q., Liu, J., Duan, J.: Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 168, 92–103 (2015)

    Article  Google Scholar 

  13. Peng, H., Long, F., Ding, C.: Feature selection based on mutual information: criterion of max-dependency, max-relevance and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 27(8), 1226–1238 (2005)

    Article  Google Scholar 

  14. Lin, Y., Hu, Q., Liu, J., Chen, J., Duan, J.: Multi-label feature selection based on neighborhood mutual information. Appl. Soft Comput. 38, 244–256 (2016)

    Article  Google Scholar 

  15. Lee, J., Kim, D.W.: Mutual information-based multi-label feature selection using interaction information. Expert Syst. Appl. 42(4), 2013–2025 (2015)

    Article  Google Scholar 

  16. Lee, J., Kim, D.: Memetic feature selection algorithm for multi-label classification. Inf. Sci. 293(293), 80–96 (2015)

    Article  Google Scholar 

  17. Lim, H., Lee, J., Kim, D.W.: Multi-label learning using mathematical programming. IEICE Trans. Inform. Syst. 98(1), 197–200 (2015)

    Article  Google Scholar 

  18. Lim, H., Lee, J., Kim, D.W.: Low-rank approximation for multi-label feature selection. Int. J. Mach. Learn. Comput. 6(1), 42–46 (2016)

    Google Scholar 

  19. Xu, J., Ma, Q.: Multi-label regularized quadratic programming feature selection algorithm with frank-wolfe method. Expert Syst. Appl. 95, 14–31 (2018)

    Article  Google Scholar 

  20. Poli, R., Kennedy, J., Blackwell, T.: Particle swarm optimization: an overview. Swarm Intell. 1(1), 33–57 (2007)

    Article  Google Scholar 

  21. Zhang, Y., Wang, S., Phillips, P., Ji, G.: Binary PSO with mutation operator for feature selection using decision tree applied to spam detection. Knowl.-Based Syst. 64, 22–31 (2014)

    Article  Google Scholar 

  22. Zhang, M., Zhou, Z.: ML-kNN: A lazy approach to multi-label learning. Pattern Recognit. 40(7), 2038–2048 (2007)

    Article  Google Scholar 

  23. Demsar, J.: Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 7, 1–30 (2006)

    MathSciNet  MATH  Google Scholar 

Download references

Acknowledgments

This work was supported by the Natural Science Foundation of China (NSFC) under Grant 61273246.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jianhua Xu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wang, X., Zhao, L., Xu, J. (2018). Multi-label Feature Selection Method Based on Multivariate Mutual Information and Particle Swarm Optimization. In: Cheng, L., Leung, A., Ozawa, S. (eds) Neural Information Processing. ICONIP 2018. Lecture Notes in Computer Science(), vol 11304. Springer, Cham. https://doi.org/10.1007/978-3-030-04212-7_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-04212-7_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-04211-0

  • Online ISBN: 978-3-030-04212-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics