Skip to main content

MLND: A Weight-Adapting Method for Multi-label Classification Based on Neighbor Label Distribution

  • Conference paper
  • First Online:
Web and Big Data (APWeb-WAIM 2020)

Abstract

In multi-label classification, each training sample is associated with a set of labels and the task is to predict the correct set of labels for the unseen instance. Learning from the multi-label samples is very challenging due to the tremendous number of possible label sets. Therefore, the key to successful multi-label learning is exploiting the label correlations effectively to facilitate the learning process. In this paper, we analyze the limitations of existing methods that add label correlations and propose MLND, a new method which extracts the label correlations from neighbors. Specifically, we take neighbor’s label distribution as new features of an instance and obtain the label’s confidence according to the new features. Nevertheless, the neighbor information is unreliable when the intersection of nearest neighbor samples is small, so we use information entropy to measure the uncertainty of the neighbor information and combine the original instance features with the new features to perform multi-label classification. Experiments on three different real-world multi-label datasets validate the effectiveness of our method against other state-of-the-art methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 109.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Berger, A.L.: A maximum entropy approach to natural language processing. Comput. Linguist. 22(1), 39–71 (1996)

    Google Scholar 

  2. Bucak, S.S., Jin, R., Jain, A.K.: Multi-label learning with incomplete class assignments. In: CVPR 2011, pp. 2801–2808 (2011)

    Google Scholar 

  3. Burkhardt, S., Kramer, S.: Multi-label classification using stacked hierarchical Dirichlet processes with reduced sampling complexity. Knowl. Inf. Syst. 59(1), 93–115 (2019). https://doi.org/10.1007/s10115-018-1204-z

    Article  Google Scholar 

  4. Cheng, W., Hüllermeier, E.: Combining instance-based learning and logistic regression for multilabel classification. Mach. Learn. 76(2–3), 211–225 (2009). https://doi.org/10.1007/s10994-009-5127-5

    Article  Google Scholar 

  5. Elisseeff, A., Weston, J.: A kernel method for multi-labelled classification. In: Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic, NIPS 2001, pp. 681–687. MIT Press, Cambridge (2001). http://dl.acm.org/citation.cfm?id=2980539.2980628

  6. Jaynes, E.: Information theory and statistical mechanics, 106(4), 620–630 (1957)

    Google Scholar 

  7. Katakis, I., Tsoumakas, G., Vlahavas, I.: Multilabel text classification for automated tag suggestion. In: Proceedings of the ECML/PKDD 2008 Discovery Challenge (2008)

    Google Scholar 

  8. Lee, C.P., Lin, C.J.: Large-scale linear rankSVM. Neural Comput. 26(4), 781–817 (2014)

    Article  MathSciNet  Google Scholar 

  9. Read, J., Pfahringer, B., Holmes, G., Frank, E.: Classifier chains for multi-label classification. Mach. Learn. 85(3), 333 (2011). https://doi.org/10.1007/s10994-011-5256-5

    Article  MathSciNet  Google Scholar 

  10. Schapire, R.E., Singer, Y.: Boostexter: a boosting-based system for text categorization. Mach. Learn. 39(2–3), 135–168 (2000). https://doi.org/10.1023/A:1007649029923

    Article  MATH  Google Scholar 

  11. Spyromitros, E., Tsoumakas, G., Vlahavas, I.: An empirical study of lazy multilabel classification algorithms. In: Darzentas, J., Vouros, G.A., Vosinakis, S., Arnellos, A. (eds.) SETN 2008. LNCS (LNAI), vol. 5138, pp. 401–406. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-87881-0_40

    Chapter  Google Scholar 

  12. Trohidis, K., Tsoumakas, G., Kalliris, G., Vlahavas, I.: Multi-label classification of music by emotion. EURASIP J. Audio Speech Music Process. 2011(1), 4 (2011). https://doi.org/10.1186/1687-4722-2011-426793

    Article  Google Scholar 

  13. Tsoumakas, G., Katakis, I., Vlahavas, I.: Mining multi-label data. In: Maimon, O., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook, pp. 667–685. Springer, Boston (2010). https://doi.org/10.1007/978-0-387-09823-4_34

    Chapter  Google Scholar 

  14. Tsoumakas, G., Vlahavas, I.: Random k-Labelsets: an ensemble method for multilabel classification. In: Kok, J.N., Koronacki, J., Mantaras, R.L., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI), vol. 4701, pp. 406–417. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-74958-5_38

    Chapter  Google Scholar 

  15. Van Linh, N., Anh, N.K., Than, K., Dang, C.N.: An effective and interpretable method for document classification. Knowl. Inf. Syst. 50(3), 763–793 (2016). https://doi.org/10.1007/s10115-016-0956-6

    Article  Google Scholar 

  16. Zhan, W., Zhang, M.L.: Multi-label learning with label-specific features via clustering ensemble, pp. 129–136 (2017)

    Google Scholar 

  17. Zhang, M.L., Zhou, Z.H.: ML-KNN: a lazy learning approach to multi-label leaming. Pattern Recognit. 40, 2038–2048 (2007)

    Article  Google Scholar 

  18. Zhang, Q.W., Zhong, Y., Zhang, M.L.: Feature-induced labeling information enrichment for multi-label learning. In: AAAI (2018)

    Google Scholar 

Download references

Acknowledgment

This work is supported by NSFC No. 61772216, 61821003, U1705261.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhan Shi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yang, L. et al. (2020). MLND: A Weight-Adapting Method for Multi-label Classification Based on Neighbor Label Distribution. In: Wang, X., Zhang, R., Lee, YK., Sun, L., Moon, YS. (eds) Web and Big Data. APWeb-WAIM 2020. Lecture Notes in Computer Science(), vol 12317. Springer, Cham. https://doi.org/10.1007/978-3-030-60259-8_47

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-60259-8_47

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-60258-1

  • Online ISBN: 978-3-030-60259-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics