Abstract
In multi-label classification, each training sample is associated with a set of labels and the task is to predict the correct set of labels for the unseen instance. Learning from the multi-label samples is very challenging due to the tremendous number of possible label sets. Therefore, the key to successful multi-label learning is exploiting the label correlations effectively to facilitate the learning process. In this paper, we analyze the limitations of existing methods that add label correlations and propose MLND, a new method which extracts the label correlations from neighbors. Specifically, we take neighbor’s label distribution as new features of an instance and obtain the label’s confidence according to the new features. Nevertheless, the neighbor information is unreliable when the intersection of nearest neighbor samples is small, so we use information entropy to measure the uncertainty of the neighbor information and combine the original instance features with the new features to perform multi-label classification. Experiments on three different real-world multi-label datasets validate the effectiveness of our method against other state-of-the-art methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Berger, A.L.: A maximum entropy approach to natural language processing. Comput. Linguist. 22(1), 39–71 (1996)
Bucak, S.S., Jin, R., Jain, A.K.: Multi-label learning with incomplete class assignments. In: CVPR 2011, pp. 2801–2808 (2011)
Burkhardt, S., Kramer, S.: Multi-label classification using stacked hierarchical Dirichlet processes with reduced sampling complexity. Knowl. Inf. Syst. 59(1), 93–115 (2019). https://doi.org/10.1007/s10115-018-1204-z
Cheng, W., Hüllermeier, E.: Combining instance-based learning and logistic regression for multilabel classification. Mach. Learn. 76(2–3), 211–225 (2009). https://doi.org/10.1007/s10994-009-5127-5
Elisseeff, A., Weston, J.: A kernel method for multi-labelled classification. In: Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic, NIPS 2001, pp. 681–687. MIT Press, Cambridge (2001). http://dl.acm.org/citation.cfm?id=2980539.2980628
Jaynes, E.: Information theory and statistical mechanics, 106(4), 620–630 (1957)
Katakis, I., Tsoumakas, G., Vlahavas, I.: Multilabel text classification for automated tag suggestion. In: Proceedings of the ECML/PKDD 2008 Discovery Challenge (2008)
Lee, C.P., Lin, C.J.: Large-scale linear rankSVM. Neural Comput. 26(4), 781–817 (2014)
Read, J., Pfahringer, B., Holmes, G., Frank, E.: Classifier chains for multi-label classification. Mach. Learn. 85(3), 333 (2011). https://doi.org/10.1007/s10994-011-5256-5
Schapire, R.E., Singer, Y.: Boostexter: a boosting-based system for text categorization. Mach. Learn. 39(2–3), 135–168 (2000). https://doi.org/10.1023/A:1007649029923
Spyromitros, E., Tsoumakas, G., Vlahavas, I.: An empirical study of lazy multilabel classification algorithms. In: Darzentas, J., Vouros, G.A., Vosinakis, S., Arnellos, A. (eds.) SETN 2008. LNCS (LNAI), vol. 5138, pp. 401–406. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-87881-0_40
Trohidis, K., Tsoumakas, G., Kalliris, G., Vlahavas, I.: Multi-label classification of music by emotion. EURASIP J. Audio Speech Music Process. 2011(1), 4 (2011). https://doi.org/10.1186/1687-4722-2011-426793
Tsoumakas, G., Katakis, I., Vlahavas, I.: Mining multi-label data. In: Maimon, O., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook, pp. 667–685. Springer, Boston (2010). https://doi.org/10.1007/978-0-387-09823-4_34
Tsoumakas, G., Vlahavas, I.: Random k-Labelsets: an ensemble method for multilabel classification. In: Kok, J.N., Koronacki, J., Mantaras, R.L., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI), vol. 4701, pp. 406–417. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-74958-5_38
Van Linh, N., Anh, N.K., Than, K., Dang, C.N.: An effective and interpretable method for document classification. Knowl. Inf. Syst. 50(3), 763–793 (2016). https://doi.org/10.1007/s10115-016-0956-6
Zhan, W., Zhang, M.L.: Multi-label learning with label-specific features via clustering ensemble, pp. 129–136 (2017)
Zhang, M.L., Zhou, Z.H.: ML-KNN: a lazy learning approach to multi-label leaming. Pattern Recognit. 40, 2038–2048 (2007)
Zhang, Q.W., Zhong, Y., Zhang, M.L.: Feature-induced labeling information enrichment for multi-label learning. In: AAAI (2018)
Acknowledgment
This work is supported by NSFC No. 61772216, 61821003, U1705261.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Yang, L. et al. (2020). MLND: A Weight-Adapting Method for Multi-label Classification Based on Neighbor Label Distribution. In: Wang, X., Zhang, R., Lee, YK., Sun, L., Moon, YS. (eds) Web and Big Data. APWeb-WAIM 2020. Lecture Notes in Computer Science(), vol 12317. Springer, Cham. https://doi.org/10.1007/978-3-030-60259-8_47
Download citation
DOI: https://doi.org/10.1007/978-3-030-60259-8_47
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60258-1
Online ISBN: 978-3-030-60259-8
eBook Packages: Computer ScienceComputer Science (R0)