Abstract
Multi-task learning is to improve the performance of the model by transferring and exploiting common knowledge among tasks. Existing MTL works mainly focus on the scenario where label sets among multiple tasks (MTs) are usually the same, thus they can be utilized for learning across the tasks. However, the real world has more general scenarios in which each task has only a small number of training samples and their label sets are just partially overlapped or even not. Learning such MTs is more challenging because of less correlation information available among these tasks. For this, we propose a framework to learn these tasks by jointly leveraging both abundant information from a learnt auxiliary big task with sufficiently many classes to cover those of all these tasks and the information shared among those partially-overlapped tasks. In our implementation of using the same neural network architecture of the learnt auxiliary task to learn individual tasks, the key idea is to utilize available label information to adaptively prune the hidden layer neurons of the auxiliary network to construct corresponding network for each task, while accompanying a joint learning across individual tasks. Extensive experimental results demonstrate that our proposed method is significantly competitive compared to state-of-the-art methods.
Similar content being viewed by others
References
Khattar A, Hegde S, Hebbalaguppe R. Cross-domain multi-task learning for object detection and saliency estimation. In: Proceedings of 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 2021, 3634–3643
Al-Qaisi L, Hassonah M A, Al-Zoubi M M, Al-Zoubi A M. A review of evolutionary data clustering algorithms for image segmentation. In: Aljarah I, Faris H, Mirjalili S, eds. Evolutionary Data Clustering: Algorithms and Applications. Singapore: Springer, 2021, 201–214
Chaturvedi I, Su C L, Welsch R E. Fuzzy aggregated topology evolution for cognitive multi-tasks. Cognitive Computation, 2021, 13(1): 96–107
Ravanelli M, Zhong J, Pascual S, Swietojanski P, Monteiro J, Trmal J, Bengio Y. Multi-task self-supervised learning for robust speech recognition. In: Proceedings of 2020 IEEE International Conference on Acoustics, Speech and Signal Processing. 2020, 6989–6993
Li C, Wang B, Zhang S, Liu Y, Song R, Cheng J, Chen X. Emotion recognition from EEG based on multi-task learning with capsule network and attention mechanism. Computers in Biology and Medicine, 2022, 143: 105303
Cheng B, Zhu B, Pu S. Multi-auxiliary domain transfer learning for diagnosis of MCI conversion. Neurological Sciences, 2022, 43(3): 1721–1739
Ruder S, Bingel J, Augenstein I, Søgaard A. Sluice networks: learning what to share between loosely related tasks. 2017, arXiv preprint arXiv: 1705.08142v1
Gong T, Zheng X, Lu X. Remote sensing scene classification with multi-task learning. In: Wang L, Wu Y, Gong J, eds. Proceedings of the 7th China High Resolution Earth Observation Conference (CHREOC 2020). Singapore: Springer, 2022, 403–418
Islam M M, Iqbal T. Mumu: cooperative multitask learning-based guided multimodal fusion. In: Proceedings of the 36th Conference on Artificial Intelligence. 2022, 1043–1051
Long M, Cao Z, Wang J, Yu P S. Learning multiple tasks with multilinear relationship networks. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 1593–1602
Ma J, Zhao Z, Yi X, Chen J, Hong L, Chi E H. Modeling task relationships in multi-task learning with multi-gate mixture-of-experts. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2018, 1930–1939
Feng Q, Yao J, Zhong Y, Li P, Pan Z. Learning twofold heterogeneous multi-task by sharing similar convolution kernel pairs. Knowledge-Based Systems, 2022, 252: 109396
Wu S, Zhang H R, Ré C. Understanding and improving information transfer in multi-task learning. In: Proceedings of the 8th International Conference on Learning Representations. 2020, 26–30
Evgeniou T, Pontil M. Regularized multi-task learning. In: Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2004, 109–117
Honorio J, Samaras D. Multi-task learning of Gaussian graphical models. In: Proceedings of the 27th International Conference on Machine Learning. 2010, 447–454
Liu Q, Li X, He Z, Fan N, Yuan D, Liu W, Liang Y. Multi-task driven feature models for thermal infrared tracking. In: Proceedings of the 34th AAAI Conference on Artificial Intelligence. 2020, 11604–11611
Wang J, Zhang S, Wang Y, Zhu Z. Learning efficient multi-task stereo matching network with richer feature information. Neurocomputing, 2021, 421: 151–160
Guo P, Deng C, Xu L, Huang X, Zhang Y. Deep multi-task augmented feature learning via hierarchical graph neural network. In: Proceedings of European Conference on Machine Learning and Knowledge Discovery in Databases. 2021, 538–553
Vandenhende S, Georgoulis S, van Gansbeke W, Proesmans M, Dai D, van Gool L. Multi-task learning for dense prediction tasks: a survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 44(7): 3614–3633
Shen Z, Cui C, Huang J, Zong J, Chen M, Yin Y. Deep adaptive feature aggregation in multi-task convolutional neural networks. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management. 2020, 2213–2216
Yadav S, Chauhan J, Sain J P, Thirunarayan K, Sheth A, Schumm J. Identifying depressive symptoms from tweets: figurative language enabled multitask learning framework. In: Proceedings of the 28th International Conference on Computational Linguistics. 2020, 696–709
Sun X, Panda R, Feris R, Saenko K. Adashare: learning what to share for efficient deep multi-task learning. In: Proceedings of the 34th Conference on Neural Information Processing Systems. 2020, 8728–8740
Sun T, Shao Y, Li X, Liu P, Yan H, Qiu X, Huang X. Learning sparse sharing architectures for multiple tasks. In: Proceedings of the 34th AAAI Conference on Artificial Intelligence. 2020, 8936–8943
Verboven S, Hafeez Chaudhary M, Berrevoets J, Verbeke W. HydaLearn: highly dynamic task weighting for multi-task learning with auxiliary tasks. 2008, arXiv preprint arXiv: 2008.11643
Sanh V, Wolf T, Ruder S. A hierarchical multi-task approach for learning embeddings from semantic tasks. In: Proceedings of the 33rd AAAI Conference on Artificial Intelligence. 2019, 6949–6956
Baxter J. A Bayesian/information theoretic model of learning to learn via multiple task sampling. Machine Learning, 1997, 28(1): 7–39
Ruder S, Bingel J, Augenstein I, Søgaard A. Latent multi-task architecture learning. In: In Proceedings of the 33rd AAAI Conference on Artificial Intelligence. 2019, 4822–4829
Strezoski G, Noord N, Worring M. Many task learning with task routing. In: Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. 2019, 1375–1384
Fernando C, Banarse D, Blundell C, Zwols Y, Ha D, Rusu A A, Pritzel A, Wierstra D. PathNet: Evolution channels gradient descent in super neural networks. 2017, arXiv preprint arXiv: 1701.08734
Pironkov G, Wood S U, Dupont S. Hybrid-task learning for robust automatic speech recognition. Computer Speech & Language, 2020, 64: 101103
Cao P, Shan X, Zhao D, Huang M, Zaiane O. Sparse shared structure based multi-task learning for MRI based cognitive performance prediction of Alzheimer’s disease. Pattern Recognition, 2017, 72: 219–235
Lee S, Son Y. Multitask learning with single gradient step update for task balancing. Neurocomputing, 2022, 467: 442–453
Søgaard A, Goldberg Y. Deep multi-task learning with low level tasks supervised at lower layers. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. 2016, 231–235
Fan J, Zhao T, Kuang Z, Zheng Y, Zhang J, Yu J, Peng J. HD-MTL: hierarchical deep multi-task learning for large-scale visual recognition. IEEE Transactions on Image Processing, 2017, 26(4): 1923–1938
Ott F, Rügamer D, Heublein L, Bischl B, Mutschler C. Joint classification and trajectory regression of online handwriting using a multi-task learning approach. In: Proceedings of 2022 IEEE/CVF Winter Conference on Applications of Computer Vision. 2022, 1244–1254
Zhang C, Li Y, Du N, Fan W, Yu P S. Joint slot filling and intent detection via capsule neural networks. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019, 5259–5267
Li X C, Zhan D C. FedRS: federated learning with restricted softmax for label distribution non-IID data. In: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 2021, 995–1005
Wang Y, Zhang Z, Hao W, Song C. Attention guided multiple source and target domain adaptation. IEEE Transactions on Image Processing, 2021, 30: 892–906
Xue W. Weighted feature-task-aware regularization learner for multitask learning. Pattern Analysis and Applications, 2020, 23(1): 253–263
Zhang J, Miao J, Zhao K, Tian Y. Multi-task feature selection with sparse regularization to extract common and task-specific features. Neurocomputing, 2019, 340: 76–89
Shao W, Peng Y, Zu C, Wang M, Zhang D, The Alzheimer’s Disease Neuroimaging Initiative. Hypergraph based multi-task feature selection for multimodal classification of Alzheimer’s disease. Computerized Medical Imaging and Graphics, 2020, 80: 101663
Li L, Pan X, Yang H, Liu Z, He Y, Li Z, Fan Y, Cao Z, Zhang L. Multitask deep learning for fine-grained classification and grading in breast cancer histopathological images. Multimedia Tools and Applications, 2020, 79(21–22): 14509–14528
Zheng Z, Wang Y, Dai Q, Zheng H, Wang D. Metadata-driven task relation discovery for multi-task learning. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence. 2019, 4426–4432
Yan C, Xu J, Xie J, Cai C, Lu H. Prior-aware CNN with multi-task learning for colon images analysis. In: Proceedings of the 17th IEEE International Symposium on Biomedical Imaging. 2020, 254–257
Misra I, Shrivastava A, Gupta A, Hebert M. Cross-stitch networks for multi-task learning. In: Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. 2016, 3994–4003
Duan R, Chen N F. Unsupervised feature adaptation using adversarial multi-task training for automatic evaluation of children’s speech. In: Proceedings of the 21st Annual Conference of the International Speech Communication Association. 2020, 3037–3041
Augenstein I, Søgaard A. Multi-task learning of keyphrase boundary classification. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017, 341–346
Rai P, Daumé III H. Infinite predictor subspace models for multitask learning. In: Proceedings of the 13th International Conference on Artificial Intelligence and Statistics. 2010, 613–620
Zhou L, Cui Z, Xu C, Zhang Z, Wang C, Zhang T, Yang J. Pattern-structure diffusion for multi-task learning. In: Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020, 4513–4522
Wang Y, Luo X, Ding L, Fu S, Hu S. Multi-task non-negative matrix factorization for visual object tracking. Pattern Analysis and Applications, 2020, 23(1): 493–507
Jeong J Y, Jun C H. Sparse tensor decomposition for multi-task interaction selection. In: Proceedings of 2019 IEEE International Conference on Big Knowledge. 2019, 105–114
Huang F, Qiu Y, Li Q, Liu S, Ni F. Predicting drug-disease associations via multi-task learning based on collective matrix factorization. Frontiers in Bioengineering and Biotechnology, 2020, 8: 218
Zhang Y, Zhang Y, Wang W. Deep multi-task learning via generalized tensor trace norm. 2002, arXiv preprint arXiv: 2002.04799
Chen Z, Lei H, Zhao Y, Huang Z, Xiao X, Lei Y, Tan E L, Lei B. Template-oriented multi-task sparse low-rank learning for parkinson’s diseases diagnosis. In: Proceedings of the 3rd International Workshop on PRedictive Intelligence in MEdicine. 2020, 178–187
Wu X, Zhang X, Cen Y. Multi-task joint sparse and low-rank representation target detection for hyperspectral image. IEEE Geoscience and Remote Sensing Letters, 2019, 16(11): 1756–1760
Zhang Y, Yang Q. A survey on multi-task learning. IEEE Transactions on Knowledge and Data Engineering, 2021
He K, Chen X, Xie S, Li Y, Dollár P, Girshick R. Masked autoencoders are scalable vision learners. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 15979–15988
Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. In: Proceedings of the 3rd International Conference on Learning Representations. 2015, 1556
Jaworek-Korjakowska J, Kleczek P, Gorgon M. Melanoma thickness prediction based on convolutional neural network with VGG-19 model transfer learning. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 2019, 2748–2756
Gao Y, Ma J, Zhao M, Liu W, Yuille A L. Nddr-CNN: layerwise feature fusing in multi-task CNNs by neural discriminative dimensionality reduction. In: Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019, 3200–3209
Heuer F, Mantowsky S, Bukhari S S, Schneider G. Multitask-centerNet (MCN): efficient and diverse multitask learning using an anchor free approach. In: Proceedings of 2021 IEEE/CVF International Conference on Computer Vision Workshops. 2021, 997–1005
Acknowledgements
This work was supported by the NSFC (Grant No. 61672281), and the Key Program of NSFC (No. 61732006).
Author information
Authors and Affiliations
Corresponding author
Additional information
Quan Feng received his B.S. degree in Information and Computing Science from Nanchang Hangkong University, China in 2008. In 2011, he completed his MSc degree in computer science and technique at Nanchang Hangkong University, China. He is currently a Ph.D candidate at the Key Laboratory of Model Analysis and Machine Intelligence, Ministry of Industry and Information Technology, Nanjing University of Aeronautics and Astronautics, China. His research interests include pattern recognition, machine learning and multi-task learning.
Songcan Chen received the BS degree in mathematics from Hangzhou University (now merged into Zhejiang University), China in 1983, and the MS degree in computer applications from Shanghai Jiao Tong University, China in 1985, and then worked with NUAA in January 1986. He received the PhD degree in communication and information systems from the Nanjing University of Aeronautics and Astronautics (NUAA), China in 1997. Since 1998, as a fulltime professor, he has been with the College of Computer Science & Technology, NUAA. His research interests include pattern recognition, machine learning, and neural computing. He is also an IAPR and CCAI Fellow.
Electronic Supplementary Material
Rights and permissions
About this article
Cite this article
Feng, Q., Chen, S. Learning multi-tasks with inconsistent labels by using auxiliary big task. Front. Comput. Sci. 17, 175342 (2023). https://doi.org/10.1007/s11704-022-2251-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11704-022-2251-x