Abstract
Source-free unsupervised domain adaptation (SFUDA) aims to achieve target domain predictions through a source model instead of source data. One of the representative ideas for the SFUDA problem is to apply a self-supervised pseudo-labeling strategy (SSPL) to achieve target domain adaptation, but it is prone to being plagued by noisy labels, which can lead to negative transfer. Therefore, many methods attempt to improve the SSPL and leverage confidence scores to weaken the impact of low-confidence samples on the model, which are potentially noisy samples. However, they are unable to completely overcome the problem of noisy labels because the pseudo-labels of high-confidence samples may also be incorrect. Besides, they rarely allow low-confidence samples to be added to training, which can lead to sample selection bias and thus limit the model generalization ability. In this work, we propose information re-exploitation based on confidence scores (RECS) for the SFUDA problem, in which we rethink the information brought by confidence scores and take advantage of them to solve the shortcomings of the improved SSPL. Specifically, we realize cross-domain target adaptation by the symmetric SSPL with dual denoising, and reduce the intra-domain distribution discrepancy by the discriminative class-balanced feature alignment. In this way, the model robustness and generalization are enhanced. Extensive experiments conducted on three standard datasets have demonstrated the effectiveness and superiority of our proposed method. The code is available at https://github.com/lingyuxuan1234/RECS.
Similar content being viewed by others
Data availability
All data generated or analyzed during this study are included in this published article [and its supplementary information files].
References
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778
Perez L, Wang J (2017) The effectiveness of data augmentation in image classification using deep learning. arXiv preprint arXiv:1712.04621
Cai L, Gao J, Zhao D (2020) A review of the application of deep learning in medical image classification and segmentation. Ann Transl Med 8(11):213–226
Lateef F, Ruichek Y (2019) Survey on semantic segmentation using deep learning techniques. Neurocomputing 338:321–348
Liu L, Ouyang W, Wang X, Fieguth P, Chen J, Liu X, Pietikäinen M (2020) Deep learning for generic object detection: a survey. Int J Comput Vision 128:261–318
Zhao Z-Q, Zheng P, Xu S-T, Wu X (2019) Object detection with deep learning: a review. IEEE Trans Neural Netw Learn Syst 30(11):3212–3232
Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359
Wilson G, Cook DJ (2020) A survey of unsupervised deep domain adaptation. ACM Trans Intell Syst Technol (TIST) 11(5):1–46
Long M, Cao Y, Wang J, Jordan M (2015) Learning transferable features with deep adaptation networks. In: International conference on machine learning, pp. 97–105. PMLR
Sun B, Saenko K (2016) Deep coral: Correlation alignment for deep domain adaptation. In: Computer Vision–ECCV 2016 workshops: Amsterdam, The Netherlands, Oct 8–10 and 15–16, 2016, Proceedings, part III 14, pp. 443–450. Springer
Ganin Y, Ustinova E, Ajakan H, Germain P, Larochelle H, Laviolette F, Marchand M, Lempitsky V (2016) Domain-adversarial training of neural networks. J Mach Learn Res 17(1):2006–2030
Tzeng E, Hoffman J, Saenko K, Darrell T (2017) Adversarial discriminative domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7167–7176
Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, Zhang C, Agarwal S, Slama K, Ray A et al (2022) Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 35:27730–27744
Liang J, Hu D, Feng J (2020) Do we really need to access the source data? Source hypothesis transfer for unsupervised domain adaptation. In: International conference on machine learning, pp. 6028–6039. PMLR
Qiu Z, Zhang Y, Lin H, Niu S, Liu Y, Du Q, Tan M (2021) Source-free domain adaptation via avatar prototype generation and adaptation. arXiv preprint arXiv:2106.15326
Chen W, Lin L, Yang S, Xie D, Pu S, Zhuang Y (2022) Self-supervised noisy label learning for source-free unsupervised domain adaptation. In: 2022 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp. 10185–10192. IEEE
Lee J, Jun, D, Yim J, Yoon S (2022) Confidence score for source-free unsupervised domain adaptation. In: International conference on machine learning, pp. 12365–12377. PMLR
Tian Q, Peng S, Ma T (2023) Source-free unsupervised domain adaptation with trusted pseudo samples. ACM Trans Intell Syst Technol 14(2):1–17
Zhang Z, Chen W, Cheng H, Li Z, Li S, Lin L, Li G (2022) Divide and contrast: source-free domain adaptation via adaptive contrastive learning. Adv Neural Inf Process Syst 35:5137–5149
Zhang H, Cisse M, Dauphin YN, Lopez-Paz D (2017) mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412
Lee C-Y, Batra T, Baig MH, Ulbricht D (2019) Sliced wasserstein discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10285–10295
Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T (2014) Deep domain confusion: maximizing for domain invariance. arXiv preprint arXiv:1412.3474
Ganin Y, Lempitsky V (2015) Unsupervised domain adaptation by backpropagation. In: International conference on machine learning, pp. 1180–1189. PMLR
Tian Q, Zhu Y, Sun H, Chen S, Yin H (2022) Unsupervised domain adaptation through dynamically aligning both the feature and label spaces. IEEE Trans Circuits Syst Video Technol 32(12):8562–8573
Ben-David S, Blitzer J, Crammer K, Kulesza A, Pereira F, Vaughan JW (2010) A theory of learning from different domains. Mach Learn 79:151–175
Courty N, Flamary R, Habrard A, Rakotomamonjy A (2017) Joint distribution optimal transportation for domain adaptation. Adv Neural Inform Process Syst 30:3730–3739
Redko I, Habrard A, Sebban M (2017) Theoretical analysis of domain adaptation with optimal transport. In: Machine learning and knowledge discovery in databases: European conference, ECML PKDD 2017, Skopje, Macedonia, Sept 18–22, 2017, Proceedings, Part II 10, pp. 737–753. Springer
Damodaran BB, Kellenberger B, Flamary R, Tuia D, Courty N (2018) Deepjdot: deep joint distribution optimal transport for unsupervised domain adaptation. In: Proceedings of the European conference on computer vision (ECCV), pp. 447–463
Vallender S (1974) Calculation of the Wasserstein distance between probability distributions on the line. Theor Probab Appl 18(4):784–786
Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139–144
Saito K, Watanabe K, Ushiku Y, Harada T (2018) Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3723–3732
Li S, Lv F, Xie B, Liu CH, Liang J, Qin C (2021) Bi-classifier determinacy maximization for unsupervised domain adaptation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35, pp. 8455–8464
Dai S, Cheng Y, Zhang Y, Gan Z, Liu J, Carin L (2020) Contrastively smoothed class alignment for unsupervised domain adaptation. In: Proceedings of the Asian conference on computer vision
Thota M, Leontidis G (2021) Contrastive domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 2209–2218
Wang R, Wu Z, Weng Z, Chen J, Qi G-J, Jiang Y-G (2022) Cross-domain contrastive learning for unsupervised domain adaptation. IEEE Trans Multimed 25:405
Kurmi VK, Subramanian VK, Namboodiri VP (2021) Domain impression: a source data free domain adaptation method. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp. 615–625
Tian Q, Ma C, Zhang F-Y, Peng S, Xue H (2021) Source-free unsupervised domain adaptation with sample transport learning. J Comput Sci Technol 36(3):606–616
Ding Y, Sheng L, Liang J, Zheng A, He R (2023) Proxymix: Proxy-based mixup training with label refinery for source-free domain adaptation. Neural Netw 167:92–103
Yang S, Weijer J, Herranz L, Jui S et al (2021) Exploiting the intrinsic neighborhood structure for source-free domain adaptation. Adv Neural Inf Process Syst 34:29393–29405
Yang S, Wang Y, Van De Weijer J, Herranz L, Jui S (2021) Generalized source-free domain adaptation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 8978–8987
Yang S, Wang Y, Wang K, Jui S et al (2022) Attracting and dispersing: a simple approach for source-free domain adaptation. Adv Neural Inform Process Syst 35:5802
Yang S, Wang Y, Weijer J, Herranz L, Jui S, Yang J (2023) Trust your good friends: source-free domain adaptation by reciprocal neighborhood clustering. In: IEEE Transactions on pattern analysis and machine intelligence
Pei J, Jiang Z, Men A, Chen L, Liu Y, Chen Q (2023) Uncertainty-induced transferability representation for source-free unsupervised domain adaptation. IEEE Trans Image Process 32:2033–2048
Chu T, Liu Y, Deng J, Li W, Duan L (2022) Denoised maximum classifier discrepancy for source-free unsupervised domain adaptation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 36, pp. 472–480
Roy S, Trapp M, Pilzer A, Kannala J, Sebe N, Ricci E, Solin A (2022) Uncertainty-guided source-free domain adaptation. In: European conference on computer vision, pp. 537–555. Springer
Reynolds DA et al (2009) Gaussian mixture models. Encycl Biomet 741:659–663
Kang G, Jiang L, Yang Y, Hauptmann AG (2019) Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4893–4902
Zhang C, Bengio S, Hardt M, Recht B, Vinyals O (2021) Understanding deep learning (still) requires rethinking generalization. Commun ACM 64(3):107–115
Wang Y, Ma X, Chen Z, Luo Y, Yi J, Bailey J (2019) Symmetric cross entropy for robust learning with noisy labels. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 322–330
Berlinet A, Thomas-Agnan C (2011) Reproducing Kernel Hilbert spaces in probability and statistics. Springer, Cham
Saenko K, Kulis B, Fritz M, Darrell T (2010) Adapting visual category models to new domains. In: Computer Vision–ECCV 2010: 11th European conference on computer vision, Heraklion, Crete, Greece, Sept 5–11, 2010, Proceedings, Part IV 11, pp. 213–226. Springer
Venkateswara H, Eusebio J, Chakraborty S, Panchanathan S (2017) Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5018–5027
Peng X, Usman B, Kaushik N, Hoffman J, Wang D, Saenko K (2017) Visda: The visual domain adaptation challenge. arXiv preprint arXiv:1710.06924
Zhou L, Li N, Ye M, Zhu X, Tang S (2023) Source-free domain adaptation with class prototype discovery. Pattern Recognit 145:109974
Long M, Cao Z, Wang J, Jordan MI (2018) Conditional adversarial domain adaptation. Adv Neural Inform Process Syst 31:1647–1657
Lu Z, Yang Y, Zhu X, Liu C, Song Y-Z, Xiang T (2020) Stochastic classifiers for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pp. 9111–9120
Wang Q, Breckon T (2020) Unsupervised domain adaptation via structured prediction based selective pseudo-labeling. In: Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp. 6243–6250
Zhou L, Ye M, Zhu X, Li S, Liu Y (2022) Class discriminative adversarial learning for unsupervised domain adaptation. In: Proceedings of the 30th ACM international conference on multimedia, pp. 4318–4326
Müller R, Kornblith S, Hinton GE (2019) When does label smoothing help? Adv Neural Inform Process Syst 32:4696–4705
Nagy ZK, Braatz RD (2007) Distributional uncertainty analysis using power series and polynomial chaos expansions. J Process Control 17(3):229–240
Grandvalet Y, Bengio Y (2004) Semi-supervised learning by entropy minimization. Adv Neural Inform Process Syst 17:529–536
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant 62176128, the Natural Science Foundation of Jiangsu Province under Grant BK20231143, the Open Projects Program of State Key Laboratory for Novel Software Technology of Nanjing University under Grant KFKT2022B06, the Fundamental Research Funds for the Central Universities No. NJ2022028, the Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund, as well as the Qing Lan Project of Jiangsu Province.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Tian, Q., Sun, C. Rethinking confidence scores for source-free unsupervised domain adaptation. Neural Comput & Applic (2024). https://doi.org/10.1007/s00521-024-09867-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00521-024-09867-9