Skip to main content
Log in

Rethinking confidence scores for source-free unsupervised domain adaptation

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Source-free unsupervised domain adaptation (SFUDA) aims to achieve target domain predictions through a source model instead of source data. One of the representative ideas for the SFUDA problem is to apply a self-supervised pseudo-labeling strategy (SSPL) to achieve target domain adaptation, but it is prone to being plagued by noisy labels, which can lead to negative transfer. Therefore, many methods attempt to improve the SSPL and leverage confidence scores to weaken the impact of low-confidence samples on the model, which are potentially noisy samples. However, they are unable to completely overcome the problem of noisy labels because the pseudo-labels of high-confidence samples may also be incorrect. Besides, they rarely allow low-confidence samples to be added to training, which can lead to sample selection bias and thus limit the model generalization ability. In this work, we propose information re-exploitation based on confidence scores (RECS) for the SFUDA problem, in which we rethink the information brought by confidence scores and take advantage of them to solve the shortcomings of the improved SSPL. Specifically, we realize cross-domain target adaptation by the symmetric SSPL with dual denoising, and reduce the intra-domain distribution discrepancy by the discriminative class-balanced feature alignment. In this way, the model robustness and generalization are enhanced. Extensive experiments conducted on three standard datasets have demonstrated the effectiveness and superiority of our proposed method. The code is available at https://github.com/lingyuxuan1234/RECS.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Data availability

All data generated or analyzed during this study are included in this published article [and its supplementary information files].

Notes

  1. http://ai.bu.edu/adaptation.html.

  2. https://www.hemanthdv.org/officeHomeDataset.html.

  3. http://ai.bu.edu/visda-2017.

References

  1. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778

  2. Perez L, Wang J (2017) The effectiveness of data augmentation in image classification using deep learning. arXiv preprint arXiv:1712.04621

  3. Cai L, Gao J, Zhao D (2020) A review of the application of deep learning in medical image classification and segmentation. Ann Transl Med 8(11):213–226

    Article  Google Scholar 

  4. Lateef F, Ruichek Y (2019) Survey on semantic segmentation using deep learning techniques. Neurocomputing 338:321–348

    Article  Google Scholar 

  5. Liu L, Ouyang W, Wang X, Fieguth P, Chen J, Liu X, Pietikäinen M (2020) Deep learning for generic object detection: a survey. Int J Comput Vision 128:261–318

    Article  Google Scholar 

  6. Zhao Z-Q, Zheng P, Xu S-T, Wu X (2019) Object detection with deep learning: a review. IEEE Trans Neural Netw Learn Syst 30(11):3212–3232

    Article  Google Scholar 

  7. Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359

    Article  Google Scholar 

  8. Wilson G, Cook DJ (2020) A survey of unsupervised deep domain adaptation. ACM Trans Intell Syst Technol (TIST) 11(5):1–46

    Article  Google Scholar 

  9. Long M, Cao Y, Wang J, Jordan M (2015) Learning transferable features with deep adaptation networks. In: International conference on machine learning, pp. 97–105. PMLR

  10. Sun B, Saenko K (2016) Deep coral: Correlation alignment for deep domain adaptation. In: Computer Vision–ECCV 2016 workshops: Amsterdam, The Netherlands, Oct 8–10 and 15–16, 2016, Proceedings, part III 14, pp. 443–450. Springer

  11. Ganin Y, Ustinova E, Ajakan H, Germain P, Larochelle H, Laviolette F, Marchand M, Lempitsky V (2016) Domain-adversarial training of neural networks. J Mach Learn Res 17(1):2006–2030

    MathSciNet  Google Scholar 

  12. Tzeng E, Hoffman J, Saenko K, Darrell T (2017) Adversarial discriminative domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7167–7176

  13. Ouyang L, Wu J, Jiang X, Almeida D, Wainwright C, Mishkin P, Zhang C, Agarwal S, Slama K, Ray A et al (2022) Training language models to follow instructions with human feedback. Adv Neural Inf Process Syst 35:27730–27744

    Google Scholar 

  14. Liang J, Hu D, Feng J (2020) Do we really need to access the source data? Source hypothesis transfer for unsupervised domain adaptation. In: International conference on machine learning, pp. 6028–6039. PMLR

  15. Qiu Z, Zhang Y, Lin H, Niu S, Liu Y, Du Q, Tan M (2021) Source-free domain adaptation via avatar prototype generation and adaptation. arXiv preprint arXiv:2106.15326

  16. Chen W, Lin L, Yang S, Xie D, Pu S, Zhuang Y (2022) Self-supervised noisy label learning for source-free unsupervised domain adaptation. In: 2022 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp. 10185–10192. IEEE

  17. Lee J, Jun, D, Yim J, Yoon S (2022) Confidence score for source-free unsupervised domain adaptation. In: International conference on machine learning, pp. 12365–12377. PMLR

  18. Tian Q, Peng S, Ma T (2023) Source-free unsupervised domain adaptation with trusted pseudo samples. ACM Trans Intell Syst Technol 14(2):1–17

    Google Scholar 

  19. Zhang Z, Chen W, Cheng H, Li Z, Li S, Lin L, Li G (2022) Divide and contrast: source-free domain adaptation via adaptive contrastive learning. Adv Neural Inf Process Syst 35:5137–5149

    Google Scholar 

  20. Zhang H, Cisse M, Dauphin YN, Lopez-Paz D (2017) mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412

  21. Lee C-Y, Batra T, Baig MH, Ulbricht D (2019) Sliced wasserstein discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10285–10295

  22. Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T (2014) Deep domain confusion: maximizing for domain invariance. arXiv preprint arXiv:1412.3474

  23. Ganin Y, Lempitsky V (2015) Unsupervised domain adaptation by backpropagation. In: International conference on machine learning, pp. 1180–1189. PMLR

  24. Tian Q, Zhu Y, Sun H, Chen S, Yin H (2022) Unsupervised domain adaptation through dynamically aligning both the feature and label spaces. IEEE Trans Circuits Syst Video Technol 32(12):8562–8573

    Article  Google Scholar 

  25. Ben-David S, Blitzer J, Crammer K, Kulesza A, Pereira F, Vaughan JW (2010) A theory of learning from different domains. Mach Learn 79:151–175

    Article  MathSciNet  Google Scholar 

  26. Courty N, Flamary R, Habrard A, Rakotomamonjy A (2017) Joint distribution optimal transportation for domain adaptation. Adv Neural Inform Process Syst 30:3730–3739

    Google Scholar 

  27. Redko I, Habrard A, Sebban M (2017) Theoretical analysis of domain adaptation with optimal transport. In: Machine learning and knowledge discovery in databases: European conference, ECML PKDD 2017, Skopje, Macedonia, Sept 18–22, 2017, Proceedings, Part II 10, pp. 737–753. Springer

  28. Damodaran BB, Kellenberger B, Flamary R, Tuia D, Courty N (2018) Deepjdot: deep joint distribution optimal transport for unsupervised domain adaptation. In: Proceedings of the European conference on computer vision (ECCV), pp. 447–463

  29. Vallender S (1974) Calculation of the Wasserstein distance between probability distributions on the line. Theor Probab Appl 18(4):784–786

    Article  Google Scholar 

  30. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139–144

    Article  MathSciNet  Google Scholar 

  31. Saito K, Watanabe K, Ushiku Y, Harada T (2018) Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3723–3732

  32. Li S, Lv F, Xie B, Liu CH, Liang J, Qin C (2021) Bi-classifier determinacy maximization for unsupervised domain adaptation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 35, pp. 8455–8464

  33. Dai S, Cheng Y, Zhang Y, Gan Z, Liu J, Carin L (2020) Contrastively smoothed class alignment for unsupervised domain adaptation. In: Proceedings of the Asian conference on computer vision

  34. Thota M, Leontidis G (2021) Contrastive domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 2209–2218

  35. Wang R, Wu Z, Weng Z, Chen J, Qi G-J, Jiang Y-G (2022) Cross-domain contrastive learning for unsupervised domain adaptation. IEEE Trans Multimed 25:405

    Google Scholar 

  36. Kurmi VK, Subramanian VK, Namboodiri VP (2021) Domain impression: a source data free domain adaptation method. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp. 615–625

  37. Tian Q, Ma C, Zhang F-Y, Peng S, Xue H (2021) Source-free unsupervised domain adaptation with sample transport learning. J Comput Sci Technol 36(3):606–616

    Article  Google Scholar 

  38. Ding Y, Sheng L, Liang J, Zheng A, He R (2023) Proxymix: Proxy-based mixup training with label refinery for source-free domain adaptation. Neural Netw 167:92–103

    Article  Google Scholar 

  39. Yang S, Weijer J, Herranz L, Jui S et al (2021) Exploiting the intrinsic neighborhood structure for source-free domain adaptation. Adv Neural Inf Process Syst 34:29393–29405

    Google Scholar 

  40. Yang S, Wang Y, Van De Weijer J, Herranz L, Jui S (2021) Generalized source-free domain adaptation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 8978–8987

  41. Yang S, Wang Y, Wang K, Jui S et al (2022) Attracting and dispersing: a simple approach for source-free domain adaptation. Adv Neural Inform Process Syst 35:5802

    Google Scholar 

  42. Yang S, Wang Y, Weijer J, Herranz L, Jui S, Yang J (2023) Trust your good friends: source-free domain adaptation by reciprocal neighborhood clustering. In: IEEE Transactions on pattern analysis and machine intelligence

  43. Pei J, Jiang Z, Men A, Chen L, Liu Y, Chen Q (2023) Uncertainty-induced transferability representation for source-free unsupervised domain adaptation. IEEE Trans Image Process 32:2033–2048

    Article  Google Scholar 

  44. Chu T, Liu Y, Deng J, Li W, Duan L (2022) Denoised maximum classifier discrepancy for source-free unsupervised domain adaptation. In: Proceedings of the AAAI conference on artificial intelligence, vol. 36, pp. 472–480

  45. Roy S, Trapp M, Pilzer A, Kannala J, Sebe N, Ricci E, Solin A (2022) Uncertainty-guided source-free domain adaptation. In: European conference on computer vision, pp. 537–555. Springer

  46. Reynolds DA et al (2009) Gaussian mixture models. Encycl Biomet 741:659–663

    Article  Google Scholar 

  47. Kang G, Jiang L, Yang Y, Hauptmann AG (2019) Contrastive adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4893–4902

  48. Zhang C, Bengio S, Hardt M, Recht B, Vinyals O (2021) Understanding deep learning (still) requires rethinking generalization. Commun ACM 64(3):107–115

    Article  Google Scholar 

  49. Wang Y, Ma X, Chen Z, Luo Y, Yi J, Bailey J (2019) Symmetric cross entropy for robust learning with noisy labels. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 322–330

  50. Berlinet A, Thomas-Agnan C (2011) Reproducing Kernel Hilbert spaces in probability and statistics. Springer, Cham

    Google Scholar 

  51. Saenko K, Kulis B, Fritz M, Darrell T (2010) Adapting visual category models to new domains. In: Computer Vision–ECCV 2010: 11th European conference on computer vision, Heraklion, Crete, Greece, Sept 5–11, 2010, Proceedings, Part IV 11, pp. 213–226. Springer

  52. Venkateswara H, Eusebio J, Chakraborty S, Panchanathan S (2017) Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5018–5027

  53. Peng X, Usman B, Kaushik N, Hoffman J, Wang D, Saenko K (2017) Visda: The visual domain adaptation challenge. arXiv preprint arXiv:1710.06924

  54. Zhou L, Li N, Ye M, Zhu X, Tang S (2023) Source-free domain adaptation with class prototype discovery. Pattern Recognit 145:109974

    Article  Google Scholar 

  55. Long M, Cao Z, Wang J, Jordan MI (2018) Conditional adversarial domain adaptation. Adv Neural Inform Process Syst 31:1647–1657

    Google Scholar 

  56. Lu Z, Yang Y, Zhu X, Liu C, Song Y-Z, Xiang T (2020) Stochastic classifiers for unsupervised domain adaptation. In: Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, pp. 9111–9120

  57. Wang Q, Breckon T (2020) Unsupervised domain adaptation via structured prediction based selective pseudo-labeling. In: Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp. 6243–6250

  58. Zhou L, Ye M, Zhu X, Li S, Liu Y (2022) Class discriminative adversarial learning for unsupervised domain adaptation. In: Proceedings of the 30th ACM international conference on multimedia, pp. 4318–4326

  59. Müller R, Kornblith S, Hinton GE (2019) When does label smoothing help? Adv Neural Inform Process Syst 32:4696–4705

    Google Scholar 

  60. Nagy ZK, Braatz RD (2007) Distributional uncertainty analysis using power series and polynomial chaos expansions. J Process Control 17(3):229–240

    Article  Google Scholar 

  61. Grandvalet Y, Bengio Y (2004) Semi-supervised learning by entropy minimization. Adv Neural Inform Process Syst 17:529–536

    Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China under Grant 62176128, the Natural Science Foundation of Jiangsu Province under Grant BK20231143, the Open Projects Program of State Key Laboratory for Novel Software Technology of Nanjing University under Grant KFKT2022B06, the Fundamental Research Funds for the Central Universities No. NJ2022028, the Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund, as well as the Qing Lan Project of Jiangsu Province.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Qing Tian.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tian, Q., Sun, C. Rethinking confidence scores for source-free unsupervised domain adaptation. Neural Comput & Applic (2024). https://doi.org/10.1007/s00521-024-09867-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s00521-024-09867-9

Keywords

Navigation