Skip to main content

Open Set Domain Adaptation with Entropy Minimization

  • Conference paper
  • First Online:
Pattern Recognition and Computer Vision (PRCV 2020)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 12307))

Included in the following conference series:

Abstract

Unsupervised domain adaptation has achieved great progress in the past few years. Nevertheless, most existing methods work in the so-called closed-set scenario, assuming that the classes depicted by the target samples are exactly the same as those of the source domain. In this paper, we tackle the more challenging scenario of open set domain adaptation with a novel end-to-end training approach, where the samples of unknown class can be present in the target domain. Our method employs entropy minimization for performing unsupervised domain adaptation, where unknown samples are aggressively used in training by forcing the classifier to output the probability of 0.5 on the unknown class. Experimental evidence demonstrates that our approach significantly outperforms the state-of-the-art in open set domain adaptation.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Baktashmotlagh, M., Faraki, M., Drummond, T., Salzmann, M.: Learning factorized representations for open-set domain adaptation. In: ICLR (2019)

    Google Scholar 

  2. Ben-David, S., Blitzer, J., Crammer, K., Pereira, F.: Analysis of representations for domain adaptation. In: NIPS (2007)

    Google Scholar 

  3. Blitzer, J., Crammer, K., Kulesza, A., Pereira, F., Wortman, J.: Learning bounds for domain adaptation. In: NIPS (2008)

    Google Scholar 

  4. Busto, P.P., Gall, J.: Open set domain adaptation. In: ICCV (2017)

    Google Scholar 

  5. Donahue, J., et al.: DeCAF: a deep convolutional activation feature for generic visual recognition. In: ICML (2013)

    Google Scholar 

  6. Fernando, B., Habrard, A., Sebban, M., Tuytelaars, T.: Unsupervised visual domain adaptation using subspace alignment. In: ICCV (2013)

    Google Scholar 

  7. Ganin, Y., Lempitsky, V.S.: Unsupervised domain adaptation by backpropagation. In: ICML (2015)

    Google Scholar 

  8. Gong, B., Shi, Y., Sha, F., Grauman, K.: Geodesic flow kernel for unsupervised domain adaptation. In: CVPR (2012)

    Google Scholar 

  9. Haeusser, P., Frerix, T., Mordvintsev, A., Cremers, D.: Associative domain adaptation. In: ICCV (2017)

    Google Scholar 

  10. Ian, G., et al.: Generative adversarial nets. In: NIPS (2014)

    Google Scholar 

  11. Jain, L.P., Scheirer, W.J., Boult, T.E.: Multi-class open set recognition using probability of inclusion. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8691, pp. 393–409. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10578-9_26

    Chapter  Google Scholar 

  12. Konstantinos, B., George, T., Nathan, S., Dilip, K., Dumitru, E.: Domain separation networks. In: NIPS (2016)

    Google Scholar 

  13. Long, M., Cao, Y., Wang, J., Jordan, M.I.: Learning transferable features with deep adaptation networks. In: ICML (2015)

    Google Scholar 

  14. Chen, M., Weinberger, K.Q., Blitzer, J.: Co-training for domain adaptation. In: NIPS (2011)

    Google Scholar 

  15. Pan, S., Tsang, I., Kwok, J., Yang, Q.: Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 22, 199–210 (2011)

    Article  Google Scholar 

  16. Pietro, M., Jacopo, C., Vittorio, M.: Minimal-entropy correlation alignment for unsupervised deep domain adaptation. In: ICLR (2018)

    Google Scholar 

  17. Saito, K., Yamamoto, S., Ushiku, Y., Harada, T.: Open set domain adaptation by backpropagation. In: CVPR (2018)

    Google Scholar 

  18. Sun, B., Saenko, K.: Deep coral: correlation alignment for deep domain adaptation. In: ICCV Workshop on Transferring and Adapting Source Knowledge in Computer Vision (2016)

    Google Scholar 

  19. Tommasi, T., Tuytelaars, T.: A testbed for cross-dataset analysis. In: Agapito, L., Bronstein, M.M., Rother, C. (eds.) ECCV 2014. LNCS, vol. 8927, pp. 18–31. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-16199-0_2

    Chapter  Google Scholar 

  20. Tzeng, E., Hoffman, J., Saenko, K., Darrell, T.: Adversarial discriminative domain adaptation. In: CVPR (2017)

    Google Scholar 

  21. Tzeng, E., Hoffman, J., Zhang, N., Saenko, K.: Deep domain confusion: maximizing for domain invariance. arXiv:1412.3474 (2014)

  22. Wu, X., Zhang, S., Zhou, Q., Yang, Z., Zhao, C., Latecki, L.J.: Minimal-entropy diversity maximization for unsupervised domain adaptation. arXiv:2002.01690 (2020)

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaofu Wu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Wu, X., Cheng, L., Zhang, S. (2020). Open Set Domain Adaptation with Entropy Minimization. In: Peng, Y., et al. Pattern Recognition and Computer Vision. PRCV 2020. Lecture Notes in Computer Science(), vol 12307. Springer, Cham. https://doi.org/10.1007/978-3-030-60636-7_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-60636-7_3

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-60635-0

  • Online ISBN: 978-3-030-60636-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics