Skip to main content
Log in

Domain adaptation for object recognition using subspace sampling demons

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Manually labeling data for training machine learning models is time-consuming and expensive. Therefore, it is often necessary to apply models built in one domain to a new domain. However, existing approaches do not evaluate the quality of intermediate features that are learned in the process of transferring from the source domain to the target domain, which results in the potential for sub-optimal features. Also, transfer learning models in existing work do not provide optimal results for a new domain. In this paper, we first propose a fast subspace sampling demons (SSD) method to learn intermediate subspace features from two domains and then evaluate the quality of the learned features. To show the applicability of our model, we test our model using a synthetic dataset as well as several benchmark datasets. Extensive experiments demonstrate significant improvements in classification accuracy over the state of the art.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Notes

  1. Inception-ResNet-v2 is a proposed architecture that performs highly on the ImageNet object recognition task.

References

  1. Belkin M, Niyogi P (2004) Semi-supervised learning on Riemannian manifolds. Mach Learn 56(1-3):209–239

    Article  Google Scholar 

  2. Ben-David S, Blitzer J, Crammer K, Kulesza A, Pereira F, Vaughan JW (2010) A theory of learning from different domains. Mach Learn 79 (1-2):151–175

    Article  MathSciNet  Google Scholar 

  3. Ben-David S, Blitzer J, Crammer K, Pereira F (2007) Analysis of representations for domain adaptation. In: Advances in neural information processing systems, pp 137–144

  4. Ben-David S, Lu T, Luu T, Pȧl D. (2010) Impossibility theorems for domain adaptation. In: International conference on artificial intelligence and statistics, pp 129–136

  5. Bergamo A, Torresani L (2010) Exploiting weakly-labeled web images to improve object classification: a domain adaptation approach. In: Advances in neural information processing systems, pp 181–189

  6. Blitzer J, Crammer K, Kulesza A, Pereira F, Wortman J (2008) Learning bounds for domain adaptation. In: Advances in neural information processing systems, pp 129–136

  7. Blitzer J, Dredze M, Pereira F (2007) Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In: Proceedings of the 45th annual meeting of the association of computational linguistics, pp 440–447

  8. Chen C, Chen Z, Jiang B, Jin X (2019) Joint domain alignment and discriminative feature learning for unsupervised deep domain adaptation. In: Proceedings of AAAI

  9. Chen L, Zhang H, Xiao J, Liu W, Chang SF (2018) Zero-shot visual recognition using semantics-preserving adversarial embedding network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, vol 2

  10. Fletcher PT, Lu C, Pizer SM, Joshi S (2004) Principal geodesic analysis for the study of nonlinear statistics of shape. IEEE Trans Med Imaging 23 (8):995–1005

    Article  Google Scholar 

  11. Ghifary M, Kleijn WB, Zhang M (2014) Domain adaptive neural networks for object recognition. In: Pacific rim international conference on artificial intelligence. Springer, New York, pp 898–904

  12. Gong B, Shi Y, Sha F, Grauman K (2012) Geodesic flow kernel for unsupervised domain adaptation. In: IEEE conference on computer vision and pattern recognition (CVPR), IEEE, pp 2066–2073

  13. Gopalan R, Li R, Chellappa R (2011) Domain adaptation for object recognition: an unsupervised approach. In: IEEE international conference on computer vision (ICCV), IEEE, pp 999–1006

  14. Jiang M, Huang W, Huang Z, Yen GG (2017) Integration of global and local metrics for domain adaptation learning via dimensionality reduction. IEEE Trans Cybern 47(1):38–51

    Article  Google Scholar 

  15. Jolliffe I (2011) Principal component analysis. In: International encyclopedia of statistical science. Springer, New York, pp 1094–1096

  16. Long M, Cao Y, Wang J, Jordan MI (2015) Learning transferable features with deep adaptation networks. arXiv:1502.02791

  17. Long M, Cao Z, Wang J, Jordan MI (2018) Conditional adversarial domain adaptation. In: Advances in neural information processing systems, pp 1647–1657

  18. Long M, Wang J, Ding G, Sun J, Yu PS (2013) Transfer feature learning with joint distribution adaptation. In: Proceedings of the IEEE international conference on computer vision, pp 2200–2207

  19. Long M, Wang J, Ding G, Sun J, Yu PS (2014) Transfer joint matching for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1410–1417

  20. Long M, Zhu H, Wang J, Jordan MI (2016) Unsupervised domain adaptation with residual transfer networks. In: Advances in neural information processing systems, pp 136–144

  21. Long M, Zhu H, Wang J, Jordan MI (2017) Deep transfer learning with joint adaptation networks. In: Proceedings of the 34th international conference on machine learning, JMLR.org, vol 70, pp 2208–2217

  22. Maaten LVD, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(Nov):2579–2605

    MATH  Google Scholar 

  23. Mansour Y, Mohri M, Rostamizadeh A (2009) Domain adaptation:, Learning bounds and algorithms. arXiv:0902.3430

  24. Pan SJ, Tsang IW, Kwok JT, Yang Q (2011) Domain adaptation via transfer component analysis. IEEE Trans Neural Netw 22(2):199–210

    Article  Google Scholar 

  25. Pan SJ, Yang Q, et al. (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359

    Article  Google Scholar 

  26. Rahman MM, Fookes C, Baktashmotlagh M, Sridharan S (2020) On minimum discrepancy estimation for deep domain adaptation. In: Domain adaptation for visual understanding. Springer, New York, pp 81–94

  27. Saenko K, Kulis B, Fritz M, Darrell T (2010) Adapting visual category models to new domains. In: European conference on computer vision. Springer, New York, pp 213–226

  28. Sun B, Saenko K (2016) Deep coral: Correlation alignment for deep domain adaptation. In: European conference on computer vision. Springer, New York, pp 443–450

  29. Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: Thirty-first AAAI conference on artificial intelligence

  30. Thirion JP (1998) Image matching as a diffusion process: an analogy with maxwell’s demons. Med Image Anal 2(3):243–260

    Article  Google Scholar 

  31. Tzeng E, Hoffman J, Saenko K, Darrell T (2017) Adversarial discriminative domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7167–7176

  32. Tzeng E, Hoffman J, Zhang N, Saenko K, Darrell T (2014) Deep domain confusion:, Maximizing for domain invariance. arXiv:1412.3474

  33. Venkateswara H, Eusebio J, Chakraborty S, Panchanathan S (2017) Deep hashing network for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5018–5027

  34. Wang C, Mahadevan S (2009) Manifold alignment without correspondence. In: IJCAI, vol 2, p 3

  35. Wang J, Chen Y, Hao S, Feng W, Shen Z (2017) Balanced distribution adaptation for transfer learning. In: Proceedings of the IEEE international conference on data mining (ICDM), pp 1129–1134

  36. Wang J, Feng W, Chen Y, Yu H, Huang M, Yu PS (2018) Visual domain adaptation with manifold embedded distribution alignment. In: Proceedings of the 26th ACM international conference on multimedia, MM ’18, pp 402–410, DOI https://doi.org/10.1145/3240508.3240512, (to appear in print)

  37. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612

    Article  Google Scholar 

  38. Zhang J, Li W, Ogunbona P (2017) Joint geometrical and statistical alignment for visual domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1859–1867

  39. Zhang W, Ouyang W, Li W, Xu D (2018) Collaborative and adversarial network for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3801–3809

  40. Zhang Y, Allem JP, Unger JB, Cruz TB (2018) Automated identification of hookahs (waterpipes) on instagram: an application in feature extraction using convolutional neural network and support vector machine classification. J Med Inter Res 20(11):e10513

    Google Scholar 

  41. Zhang Y, Davison BD (2019) Modified distribution alignment for domain adaptation with pre-trained inception resnet. arXiv:1904.02322

  42. Zhang Y, Davison BD (2020) Impact of imagenet model selection on domain adaptation. In: Proceedings of the IEEE winter conference on applications of computer vision workshops, pp 173–182

  43. Zhang Y, Xie S, Davison BD (2019) Transductive learning via improved geodesic sampling. In: Proceedings of the 30th british machine vision conference

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Youshan Zhang.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, Y., Davison, B.D. Domain adaptation for object recognition using subspace sampling demons. Multimed Tools Appl 80, 23255–23274 (2021). https://doi.org/10.1007/s11042-020-09336-0

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-020-09336-0

Keywords

Navigation