Unsupervised Domain Adaptation for Object Detection Using Distribution Matching in Various Feature Level
As the research on deep learning has become more active, the need for a lot of data has emerged. However, there are limitations in acquiring real data such as digital forensics, so domain adaptation technology is required to overcome this problem. This paper considers distribution matching in various feature level for unsupervised domain adaptation for object detection with a single stage detector. The object detection task assumes that training and test data are drawn from the same distribution; however, in a real environment, there is a domain gap between training and test data which leads to degrading performance significantly. Therefore, we aim to learn a model to generalize well in target domain of object detection by using maximum mean discrepancy (MMD) in various feature levels. We adjust MMD based on single shot multibox detector (SSD) model which is a single stage detector that learns to localize objects with various size using a multi-layer design of bounding box regression and infers object class simultaneously. The MMD loss on high-level features between source and target domain effectively reduces the domain discrepancy to learn a domain-invariant feature in SSD model. We evaluate the approaches using Syn2real object detection dataset. Experimental results show that reducing the domain shift in high-level features improves the cross-domain robustness of object detection, and domain adaptation works better with simple MMD method than complex method as GAN.
KeywordsObject detection Unsupervised domain adaptation Maximum mean discrepancy
This research was supported by the Korea Electric Power Research Institute (KEPRI) of the Korea Electric Power Corporation (KEPCO).
- 1.Girshick, R., et al.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2014)Google Scholar
- 2.Girshick, R.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision (2015)Google Scholar
- 3.Ren, S., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems (2015)Google Scholar
- 4.Redmon, J., et al.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)Google Scholar
- 6.Peng, X., et al.: Syn2Real: A New Benchmark for Synthetic-to-Real Visual Domain Adaptation. arXiv preprint arXiv:1806.09755 (2018)
- 7.Quiñonero-Candela, J., et al.: Covariate shift and local learning by distribution matching (2008)Google Scholar
- 8.Chen, Y., et al.: Domain adaptive faster R-CNN for object detection in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)Google Scholar
- 9.Tzeng, E., et al.: Simultaneous deep transfer across domains and tasks. In: Proceedings of the IEEE International Conference on Computer Vision (2015)Google Scholar
- 11.Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2096–2030 (2016)Google Scholar
- 12.Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
- 13.Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems (2014)Google Scholar