Advertisement

Unsupervised Domain Adaptation for Object Detection Using Distribution Matching in Various Feature Level

  • Hyoungwoo ParkEmail author
  • Minjeong Ju
  • Sangkeun Moon
  • Chang D. Yoo
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11378)

Abstract

As the research on deep learning has become more active, the need for a lot of data has emerged. However, there are limitations in acquiring real data such as digital forensics, so domain adaptation technology is required to overcome this problem. This paper considers distribution matching in various feature level for unsupervised domain adaptation for object detection with a single stage detector. The object detection task assumes that training and test data are drawn from the same distribution; however, in a real environment, there is a domain gap between training and test data which leads to degrading performance significantly. Therefore, we aim to learn a model to generalize well in target domain of object detection by using maximum mean discrepancy (MMD) in various feature levels. We adjust MMD based on single shot multibox detector (SSD) model which is a single stage detector that learns to localize objects with various size using a multi-layer design of bounding box regression and infers object class simultaneously. The MMD loss on high-level features between source and target domain effectively reduces the domain discrepancy to learn a domain-invariant feature in SSD model. We evaluate the approaches using Syn2real object detection dataset. Experimental results show that reducing the domain shift in high-level features improves the cross-domain robustness of object detection, and domain adaptation works better with simple MMD method than complex method as GAN.

Keywords

Object detection Unsupervised domain adaptation Maximum mean discrepancy 

Notes

Acknowledgement

This research was supported by the Korea Electric Power Research Institute (KEPRI) of the Korea Electric Power Corporation (KEPCO).

References

  1. 1.
    Girshick, R., et al.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2014)Google Scholar
  2. 2.
    Girshick, R.: Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision (2015)Google Scholar
  3. 3.
    Ren, S., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems (2015)Google Scholar
  4. 4.
    Redmon, J., et al.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)Google Scholar
  5. 5.
    Liu, W., et al.: SSD: single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016, Part I. LNCS, vol. 9905, pp. 21–37. Springer, Cham (2016).  https://doi.org/10.1007/978-3-319-46448-0_2CrossRefGoogle Scholar
  6. 6.
    Peng, X., et al.: Syn2Real: A New Benchmark for Synthetic-to-Real Visual Domain Adaptation. arXiv preprint arXiv:1806.09755 (2018)
  7. 7.
    Quiñonero-Candela, J., et al.: Covariate shift and local learning by distribution matching (2008)Google Scholar
  8. 8.
    Chen, Y., et al.: Domain adaptive faster R-CNN for object detection in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)Google Scholar
  9. 9.
    Tzeng, E., et al.: Simultaneous deep transfer across domains and tasks. In: Proceedings of the IEEE International Conference on Computer Vision (2015)Google Scholar
  10. 10.
    Sun, B., Saenko, K.: Deep CORAL: correlation alignment for deep domain adaptation. In: Hua, G., Jégou, H. (eds.) ECCV 2016, Part III. LNCS, vol. 9915, pp. 443–450. Springer, Cham (2016).  https://doi.org/10.1007/978-3-319-49409-8_35CrossRefGoogle Scholar
  11. 11.
    Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(1), 2096–2030 (2016)Google Scholar
  12. 12.
    Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)
  13. 13.
    Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems (2014)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  • Hyoungwoo Park
    • 1
    Email author
  • Minjeong Ju
    • 1
  • Sangkeun Moon
    • 2
  • Chang D. Yoo
    • 1
  1. 1.Korea Advanced Institute of Science and TechnologyDaejeonSouth Korea
  2. 2.Korea Electric Power Corporation (KEPCO)DaejeonSouth Korea

Personalised recommendations