Skip to main content

Self-learning Framework with Temporal Filtering for Robust Maritime Vessel Detection

  • Conference paper
  • First Online:
Representations, Analysis and Recognition of Shape and Motion from Imaging Data (RFMI 2017)

Abstract

With the recent development in ConvNet-based detectors, a successful solution for vessel detection becomes possible. However, it is essential to access a comprehensive annotated training set from different maritime environments. Creating such a dataset is expensive and time consuming. To automate this process, this paper proposes a novel self learning framework which automatically finetunes a generic pre-trained model to any new environment. With this, the framework enables automated labeling of new dataset types. The method first explores the video frames captured from a new target environment to generate the candidate vessel samples. Afterwards, it exploits a temporal filtering concept to verify the correctly generated candidates as new labels for learning, while removing the false positives. Finally, the system updates the vessel model using the provided self-learning dataset. Experimental results on our real-world evaluation dataset show that generalizing a finetuned Single Shot Detector to a new target domain using the proposed self-learning framework increases the average precision and the F1-score by 12% and 5%, respectively. Additionally, the proposed temporal filter reduced the noisy detections in a sensitive setting from 58% to only 5%.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 74.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Khurana, P., Sharma, A., Singh, S.N., Singh, P.K.: A survey on object recognition and segmentation techniques. In: 3rd International IEEE Conference on Computing for Sustainable Global Development (INDIACom) (2016)

    Google Scholar 

  2. Shantaiya, S., Verma, K., Mehta, K.: A survey on approaches of object detection. Int. J. Comput. Appl. (0975–8887), 65(18) (2013)

    Google Scholar 

  3. Bidyalakshmi Devi, R.B., Jina Chanu, Y., Singh, K.M.: A survey on different background subtraction method for moving object detection. Int. J. Res. Emerg. Sci. Technol. 3(10) (2016)

    Google Scholar 

  4. Abdul Malik, A., Khalil, A., Ullah Khan, H.: Object detection and tracking using background subtraction and connected component labeling. Int. J. Comput. Appl. (0975–8887), 75(13) (2013)

    Google Scholar 

  5. Ghahremani, A., Bondarev, E., de With, P.H.N.: Water region extraction in thermal and RGB sequences using spatiotemporally-oriented energy features. IS&T Electronic Imaging - Algorithms and Systems, USA (2017)

    Google Scholar 

  6. Druzhkov, P.N., Kustikova, V.D.: A survey of deep learning methods and software tools for image classification and object detection. Pattern Recognit. Image Anal. 26(1), 9–15 (2016)

    Article  Google Scholar 

  7. Cabrera-Vives, G., Reyes, I., Forstert, F., Estevez, P.A., Maureira, J.C.: Supernovae detection by using convolutional neural networks. In: International Joint Conference on Neural Networks, JCNN (2016)

    Google Scholar 

  8. Liu, W., et al.: SSD: single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21–37. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2

    Chapter  Google Scholar 

  9. Maâmatou, H., Chateau, T., Gazzah, S., Goyat, Y., Amara, N.E.B.: Transductive transfer learning to specialize a generic classifier towards a specific scene. In: VISIGRAPP (4: VISAPP) (2016)

    Google Scholar 

  10. Mhalla, A., Chateaub, T., Maâmatoua, H., Gazzaha, S., Amara, N.E.B.: SMC faster R-CNN: toward a scene-specialized multi-object detector. Comput. Vis. Image Underst. 164, 1–13 (2017)

    Article  Google Scholar 

  11. All, K., Hasler, D., Fleuret, F.: Flowboost—appearance learning from sparsely annotated video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1433–1440 (2011)

    Google Scholar 

  12. Wang, M., Li, W., Wang, X.: Transferring a generic pedestrian detector towards specific scenes. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3274–3281 (2012)

    Google Scholar 

  13. Aytar, Y., Zisserman, A., Rasa, T.: Model transfer for object category detection. In: International IEEE Conference on Computer Vision, pp. 2252–2259 (2011)

    Google Scholar 

  14. Pan, S.J., Tsang, I.W., Kwok, J.T., Yang, Q.: Domain adaptation via transfer component analysis. IEEE Trans Neural Netw. 22, 199–210 (2011)

    Article  Google Scholar 

  15. Li, X., Ye, M., Fu, M., Xu, P., Li, T.: Domain adaption of vehicle detector based on convolutional neural networks. Int. J. Control. Autom. Syst. 13(4), 1020–1031 (2015)

    Article  Google Scholar 

  16. Wang, X., Hua, G., Han, T.X.: Detection by detections: non-parametric detector adaptation for a video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 350–357 (2012)

    Google Scholar 

  17. Wang, X., Wang, M., Li, W.: Scene-specific pedestrian detection for static video surveillance. IEEE Trans. Pattern Anal. Mach. Intell. 36(2), 361–374 (2014)

    Article  MathSciNet  Google Scholar 

  18. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: NIPS (2015)

    Google Scholar 

Download references

Acknowledgement

This work is supported by the European ITEA APPS project. We thank the company Vinotion for providing the Botlek dataset to us. We also show our gratitude to the company NVIDIA for granting us a “TITAN X PASCAL” GPU.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Amir Ghahremani .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ghahremani, A., Bondarev, E., de With, P.H.N. (2019). Self-learning Framework with Temporal Filtering for Robust Maritime Vessel Detection. In: Chen, L., Ben Amor, B., Ghorbel, F. (eds) Representations, Analysis and Recognition of Shape and Motion from Imaging Data. RFMI 2017. Communications in Computer and Information Science, vol 842. Springer, Cham. https://doi.org/10.1007/978-3-030-19816-9_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-19816-9_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-19815-2

  • Online ISBN: 978-3-030-19816-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics