Abstract
With the recent development in ConvNet-based detectors, a successful solution for vessel detection becomes possible. However, it is essential to access a comprehensive annotated training set from different maritime environments. Creating such a dataset is expensive and time consuming. To automate this process, this paper proposes a novel self learning framework which automatically finetunes a generic pre-trained model to any new environment. With this, the framework enables automated labeling of new dataset types. The method first explores the video frames captured from a new target environment to generate the candidate vessel samples. Afterwards, it exploits a temporal filtering concept to verify the correctly generated candidates as new labels for learning, while removing the false positives. Finally, the system updates the vessel model using the provided self-learning dataset. Experimental results on our real-world evaluation dataset show that generalizing a finetuned Single Shot Detector to a new target domain using the proposed self-learning framework increases the average precision and the F1-score by 12% and 5%, respectively. Additionally, the proposed temporal filter reduced the noisy detections in a sensitive setting from 58% to only 5%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Khurana, P., Sharma, A., Singh, S.N., Singh, P.K.: A survey on object recognition and segmentation techniques. In: 3rd International IEEE Conference on Computing for Sustainable Global Development (INDIACom) (2016)
Shantaiya, S., Verma, K., Mehta, K.: A survey on approaches of object detection. Int. J. Comput. Appl. (0975–8887), 65(18) (2013)
Bidyalakshmi Devi, R.B., Jina Chanu, Y., Singh, K.M.: A survey on different background subtraction method for moving object detection. Int. J. Res. Emerg. Sci. Technol. 3(10) (2016)
Abdul Malik, A., Khalil, A., Ullah Khan, H.: Object detection and tracking using background subtraction and connected component labeling. Int. J. Comput. Appl. (0975–8887), 75(13) (2013)
Ghahremani, A., Bondarev, E., de With, P.H.N.: Water region extraction in thermal and RGB sequences using spatiotemporally-oriented energy features. IS&T Electronic Imaging - Algorithms and Systems, USA (2017)
Druzhkov, P.N., Kustikova, V.D.: A survey of deep learning methods and software tools for image classification and object detection. Pattern Recognit. Image Anal. 26(1), 9–15 (2016)
Cabrera-Vives, G., Reyes, I., Forstert, F., Estevez, P.A., Maureira, J.C.: Supernovae detection by using convolutional neural networks. In: International Joint Conference on Neural Networks, JCNN (2016)
Liu, W., et al.: SSD: single shot multibox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21–37. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2
Maâmatou, H., Chateau, T., Gazzah, S., Goyat, Y., Amara, N.E.B.: Transductive transfer learning to specialize a generic classifier towards a specific scene. In: VISIGRAPP (4: VISAPP) (2016)
Mhalla, A., Chateaub, T., Maâmatoua, H., Gazzaha, S., Amara, N.E.B.: SMC faster R-CNN: toward a scene-specialized multi-object detector. Comput. Vis. Image Underst. 164, 1–13 (2017)
All, K., Hasler, D., Fleuret, F.: Flowboost—appearance learning from sparsely annotated video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1433–1440 (2011)
Wang, M., Li, W., Wang, X.: Transferring a generic pedestrian detector towards specific scenes. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3274–3281 (2012)
Aytar, Y., Zisserman, A., Rasa, T.: Model transfer for object category detection. In: International IEEE Conference on Computer Vision, pp. 2252–2259 (2011)
Pan, S.J., Tsang, I.W., Kwok, J.T., Yang, Q.: Domain adaptation via transfer component analysis. IEEE Trans Neural Netw. 22, 199–210 (2011)
Li, X., Ye, M., Fu, M., Xu, P., Li, T.: Domain adaption of vehicle detector based on convolutional neural networks. Int. J. Control. Autom. Syst. 13(4), 1020–1031 (2015)
Wang, X., Hua, G., Han, T.X.: Detection by detections: non-parametric detector adaptation for a video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 350–357 (2012)
Wang, X., Wang, M., Li, W.: Scene-specific pedestrian detection for static video surveillance. IEEE Trans. Pattern Anal. Mach. Intell. 36(2), 361–374 (2014)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: NIPS (2015)
Acknowledgement
This work is supported by the European ITEA APPS project. We thank the company Vinotion for providing the Botlek dataset to us. We also show our gratitude to the company NVIDIA for granting us a “TITAN X PASCAL” GPU.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Ghahremani, A., Bondarev, E., de With, P.H.N. (2019). Self-learning Framework with Temporal Filtering for Robust Maritime Vessel Detection. In: Chen, L., Ben Amor, B., Ghorbel, F. (eds) Representations, Analysis and Recognition of Shape and Motion from Imaging Data. RFMI 2017. Communications in Computer and Information Science, vol 842. Springer, Cham. https://doi.org/10.1007/978-3-030-19816-9_10
Download citation
DOI: https://doi.org/10.1007/978-3-030-19816-9_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-19815-2
Online ISBN: 978-3-030-19816-9
eBook Packages: Computer ScienceComputer Science (R0)