Abstract
Recent years have witnessed the widespread success of convolutional neural networks (CNNs) in computer vision and multimedia. The CNNs based background subtraction methods, which are effective for addressing the challenges (such as shadows, dynamic backgrounds, illumination changes) existing in real-world applications, have attracted much attention. However, these methods usually require a large amount of densely labeled video training data, which are hardly collected in the real-world. To address this problem, in this paper, we propose a multiscale cascaded scene-specific CNNs based background subtraction method equipped with a novel training strategy, which takes advantage of the balance of positive and negative training samples. The proposed method can rely on a small number of training samples to effectively train the robust neural network models. Experimental results on the CDnet-2014 dataset show that the proposed method obtains better performance with much less training samples compared with the state-of-the-art methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Maddalena, L., Petrosino, A.: A self-organizing approach to background subtraction for visual surveillance applications. IEEE Trans. Image Process. 17(7), 1168–1177 (2008)
Biswas, K.K., Basu, S.K.: Gesture recognition using microsoft kinect®. In: IEEE International Conference on Automation, Robotics and Applications, pp. 100–103 (2011)
Paul, M., Lin, W., Lau, C.T., Lee, B.S.: Pattern-based video coding with dynamic background modeling. EURASIP J. Adv. Signal Process. 2013(1), 138–153 (2013)
Stauffer, C., Grimson, W.E.L.: Adaptive background mixture models for real-time tracking. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 246–252 (1999)
Bianco, S., Ciocca, G., Schettini, R.: Combination of video change detection algorithms by genetic programming. IEEE Trans. Evol. Comput. 21(6), 914–928 (2017)
Jiang, S., Lu, X.: WeSamBE: a weight-sample-based method for background subtraction. IEEE Trans. Circuits Syst. Video Technol. (2017). https://doi.org/10.1109/TCSVT.2017.2711659
Liao, J., Wang, H., Yan, Y., Zheng, J.: A novel background subtraction method based on ViBe. In: Zeng, B., Huang, Q., El Saddik, A., Li, H., Jiang, S., Fan, X. (eds.) PCM 2017, Part II. LNCS, vol. 10736, pp. 428–437. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-77383-4_42
Wang, Y., Jodoin, P.M., Porikli, F., Konrad, J., Ishwar, P.: CDnet 2014: an expanded change detection benchmark dataset. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 387–394 (2014)
Braham, M., Droogenbroeck, M.V.: Deep background subtraction with scene-specific convolutional neural networks. In: International Conference on Systems, Signals and Image Processing, pp. 1–4 (2016)
Wang, Y., Luo, Z., Jodoin, P.M.: Interactive deep learning method for segmenting moving objects. Pattern Recognit. Lett. 96, 66–75 (2017)
Tieleman, T., Hinton, G.: Lecture 6.5-RmsProp: divide the gradient by a running average of its recent magnitude. In: COURSERA: Neural Networks for Machine Learning (2012)
Braham, M., Pierard, S., Droogenbroeck, M.V.: Semantic background subtraction. In: IEEE International Conference on Image Processing, pp. 4552–4556 (2017)
Babaee, M., Dinh, D.T., Rigoll, G.: A deep convolutional neural network for video sequence background subtraction. Pattern Recognit. 76, 635–649 (2018)
Zhao, X., Chen, Y., Tang, M., Wang, J.: Joint background reconstruction and foreground segmentation via a two-stage convolutional neural network. In: IEEE International Conference on Multimedia and Expo., pp. 343–348 (2017)
Chen, Y., Wang, J., Zhu, B., Tang, M., Lu, H.: Pixel-wise deep sequence learning for moving object detection. IEEE Trans. Circuits Syst. Video Technol. (2017). https://doi.org/10.1109/TCSVT.2017.2770319
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grants U1605252, 61472334, 61571379, and by the Natural Science Foundation of Fujian Province of China under Grant 2017J01127.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Liao, J., Guo, G., Yan, Y., Wang, H. (2018). Multiscale Cascaded Scene-Specific Convolutional Neural Networks for Background Subtraction. In: Hong, R., Cheng, WH., Yamasaki, T., Wang, M., Ngo, CW. (eds) Advances in Multimedia Information Processing – PCM 2018. PCM 2018. Lecture Notes in Computer Science(), vol 11164. Springer, Cham. https://doi.org/10.1007/978-3-030-00776-8_48
Download citation
DOI: https://doi.org/10.1007/978-3-030-00776-8_48
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-00775-1
Online ISBN: 978-3-030-00776-8
eBook Packages: Computer ScienceComputer Science (R0)