Abstract
Visual saliency is an important research topic in the field of computer vision due to its numerous possible applications. It helps to focus on regions of interest instead of processing the whole image or video data. Detecting visual saliency in still images has been widely addressed in literature with several formulations. However, visual saliency detection in videos has attracted little attention, and is a more challenging task due to additional temporal information. A common approach for obtaining a spatio-temporal saliency map is to combine a static saliency map and a dynamic saliency map. In our work, we model the dynamic textures in a dynamic scene with local binary patterns to compute the dynamic saliency map, and we use color features to compute the static saliency map. Both saliency maps are computed using a bio-inspired mechanism of human visual system with a discriminant formulation known as center surround saliency, and are fused in a proper way. The proposed model has been extensively evaluated with diverse publicly available datasets which contain several videos of dynamic scenes, and comparison with state-of-the art methods shows that it achieves competitive results.
Similar content being viewed by others
References
Achanta R, Estrada F, Susstrunk S, Hemami S Frequency-tuned salient region detection, Computer Vision and Pattern Recognition, 2009 pp 1597–1604
Borji A, Itti L (2013) State-of-the-art in visual attention modeling. IEEE Trans Pattern Anal Mach Intell 35(1):185–207
Borji A, Cheng MM, Jiang H, Li J (2015) Salient object detection: A benchmark. IEEE Trans Image Process 24(12):5706–5722
Chang L, Yuen PC, Qiu G (2009) Object motion detection using information theoretic spatio-temporal saliency. Pattern Recogn 42(11):2897–2906
Chetverikov D, Péteri R. (2005) A brief survey of dynamic texture description and recognition Computer Recognition Systems. Springer, , pp 17–26
Fawcett T (2006) An introduction to roc analysis. Pattern Recogn Lett 27 (8):861–874
Frintrop S (2011) Computational visual attention. Springer,
Fu K, Gu IYH, Yun Y, Gong C, Yang J Graph construction for salient object detection in videos, In: ICPR, 2014, pp 2371–2376
Goferman S, Zelnik-manor L, Tal A (2010) Context-aware saliency detection, In: IEEE CVPR
Guo CL, Zhang L (2010) A novel multiresolution spatiotemporal saliency detection model and its applications in image and video compression. IEEE TIP 19(1):185–198
Hou X, Zhang L Dynamic visual attention: searching for coding length increments, In: NIPS, Vol. 5, 2008, p 7
Kim W, Jung C, Kim C (2011) Spatiotemporal saliency detection and its applications in static and dynamic scenes. IEEE Trans Circuits Syst Video Techn 21 (4):446–456
Le Meur O, Baccino T (2013) Methods for comparing scanpaths and saliency maps: strengths and weaknesses. Behav Res Methods 45(1):251–266
Lu T, Yuan Z, Huang Y, Wu D, Yu H (2010) Video retargeting with nonlinear spatial-temporal saliency fusion, In: ICIP
Mahadevan V, Vasconcelos N (2010) Spatiotemporal saliency in dynamic scenes. IEEE Trans Pattern Anal Mach Intell 32(1):171–177
Mancas M, Riche N, Leroy J, Gosselin B Abnormal motion selection in crowds using bottom-up saliency. In: IEEE ICIP 2011, pp 229–232
Marat S, Phuoc TH, Granjon L, Guyader N, Pellerin D, Guérin-Dugué A (2009) Modelling spatio-temporal saliency to predict gaze direction for short videos. IJCV 82(3):231–243
Muddamsetty SM, Sidibé D, Trémeau A, Mériaudeau F (2013) A performance evaluation of fusion techniques for spatio-temporal saliency detection in dynamic scenes. In: ICIP
Pinto Y, Van der Leij AR, Sligte IG, Lamme VAF, Scholte HS (2013) Bottom-up and top-down attention are independent. J Vis 13(3):16
Riche N, Mancas M, Culibrk D, Crnojevic V, Gosselin B, Dutoit T (2013) Dynamic saliency models and human attention: A comparative study on videos Computer Vision–ACCV 2012. Springer, , pp 586–598
Siagian C, Itti L (2009) Biologically inspired mobile robot vision localization. IEEE Trans Robot 25(4):861–873
Sidibé D., Fofi D, Mériaudeau F. (2010) Using visual saliency for object tracking with particle filters. In: EUSIPCO
Seo HJJ, Milanfar P (2009) Static and space-time visual saliency detection by self-resemblance. J Vis 9(12):
Sun S (2013) A survey of multi-view machine learning. Neural Comput Appl 23 (7):2013–2038
Xu C, Tao D, Xu C A survey on multi-view learning, arXiv preprint (2013), pp 1304.5634
Xu C, Tao D, Xu C (2015) Multi-view intact space learning. IEEE Trans PAMI 37(12):2531–2544
Yubing T, Cheikh FA, Guraya FFE, Konik H, Trémeau A (2011) A spatiotemporal saliency model for video surveillance. Cogn Compuation 3(1):241–263
Zhang L, Tong MH, Marks TK, Shan H, Cottrell GW (2008) Sun: A bayesian framework for saliency using natural statistics. J. Vis 8(7):32
Zhang L, Tong MH, Cottrell GW Sunday: Saliency using natural statistics for dynamic analysis of scenes, In: Proceedings of the 31st Annual Cognitive Science Conference (2009) pp 2944–2949
Zhao G, Pietikäinen M (2007) Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans Pattern Anal Mach Intell 29(6):915–928
Zhou B, Hou X, Zhang L A phase discrepancy analysis of object motion. In: Proceeding of the 10th Asian Confernce of Computer Vision, 2011, pp 225–238
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Muddamsetty, S.M., Sidibé, D., Trémeau, A. et al. Salient objects detection in dynamic scenes using color and texture features. Multimed Tools Appl 77, 5461–5474 (2018). https://doi.org/10.1007/s11042-017-4462-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-017-4462-y