Flame Image Segmentation Algorithm Based on Motion and Color Saliency
This paper proposed a flame segmentation algorithm based on the saliency of motion and color. First, feature point detection is performed on the video image using the scale-invariant feature transform (SIFT) algorithm, and the optical flow field of moving object in the adjacent frame is acquired by the optical flow method. According to the optical flow vector difference between the target pixel point and the surrounding neighborhood pixels, the motion saliency map is obtained based on the Munsell color system. Then, the LSI flame color statistical model based on the Lab and HSI space is used to extract color saliency map of video images. Finally, under the Bayes framework, the motion saliency map and the color saliency map are fused in an interactive manner to obtain the final flame segmentation map. Experimental results show that the proposed algorithm can effectively segment the flame image in different scenarios.
KeywordsVisual saliency Flame segmentation Optical flow LSI color statistics model
This paper was supported by the Project of Applied Basic Research Project of Ministry of Transport of the People’s Republic of China (Grant No. 2015 319 225 210).
- 1.Estrada FJ, Jepson AD. Benchmarking image segmentation algorithms. IJCV. 2009;85:167–81.Google Scholar
- 2.Tsotsos JK, Culhane SM, Wai WYK, Lai Y, Davis N, Nuflo F. Modelling visual attention via selective tuning. ArtifIntell. 1995;78(1–2):507–45.Google Scholar
- 4.Hou X, Zhang L. Saliency detection: A spectral residual approach. In: CVPR; 2007.Google Scholar
- 5.Rahtu E, Kannala J, Salo M, Heikkilä J. Segmenting salient objects from images and videos. In: ECCV; 2010. p. 366–79.Google Scholar
- 7.Xie Y, Lu H. Visual saliency detection based on bayesian model. In: ICIP; 2011. p. 645–8.Google Scholar
- 8.Li X, Lu H, Zhang L, Ruan X, Yang M-H. Saliency detection via dense and sparse reconstruction. In: ICCV; 2013.Google Scholar
- 10.Bruce N, Tsotsos J. Saliency based on information maximization. In: NIPS; 2006.Google Scholar