Abstract
Visual saliency is very helpful for image detection and image processing. This paper proposes a novel visual saliency model. First, the proposed model can extract a saliency map with high precision and compound the linear combination of saliency map. Second, based on two-dimensional fractional Fourier transform, the proposed model generates a robust saliency map from the input image with Gaussian or salt-and-pepper noise. In order to reveal the noise influence from the given image, we provide a concept called the noise sensitivity scale (NSS). Third, using the image database from MSRA10K, we analyze the precision-recall and ROC curve and experimentally demonstrate that the proposed model can evaluate human fixation to some extent.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alpern, M.: Eye movements, vision and behavior. Am. J. Ophthalmol. 80(2), 307–308 (1975)
Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20(11), 1254–1259 (1998)
Kadir, T., Brady, M.: Saliency, scale and image description. Int. J. Comput. Vis. 45(2), 83–105 (2001)
Bruce, N.D.B., Tsotsos, J.K.: Saliency based on information maximization. In: International Conference on Neural Information Processing Systems (2005)
Yantis, S.: How visual salience wins the battle for awareness. Nat. Neurosci. 8(8), 975–977 (2005)
De Brecht, M., Saiki, J.: A neural network implementation of a saliency map model. Neural Netw. Off. J. Int. Neural Netw. Soc. 19(10), 1467 (2006)
Harel, J., Koch, C., Perona, P.: Graph-based visual saliency, vol. 19, pp. 545–552 (2006)
Meur, O.L., Callet, P.L., Barba, D.: A coherent computational approach to model bottom-up visual attention. IEEE Trans. Pattern Anal. Mach. Intell. 28(5), 802–817 (2006)
Gao, D., Mahadevan, V., Vasconcelos, N.: The discriminant center-surround hypothesis for bottom-up saliency. In: Advances in Neural Information Processing Systems, vol. 20, pp. 497–504 (2007)
Gao, D., Vasconcelos, N.: Bottom-up saliency is a discriminant process. In: IEEE International Conference on Computer Vision (2007)
Schölkopf, B., Platt, J., Hofmann, T.: A nonparametric approach to bottom-up visual saliency. In: Conference on Advances in Neural Information Processing Systems (2007)
Yu, Z., Wong, H.S.: A rule based technique for extraction of visual attention regions based on real-time clustering. IEEE Trans. Multimed. 9(4), 766–784 (2007)
Cerf, M., Harel, J., Einhäuser, W., Koch, C.: Predicting human gaze using low-level saliency combined with face detection. In: Advances in Neural Information Processing Systems, vol. 20, pp. 241–248 (2008)
Hou, X., Zhang, L.: Dynamic visual attention: searching for coding length increments. In: Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, December 2008
Zhang, L., Tong, M.H., Marks, T.K., Shan, H., Cottrell, G.W.: SUN: a Bayesian framework for saliency using natural statistics. Journal of Vision 8(7), 32.1 (2008)
Gao, D., Han, S., Vasconcelos, N.: Discriminant saliency, the detection of suspicious coincidences, and applications to visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 31(6), 989 (2009)
Itti, L., Baldi, P.: Bayesian surprise attracts human attention. Vis. Res. 49(10), 1295–1306 (2009)
Khan, F.S., Weijer, J.V.D., Vanrell, M.: Top-down color attention for object recognition. In: IEEE International Conference on Computer Vision (2009)
Avraham, T., Lindenbaum, M.: Esaliency (extended saliency): meaningful attention using stochastic image modeling. IEEE Trans. Softw. Eng. 32(4), 693–708 (2010)
Chikkerur, S., Serre, T., Tan, C., Poggio, T.: What and where: a Bayesian inference theory of attention. Vis. Res. 50(50), 2233–2247 (2010)
Mahadevan, V., Vasconcelos, N.: Spatiotemporal saliency in dynamic scenes. IEEE Trans. Softw. Eng. 32(1), 171–177 (2010)
Zhang, Q., Liu, H., Shen, J., Gu, G.: An improved computational approach for salient region detection. J. Comput. 5, 1011–1018 (2010). [cited 5 7]
Cheng, M., Mitra, N.J., Huang, X., Torr, P.H.S., Hu, S.: Global contrast based salient region detection. In: Computer Vision and Pattern Recognition (2011)
Kim, W., Jung, C., Kim, C.: Spatiotemporal saliency detection and its applications in static and dynamic scenes. IEEE Trans. Circuits Syst. Video Technol. 21(4), 446–456 (2011)
Guo, C., Ma, Q., Zhang, L.: Spatio-temporal Saliency detection using phase spectrum of quaternion fourier transform. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2008 (2008)
Achanta, R., Hemami, S., Estrada, F., Susstrunk, S.: Frequency-tuned salient region detection. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2009 (2009)
Li, J., Levine, M.D., An, X., Xu, X., He, H.: Visual saliency based on scale-space analysis in the frequency domain. IEEE Trans. Pattern Anal. Mach. Intell. 35(4), 996–1010 (2013)
Jiang, B., Zhang, L., Lu, H., Yang, C., Yang, M.: Saliency detection via absorbing Markov chain. In: IEEE International Conference on Computer Vision (2013)
Wang, J., Jiang, H., Yuan, Z., Cheng, M., Hu, X.: Salient object detection: a discriminative regional feature integration approach. In: IEEE Conference on Computer Vision and Pattern Recognition (2013)
Li, G., Shi, J., Luo, H., Tang, M.: A computational model of vision attention for inspection of surface quality in production line. Mach. Vis. Appl. 24(4), 835–844 (2013)
Qi, L., Chen, E., Mu, X., Guang, L., Zhang, S.: Recognizing human emotional state based on the 2D-FrFT and FLDA. In: International Congress on Image and Signal Processing (2009)
Singh, A.K., Saxena, R.: On convolution and product theorems for FRFT. Wireless Pers. Commun. 65(1), 189–201 (2012)
Gao, L., Qi, L., Wang, Y., Chen, E., Yang, S.: Rotation invariance in 2D-FRFT with application to digital image watermarking. J. Signal Process. Syst. 72(2), 133–148 (2013)
Wang, P., Tian, H., Zheng, W.: A novel image fusion method based on FRFT-NSCT. Math. Probl. Eng. 2013(11), 1–9 (2013)
Wu, J., Luo, X., Zhou, N.: Four-image encryption method based on spectrum truncation, chaos and the MODFrFT. Opt. Laser Technol. 45(1), 571–577 (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Ethics declarations
The authors declare that there are no conflicts of interest regarding the publication of this article.
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Xu, H., Jiang, C. (2019). Visual Saliency Based on Two-Dimensional Fractional Fourier Transform. In: Jin, H., Lin, X., Cheng, X., Shi, X., Xiao, N., Huang, Y. (eds) Big Data. BigData 2019. Communications in Computer and Information Science, vol 1120. Springer, Singapore. https://doi.org/10.1007/978-981-15-1899-7_17
Download citation
DOI: https://doi.org/10.1007/978-981-15-1899-7_17
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-1898-0
Online ISBN: 978-981-15-1899-7
eBook Packages: Computer ScienceComputer Science (R0)