Skip to main content

Learning Visual Saliency Based on Object’s Relative Relationship

  • Conference paper
  • 3993 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 7667))

Abstract

As a challenging issue in both computer vision and psychological research, visual attention has arouse a wide range of discussions and studies in recent years. However, conventional computational models mainly focus on low-level information, while high-level information and their interrelationship are ignored. In this paper, we stress the issue of relative relationship between high-level information, and a saliency model based on low-level and high-level analysis is also proposed. Firstly, more than 50 categories of objects are selected from nearly 800 images in MIT data set[1], and concrete quantitative relationship is learned based on detail analysis and computation. Secondly, using the least square regression with constraints method, we demonstrate an optimal saliency model to produce saliency maps. Experimental results indicate that our model outperforms several state-of-art methods and produces better matching to human eye-tracking data.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Judd, T., Ehinger, K., Torralba, A.: Learning to predict where humans look. In: International Conference on Computre Vision, pp. 2106–2113 (2009)

    Google Scholar 

  2. Itti, L., Koch, C., Niebur, E.: A Model of Saliency-based Visual Attention for Rapid Scene Analysis. Pattern Analysis and Machine Intelligence 20, 1254–1259 (1998)

    Article  Google Scholar 

  3. Zhao, Q., Koch, C.: Learning visual saliency. In: Conference on Information Sciences and Systems (CISS), pp. 1–6 (2011)

    Google Scholar 

  4. Bruce, N.D.B., Tsotsos, J.K.: Saliency, attention, and visual search: An information theoretic approach. Journal of Vision 9, 1–24 (2009)

    Article  Google Scholar 

  5. Hou, X., Zhang, L.: Dynamic visual attention:Searching for coding length increments. In: Neural Information Processing Systems, vol. 21, pp. 681–688 (2008)

    Google Scholar 

  6. Cerf, M., Frady, E., Koch, C.: Faces and text attract gaze independent of the task: Experimental data and computer model. Journal of Vision 9, 1–15 (2009)

    Article  Google Scholar 

  7. Goferman, S., Zelnik-Manor, L., Tal, A.: Contextaware saliency detection. In: Computer Vision and Pattern Recognition, pp. 2376–2383 (2010)

    Google Scholar 

  8. Liu, T., Yuan, Z., Sun, J., Wang, J., Zheng, N., Tang, X., Shum, H.: Learning to detect a salient object. Pattern Analysis and Machine Intelligence 33, 353–367 (2011)

    Article  Google Scholar 

  9. Itti, L., Koch, C.: Computational modelling of visual attention. Nature Reviews. Neuroscience 2, 194–203 (2001)

    Article  Google Scholar 

  10. Goferman, S., Zelnik-Manor, L., Tal, A.: Context-aware saliency detection. In: Computer Vision and Pattern Recognition, pp. 2376–2383 (2010)

    Google Scholar 

  11. Yang, Y., Song, M., Li, N., Bu, J., Chen, C.: What Is the Chance of Happening: A New Way to Predict Where People Look. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part V. LNCS, vol. 6315, pp. 631–643. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  12. Viola, P., Jones, M.: Robust real-time face detection. International Journal of Computer Vision 57, 137–154 (2004)

    Article  Google Scholar 

  13. Felzenszwalb, P., McAllester, D., Ramanan, D.: A discriminatively trained, multiscale,deformable part model. In: Computer Vision and Pattern Recognition, pp. 1–8 (2008)

    Google Scholar 

  14. Wang, K., Belongie, S.: Word Spotting in the Wild. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part I. LNCS, vol. 6311, pp. 591–604. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Wang, S., Zhao, Q., Song, M., Bu, J., Chen, C., Tao, D. (2012). Learning Visual Saliency Based on Object’s Relative Relationship. In: Huang, T., Zeng, Z., Li, C., Leung, C.S. (eds) Neural Information Processing. ICONIP 2012. Lecture Notes in Computer Science, vol 7667. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-34500-5_38

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-34500-5_38

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-34499-2

  • Online ISBN: 978-3-642-34500-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics