Skip to main content

Stereoscopic Visual Attention Model for 3D Video

  • Conference paper
Advances in Multimedia Modeling (MMM 2010)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 5916))

Included in the following conference series:

Abstract

Compared with traditional mono-view video, three-dimensional video (3DV) provides user interactive functionalities and stereoscopic perception, which makes people more interested in pop-out regions or the regions with small depth value. Thus, traditional visual attention model for mono-view video can hardly be directly applied to stereoscopic visual attention (SVA) analysis for 3DV. In this paper, we propose a bottom-up SVA model to simulate human visual system with stereoscopic vision more accurately. The proposed model is based on multiple perceptual stimuli including depth information, luminance, color, orientation and motion contrast. Then, a depth based dynamic fusion is proposed to integrate these features. The experimental results on multi-view video test sequences show that the proposed model maintains high robustness and is able to efficiently simulate SVA of human eyes.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Tanimoto, M.: Overview of Free Viewpoint Television. Signal Proc.: Image Commun. 21(6), 454–461 (2006)

    Article  Google Scholar 

  2. Muller, K., Merkle, P., Wiegand, T.: Compressing time- varying visual content. IEEE Signal Processing Magazine 24(6), 58–65 (2007)

    Article  Google Scholar 

  3. Smolic, A., Mueller, K., Merkle, P., et al.: Multi-view video plus depth (MVD) format for advanced 3D video systems. MPEG and ITU-T SG16 Q.6, JVT-W100, San Jose, USA (April 2007)

    Google Scholar 

  4. Han, J., Ngan, K.N., Li, M., Zhang, H.: Unsupervised extraction of visual attention objects in color images. IEEE Trans. CSVT 16(1), 141–145 (2006)

    Google Scholar 

  5. Ma, Y.F., Hua, X.S., Lu, L., et al.: A generic framework of user attention model and its application in video summarization. IEEE Trans. Multimedia 7(5), 907–919 (2005)

    Article  Google Scholar 

  6. Itti, L., Koch, C.: Computational Modeling of Visual Attention. Nature Reviews Neuroscience 2(3), 194–203 (2001)

    Article  Google Scholar 

  7. Itti, L., Koch, C.: Feature combination strategies for saliency-based visual attention system. J. Electron. Imaging 10, 161–169 (2001)

    Article  Google Scholar 

  8. Zhai, G.T., Chen, Q., Yang, X.K., Zhang, W.J.: Scalable visual sensitivity profile estimation. In: ICASSP, Las Vegas, Nevada, USA, April 2008, pp. 876–879 (2008)

    Google Scholar 

  9. Zhai, Y., Shah, M.: Visual attention detection in video sequences using spatiotemporal cues. In: Proceedings of the 14th ACM Multimedia, Santa Barbara, CA, USA, pp. 815–824 (2006)

    Google Scholar 

  10. Wang, P.P., Zhang, W., Li, J., Zhang, Y.: Real-time detection of salient moving object: a multi-core solution. In: ICASSP, Las Vegas, Nevada, USA, April 2008, pp. 1481–1484 (2008)

    Google Scholar 

  11. Lu, Z., Lin, W., Yang, X., Ong, E.P., Yao, S.: Modeling Visual Attention’s Modulatory Aftereffects on Visual Sensitivity and Quality Evaluation. IEEE Trans Image Proc. 14(11), 1928–1942 (2005)

    Article  Google Scholar 

  12. Kolmogorov, V., Zabih, R.: Multi-camera scene reconstruction via graph cuts. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002. LNCS, vol. 2352, pp. 82–96. Springer, Heidelberg (2002)

    Chapter  Google Scholar 

  13. Tanimoto, M., Fujii, T., Suzuki, K.: Improvement of Depth Map Estimation and View Synthesis, ISO/IEC JTC1/SC29/WG11 M15090, Antalya, Turkey (January 2008)

    Google Scholar 

  14. Vetro, A., McGuire, M., Matusik, W., et al.: Multiview Video Test Sequences from MERL, ISO/IEC JTC1/SC29/WG11, MPEG05/m12077, Busan, Korea (April 2005)

    Google Scholar 

  15. Feldmann, I., Mueller, M., Zilly, F., et al.: HHI Test Material for 3D Video, ISO/IEC JTC1/SC29/WG11, M15413, Archamps, France (April 2008)

    Google Scholar 

  16. Zitnick, C.L., Kang, S.B., Uyttendaele, M., et al.: High-quality video view interpolation using a layered representation. In: ACM SIGGRAPH and ACM Trans. on Graphics, Los Angeles, CA, August 2004, pp. 600–608 (2004)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zhang, Y., Jiang, G., Yu, M., Chen, K. (2010). Stereoscopic Visual Attention Model for 3D Video. In: Boll, S., Tian, Q., Zhang, L., Zhang, Z., Chen, YP.P. (eds) Advances in Multimedia Modeling. MMM 2010. Lecture Notes in Computer Science, vol 5916. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-11301-7_33

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-11301-7_33

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-11300-0

  • Online ISBN: 978-3-642-11301-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics