Abstract
In the heart of the computer model of visual attention, an interest or saliency map is derived from an input image in a process that encompasses several data combination steps. While several combination strategies are possible and the choice of a method influences the final saliency substantially, there is a real need for a performance comparison for the purpose of model improvement. This paper presents contributing work in which model performances are measured by comparing saliency maps with human eye fixations. Four combination methods are compared in experiments involving the viewing of 40 images by 20 observers. Similarity is evaluated qualitatively by visual tests and quantitatively by use of a similarity score. With similarity scores lying 100% higher, non-linear combinations outperform linear methods. The comparison with human vision thus shows the superiority of non-linear over linear combination schemes and speaks for their preferred use in computer models.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Ahmed, S.: VISIT: An Efficient Computational Model of Human Visual Attention. PhD thesis, University of Illinois at Urbana-Champaign (1991)
Milanese, R.: Detecting Salient Regions in an Image: from Biological Evidence to Computer implementation. PhD thesis, Dept. Computer Science, Univ. of Geneva, Switzerland (1993)
Tsotsos, J.K.: Toward a computational model of visual attention. In: Papathomas, T.V., Chubb, C., Gorea, A., Kowler, E. (eds.) Early vision and beyond, pp. 207–226. MIT Press, Cambridge (1995)
Treisman, A.M., Gelade, G.: A feature-integration theory of attention. Cognitive Psychology, 97–136 (1980)
Koch, C., Ullman, S.: Shifts in selective visual attention: Towards the under- lying neural circuitry. Human Neurobiology 4, 219–227 (1985)
Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 20(11), 1254–1259 (1998)
Ouerhani, N., Hügli, H.: Real-time visual attention on a massively parallel SIMD architecture. International Journal of Real Time Imaging 9(3), 189–196 (2003)
Clark, J.J., Ferrier, N.J.: Control of visual attention in mobile robots. In: IEEE Conference on Robotics and Automation, pp. 826–831 (1989)
Ouerhani, N., Bur, A., Hügli, H.: Visual attention-based robot self-localization. In: European Conference on Mobile Robotics (ECMR 2005), Ancona, Italy, September 7-10, pp. 8–13 (2005)
Ouerhani, N., Hügli, H.: MAPS: Multiscale attention-based presegmentation of color images. In: Griffin, L.D., Lillholm, M. (eds.) Scale-Space 2003. LNCS, vol. 2695, pp. 537–549. Springer, Heidelberg (2003)
Walther, D., Rutishauser, U., Koch, C., Perona, P.: Selective visual attention enables learning and recognition of multiple objects in cluttered scenes. Computer Vision and Image Understanding 100(1-2), 41–63 (2005)
Itti, L., Koch, C.: A comparison of feature combination strategies for saliency- based visual attention systems. In: SPIE Human Vision and Electronic Imaging IV (HVEI 1999), vol. 3644, pp. 373–382 (1999)
Ouerhani, N., Jost, T., Bur, A., Hügli, H.: Cue normalization schemes in saliency- based visual attention models. In: Proc. International Cognitive Vision Workshop, Graz, Austria (2006)
Jost, T., Ouerhani, N., von Wartburg, R., Mueri, R., Hügli, H.: Assessing the contribution of color in visual attention. International Journal of Computer Vision and Image Understanding (CVIU) 100, 107–123 (2005)
Parkhurst, D., Law, K., Niebur, E.: Modeling the role of salience in the allocation of overt visual attention. Vision Research 42(1), 107–123 (2002)
Itti, L.: Quantitative modeling of perceptual salience at human eye position. Visual Cognition (in press, 2005)
Ouerhani, N., von Wartburg, R., Hügli, H., Müri, R.M.: Empirical validation of Saliency-based model of visual attention. Electronic Letters on Computer Vision and Image Analysis 3(1), 13–24 (2003)
Le Meur, O., Le Callet, P., Barba, D., Thoreau, D.: A coherent computational approach to model bottom-up visual attention. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI) 28(5) (May 2006)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Ouerhani, N., Bur, A., Hügli, H. (2006). Linear vs. Nonlinear Feature Combination for Saliency Computation: A Comparison with Human Vision. In: Franke, K., Müller, KR., Nickolay, B., Schäfer, R. (eds) Pattern Recognition. DAGM 2006. Lecture Notes in Computer Science, vol 4174. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11861898_32
Download citation
DOI: https://doi.org/10.1007/11861898_32
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44412-1
Online ISBN: 978-3-540-44414-5
eBook Packages: Computer ScienceComputer Science (R0)