Skip to main content

Image Similarities on the Basis of Visual Content – An Attempt to Bridge the Semantic Gap

  • Conference paper
Intelligent Information and Database Systems (ACIIDS 2011)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 6591))

Included in the following conference series:

Abstract

Image similarities is a useful concept regarding to the image retrieval on the basis of visual content of the images (CBIR - Content Based Image Retrieval). Because an image can have far more interpretations than text, visual similarity can be totally different from semantic similarity. We have developed similar images searching tools using global approaches as well as local approaches to find near similar images. In this paper we propose a method of bridging local and global levels, what should solve the problem of limited, non-adaptable dictionary when we use automatic annotations in a similar images retrieving task. Our faraway goal is to face the difficult problem with all current approaches to CBIR systems, connected with visual similarity: the semantic gap between low-level content and higher-level concepts.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Broda, B., Kwasnicka, H., Paradowski, M., Stanek, M.: MAGMA: efficient method for image annotation in low dimensional feature space based on Multivariate Gaussian Models. In: Ganzha, M., Paprzycki, M. (eds.) Proc. of the IMCSIT. Polish Information Processing Society 2009, pp. 131–138 (2009)

    Google Scholar 

  2. Brown, M., Lowe, D.G.: Recognising panoramas. In: Ninth IEEE International Conference on Computer Vision, vol. 2, pp. 12–18 (2003)

    Google Scholar 

  3. Carneiro, G., Chan, A., Moreno, P., Vasconcelos, N.: Supervised learning of semantic classes for image annotation and retrieval. IEEE Trans. on Pattern Analysis and Machine Intelligence 29(3), 394–410 (2007)

    Article  Google Scholar 

  4. Chatzichristofis, S., Boutalis, Y.: Cedd: Color and edge directivity descriptor: A compact descriptor for image indexing and retrieval. Computer Vision Systems, 312–322 (2008)

    Google Scholar 

  5. Chatzichristofis, S.A., Boutalis, Y.S.: Fcth: Fuzzy color and texture histogram – a low level feature for accurate image retrieval. In: Interactive Services, Intern. Workshop on Image Analysis for Multimedia, pp. 191–196 (2008)

    Google Scholar 

  6. Datta, R., Joshi, D., Li, J., Wang, J.: Image retrieval: Ideas, influences, and trends of the new age. ACM Transactions on Computing 40(2) (2008)

    Google Scholar 

  7. Dickinson, S.J., Leonardis, A., Schiele, B., Tarr, M.J. (eds.): Object Categorization: Computer and Human Vision Perspectives. Cambridge University Press, Cambridge (2009)

    Google Scholar 

  8. Duygulu, P., Barnard, K., de Freitas, J.F.G., Forsyth, D.A.: Object recognition as machine translation: Learning a lexicon for a fixed image vocabulary. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002. LNCS, vol. 2353, pp. 97–112. Springer, Heidelberg (2002)

    Chapter  Google Scholar 

  9. Feng, S.L., Manmatha, R., Lavrenko, V.: Multiple bernoulli relevance models for image and video annotation. In: Proc. of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2, pp. 1002–1009 (2004)

    Google Scholar 

  10. Hironobu, Y.M., Takahashi, H., Oka, R.: Image-to-word transformation based on dividing and vector quantizing images with words. Neural Networks in Boltzmann Machines 4 (1999)

    Google Scholar 

  11. Inoue, M.: On the need for annotation-based image retrieval. In: Proc. of the Information Retrieval in Context (IRiX), A Workshop at SIGIR 2004, pp. 44–46 (2004)

    Google Scholar 

  12. Jin, Y., Khan, L., Wang, L., Awad, M.: Image annotations by combining multiple evidence and wordnet. In: Proc. of the 13th Annual ACM Intern. Conf. on Multimedia (2005)

    Google Scholar 

  13. Kwasnicka, H., Paradowski, M.: Resulted word counts optimization-a new approach for better automatic image annotation. Pattern Recognition 41(12) (2008)

    Google Scholar 

  14. Lavrenko, V., Manmatha, R., Jeon, J.: A model for learning the semantics of pictures. In: Proc. of Neural Information Processing Systems (NIPS). MIT Press, Cambridge (2003)

    Google Scholar 

  15. Llorente, A., Motta, E., Roger, S.: Image annotation refinement using web-based keyword correlation. In: Proc. of the 4th International Conference on Semantic and Digital Media Technologies, Berlin, pp. 188–191 (2009)

    Google Scholar 

  16. Lowe, D.G.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2), 91–110 (2004)

    Article  Google Scholar 

  17. Maier, O., Stanek, M., Kwasnicka, H.: PATSI - photo annotation through similar images with annotation length optimization. In: Klopotek, M.A., et al. (eds.) Intelligent information systems 2010, pp. 219–232. Pub. House of Univ. of Podlasie (2010)

    Google Scholar 

  18. Matas, J., Chum, O., Urban, M., Pajdla, T.: Robust Wide Baseline Stereo from Maximally Stable Extremal Regions. In: Proc. of British Machine Vision Conference, pp. 384–393 (2002)

    Google Scholar 

  19. Makadia, A., Pavlovic, V., Kumar, S.: A new baseline for image annotation. In: Proc. of the 10th European Conference on Computer Vision, pp. 316–329 (2008)

    Google Scholar 

  20. Mikolajczyk, K., Schmid, C.: Scale and Affine Invariant Interest Point Detectors. International Journal of Computer Vision 60, 63–86 (2004)

    Article  Google Scholar 

  21. Monay, F., Quelhas, P., Odobez, J.M., Gatica-Perez, D.: Contextual Classification of Image Patches with Latent Aspect Models. EURASIP Journal on Image and Video Processing, Article ID 602920 (2009), doi:10.1155/2009/602920

    Google Scholar 

  22. Paradowski, M.: Methods of automatic annotation as an efficient tool for images collections describing. PhD thesis, Wroclaw Univ. of Technology (2008) (in Polish)

    Google Scholar 

  23. Paradowski, M., Sluzek, A.: Automatic Visual Object Formation using Image Fragment Matching. In: Proc. of 2010 Intern. Multiconf. on Computer Science and Information Technology, vol. 5, pp. 45–51, IEEE Catalog CFP1064E-CDR (2010)

    Google Scholar 

  24. Russell, R., Sinha, P.: Perceptually-based comparison of image similarity metrics. Technical report, AIM-2001-014, CBCL-201 (2001)

    Google Scholar 

  25. Sluzek, A., Paradowski, M.: A Vision-based Technique for Assisting Visually Impaired People and Autonomous Agents. In: Proc. of 3th Intern. Conference on Human System Interaction, pp. 653–660 (2010)

    Google Scholar 

  26. Smeulders, A.W.M., Gupta, A.: Content-Based Image Retrieval at the End of the Early Years. Transactions on Pattern Analysis and Machine Intelligence 22(12), 1349–1380 (2000)

    Article  Google Scholar 

  27. Stanek, M., Broda, B., Kwasnicka, H.: PATSI — Photo Annotation through Finding Similar Images with Multivariate Gaussian Models. In: Bolc, L., Tadeusiewicz, R., Chmielewski, L.J., Wojciechowski, K. (eds.) ICCVG 2010. LNCS, vol. 6375, pp. 284–291. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  28. Tadeusiewicz, R., Ogiela, M.R.: The New Concept In Computer Vision: Automatic Understanding of the Images. Proc. of Artificial Intelligence and Soft Computing, 133–144 (2004)

    Google Scholar 

  29. The American Heritage® Dictionary of the English Language. 4th edn. copyright ©2000, by Houghton Mifflin Company. Updated in 2009, Houghton Mifflin Company Pub. (2000)

    Google Scholar 

  30. Yang, D., Sluzek, A.: A low-dimensional local descriptor incorporating TPS warping for image matching. Image and Vision Computing 28(8) (2010)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kwasnicka, H., Paradowski, M., Stanek, M., Spytkowski, M., Sluzek, A. (2011). Image Similarities on the Basis of Visual Content – An Attempt to Bridge the Semantic Gap. In: Nguyen, N.T., Kim, CG., Janiak, A. (eds) Intelligent Information and Database Systems. ACIIDS 2011. Lecture Notes in Computer Science(), vol 6591. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-20039-7_2

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-20039-7_2

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-20038-0

  • Online ISBN: 978-3-642-20039-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics