Abstract
In the computer vision research field, the “Bag of Words” model is known as a popular method for image representation. The Information Bottleneck principle derived from the rate-distortion theory in basic information theory has been applied to many applications in machine learning. In this paper, we introduce a method which combines the two state-of-the-art techniques for image clustering. Images are firstly represented using the “Bag of Words” model, and in the process of clustering based on Information Bottleneck principle, we utilize the Bregman divergence algorithm which works like k-means to get the optimal clustering result. Through the experimental results, we present several points of improvement obtained by the proposed method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Fei-Fei, L., Perona, P.: A bayesian hierarchical model for learning natural scene categories. In: Proc. of IEEE International Conf. on Computer Vision and Pattern Recognition, pp. 524–531 (2005)
Sivic, J., Russell, B.C., Efros, A.A., Zisserman, A., Freeman, W.T.: Discovering object categories in image collections. In: Proc. IEEE International Conf. on Computer Vision (2005)
Cover, T.M., Thomas, J.A.: Elements of Information Theory. John Wiley and Sons, New York (1991)
Tishby, N., Pereira, F.C., Bialek, W.: Information bottleneck method. In: Proc. 37th Annual Allerton Conf. on Communication, Control and Computing, pp. 368–377 (1999)
Goldberger, J., Gordon, S., Greenspan, H.: Unsupervised image-set clustering using an information theoretic framework. IEEE Trans. Image Processing 15(2), 449–458 (2006)
Goldberger, J., Greenspan, H., Gordon, S.: Unsupervised image clustering using the information bottleneck method. In: Proc. 24th DAGM Symposium on Pattern Recognition, pp. 158–165 (2002)
Gordon, S., Greenspan, H., Goldberger, J.: Applying the information bottleneck principle to unsupervised clustering of discrete and continuous image representations. In: Proc. 9th IEEE International Conf. on Computer Vision (2003)
Slonim, N., Tishby, N.: Agglomerative information bottleneck. In: Advances in Neural Information Processing Systems, pp. 617–623 (1999)
Banerjee, A., Merugu, S., Dhillon, I.S., Ghosh, J.: Clustering with bregman divergences. Journal of Machine Learning Research 6, 1705–1749 (2005)
Dhillon, I.S., Mallela, S., Kumar, R.: A divisive information-theoretic feature clustering algorithm for text classification. Journal of Machine Learning Research 3, 1265–1287 (2003)
Computational Vision: Archive (2005), http://www.vision.caltech.edu/html-files/archive.html
Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental bayesian approach tested on 101 object categories. In: IEEE Workshop on Generative-Model Based Vision, CVPR 2004 (2004)
Lowe, D.G.: Object recognition from local scale-invariant features. In: Proc. IEEE International Conf. on Computer Vision (1999)
Affine Covariant Features (2007), http://www.robots.ox.ac.uk/~vgg/research/affine/
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Bai, X., Luo, S. (2010). Combining Bag of Words Model and Information Theoretic Method for Image Clustering. In: Zeng, Z., Wang, J. (eds) Advances in Neural Network Research and Applications. Lecture Notes in Electrical Engineering, vol 67. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-12990-2_48
Download citation
DOI: https://doi.org/10.1007/978-3-642-12990-2_48
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-12989-6
Online ISBN: 978-3-642-12990-2
eBook Packages: EngineeringEngineering (R0)