Abstract
In modern Information Retrieval, traditional relevance feedback techniques, which utilize the terms in the relevant documents to enrich the user’s initial query, is an effective method to improve retrieval performance. In this paper, we re-examine this method and show that it does not hold in reality – many expansion terms identified in traditional approaches are indeed unrelated to the query and harmful to the retrieval. We then propose a Text Classification Based method for relevance feedback. The classifier trained on the feedback documents can classify the rest of the documents. Thus, in the result list, the relevant documents will be in front of the non-relevant documents. This new approach avoids modifying the query via text classification algorithm in the relevance feedback, and it is a new direction for the relevance feedback techniques. Our Experiments on TREC dataset demonstrate that retrieval effectiveness can be much improved when text classification is used.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Amo, P., Ferreras, F.L., Cruz, F., et al.: Smoothing functions for automatic relevance feedback in information retrieval. In: Proc of the 11th International Workshop on Database and Expert Systems Applications, pp. 115–119 (2000)
Spink, A., Jansen, B.J., Wolfram, D., Saracevic, T.: From e-sex to e-commerce: Web search changes. IEEE Computer 35(3), 107–109 (2002)
Bishop, C.: Patten recognition and machine learning. Springer, Berlin (2006)
Salton, G., McGill, M.J.: Introduction to Modern Information Retrieval. McGraw-Hill, New York (1983)
Cao, G., Nie, J.-Y., Gao, J., Robertson, S.: Selecting good expansion terms for pseudo-relevance feedback. In: Proc of ACM SIGIR Conference on Research and Development in Information Retrieval (2008)
Hsu, C.W., Chang, C.C., Lin, C.J.: A practical guide to support vector classification (2003-08-10/2004-11-10), http://www.csie.ntu.edu.tw/-cjlin/papers/guide/guide.pdf
Iwayama, M.: Relevance feedback with a small number of relevance judgements. In: Proc of ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 10–16 (2000)
Rocchio, J.J.: Relevance feedback in information retrieval//Salton G. In: The Smart Retrieval System: Experiments in Automatic Document Processing, pp. 313–323. Prentice Hall, New Jersey (1971)
Tan, B., Velivellia, Fang, H., et al.: Term feedback for information retrieval with language models. In: Proc of ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 263–270 (2007)
Vapnik, V.: Statistical Learning Theory. John Wiley & Sons Inc., New York (1998)
Shen, X., Zhai, C.: Active feedback in ad hoc information retrieval. In: Proc. of ACM SIGIR conference on Research and development in information retrieval, March 2005, pp. 55–66 (2005)
Wang, X., Fang, H., Zhai, C.: Improve retrieval accuracy for difficult queries using negative feedback. In: CIKM, pp. 991–994 (2007)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Chen, Z., Lu, Y. (2010). Using Text Classification Method in Relevance Feedback. In: Nguyen, N.T., Le, M.T., Świątek, J. (eds) Intelligent Information and Database Systems. ACIIDS 2010. Lecture Notes in Computer Science(), vol 5991. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-12101-2_45
Download citation
DOI: https://doi.org/10.1007/978-3-642-12101-2_45
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-12100-5
Online ISBN: 978-3-642-12101-2
eBook Packages: Computer ScienceComputer Science (R0)