Skip to main content

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2888))

Abstract

The k-Nearest-Neighbours (kNN) is a simple but effective method for classification. The major drawbacks with respect to kNN are (1) its low efficiency – being a lazy learning method prohibits it in many applications such as dynamic web mining for a large repository, and (2) its dependency on the selection of a “good value” for k. In this paper, we propose a novel kNN type method for classification that is aimed at overcoming these shortcomings. Our method constructs a kNN model for the data, which replaces the data to serve as the basis of classification. The value of k is automatically determined, is varied for different data, and is optimal in terms of classification accuracy. The construction of the model reduces the dependency on k and makes classification faster. Experiments were carried out on some public datasets collected from the UCI machine learning repository in order to test our method. The experimental results show that the kNN based model compares well with C5.0 and kNN in terms of classification accuracy, but is more efficient than the standard kNN.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hand, D., Mannila, H., Smyth, P.: Principles of Data Mining. MIT Press, Cambridge (2001)

    Google Scholar 

  2. Wang, H.: Nearest Neighbours without k: A Classification Formalism based on Probability, technical report, Faculty of Informatics, University of Ulster, N.Ireland, UK (2002)

    Google Scholar 

  3. Sebastiani, F.: Machine Learning in Automated Text Categorization. ACM Computing Surveys 34(1), 1–47 (2002)

    Article  Google Scholar 

  4. Wang, H., Duntsch, I., Bell, D.: Data Reduction Based on Hyper Relations. In: Proceedings of KDD 1998, New York, pp. 349–353 (1998)

    Google Scholar 

  5. Hart, P.: The Condensed Nearest Neighbour Rule. IEEE Transactions on Information Theory 14, 515–516 (1968)

    Article  Google Scholar 

  6. Gates, G.: The Reduced Nearest Neighbour Rule. IEEE Transactions on Information Theory 18, 431–433 (1972)

    Article  Google Scholar 

  7. Alpaydin, E.: Voting Over Multiple Condensed Nearest Neoghbors. Artificial Intelligence Review 11, 115–132 (1997); © Kluwer Academic Publishers (1997)

    Article  Google Scholar 

  8. Kubat, M., Jr., M.: Voting Nearest-Neighbour Subclassifiers. In: Proceedings of the 17th International Conference on Machine Learning, ICML 2000, pp. 503–510, Stanford, CA, June 29-July 2 (2000)

    Google Scholar 

  9. Wilson, D.R., Martinez, T.R.: Reduction Techniques for Exemplar-Based Learning Algorithms. Machine learning 38(3), 257–286 (2000)

    Article  MATH  Google Scholar 

  10. Mitchell, T.: Machine Learning. MITPress/McGraw-Hill (1997)

    Google Scholar 

  11. Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, UK (1995)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Guo, G., Wang, H., Bell, D., Bi, Y., Greer, K. (2003). KNN Model-Based Approach in Classification. In: Meersman, R., Tari, Z., Schmidt, D.C. (eds) On The Move to Meaningful Internet Systems 2003: CoopIS, DOA, and ODBASE. OTM 2003. Lecture Notes in Computer Science, vol 2888. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-39964-3_62

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-39964-3_62

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-20498-5

  • Online ISBN: 978-3-540-39964-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics