Skip to main content

A Non-parametric Dimensionality Reduction Technique Using Gradient Descent of Misclassification Rate

  • Conference paper
Pattern Recognition and Image Analysis (ICAPR 2005)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 3687))

Included in the following conference series:

  • 1898 Accesses

Abstract

We present a technique for dimension reduction. The technique uses a gradient descent approach to attempt to sequentially find orthogonal vectors such that when the data is projected onto each vector the classification error is minimised. We make no assumptions about the structure of the data and the technique is independent of the classifier model used. Our approach has advantages over other dimensionality reduction techniques, such as Linear Discriminant Analysis (LDA), which assumes unimodal gaussian distributions, and Principal Component Analysis (PCA) which is ignorant of class labels. In this paper we present the results of a comparison of our technique with PCA and LDA when applied to various 2-dimensional distributions and the two class cancer diagnosis task from theWisconsin Diagnostic Breast Cancer Database, which contains 30 features.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Juang, B.H., Katagiri, S.: Discriminative learning for minimum error classification [pattern recognition]. IEEE Transactions on Signal Processing [see also IEEE Transactions on Acoustics, Speech, and Signal Processing] 40(12), 3043–3054 (1992)

    MATH  Google Scholar 

  2. Wang, X., Paliwal, K.K.: Using minimum classification error training in dimensionality reduction, vol. 1, pp. 338–345 (2000)

    Google Scholar 

  3. Witten, I.H., Frank, E.: Data Mining: Practical MachineLearning tools andtechniques with Java Implementations. Morgan Kaufman, San Francisco (1999)

    Google Scholar 

  4. Jolliffe, I.T.: Principal Component Analysis. Springer, New York (2002)

    MATH  Google Scholar 

  5. Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annu. Eugenics 7, 179–188 (1936)

    Google Scholar 

  6. Okada, T., Tomita, S.: An optimal orthonormal system for discriminant analysis. Pattern Recognition 18(2), 139–144 (1985)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Redmond, S., Heneghan, C. (2005). A Non-parametric Dimensionality Reduction Technique Using Gradient Descent of Misclassification Rate. In: Singh, S., Singh, M., Apte, C., Perner, P. (eds) Pattern Recognition and Image Analysis. ICAPR 2005. Lecture Notes in Computer Science, vol 3687. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11552499_18

Download citation

  • DOI: https://doi.org/10.1007/11552499_18

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-28833-6

  • Online ISBN: 978-3-540-31999-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics