Abstract
Dimensionality reduction technique involves finding out the transformation matrix that maps from the random vector in the higher dimensional space to the lower dimensional space. This is obtained by identifying the orthonormal basis using PCA, LDA, KLDA, and ICA. In PCA, the basis vectors are identified in the direction of the maximum variance of the data. PCA doesn’t care about the class index associated with the training vectors. This is circumvented using LDA. In this case, the distance between the vectors within the class is minimized and the distance between the centroids of various classes is made apart. After subjected to PCA, the data are made uncorrelated. These are further made independent by projecting the data using ICA basis. In almost all the techniques in pattern recognition techniques, the individual classes are assumed as Gaussian distributed. But in real-time applications, data aren’t Gaussian distributed. Thus the Gaussianity of the data can be minimizing the absolute of the kurtosis measured using the given data or by maximizing the neg entropy of the given data.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsAuthor information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Gopi, E.S. (2020). Dimensionality Reduction Techniques. In: Pattern Recognition and Computational Intelligence Techniques Using Matlab. Transactions on Computational Science and Computational Intelligence. Springer, Cham. https://doi.org/10.1007/978-3-030-22273-4_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-22273-4_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-22272-7
Online ISBN: 978-3-030-22273-4
eBook Packages: EngineeringEngineering (R0)