Abstract
Probability theory is the basis for understanding and modeling random error. Conditional probability leads to Bayes’ rule and allows us to incorporate measured class-conditional probability density functions into a statistical classifier. The naïve Bayes classifier is a simple probabilistic classifier that assumes independent features. The multivariate Gaussian describes normal distributions in multi-dimensional feature space and is characterized by elliptical isocontours. The covariance matrix summarizes the relationships between the various features. Diagonalization of the covariance matrix by Principal Component Analysis results in independent feature combinations. The Mahalanobis distance is a generalization of the Euclidean distance and can be used to classify objects on the basis of the distance to the class center.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Fisher, R.A.: The use of multiple measurements in taxonomic problems. Ann. Eugen. 7, 179–188 (1936)
Gonick, L., Smith, W.: The Cartoon Guide to Statistics. Harper Collins, New York (1993)
Zhang, H.: The optimality of Naïve Bayes. Proceedings of 17th International FLAIRS (Florida Artificial Intelligence Research Society) Conference. AAAI Press (2004)
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media New York
About this chapter
Cite this chapter
Dougherty, G. (2013). Statistical Pattern Recognition. In: Pattern Recognition and Classification. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-5323-9_4
Download citation
DOI: https://doi.org/10.1007/978-1-4614-5323-9_4
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-5322-2
Online ISBN: 978-1-4614-5323-9
eBook Packages: Computer ScienceComputer Science (R0)