Abstract
The earliest attempts to predict an example’s class based on the known attribute values go back to well before World War II—prehistory, by the standards of computer science. Of course, nobody used the term “machine learning,” in those days, but the goal was essentially the same as the one addressed in this book.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
We assume here that 100 is the maximum value observed in the training set. Alternatively, our background knowledge may inform us that the given attribute’s value cannot exceed 100.
References
Chow, C. K. (1957). An optimum character recognition system using decision functions. IRE Transactions on Computers, EC-6, 247–254.
Fisher, R. A. (1936). The use of multiple measurement in taxonomic problems. Annals of Eugenics, 7, 111–132.
Good, I. J. (1965). The estimation of probabilities: An essay on modern Bayesian methods. Cambridge: MIT.
Neyman, J. & Pearson E. S. (1928). On the use and interpretation of certain test criteria for purposes of statistical inference. Biometrica, 20A, 175–240.
Parzen E. (1962). On estimation of a probability density function and mode. Annals of Mathematical Statistics, 33, 1065–1076.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this chapter
Cite this chapter
Kubat, M. (2017). Probabilities: Bayesian Classifiers. In: An Introduction to Machine Learning. Springer, Cham. https://doi.org/10.1007/978-3-319-63913-0_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-63913-0_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-63912-3
Online ISBN: 978-3-319-63913-0
eBook Packages: Computer ScienceComputer Science (R0)