Abstract
Generalisation (test set) and empirical (training-set) classification errors are meaningful characteristics of any pattern classification system. Generally, one needs to know both these error rates and their relationship with the training-set sizes, the number of features, and the type of the classification rule. This knowledge can help one to choose a classifier of the proper complexity, with an optimal number of features, and to determine a sufficient number of training vectors. While training the non-linear SLP, one initially begins with the Euclidean distance classifier and then moves dynamically towards six increasingly complex statistical classifiers. Therefore, utilisation of theoretical generalisation error results obtained for these seven statistical classifiers becomes a guide for analysing the small sample properties of neural net generated classification algorithms.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag London Limited
About this chapter
Cite this chapter
Raudys, Š. (2001). Performance and the Generalisation Error. In: Statistical and Neural Classifiers. Advances in Pattern Recognition. Springer, London. https://doi.org/10.1007/978-1-4471-0359-2_3
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0359-2_3
Publisher Name: Springer, London
Print ISBN: 978-1-85233-297-6
Online ISBN: 978-1-4471-0359-2
eBook Packages: Springer Book Archive