Abstract
The generalization error bounds found by current error models using the number of effective parameters of a classifier and the number of training samples are usually very loose. These bounds are intended for the entire input space. However, support vector machines, radial basis function neural networks and multilayer Perceptron neural networks are local learning machines for solving problems and thus treat unseen samples near the training samples as more important. In this chapter, we describe a localized generalization error model which bounds the generalization error from above within a neighborhood of the training samples using a stochastic sensitivity measure (Yeung et al., 2007 and Ng et al., 2007). This model is then used to develop an architecture selection technique for a classifier with maximal coverage of unseen samples by specifying a generalization error threshold.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Yeung, D.S., Cloete, I., Shi, D., Ng, W.W. (2009). Localized Generalization Error Model. In: Sensitivity Analysis for Neural Networks. Natural Computing Series. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02532-7_5
Download citation
DOI: https://doi.org/10.1007/978-3-642-02532-7_5
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02531-0
Online ISBN: 978-3-642-02532-7
eBook Packages: Computer ScienceComputer Science (R0)