Overview
The model selection procedure we employ as part of this methodology is a flexible framework for acquiring consistent estimators; it divides model selection into well-defined tasks, without restraining the implementation choices (see Figure 4.1). Asymptotic consistency for the ‘true’ parameter vector w0 is an essential requirement of the neural model identification methodology presented. In many respects it is quite similar to Vapnik’s and (in particular) Moody’s approaches, which use a heuristic search (a hierarchy of nested model classes) and are based on prediction risk estimates (bounds for prediction risk in the case of Vapnik). For prediction risk e use a somewhat different and more frequently used analytical formulation from Moody’s, along the lines of Murata et al (1993) and Amari (1995), and also a different empirical estimate (bootstrap). Our notational conventions are as in Amari (1995).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London
About this chapter
Cite this chapter
Zapranis, A., Refenes, AP.N. (1999). Neural Model Selection: the Minimum Prediction Risk Principle. In: Principles of Neural Model Identification, Selection and Adequacy. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0559-6_4
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0559-6_4
Publisher Name: Springer, London
Print ISBN: 978-1-85233-139-9
Online ISBN: 978-1-4471-0559-6
eBook Packages: Springer Book Archive