Abstract
Suppose that the sample \( {D_n} = \left\{ {({x_{i,1}},{x_{i,2}},...,{x_{i,m}},{y_i})} \right\}_{i = 1}^n = \left\{ {\left( {{x_{i,}}{y_i}} \right)} \right\}_{i = 1}^n \) comprises n independent observations on m explanatory variables x j , j = 1,…, m and one dependent variable y and that each observation can be regarded as a realization of an (m + 1)-dimensional distribution function Ξ(x, y) =Ψ(y|x)Ω(x) (the operating model) which will sometimes also be denoted by F for simplicity. We view the observations as being generated by an unknown function ø(x) with the addition of a stochastic component, commonly taken to be independently and identically distributed (i.i.d.) with zero mean and constant variance σ2, i.e.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag London
About this chapter
Cite this chapter
Zapranis, A., Refenes, AP.N. (1999). Neural Model Identification. In: Principles of Neural Model Identification, Selection and Adequacy. Perspectives in Neural Computing. Springer, London. https://doi.org/10.1007/978-1-4471-0559-6_2
Download citation
DOI: https://doi.org/10.1007/978-1-4471-0559-6_2
Publisher Name: Springer, London
Print ISBN: 978-1-85233-139-9
Online ISBN: 978-1-4471-0559-6
eBook Packages: Springer Book Archive