Summary
We study the consistency and asymptotic normality of the LS estimator of a function h(θ) of the parameters θ in a nonlinear regression model with observations \(y_i=\eta(x_i,\theta) +\varepsilon_i\), \(i=1,2\ldots\) and independent errors ε i . Optimum experimental design for the estimation of h(θ) frequently yields singular information matrices, which corresponds to the situation considered here. The difficulties caused by such singular designs are illustrated by a simple example: depending on the true value of the model parameters and on the type of convergence of the sequence of design points \(x_1,x_2\ldots\) to the limiting singular design measure ξ, the convergence of the estimator of h(θ) may be slower than \(1/\sqrt{n}\), and, when convergence is at a rate of \(1/\sqrt{n}\) and the estimator is asymptotically normal, its asymptotic variance may differ from that obtained for the limiting design ξ (which we call irregular asymptotic normality of the estimator). For that reason we focuss our attention on two types of design sequences: those that converge strongly to a discrete measure and those that correspond to sampling randomly from ξ. We then give assumptions on the limiting expectation surface of the model and on the estimated function h which, for the designs considered, are sufficient to ensure the regular asymptotic normality of the LS estimator of h(θ).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Atkinson, A. Donev, A. (1992). Optimum Experimental Design. Oxford University Press, NY, USA.
Bierens, H. (1994). Topics in Advanced Econometrics. Cambridge University Press, Cambridge.
Billingsley, P. (1971). Weak Convergence of Measures: Applications in Probability. SIAM, Philadelphia.
Elfving, G. (1952). Optimum allocation in linear regression. The Annals of Mathematical Statistics, 23, 255–262.
Fedorov, V. (1972). Theory of Optimal Experiments. Academic Press, New York.
Gallant, A. (1987). Nonlinear Statistical Models. Wiley, New York.
Hero, A., Fessler, J., Usman, M. (1996). Exploring estimator bias-variance tradeoffs using the uniform CR bound. IEEE Transactions on Signal Processing, 44, 2026–2041.
Ivanov, A. (1997). Asymptotic Theory of Nonlinear Regression. Kluwer, Dordrecht.
Jennrich, R. (1969). Asymptotic properties of nonlinear least squares estimation. The Annals of Mathematical Statistics, 40, 633–643.
Kiefer, J. Wolfowitz, J. (1959). Optimum designs in regression problems. The Annals of Mathematical Statistics, 30, 271–294.
Lehmann, E. Casella, G. (1998). Theory of Point Estimation. Springer, Heidelberg.
Pázman, A. (1980). Singular experimental designs. Math. Operationsforsch. Statist., Ser. Statistics, 16, 137–149.
Pázman, A. (1986). Foundations of Optimum Experimental Design. Reidel (Kluwer group), Dordrecht (co-pub. VEDA, Bratislava).
Pázman, A. Pronzato, L. (1992). Nonlinear experimental design based on the distribution of estimators. Journal of Statistical Planning and Inference, 33, 385–402.
Pázman, A. Pronzato, L. (2006). On the irregular behavior of LS estimators for asymptotically singular designs. Statistics & Probability Letters, 76, 1089–1096.
Pronzato, L. Pázman, A. (1994). Second-order approximation of the entropy in nonlinear least-squares estimation. Kybernetika, 30, (2)187–198. Erratum. 32(1):104, 1996.
Shiryaev, A. (1996). Probability. Springer, Berlin.
Silvey, S. (1980). Optimal Design. Chapman & Hall, London.
Sjöberg, J., Zhang, Q., Ljung, L., Benveniste, A., Delyon, B., Glorennec, P.-Y., Hjalmarsson, H., Juditsky, A. (1995). Nonlinear black-box modeling in system identification: a unified overview. Automatica, 31, (12)1691–1724.
Spivak, M. (1965). Calculus on Manifolds. A Modern Approach to Classical Theorems of Advanced Calculus. W. A. Benjamin, Inc., New York.
Stoica, P. (2001). Parameter estimation problems with singular information matrices. IEEE Transactions on Signal Processing, 49, 87–90.
Wu, C.-F. (1980). Characterizing the consistent directions of least squares estimates. The Annals of Statistics, 8, (4)789–801.
Wu, C.-F. (1981). Asymptotic theory of nonlinear least squares estimation. The Annals of Statistics, 9, (3)501–513.
Wu, C.-F. (1983). Further results on the consistent directions of least squares estimators. The Annals of Statistics, 11, (4)1257–1262.
Wynn, H. (1972). Results in the theory and construction of D. -optimum experimental designs Journal of the Royal Statistical Society B, 34, 133–147.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix. Proofs of Lemmas 1–3
Appendix. Proofs of Lemmas 1–3
Proof of Lemma 1. We can write
where N(x)/N is the relative frequency of the point x in the sequence \(x_1,x_2,\ldots,x_N\). The last sum for \(x \in S_\xi\) tends to zero a.s. and uniformly on Θ, since N(x)/N tends to \(\xi(\{x\})\), and \([1/N(x)] \sum_{k=1, \ x_k=x}^N\alpha_k\) converges a.s. to \({\rm I}\!{\rm E}\{\alpha_1\}\). The first sum on the right-hand side is bounded by
This expression tends a.s. to zero, since \(N({\mathcal X} \setminus S_\xi)/N\) tends to zero, and the law of large numbers applies for the remaining part in case \(N({\mathcal X}\setminus S_\xi)\rightarrow \infty\).
Proof of Lemma 2. We use a construction similar to that in Bierens (1994, p. 43). Take some fixed \(\theta^1 \in \Theta\) and consider the set
Define \(\bar {a}_\delta(z)\) and \(\underline{a}_\delta(z)\) as the maximum and the minimum of \(a(z,\theta)\) over the set \({\mathcal B}(\theta^1,\delta)\).
The expectations \({\rm I}\!{\rm E}\{|\underline{a}_\delta(z)|\}\) and \({\rm I}\!{\rm E}\{ |\bar{a}_\delta(z)|\}\) are bounded by
Also, \(\bar{a}_\delta(z) - \underline{a}_\delta(z)\) is an increasing function of δ. Hence, we can interchange the order of the limit and expectation in the following expression
which proves the continuity of \({\rm I}\!{\rm E}\{ a(z,\theta)\}\) at θ1 and implies
Hence we can write for every \(\theta \in {\mathcal B}(\theta^1,\delta(\beta))\)
From the strong law of large numbers, we have that \(\forall \gamma>0\), \(\exists N_1(\beta,\gamma)\) such that
Combining with previous inequalities, we obtain
It only remains to cover Θ with a finite numbers of sets \({\mathcal B}(\theta^i,\delta(\beta))\), \(i=1,\ldots,n(\beta)\), which is always possible from the compactness assumption. For any \(\alpha > 0, \beta > 0\), take \(\gamma = \alpha /n(\beta)\), \(N(\beta) =\max_iN_i(\beta,\gamma)\). We obtain
which completes the proof.
Proof of Lemma 3. Since P θ is the orthogonal projector onto \({\mathcal L}_\theta\) it is sufficient to prove that \(\bar{\alpha}{\mathop {\sim}\limits^{\xi}} \bar{\theta}\) implies that any element of \({\mathcal L}_{\bar{\alpha}}\) is in \({\mathcal L}_{\bar{\theta}}\).
From \(\{{\bf f}_{{\bar{\theta}}}\}_1,\ldots,\{{\bf f}_{{\bar{\theta}}}\}_p\) we choose r functions that form a linear basis of \({\mathcal L}_{\bar{\theta}}\). Without any loss of generality we can suppose that they are the first r ones. Decompose θ into \(\theta = (\beta, \gamma)\), where β corresponds to the first r components of θ and γ to the p – r remaining ones. Define similarly \({\bar{\theta}}=(\bar{\beta},\bar{\gamma})\). From A4, the components of \(\partial \eta[x,(\beta,\gamma)] / \partial\gamma\) are linear combinations of components of \(\partial\eta[x,(\beta,\gamma)]/\partial \beta\) not only for \(\theta = \bar{\theta}\) but also for θ in some neighborhood of \(\bar{\theta}\).
Define the following mapping G from \({\mathbb R}^{r+p}\) to \({\mathbb R}^r\) by
From \(\bar{\alpha}{\mathop {\sim}\limits^{\xi}} \bar{\theta}\) we obtain \(G(\bar{\beta},\bar{\alpha})=0\). The matrix
is a nonsingular r × r submatrix of \({\bf M}(\xi,\bar{\theta})\), with rank \([{\bf M}(\xi,\theta)] =r\) for θ in a neighborhood of \(\bar{\theta}\). From the Implicit Function Theorem, see Spivak (1965, Th. 2–12, p. 41), there exist neighborhoods \({\mathcal V}({\bar{\alpha}})\), \({\mathcal W}({\bar{\beta}})\) and a differentiable mapping \(\psi: {\mathcal V}({\bar{\alpha}}) \rightarrow {\mathcal W}({\bar{\beta}})\) such that \(\psi(\bar{\alpha})=\bar{\beta}\) and that \(\alpha \in {\mathcal V}({\bar{\alpha}})\) implies \(G[\psi(\alpha),\alpha]=0\). It follows that
Since the components of \(\partial \eta[x,(\beta,\gamma)]/\partial \gamma\) are linear combinations of the components of \(\partial \eta[x,(\beta,\gamma)] /\partial \beta\) for any \(\theta = (\beta,\gamma)\) in some neighborhood of \(\bar{\theta}\), we obtain from (8.24)
Combining with (8.24) we obtain that
for all α belonging to some neighborhood \({\mathcal U}({\bar{\alpha}})\). We can make \({\mathcal U}({\bar{\alpha}})\) small enough to satisfy the inequality \(\left\| \eta[x,(\psi(\alpha),\bar{\gamma})] -\eta(x,\bar{\theta}) \right\|_\xi^2<\epsilon\) required in A5. It follows that \((\psi(\alpha),\bar{\gamma}) {\mathop {\sim}\limits^{\xi}} \alpha\), that is, \(\eta(\cdot,\alpha) {\mathop {=}\limits^{\xi}} \eta[\cdot,(\psi(\alpha),\bar{\gamma})]\) for all α in a neighborhood of \(\bar{\alpha}\). By taking derivatives we then obtain
that is, \({\mathcal L}_{\bar{\alpha}}{\mathop {\subset}\limits^{\xi}} {\mathcal L}_{(\psi( \bar{ \alpha}), \bar{\gamma}) } = {\mathcal L}_{\bar{\theta}}\).
By interchanging \(\bar{\alpha}\) with \(\bar{\theta}\) we obtain \({\mathcal L}_{\bar{\theta}}{\mathop {\subset}\limits^{\xi}} {\mathcal L}_{\bar{\alpha}}\).
Rights and permissions
Copyright information
© 2009 Springer Science+Business Media LLC
About this chapter
Cite this chapter
Pázman, A., Pronzato, L. (2009). Asymptotic Normality of Nonlinear Least Squares under Singular Experimental Designs. In: Pronzato, L., Zhigljavsky, A. (eds) Optimal Design and Related Areas in Optimization and Statistics. Springer Optimization and Its Applications, vol 28. Springer, New York, NY. https://doi.org/10.1007/978-0-387-79936-0_8
Download citation
DOI: https://doi.org/10.1007/978-0-387-79936-0_8
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-79935-3
Online ISBN: 978-0-387-79936-0
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)