Abstract
In Hettmansperger and Randles (Biometrika 89:851–860, 2002) spatial sign vectors were used to derive simultaneous estimators of multivariate location and shape. Oja (Multivariate nonparametric methods with R. Springer, New York, 2010) proposed a similar approach for the multivariate linear regression case. These estimators are highly robust and have under general assumptions a joint limiting multinormal distribution. The estimates are easy to compute using fixed-point algorithms. There are however no exact proofs for the convergence of these algorithms. The existence and uniqueness of the solutions also still remain unproven although we believe that they hold under general conditions. To circumvent these problems, we consider in this paper k-step versions of Hettmansperger and Randles (HR) location and shape estimators and their extensions to the linear regression problem. The influence functions, limiting distributions and asymptotical efficiencies of the estimators are derived at the multivariate elliptical case.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Arcones, M. A. (1998). Asymptotic theory for M-estimators over a convex kernel. Economic Theory, 14, 387–422.
Bai, Z. D., Chen, R., Miao, B. Q., & Rao, C. R. (1990). Asymptotic theory of least distances estimate in the multivariate linear models. Statistics, 21, 503–519.
Brown, B. M. (1983). Statistical uses of the spatial median. Journal of the Royal Statistical Society, Series B, 45, 25–30.
Chakraborty, B., Chaudhuri, P., & Oja, H. (1998). Operating transformation retransformation on spatial median and angle test. Statistica Sinica, 8, 767–784.
Croux, C., Dehon, C., & Yadine, A. (2010). The k-step spatial sign covariance matrix. Advances in Data Analysis and Classification, 4, 137–150.
Davies, P. L. (1987). Asymptotic behaviour of S-estimates of multivariate location and dispersion matrices. Annals of Statistics, 15, 1269–1292.
Dümbgen, L., & Tyler, D. (2005). On the breakdown properties of some multivariate M-functionals. Scandinavian Journal of Statistics, 32, 247–264.
Fang, K. T., Kotz, S., & Ng, K. W. (1990). Symmetric multivariate and related distributions. London: Chapman and Hall.
Frahm, G. (2009). Asymptotic distributions of robust shape matrices and scales. Journal of Multivariate Analysis, 100, 1329–1337.
Hampel, F. R., Ronchetti, E. M., Rousseeuw, P. J., & Stahel, W. J. (1986). Robust statistics: The approach based on influence functions. New York: Wiley.
Hettmansperger, T. P., & McKean, J. W. (2011). Robust nonparametric statistical methods (2nd ed.). London: Arnold.
Hettmansperger, T. P., & Randles, R. H. (2002). A practical affine equivariant multivariate median. Biometrika, 89, 851–860.
Ilmonen, P., Serfling, R., & Oja, H. (2012). Invariant coordinate selection (ICS) functionals. International Statistical Review, 80, 93–110.
Locantore, N., Marron, J. S., Simpson, D. G., Tripoli, N., Zhang, J. T., Cohen, K. L. (1999). Robust principal components for functional data. Test, 8, 1–28.
Marden, J. I. (1999). Some robust estimates of principal components. Statistics & Probability Letters, 43, 349–359.
Maronna, R. A. (1976). Robust M-estimators of multivariate location and scatter. Annals of Statistics, 4, 51–67.
Oja, H. (1999). Affine invariant multivariate sign and rank tests and corresponding estimates: A review. Scandinavian Journal of Statistics, 26, 319–343.
Oja, H. (2010). Multivariate nonparametric methods with R. New York: Springer.
Ollila, E., Hettmansperger, T. P., Oja, H. (2004). Affine equivariant multivariate sign methods. University of Jyvaskyla. Technical report.
Paindaveine, D. (2008). A canonical definition of shape. Statistics & Probability Letters, 78, 2240–2247.
Puri, M. L., & Sen, P. K. (1971). Nonparametric methods in multivariate analysis. New York: Wiley.
Taskinen, S., Sirkiä, S., & Oja, H. (2010). k-Step estimators of shape based on spatial signs and ranks. Journal of Statistical Planning and Inference, 140, 3376–3388.
Tyler, D., Critchley, F., Dümbgen, L., & Oja, H. (2009). Invariant coordinate selection. Journal of Royal Statistical Society B, 71, 549–592.
Tyler, D. E. (1987). A distribution-free M-estimator of multivariate scatter. Annals of Statistics, 15, 234–251.
Visuri, S., Oja, H., & Koivunen, V. (2000). Sign and rank covariance matrices. Journal of Statistical Planning and Inference, 91, 557–575.
Acknowledgements
The authors wish to thank a referee for several helpful comments and suggestions. The Research was funded by the Academy of Finland (grants 251965 and 268703).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix
Appendix
Proof (Theorem 11.1).
The functional (11.7) solves
where \(\mathbf{z} = \mathbf{V}_{k-1}^{-1/2}(F)(\mathbf{y} -\boldsymbol{\mu }_{k-1}(F))\). Write \(F_{\epsilon } = (1-\epsilon )F_{0} +\epsilon \varDelta _{\mathbf{y}_{0}}\). Then
and, further,
Substituting these in (11.15) and having the expectation at F ε gives
Find next the influence function of V k (F). Write (11.8) as
where again \(\mathbf{z} = \mathbf{V}_{k-1}^{-1/2}(F)(\mathbf{y} -\mathbf{\boldsymbol{\mu }}_{k-1}(F))\). Proceeding then as in the proof for \(\boldsymbol{\mu }_{k}(F)\), we get
The result then follows from the above recursive formulas for \(IF(\mathbf{y};\boldsymbol{\mu }_{k},F_{0})\) and IF(y; V k , F 0). □
Proof (Theorem 11.2).
Consider first the limiting distribution of 1-step HR location estimator. Let \(\mathbf{y}_{i},\ldots,\mathbf{y}_{n}\) be a sample from a spherically symmetric distribution F 0 and write \(r_{i} =\Vert \mathbf{y}_{i}\Vert\) and \(\mathbf{u}_{i} = r_{i}^{-1}\mathbf{y}_{i}\). Further, as we assume that \(\hat{\boldsymbol{\mu }}_{0}\) and \(\hat{\mathbf{V}} _{0}\) are \(\sqrt{n}\)-consistent, we write \(\boldsymbol{\mu }_{0}^{{\ast}}:= \sqrt{n}\hat{\boldsymbol{\mu }}_{0}\) and \(\mathbf{V}_{0}^{{\ast}}:= \sqrt{n}(\hat{\mathbf{V}} _{0} -\mathbf{I}_{p})\), where \(\boldsymbol{\mu }_{0}^{{\ast}} = O_{p}(1)\) and V 0 ∗ = O p (1). Now using the delta-method as in Taskinen et al. (2010), we get
As \(\sqrt{n}\hat{\boldsymbol{\mu }}_{0} = \sqrt{n}\,ave\{\gamma _{0}(r_{i})\mathbf{u}_{i}\} + o_{p}(1)\), the asymptotic normality of \(\sqrt{n}\hat{\boldsymbol{\mu }}_{1}\) follows from the Slutsky’s theorem and joint limiting multivariate normality of \(\sqrt{n}\,ave\{\mathbf{u}_{ i}\}\) and \(\boldsymbol{\mu }_{0}^{{\ast}} = \sqrt{n}\hat{\boldsymbol{\mu }}_{0}\) (and \(E[\mathbf{u}_{i}^{T}\mathbf{V}^{{\ast}}\mathbf{u}_{i}] = Tr(\mathbf{V}^{{\ast}}) = 0\)). Equation (11.16) reduces to
Continuing in a similar way with \(\sqrt{n}\hat{\boldsymbol{\mu }}_{2}\), \(\sqrt{n}\hat{\boldsymbol{\mu }}_{ 3}\), and so on, we finally get
Thus \(\sqrt{n}\,\hat{\boldsymbol{\mu }}_{ k} = \sqrt{n}\,ave\{\gamma _{k}(r_{i})\mathbf{u}_{i}\} + o_{p}(1).\) and the limiting covariance matrix of \(\sqrt{n}\hat{\boldsymbol{\mu }}_{k}\) equals to \(E[\gamma _{k}^{2}(r)\mathbf{u}\mathbf{u}^{T}] = p^{-1}E[\gamma _{k}^{2}(r)]\mathbf{I}_{p}.\)
The limiting distribution for k-step HR shape estimator can be computed as above starting from 1-step estimator
Note that the estimator is scaled so that \(Tr(\hat{\mathbf{V}} _{1}) = p\). After some straightforward derivations,
As the joint limiting distribution of \(\sqrt{n}\,(ave\{\mathbf{u}_{ i}\mathbf{u}_{i}^{T}\} - p^{-1}\mathbf{I}_{ p})\) and \(\sqrt{n}(\hat{\mathbf{V}} _{ 0} -\mathbf{I}_{p}) = \sqrt{n}ave\{\alpha _{0}(r_{i})(\mathbf{u}_{i}\mathbf{u}_{i}^{T} - p^{-1}\mathbf{I}_{ p})\} + o_{p}(1)\) is multivariate normal, the asymptotic normality of \(\sqrt{n}(\hat{\mathbf{V}} _{ 1} -\mathbf{I}_{p})\) follows and
Continuing in the same way, we obtain
The limiting covariance matrix of \(\sqrt{n}\,vec(\hat{\mathbf{V}} _{k} -\mathbf{I}_{p})\) is then
□
Proof (Theorem 11.3).
First note that (11.11) is equivalent to
where \(\mathbf{e} = \mathbf{V}_{k-1}^{-1/2}(\mathbf{y} -\mathbf{B}_{k}^{T}\mathbf{x})\). Proceeding as in the Proof of Theorem 11.1, and assuming (without loss of generality) the spherical case with B = 0 and \(\boldsymbol{\varSigma }= \mathbf{I}_{p}\), we end up after some tedious derivations to
where y = r u with r = | | y | | and \(\mathbf{u} = \mathbf{y}/r\). As \(E[\mathbf{u}\mathbf{u}^{T}] = p^{-1}\mathbf{I}_{p}\), this simplifies to
and as the influence functions for all k are of the same type, we get
□
Proof (Theorem 11.4).
Consider first the general case, where \(\hat{\mathbf{B}} _{0}\) and \(\hat{\mathbf{V}} _{0}\) are assumed to be any \(\sqrt{n}\)-consistent estimators and write \(\mathbf{B}_{0}^{{\ast}} = \sqrt{n}\hat{\mathbf{B}} _{0}\) and \(\mathbf{V}_{0}^{{\ast}} = \sqrt{n}(\hat{\mathbf{V}} _{0} -\mathbf{I}_{p})\), where B 0 ∗ = O p (1) and V 0 ∗ = O p (1).
Without loss of generality, assume that B = 0 and \(\boldsymbol{\varSigma }= \mathbf{I}_{p}\) so that \(\mathbf{y}_{1},\ldots,\mathbf{y}_{n}\) is a random sample from a spherical distribution with zero mean vector zero and identity covariance matrix. Write \(r_{i} =\Vert \mathbf{y}_{i}\Vert\) and \(\mathbf{u}_{i} = \mathbf{y}_{i}/r_{i}\). Now as in the proof of Theorem 11.2, the 1-step HR regression estimator may be written as
The limiting multivariate normality of \(\sqrt{n}\hat{\mathbf{B}}_{ 1}\) then follows from the joint limiting multivariate normality of \(\mathbf{B}_{0}^{{\ast}} = \sqrt{n}\hat{\mathbf{B}} _{0}\) and \(\sqrt{n}ave\{\mathbf{x}_{ i}\mathbf{u}_{i}^{T}\}\) and the Slutsky’s theorem. The above equation then reduces to
where D = E[xx T], and for the k-step HR regression estimator we get
again with a limiting multivariate normality.
Let us next consider the simple case, where the initial estimator is of the type
where η k is given in (11.14). The covariance matrix of \(\sqrt{n}\,vec(\hat{\mathbf{B}} _{k})\) then equals to
□
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Taskinen, S., Oja, H. (2016). Influence Functions and Efficiencies of k-Step Hettmansperger–Randles Estimators for Multivariate Location and Regression. In: Liu, R., McKean, J. (eds) Robust Rank-Based and Nonparametric Methods. Springer Proceedings in Mathematics & Statistics, vol 168. Springer, Cham. https://doi.org/10.1007/978-3-319-39065-9_11
Download citation
DOI: https://doi.org/10.1007/978-3-319-39065-9_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-39063-5
Online ISBN: 978-3-319-39065-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)