Abstract
In this chapter we assume that x ∈ ℝp with E x = µ and var x = Σ = (σij). When the dimension p is large, the principal components method seeks to replace x by y ∈ ℝk, where k < p (and hopefully much smaller), without losing too much “information.” This is sometimes particularly useful for a graphical description of the data since it is much easier to view vectors of low dimension. Section 10.2 defines principal components and gives their interpretation as normalized linear combinations with maximum variance. In Section 10.3, we explain an optimal property of principal components as best approximating subspace of dimension k in terms of squared prediction error. Section 10.4 introduces the sample principal components; they give the coordinates of the projected data which is closest, in terms of euclidian distance, to the original data. Section 10.5 treats the sample principal components calculated from the correlation matrix. Finally, Section 10.6 presents a simple test for multivariate normality which generalizes the univariate Shapiro and Wilk’s statistic. A book entirely devoted to principal component analysis is that of Jolliffe (1986).
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 1999 Springer-Verlag New York, Inc.
About this chapter
Cite this chapter
(1999). Principal components. In: Theory of Multivariate Statistics. Springer Texts in Statistics. Springer, New York, NY. https://doi.org/10.1007/978-0-387-22616-3_10
Download citation
DOI: https://doi.org/10.1007/978-0-387-22616-3_10
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-98739-2
Online ISBN: 978-0-387-22616-3
eBook Packages: Springer Book Archive