Skip to main content

Dimension Reduction and Clustering

  • Chapter
  • First Online:
  • 1777 Accesses

Part of the book series: Springer Series in Astrostatistics ((SSIA,volume 3))

Abstract

For multivariate analysis with p variables the problem that often arises is the ambiguous nature of the correlation or covariance matrix. When p is moderately or very large it is generally difficult to identify the true nature of relationship among the variables as well as observations from the covariance or correlation matrix. Under such situations a very common way to simplify the matter is to reduce the dimension by considering only those variables (actual or derived) which are truly responsible for the overall variation. Important and useful dimension reduction techniques are Principal Component Analysis (PCA), Factor Analysis, Multidimensional Scaling, Independent Component Analysis (ICA), etc. Among them PCA is the most popular one. One may look at this method in three different ways. It may be considered as a method of transforming correlated variables into uncorrelated one or a method of finding linear combinations with relatively small or large variability or a tool for data reduction. The third criterion is more data oriented. In PCA primarily it is not necessary to make any assumption regarding the underlying multivariate distribution but if we are interested in some inference problems related to PCA then assumption of multivariate normality is necessary. The eigen values and eigen vectors of the covariance or correlation matrix are the main contributors of a PCA. The eigen vectors determine the directions of maximum variability whereas the eigen values specify the variances. In practice, decisions regarding the quality of the principal component approximation should be made on the basis of eigen value–eigen vector pairs.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   119.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  • Albazzaz, H., and X.Z. Wang. 2004. Industrial and Engineering Chemistry Research 43(21):6731.

    Google Scholar 

  • Babu, J., et al. 2009. The Astrophysical Journal 700:1768.

    Article  Google Scholar 

  • Chattopadhyay, A.K., T. Chattopadhyay, E. Davoust, S. Mondal, and M. Sharina. 2009. The Astrophysical Journal 705:1533.

    Article  Google Scholar 

  • Chattopadhyay A.K., S. Mondal, and T. Chattopadhyay, 2013. Computational Statistics & Data Analysis 57:17.

    Article  MathSciNet  Google Scholar 

  • Comon, P. 1994. Signal Processing 36:287.

    Article  MATH  Google Scholar 

  • Dickens, R.J. 1972. Monthly Notices of Royal Astronomical Society 157:281

    Article  Google Scholar 

  • Fusi Pecci, F., et al. 1993. Astronomical Journal 105:1145.

    Article  Google Scholar 

  • Gabriel, K.R. 1971. Biometrika 5:453.

    Article  MathSciNet  Google Scholar 

  • Hastie, T., and R. Tibshirani. 2003. In Independent component analysis through product density estimation in advances in neural information processing system, vol. 15, ed. Becker, S., and K. Obermayer, 649–656. Cambridge, MA: MIT Press.

    Google Scholar 

  • Hyvarinen, A., and E. Oja. 2000. Neural Networks 13(4–5):411.

    Google Scholar 

  • Hyvarinen, A., J. Karhunen, and E. Oja. 2001. Independent component analysis. New York: Wiley.

    Book  Google Scholar 

  • King, I.R. 2002. Introduction to Classical Stellar Dynamics. Moscow: URSS.

    Google Scholar 

  • McLaughlin, D.E., et al. 2008. Monthly Notices of the Royal Astronomical Society 384:563.

    Article  Google Scholar 

  • Qiu, D., and A.C. Tamhane. 2007. Journal of Statistical Planning and Inference 137:3722

    Article  MathSciNet  MATH  Google Scholar 

  • Recio-Blanco, A., et al. ( 2006). Astronomy & Astrophysics 452:875

    Article  Google Scholar 

  • Salaris, M., et al. 2004. Astronomy & Astrophysics 420:911.

    Article  Google Scholar 

  • Shapiro, S.S., and M.B. Wilk. 1965. Biometrika 52(3–4):591.

    Google Scholar 

  • Stones, V. 2004. Independent component analysis: a tutorial introduction. Bradford Books. Cambridge: The MIT Press.

    Google Scholar 

  • Sugar, A.S., and G.M. James. 2003. Journal of the American Statistical Association 98:750.

    Article  MathSciNet  MATH  Google Scholar 

  • Woodley, K.A., et al. 2007. Astronomical Journal 134:494.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

7.1 Electronic Supplementary material

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer Science+Business Media New York

About this chapter

Cite this chapter

Chattopadhyay, A.K., Chattopadhyay, T. (2014). Dimension Reduction and Clustering. In: Statistical Methods for Astronomical Data Analysis. Springer Series in Astrostatistics, vol 3. Springer, New York, NY. https://doi.org/10.1007/978-1-4939-1507-1_7

Download citation

Publish with us

Policies and ethics