Summary
Independent component analysis (ICA) was developed in the signal processing and neural computation communities. Its original purpose was to solve what is called the blind source separation problem: when linear mixtures of some original source signals are observed, the goal is to recover the source signals, using minimum assumptions on the mixing matrix (i.e. blindly). This leads to a linear model that is very similar to the one used in factor analysis. What makes ICA fundamentally different from conventional factor analysis is that the source signals are assumed to be non-Gaussian,in addition to the basic assumption of their independence. In fact, this implies that the model can be uniquely estimated from the data, using supplementary information that is not contained in the covariance matrix. Interestingly, a very close connection can be found with projection pursuit: The basic form of the ICA model can be estimated by finding the projections that are maximally non-Gaussian, which is the goal of projection pursuit as well. On the other hand, the dimension of the observed data vector is often first reduced by principal component analysis, in which case ICA can be viewed as a method of determining the factor rotation using the non-Gaussianity of the factors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bell, A. and Sejnowski, T. (1995). An information-maximization approach to blind separation and blind deconvolution. Neural Computation, 7: 1129–1159.
Cardoso, J.-F. (1989). Source separation using higher order moments. In Proc. IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP’89), pages 2109–2112, Glasgow, UK.
Comon, P. (1994). Independent component analysis—a new concept? Signal Processing, 36: 287–314.
Delfosse, N. and Loubaton, P. (1995). Adaptive blind separation of independent sources: a deflation approach. Signal Processing, 45: 59–83.
Hérault, J. and tins, B. (1984). Circuits neuronaux à synapses modifiables: décodage de messages composites par apprentissage non supervisé. Comptes.-Rendus de l’Académie des Sciences, 299 (III-13): 525–528.
Huber, P. (1985). Projection pursuit. The Annals of Statistics, 13 (2): 435–475.
Hyvärinen, A. (1999). Fast and robust fixed-point algorithms for independent com-ponent analysis. IEEE Trans. on Neural Networks, 10 (3): 626–634.
Hyvärinen, A., Karhunen, J., and Oja, E. (2001). Independent Component Anal-ysis. Wiley Interscience.
Jutten, C. and Hérault, J. (1991). Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture. Signal Processing, 24: 1–10.
Lawley, D. and Maxwell, A. (1971). Factor Analysis as a Statistical Method.
London: Butterworths. Mooijaart, A. (1985). Factor analysis for non-normal variables. Psychometrica, 50: 323–342.
Olshausen, B. A. and Field, D. J. (1996). Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381: 607–609.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2003 Springer Japan
About this paper
Cite this paper
Hyvärinen, A., Kano, Y. (2003). Independent Component Analysis for Non-Normal Factor Analysis. In: Yanai, H., Okada, A., Shigemasu, K., Kano, Y., Meulman, J.J. (eds) New Developments in Psychometrics. Springer, Tokyo. https://doi.org/10.1007/978-4-431-66996-8_75
Download citation
DOI: https://doi.org/10.1007/978-4-431-66996-8_75
Publisher Name: Springer, Tokyo
Print ISBN: 978-4-431-66998-2
Online ISBN: 978-4-431-66996-8
eBook Packages: Springer Book Archive