Abstract
It is common practice to apply linear or nonlinear feature extraction methods before classification. Usually linear methods are faster and simpler than nonlinear ones but an idea successfully employed in the nonlinearization of Support Vector Machines permits a simple and efective extension of several statistical methods to their nonlinear counterparts. In this paper we follow this general nonlinearization approach in the context of Independent Component Analysis, which is a general purpose statistical method for blind source separation and feature extraction. In addition, nonlinearized formulae are furnished along with an illustration of the usefulness of the proposed method as an unsupervised feature extractor for the classification of Hungarian phonemes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Aizerman, M.A., Braverman, E.M.and Rozonoer L.I., Theoretical foundation of the potential function method in pattern recognition learning, Automat. Remote Cont., vol.25, pp.821–837, 1964.
Albesano, D., De Mori, R., Gemello, R., and Mana, F., study on the efect of adding new dimensions to trajectories in the acoustic space, Proc. of EuroSpeech’ 99, pp.1503–1506, 1999.
Bishop, C.M., NeuralNetworks for Pattern Recognition, Oxford Univ. Press, 1995.
Boser, B.E., Guyon, I.M.and Vapnik V.N, A training algorithm for optimal margin classifier, in Proc.5th Annu. ACM Workshop Computat. Learning Theory, D. Haussler, ed., Pittsburgh, P, pp.144–152, 1992.
Cristianini, N. and Shawe-Taylor, J.An Introduction to Support Vector Machines and other kernel-based learning methods Cambridge University Press, 2000.
Comon, P.Independent component analysis, new concept? Signal Processing, 36:287–314, 1994.
Fastica, http://www.cis.hut.fi/projects/ica/fastica/index.shtml1997.
Hyvärinen, A.family of fixed-point algorithms for independent component analysis In Proceedings of ICASSP, Munich, Germany, 1997.
Hyvärinen, A.New Approximations of Differential Entropy for Independent Component Analysis and Projection Pursuit. In Advances in Neural Information Processing Systems, 10:273–279, MIT Press, 1998.
Jolliffe, I.J.Principal Component Analysis, Springer-Verlag, New York, 1986.
Kernel Machines Web Page,http://www.kernel-machines.org, 2001.
Kocsor, A., Tóth, L., Kuba, A.Jr., Koács, K., Jelasity, M., Gyimóthy, T.and Csirik, J., Comparative Study of Several Feature Transformation and Learning Methods for Phoneme Classification, Int. Journal of Speech Technology, Vol.3., No.3/4, pp.263–276, 2000.
Kocsor, A., Kuba, A.Jr. and Tóth, L.Phoneme Classification Using Kernel Principal Component Analysis, Priodica Polytechnica, in print, 2001.
Kocsor, A., Tóth, L. and Paczolay, D., Nonlinearized Discriminant Analysis and its Application to Speech Impediment Therapy, in V. Matousek et al. (eds.):Text, Speech and Dialogue, Proc. of TSD 2001, Springer Verlag LNAI, in print, 2001.
Müller, K.-R., Mika, S., Rätsch, G., Tsuda, K. and Schölkopf, B., An Introduction to Kernel-Based Learning Algorithms, IEEE T ansactions on Neural Networks, Vol.12, No.2, 2001.
Mika, S., Rätsch, G., Weston, J., Schölkopf, B. and Müller, K.-R., Fisher Discriminant Analysis with kernels, in Neural Networks for Signal Processing IX, Hu, J. Larsen, E. Wilson and S. Douglas, Eds. Piscataway, NJ:IEEE, pp.41–48, 1999.
Rabiner, L. and Juang, B.-H.Fundamentals of Speech Recognition, Prentice Hall, 1993.
Roth, V. and Steinhage, V., Nonlinear discriminant analysis using kernel func-tions, In Advances in Neural Information Processing Systems 12, S.. Solla, T.K. Leen and K.-R. Müeller, Eds. Cambridge, MA:MIT Press, pp.526–532, 2000.
Schölkopf, B., Smola, A. J. and Müller K.-R., Nonlinear component analysis as a kernel eigenvalue problem, neural comput., vol. 10, pp. 1299-1319, 1998.
Toth, L., Kocsor, A., and Koács, K., Discriminative Segmental Speech Model and Its Application to Hungarian Number Recognition, in Sojka, P.et al.(eds.):Text, Speech and DialogueProceedings of TSD 2000, Springer Verlag LNAI series, vol.1902, pp.307–313, 2000.
Vapnik, V.N., Statistical Learning Theory, John Wiley &Sons Inc., 1998.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kocsor, A., Csirik, J. (2001). Fast Independent Component Analysis in Kernel Feature Spaces. In: Pacholski, L., Ružička, P. (eds) SOFSEM 2001: Theory and Practice of Informatics. SOFSEM 2001. Lecture Notes in Computer Science, vol 2234. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45627-9_24
Download citation
DOI: https://doi.org/10.1007/3-540-45627-9_24
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42912-8
Online ISBN: 978-3-540-45627-8
eBook Packages: Springer Book Archive