Recognition of Emotions in German Speech Using Gaussian Mixture Models

  • Martin Vondra
  • Robert Vích
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 5398)


The contribution describes experiments with recognition of emotions in German speech signal based on the same principle as recognition of speakers. The most robust algorithm for speaker recognition is based on Gaussian Mixture Models (GMM). We examine three parameter sets: the first contains suprasegmental features, in the second are segmental features and the last is a combination of the two previous parameter sets. Further we want to explore the dependency of the classification accuracy on the number of GMM model components. The aim of this contribution is a recommendation for the number of GMM components and the optimal selection of speech parameters for emotion recognition in German speech.


speech emotions emotion recognition Gaussian mixture models 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Reynolds, D.A.: Speaker identification and verification using Gaussian mixture speaker models. Speech Communication 17, 91–108 (1995)CrossRefGoogle Scholar
  2. 2.
    Burkhardt, F., Paeschke, A., Rolfes, M., Sendlmeier, W., Weiss, B.: A Database of German Emotional Speech. In: Proc. Interspeech 2005, Lisbon, Portugal, September 4-8 (2005)Google Scholar
  3. 3.
    Truong, K.P., Leeuven, D.A.: An ‘open-set’ detection evaluation methology for automatic emotion recognition in speech. In: ParaLing 2007: Workshop on Paralinguistic Speech - between models and data, Saarbrücken, Germany (2007)Google Scholar
  4. 4.
    Morrison, D., Wang, R., De Silva, L.C.: Ensemble methods for spoken emotion recognition in call-centers. Speech Communication 49 (2007)Google Scholar
  5. 5.
    Sjölander, K., Beskow, J.: Wavesurfer,
  6. 6.
    Brookes, M.: VOICEBOX: Speech Processing Toolbox for MATLAB,
  7. 7.
    Nwe, T.L., Foo, S.W., De Silva, L.C.: Speech emotion recognition using hidden markovov models. Speech Communication 41, 603–623 (2003)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2009

Authors and Affiliations

  • Martin Vondra
    • 1
  • Robert Vích
    • 1
  1. 1.Institute of Photonics and ElectronicsAcademy of Sciences of the Czech RepublicPrague 8Czech Republic

Personalised recommendations