Skip to main content

Detection of Negative Emotion Using Acoustic Cues and Machine Learning Algorithms in Moroccan Dialect

  • Conference paper
  • First Online:
  • 420 Accesses

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 737))

Abstract

The speech signal provides rich information about the speaker’s emotional state. Therefore, recognition of the emotion in speech has become one of the research themes in the processing of speech and applications based on human-computer interaction. This article provides an experimental study and examines the detection of negative emotions such as fear and anger with regard to the neutral emotional state. The data set is collected from speeches recorded in the Moroccan Arabic dialect. Our aim is first to study the effects of emotion on the selected acoustic characteristics, namely the first four formants F1, F2, F3, F4, the fundamental frequency F0, Intensity, Number of pulses, Jitter and Shimmer and then compare our results to previous works. We also study the influence of phonemes and speaker gender on the relevance of these characteristics in the detection of emotion. To this aim, we performed classification tests using the WEKA software. We found that F0, Intensity, Number of Pulses have the best rates of recognition regardless speaker gender and phonemes. Moreover the second and third formant are the features that highlighted phoneme’s effect.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Aggarwal, R.K., Dave, M.: Performance evaluation of sequentially combined heterogeneous feature streams for Hindi speech recognition system. Telecommun. Syst. 52(3), 1457–1466 (2013)

    Article  Google Scholar 

  2. Albornoz, E.M., Milone, D.H., Rubner, H.L.: Spoken emotion recognition using hierarchical classifiers. Comput. Speech Lang. 25(3), 556–570 (2011)

    Article  Google Scholar 

  3. Banse, R., Scherer, K.R.: Acoustic profiles in vocal emotion expression. J. Personal. Soc. Psychol. 70(3), 614–636 (1996)

    Article  Google Scholar 

  4. Batliner, A., Schuller, B., Seppi, D., Steidl, S., Devillers, L., Vidrascu, L., Vogt, T., Aharonson, V., Amir, N.: The automatic recognition of emotions in speech. In: Emotion-Oriented Systems, pp. 71–99. Springer, Heidelberg (2011)

    Google Scholar 

  5. Boersma, P.: Praat, a system for doing phonetics by computer. Glot Int. 5(9/10), 341–345 (2001)

    Google Scholar 

  6. Burkhardt, F.: Simulation emotionaler Sprechweise mit Sprachsynthese verfahren. Ph.D. thesis, TU Berlin (2001)

    Google Scholar 

  7. Busso, C., Narayanan, S.S.: Interrelation between speech and facial gestures in emotional utterances: a single subject study. IEEE Trans. Audio Speech Lang. Process. 10(20), 1–16 (2007)

    Google Scholar 

  8. Busso, C., Narayanan, S.S.: Joint analysis of the emotional fingerprint in the face and speech: a single subject study. In: International Workshop on Multimedia Signal Processing (MMSP), Chanée, Grèce, pp. 43–47. IEEE, Octobre 2007

    Google Scholar 

  9. Wu, C.H., Liang, W.B.: Emotion recognition of affective speech based on multiple classifiers using acoustic-prosodic information and semantic labels. IEEE Trans. Affect. Comput. 2(1), 1–21 (2012)

    Google Scholar 

  10. Clore, G.L.: Why emotions are felt. In: Ekman, P., Davidson, R.J. (eds.) The Nature of Emotion: Fundamental Questions, pp. 103–111. Oxford University Press, New York (1994)

    Google Scholar 

  11. Cowie, R., Douglas-Cowie, E., Tsapatsoulis, N., Votsis, G., Kollias, S., Fellenz, W., Taylor, J.G.: Emotion recognition in human-computer interaction. IEEE Signal Process. Mag. 18(1), 32–80 (2001)

    Article  Google Scholar 

  12. Damasio, A.: L’erreur de Descartes. Grosset/Putnam, New York (1994)

    Google Scholar 

  13. Davletcharova, A., Sugathan, S., Abraham, B., James, A.P.: Detection and analysis of emotion from speech signals. Procedia Comput. Sci. 58, 91–96 (2015)

    Article  Google Scholar 

  14. Ekman, P.: Expression and the nature of emotion. In: Scherer, K.R., Ekman, P. (eds.) Approaches to Emotion, pp. 319–343. Lawrence Erlbaum Associates, Hillsdale (1984)

    Google Scholar 

  15. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA data mining software: an update. SIGKDD Explor. 11(1), 10–18 (2009)

    Article  Google Scholar 

  16. Hartmann, K., Siegert, I., Philippou-Hübner, D., Wendemuth, A.: Emotion detection in HCI: from speech features to emotion space. In: 12th IFAC Symposium on Analysis, Design, and Evaluation of Human-Machine Systems, Las Vegas, NV, USA, 11–15 August 2013

    Google Scholar 

  17. Hosmer, D.W., Lemeshow, S., Sturdivant, R.X.: Applied Logistic Regression, 3rd edn. Wiley, Hoboken (2013). ISBN 978-0470-58247-3

    Book  MATH  Google Scholar 

  18. Huang, C., Gong, W., Fu, W., Feng, D.: A research of speech emotion recognition based on deep belief network and SVM. Math. Probl. Eng. (2014)

    Google Scholar 

  19. Johnstone, T., Scherer, K.R.: Vocal communication of emotion. In: Lewis, M., Haviland-Jones, J.M. (eds.) Handbook of Emotions, pp. 220–235. Guilford, New York (2000)

    Google Scholar 

  20. Juslin, P.N., Laukka, P.: Communication of emotions in vocal expression and music performance: different channels, same code. Psychol. Bull. 129(5), 770–814 (2003)

    Article  Google Scholar 

  21. Koolagudi, S.G., Rao, K.S.: Emotion recognition from speech: a review. Int. J. Speech Technol. 15, 99–117 (2012)

    Article  Google Scholar 

  22. Mower, E., Matarić, M.J., Narayanan, S.: A framework for automatic human emotion classification using emotion profiles. IEEE Trans. Audio Speech Lang. Process. 19(5), 1057–1070 (2011)

    Article  Google Scholar 

  23. Murray, I.R., Arnott, J.L.: Toward the simulation of emotion in synthetic speech: a review of the litterature on human vocal emotion. J. Acoust. Soc. Am. 93(2), 1097–1108 (1993)

    Article  Google Scholar 

  24. Oudeyer, P.: The production and recognition of emotions in speech: features and algorithms. Int. J. Hum.-Comput. Stud. 59(1–2), 157–183 (2003)

    Google Scholar 

  25. Paeschke, A., Sendlmeier, W.: Prosodic characteristics of emotional speech: measurements of fundamental frequency movements. In: Speech Emotion, pp. 75–80 (2000)

    Google Scholar 

  26. Philippou-Hubner, D., Vlasenko, B., Bock, R., Wendemuth, A.: The performance of the speaking rate parameter in emotion recognition from speech. In: Proceedings of IEEE ICME, pp. 248–253 (2012)

    Google Scholar 

  27. Scherer, K.R.: Vocal affect expression: a review and a model for future research. Psychol. Bull. 99(2), 143–165 (1986)

    Article  Google Scholar 

  28. Scherer, K.R.: How emotion is expressed in speech and singing. In: Proceedings of 1995 ICPhS, Stockholm, pp. 90–96 (1995)

    Google Scholar 

  29. Stibbard, R.: Vocal Expression of Emotions in Non-Laboratory Speech: An Investigation of the Reading/Leeds Emotion in Speech Project Annotation Data, 245 p. Linguistics and Applied Language Studies, University of Reading, Reading, RoyaumeUni (2001)

    Google Scholar 

  30. Nwe, T.L., Foo, S.W., De Silva, L.C.: Detection of stress and emotion in speech using traditional and FFT based log energy features. In: Proceedings of the 4th International Conference on Information, Communications and Signal Processing (2009)

    Google Scholar 

  31. Vlasenko, B., Prylipko, D., Philippou-Hubner, D., Wendemuth, A.: Vowels formants analysis allows straightforward detection of high arousal acted and spontaneous emotions. In: Proceedings of INTERSPEECH 2011, Florence, Italy, pp. 1577–1580 (2011)

    Google Scholar 

  32. Vogt, T., Andre, E.: Comparing feature sets for acted and spontaneous speech in view of automatic emotion recognition. In: IEEE International Conference on Multimedia and Expo, pp. 474–477 (2005)

    Google Scholar 

  33. Yüncü, E., Hacıhabiboğluy, H., Bozşahin, C.: Automatic speech emotion recognition using auditory models with binary decision tree and SVM. In: Proceedings of the 2014 22nd International Conference on Pattern Recognition, pp. 773–778. Computer Society Washington, D.C. (2014). ISBN 978-1-4799-5209-0

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Abdellah Agrima .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Agrima, A., Elmazouzi, L., Mounir, I., Farchi, A. (2018). Detection of Negative Emotion Using Acoustic Cues and Machine Learning Algorithms in Moroccan Dialect. In: Abraham, A., Haqiq, A., Muda, A., Gandhi, N. (eds) Proceedings of the Ninth International Conference on Soft Computing and Pattern Recognition (SoCPaR 2017). SoCPaR 2017. Advances in Intelligent Systems and Computing, vol 737. Springer, Cham. https://doi.org/10.1007/978-3-319-76357-6_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-76357-6_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-76356-9

  • Online ISBN: 978-3-319-76357-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics