Skip to main content

Comparing Affect Recognition in Peaks and Onset of Laughter

  • Conference paper
  • First Online:
Principles and Practice of Multi-Agent Systems (CMNA 2015, IWEC 2015, IWEC 2014)

Abstract

Laughter is an important social signal that conveys different emotions like happiness, sadness, anger, fear, surprise, and disgust. Therefore, detecting emotions in the laughter is useful for estimating the emotional state of the user. This paper presents work that detects the emotions in Iranian laughter by using audio features and running four machine learning algorithms, namely, Sequential Minimal Optimization (SMO), Multilayer Perceptron (MLP), Logistic, and Radial Basis Function Network (RBFNetwork). We extracted features such as intensity (minimum, maximum, mean, and standard deviation), energy, power, first 3 formants, and the first thirteen Mel Frequency Cepstral Coefficients. Two datasets are used: one that contains segments of full laughter episodes and one that contains only laughter onsets. Results indicate that MLP algorithm produce the highest rate of accuracy which is 86.1372% for first dataset and 85.0123% for second dataset. Besides, using the combination of MFCC and prosodic features led to better results. This means that recognition of emotions is possible at the start of laughter, which is useful for real-time applications.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://www.irangrand.ir.

  2. 2.

    http://www.seemorgh.com.

  3. 3.

    http://tandorosti.hmg.ir.

  4. 4.

    http://manual.audacityteam.org/o/man/noise_removal.html.

  5. 5.

    http://web.stanford.edu/dept/linguistics/corpora/material/PRAAT_workshop_manual_v421.pdf.

  6. 6.

    http://www.cs.waikato.ac.nz/ml/weka.

  7. 7.

    http://cs229.stanford.edu/notes/cs229-notes1.pdf.

References

  1. Eyben, F., Petridis, S., Schuller, B., Tzimiropoulos, G.: Audiovisual classification of vocal outbursts in human conversation using long-short-term memory networks. In: 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5844–5847. IEEE, Prague (2001)

    Google Scholar 

  2. Galvan, C., Manangan, D., Sanchez, M., Wong, J., Cu, J.: Audiovisual affect recognition in spontaneous filipino laughter. In: 2011 3rd International Conference on Knowledge and Systems Engineering (KSE), pp. 266–271. IEEE, Hanoi (2011)

    Google Scholar 

  3. Hamidi, M., Mansoorizade, M.: Emotion recognition from persian speech with neural network. Int. J. Artif. Intell. Appl. 5, 107–112 (2012)

    Google Scholar 

  4. Iliev, A.I., Zhang, Y., Scordilis, M.S.: Spoken emotion classification using ToBI features and GMM. In: 14th International Workshop on 2007 and 6th EURASIP Conference Focused on Speech and Image Processing, Multimedia Communications and Services, Systems, Signals and Image Processing, pp. 495–498. IEEE, Maribor (2007)

    Google Scholar 

  5. Kennedy, L.S., Ellis, D. P.: Laughter detection in meetings. In: NIST ICASSP (2004)

    Google Scholar 

  6. Kudiri, K.M., Verma, G.K., Gohel, B.: Relative amplitude based features for emotion detection from speech. In: 2010 International Conference on Signal and Image Processing (ICSIP), pp. 301–304. IEEE, Chennai (2010)

    Google Scholar 

  7. Miranda, M., Alonzo, J. A., Campita, J., Lucila, S., Suarez, M.: Discovering emotions in Filipino laughter using audio features. In: 2010 3rd International Conference on Human-Centric Computing (HumanCom), pp. 1–6. IEEE, Cebu (2010)

    Google Scholar 

  8. Oyang, Y.-J., Hwang, W., Ou, Y.-Y., Chen, C.-Y., Chen, Z.-W.: Data classification with radial basis function networks based on a novel kernel density estimation algorithm. IEEE Trans. Neural Netw. 16, 225–236 (2005). IEEE

    Article  Google Scholar 

  9. Rodriguez, R.L., Ataollahi, F., Cabredo, R.: Modelling and detecting emotions of Filipino laughter using audio signal. In: 14th Philippine Computing Science Congress, pp. 168–174. Computing Society of the Philippines (2014)

    Google Scholar 

  10. Szameitat, D.P., Alter, K., Szameitat, A.J., Wildgruber, D., Sterr, A., Darwin, C.J.: Acoustic profiles of distinct emotional expressions in laughter. J. Acoust. Soc. Am. 126, 354–366 (2009)

    Article  Google Scholar 

  11. Tanaka, H., Campbell, N.: Acoustic features of four types of laughter in natural conversational speech. In: ICPHS XVII, Hong Kong, pp. 1958–1961 (2011)

    Google Scholar 

  12. Truong, K.P., Van Leeuwen, D.A.: Automatic detection of laughter. In: INTERSPEECH, pp. 485–488 (2005)

    Google Scholar 

  13. Urbain, J., Bevacqua, E., Dutoit, T., Moinet, A., Niewiadomski, R.: The AVLaughterCycle database. In: LREC 2010, Malta, pp. 2996–3001 (2010)

    Google Scholar 

  14. Urbain, J., Niewiadomski, R., Bevacqua, E., Dutoit, T., Moinet, A.: The AVLaughterCycle. Multimodal User Interfaces 4, 47–58 (2010). Springer

    Article  Google Scholar 

  15. Witten, I.H., Frank, E.: Data Mining – Practical Machine Learning Tools and Techniques, 3rd edn. Morgan Kaufmann, Burlington (2011)

    MATH  Google Scholar 

  16. Yap, B.W., Rani, K.A., Rahman, H.A.A., Fong, S., Khairudin, Z., Abdullah, N.N.: An application of oversampling, undersampling, bagging and boosting in handling imbalanced datasets. In: 1st International Conference on Advanced Data and Information Engineering (DaEng-2013), vol. 285, pp. 13–22. Springer, Singapore (2013)

    Google Scholar 

  17. Zare, S.: Home and away: blogging emotions in a Persian virtual Dowreh: a thesis presented in fulfilment of the requirements for the degree of Doctor of Philosophy in Linguistics and Second Language Teaching at Massey University. Massey University, Palmerston North (2011)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Faramarz Ataollahi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Ataollahi, F., Suarez, M.T. (2016). Comparing Affect Recognition in Peaks and Onset of Laughter. In: Baldoni, M., et al. Principles and Practice of Multi-Agent Systems. CMNA IWEC IWEC 2015 2015 2014. Lecture Notes in Computer Science(), vol 9935. Springer, Cham. https://doi.org/10.1007/978-3-319-46218-9_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-46218-9_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-46217-2

  • Online ISBN: 978-3-319-46218-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics