Skip to main content

Analysis of Music Timbre Features for the Construction of User-Specific Affect Model

  • Conference paper
Theory and Practice of Computation

Abstract

Music emotion research has led to identifying timbre as a feature influencing human affect. This work constructs a user-specific affect model identifying music induced emotion using several timbre features. A corpora of music-emotion data was collected, which includes 150 30-second long instrumental segments and self-annotated emotion labels. Several pieces were found whose timbral content induces a consistent emotional response. To find the relationship between emotion and timbre, 60 timbre feature derivatives were used along with 13 MFCC features. Experiments using four classifiers yielded accuracy between 44% to 72%.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Yi-Husan, Y., Ya-Fan, S., Yu-Ching, L., Chen, H.: Music emotion recognition: The role of individuality. In: Proceedings of the International Workshop on Human-Centered Multimedia, pp. 13–22 (2007)

    Google Scholar 

  2. Scherer, K.R., Zentner, M.R.: Emotional effects of music: production rules, ch. 16, pp. 361–392. Oxford University Press, Oxford (2001)

    Google Scholar 

  3. Krumhansl, C.: Music: A link between cognition and emotion. Current Directions in Psychological Science 11(2), 45–50 (2002)

    Article  Google Scholar 

  4. Theimer, W., Vatolkin, I.: Introduction to methods for music classification based on audio data. Tech. Rep., Nokia Research Center (2007)

    Google Scholar 

  5. Levy, M., Sandler, M.: Lightweight measures of timbral similarity of musical audio. In: Proceedings of the 1st ACM Workshop on Audio Music Computing Multimedia, pp. 27–36 (2006)

    Google Scholar 

  6. Savard, A.: Content-based music classification based on timbre similarity. MUMT611: Music Information Acquisition, Preservation and Retrieval Course notes (2006)

    Google Scholar 

  7. Gouyon, F., Pachet, F., Delerue, O.: On the use of zero-crossing rate for an application of classification of percussive sounds. In: Proceedings of the COST G-6 Conference on Digital Audio Effects, DAFX 2000 (2000)

    Google Scholar 

  8. Tindale, A., Kapur, A., Tzanetakis, G., Fujinaga, I.: Retrieval of percussion gestures using timber classification techniques. In: Proceedings of the International Conference on Music Information Retrieval, pp. 541–544 (2004)

    Google Scholar 

  9. Aucouturier, J., Pachet, F.: Music similarity measures: What’s the use. In: Proceedings of the International Conference on Music Information Retrieval (2002)

    Google Scholar 

  10. Weng, C., Lin, C., Jang, J.: Music instrument identification using mfcc: Ehru as an example. Tech. Rep., TainanNationalCollege of the Arts, Taiwan (2004)

    Google Scholar 

  11. Mckay, C.: jaudio: Towards a standardized extensible audio music feature extraction syste (2006)

    Google Scholar 

  12. Agostini, G., Longair, M., Pollastri, E.: Musical instrument timbers classification with spectral features. EURASIP Journal on Applied Signal Processing, 5–14 (2003)

    Google Scholar 

  13. Sato, N., Obuchi, Y.: Emotion recognition using mel-frequency ceptral coefficients. Information and Media Technologies 2(3), 835–848 (2007)

    Google Scholar 

  14. Molau, S., Pitz, M., Schluter, R., Ney, H.: Computing mel-frequency cepstral coefficients on the power spectrum. Tech. Rep., University of Technology, Germany (2001)

    Google Scholar 

  15. Curtis, C.: A brief view of music similarity analysis techniques (2008)

    Google Scholar 

  16. Bigand, E., Veillard, S., Madurell, F., Marozeau, J., Dacquet, A.: Multidimensional scaling of emotional responses to music: The effect of musical expertise and of the duration of the excerpts. Cognition and Emotion 19(8), 1113–1139 (2005)

    Article  Google Scholar 

  17. van de Laar, B.: Emotion detection using music, a survey. In: Twente Student Conference on IT (2006)

    Google Scholar 

  18. Li, T., Tzanetakis, G.: Factors in automatic musical genre classification of audio. In: 2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (2003)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer Tokyo

About this paper

Cite this paper

Avisado, H.G., Cocjin, J.V., Gaverza, J.A., Cabredo, R., Cu, J., Suarez, M. (2012). Analysis of Music Timbre Features for the Construction of User-Specific Affect Model. In: Nishizaki, Sy., Numao, M., Caro, J., Suarez, M.T. (eds) Theory and Practice of Computation. Proceedings in Information and Communications Technology, vol 5. Springer, Tokyo. https://doi.org/10.1007/978-4-431-54106-6_3

Download citation

  • DOI: https://doi.org/10.1007/978-4-431-54106-6_3

  • Publisher Name: Springer, Tokyo

  • Print ISBN: 978-4-431-54105-9

  • Online ISBN: 978-4-431-54106-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics