Abstract
Investigating subjective values of audio data is both interesting and pleasant topic for research, gaining attention and popularity among researchers recently. We focus on automatic detection of emotions in songs/audio files, using features based on spectral contents. The data set, containing a few hundreds of music pieces, was used in experiments. The emotions are grouped into 13 or 6 classes. We compare our results with tests on human subjects. One of the main conclusions is that multi-label classification is required.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Berger, A.: Error-correcting output coding for text classification. In: IJCAI 1999: Workshop on machine learning for information filtering. Stockholm, Sweden (1999), Available at http://www-2.cs.cmu.edu/aberger/pdf/ecoc.pdf
Boutell, M., Shen, X., Luo, J., Brown, C.: Multi-label Semantic Scene Classification. Technical Report, Dept. of Computer Science, U. Rochester (2003)
Burges, C.J.: A tutorial on support vector machines for pattern recognition. Data mining and knowledge discovery 2(2), 121–167 (1998)
Clare, A., King, R.D.: Knowledge Discovery in Multi-label Phenotype Data. In: Siebes, A., De Raedt, L. (eds.) PKDD 2001. LNCS (LNAI), vol. 2168, pp. 42–53. Springer, Heidelberg (2001)
Fujinaga, I., McMillan, K.: Realtime recognition of orchestral instruments. In: Proceedings of the International Computer Music Conference, pp. 141–143 (2000)
Kostek, B., Wieczorkowska, A.: Parametric Representation Of Musical Sounds. Archives of Acoustics 22(1), 3–26 (1997)
Li, T., Ogihara, M.: Detecting emotion in music. In: 4th International Conference on Music Information Retrieval ISMIR, Washington, D.C., and Baltimore, MD (2003), Available at http://ismir2003.ismir.net/papers/Li.PDF
McCallum, A.: Multi-label Text Classification with a Mixture Model Trained by EM. In: AAAI 1999 Workshop on Text Learning (1999)
Peeters, G., Rodet, X.: Automatically selecting signal descriptors for Sound Classification. In: ICMC 2002 Goteborg, Sweden (2002)
Pollard, H.F., Jansson, E.V.: A Tristimulus Method for the Specification of Musical Timbre. Acustica 51, 162–171 (1982)
Schapire, R., Singer, Y.: Boostexter: a boosting-based system for text categorization. Machine Learning 39(2/3), 135–168 (2000)
Tato, R., Santos, R., Kompe, R., Pardo, J.M.: Emotional Space Improves Emotion Recognition. In: 7th International Conference on Spoken Language Processing ICSLP 2002, Denver, Colorado (2002)
Tzanetakis, G., Cook, P.: Marsyas: A framework for audio analysis. Organized Sound 4(3), 169–175 (2000), Available at http://www-2.cs.cmu.edu/gtzan/work/pubs/organised00gtzan.pdf
Wieczorkowska, A., Wroblewski, J., Synak, P., Slezak, D.: Application of temporal descriptors to musical instrument sound recognition. Journal of Intelligent Information Systems 21(1), 71–93 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Synak, P., Wieczorkowska, A. (2005). Some Issues on Detecting Emotions in Music. In: Ślęzak, D., Yao, J., Peters, J.F., Ziarko, W., Hu, X. (eds) Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing. RSFDGrC 2005. Lecture Notes in Computer Science(), vol 3642. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11548706_33
Download citation
DOI: https://doi.org/10.1007/11548706_33
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-28660-8
Online ISBN: 978-3-540-31824-8
eBook Packages: Computer ScienceComputer Science (R0)