Abstract
Emotion-based music retrieval provides a natural and humanized way to help people experience music. In this paper, we utilize the three-dimensional Resonance-Arousal-Valence emotion model to represent the emotions invoked by music, and the relationship between acoustic features and their emotional impact based on this model is established. In addition, we also consider the emotional tag features for music, and then represent acoustic features and emotional tag features jointly in a low dimensional embedding space for music emotion, while the joint emotion space is optimized by minimizing the joint loss of acoustic features and emotional tag features through dimension reduction. Finally we construct a unified framework for music retrieval in joint emotion space by the means of query-by-music or query-by-tag or together, and then we utilize our proposed ranking algorithm to return an optimized ranked list that has the highest emotional similarity. The experimental results show that the joint emotion space and unified framework can produce satisfying results for emotion-based music retrieval.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Argyriou, A., Evgeniou, T., Pontil, M.: Convex multi-task feature learning. Machine Learning 73(3), 243–272 (2006)
Bigand, E., Vieillard, S., Madurell, F., Marozeau, J., Dacquet, A.: Multidimensional scaling of emotional responses to music: The effect of musical expertise and of the duration of the excerpts. Cognition & Emotion 19(8), 1113–1139 (2005)
Csiszár, I., Tusnády, G.: Information geometry and alternating minimization procedures. Statistics and Decisions suppl (1), 205–237 (1984)
Evangelista, A.J.: Google distance between words. Frontiers A Journal of Women Studies, 1–3 (2006)
Gebhard, P.: Alma: a layered model of affect. In: Autonomous Agents & Multiagent Systems/Agent Theories, Architectures, and Languages, pp. 29–36 (2005)
Han, B.-J., Rho, S., Dannenberg, R.B., Hwang, E.: SMERS: Music Emotion Recognition Using Support Vector Regression. In: International Society for Music Information Retrieval, Number Ismir, pp. 651–656 (2009)
Hu, X., Downie, J.S.: Exploring mood metadata: Relationships with genre, artist and usage metadata. In: International Symposium on Music Information Retrieval (2007)
Hu, X., Downie, J.S., Ehmann, A.F.: Lyric text mining in music mood classification. Information Retrieval 183(Ismir), 411–416 (2009)
Kim, Y.E., Schmidt, E.M., Migneco, R., Morton, B.G., Richardson, P., Scott, J., Speck, J.A., Turnbull, D.: Music emotion recognition: a state of the art review. Information Retrieval (Ismir), 255–266 (2010)
Lartillot, O., Toiviainen, P.: Mir in matlab (ii): A toolbox for musical feature extraction from audio. Spectrum (Ii), 127–130 (2007)
Lu, L.L.L., Liu, D., Zhang, H.-J.Z.H.-J.: Automatic mood detection and tracking of music audio signals (2006)
Nanopoulos, A., Karydis, I.: Know thy neighbor: Combining audio features and social tags for effective music similarity. In: International Conference on Acoustics, Speech, and Signal Processing, pp. 165–168 (2011)
Ortony, A., Clore, G.L., Collins, A.: The Cognitive Structure of Emotions, vol. 18. Cambridge University Press (1988)
Ruxanda, M.M., Chua, B.Y., Nanopoulos, A., Jensen, C.S.: Emotion-based music retrieval on a well-reduced audio feature space. In: International Conference on Acoustics, Speech, and Signal Processing, pp. 181–184 (2009)
Scherer, K.: Which emotions can be induced by music? what are the underlying mechanisms? and how can we measure them? Journal of New Music Research 33(3), 239–251 (2004)
Schimmack, U., Reisenzein, R.: Experiencing activation: energetic arousal and tense arousal are not mixtures of valence and activation. Emotion 2(4) (2002)
Schmidt, E.M., Turnbull, D., Kim, Y.E.: Feature selection for content-based, time-varying musical emotion regression categories and subject descriptors. Spectrum, 267–273 (2010)
Thayer, R.: The biopsychology of mood and arousal. Oxford University Press (1989)
Turnbull, D.R., Barrington, L., Lanckriet, G.R.G., Yazdani, M.: Combining audio content and social context for semantic music discovery. In: Research and Development in Information Retrieval, pp. 387–394 (2009)
Weston, J., Bengio, S., Hamel, P.: Large-scale music annotation and retrieval: Learning to rank in joint semantic spaces. CoRR, abs/1105.5196 (2011)
Yan, S., Xu, D., Zhang, B., Zhang, H.-J., Yang, Q., Lin, S.: Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Transactions on Pattern Analysis and Machine Intelligence 29(1), 40–51 (2007)
Yang, Y., Chen, H.: Ranking-based emotion recognition for music organization and retrieval. IEEE Transactions on Audio Speech and Language (2010)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Deng, J.J., Leung, C.H.C. (2013). Music Retrieval in Joint Emotion Space Using Audio Features and Emotional Tags. In: Li, S., et al. Advances in Multimedia Modeling. MMM 2013. Lecture Notes in Computer Science, vol 7732. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35725-1_48
Download citation
DOI: https://doi.org/10.1007/978-3-642-35725-1_48
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-35724-4
Online ISBN: 978-3-642-35725-1
eBook Packages: Computer ScienceComputer Science (R0)