Abstract
Expressive interpretation forms an important but complex aspect of music, in particular in certain forms of classical music. Modeling the relation between musical expression and structural aspects of the score being performed, is an ongoing line of research. Prior work has shown that some simple numerical descriptors of the score (capturing dynamics annotations and pitch) are effective for predicting expressive dynamics in classical piano performances. Nevertheless, the features have only been tested in a very simple linear regression model. In this work, we explore the potential of a non-linear model for predicting expressive dynamics. Using a set of descriptors that capture different types of structure in the musical score, we compare the predictive accuracies of linear and non-linear models. We show that, in addition to being (slightly) more accurate, non-linear models can better describe certain interactions between numerical descriptors than linear models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In the machine learning literature \(\{w_1,\dots ,w_{D_l}^{(l)}\}\) and \(w_{0}^{(l)}\) are respectively referred to as the set of weights and the bias of the l-th layer.
- 2.
See Table 1 in [10] for an overview of the rules of the KTH model.
References
Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006). Microsoft Research Ltd
Bishop, L., Bailes, F., Dean, R.T.: Performing musical dynamics. Music Percept. 32(1), 51–66 (2014)
Bresin, R.: Artificial neural networks based models for automatic performance of musical scores. J. New Music Res. 27(3), 239–270 (1998)
Cancino Chacón, C.E., Grachten, M., Widmer, G.: Bayesian linear basis models with gaussian priors for musical expression. Technical report, October 2014
Clarke, E.F.: Generative principles in music. In: Sloboda, J. (ed.) Generative Processes in Music: The Psychology of Performance, Improvisation, and Composition. Oxford University Press, New York (1988)
Dauphin, Y.N., de Vries, H., Chung, J., Bengio, Y.: RMSProp and equilibrated adaptive learning rates for non-convex optimization. arXiv 1502, 4390 (2015)
De Poli, G.S.C., Rodà, A., Vidolin, A., Zanon, P.: Analysis and modeling of expressive intentions in music performance. In: Proceedings of the International Workshop on Human Supervision and Control in Engineering and Music, Kassel, Germany, 21–24, September 2001
Eck, D.: Beat tracking using an autocorrelation phase matrix. In: IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP 2007, pp. 1313–1316. IEEE, January 2007
Flossmann, S., Goebl, W., Grachten, M., Niedermayer, B., Widmer, G.: The magaloff project: an interim report. J. new music Res. 39(4), 363–377 (2010)
Friberg, A., Bresin, R., Sundberg, J.: Overview of the kth rule system for musical performance. Adv. Cogn. Psychol. 2(2–3), 145–161 (2006)
Gabrielsson, A.: Music performance research at the millennium. Psychol. Music 31(3), 221–272 (2003)
Goebl, W.: Melody lead in piano performance: expressive device or artifact? J. Acoust. Soc. Am. 110(1), 563–572 (2001)
Grachten, M., Widmer, G.: Linear basis models for prediction and analysis of musical expression. J. New Music Res. 41(4), 311–322 (2012)
Grachten, M.: Summary of the music performance panel, MOSART workshop 2001, Barcelona. In: MOSART Workshop, pp. 1–17, March 2002
Grachten, M., Cancino Chacón, E.C., Widmer, G.: Analysis and prediction of expressive dynamics using Bayesian linear models. In: Proceedings of the 1st International Workshop on Computer and Robotic Systems for Automatic Music Performance, pp. 545–552, July 2014
Graves, A.: Generating Sequences With Recurrent Neural Networks. arXiv 1308, 850 (2013)
Grindlay, G., Helmbold, D.: Modeling, analyzing, and synthesizing expressive piano performance with graphical models. Mach. Learn. 65(2–3), 361–387 (2006)
van Herwaarden, S., Grachten, M., de Haas, W.B.: Predicting expressive dynamics using neural networks. In: Proceedings of the 15th Conference of the International Society for Music Information Retrieval, pp. 47–52, July 2014
Hinton, G.E., Srivastava, N., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Improving neural networks by preventing co-adaptation of feature detectors. arXiv 1207, 580 (2012)
Juslin, P.: Communicating emotion in music performance: a review and a theoretical framework. In: Juslin, P., Sloboda, J. (eds.) Music and Emotion: Theory and Research, pp. 309–337. Oxford University Press, New York (2001)
Palmer, C.: Anatomy of a performance: sources of musical expression. Music Percept. 13(3), 433–453 (1996)
Palmer, C.: Music performance. Annu. Rev. Psychol. 48, 115–138 (1997)
Ramirez, R., Hazan, A.: Rule induction for expressive music performance modeling. In: ECML Workshop Advances in Inductive Rule Learning, September 2004
Saltelli, A., Annoni, P., Azzini, I., Campolongo, F., Ratto, M., Tarantola, S.: Variance based sensitivity analysis of model output. Design and estimator for the total sensitivity index. Comput. Phys. Commun. 181(2), 259–270 (2010)
Sloboda, J.A.: The communication of musical metre in piano performance. Q. J. Exp. Psychol. 35A, 377–396 (1983)
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014(15), 1929–1958 (2014)
Temperley, D.: Music and Probability. Mit Press, Cambridge (2007)
Teramura, K., Okuma, H.: Gaussian process regression for rendering music performance. In: Proceedings of the 10th International Conference on Music Perception and Cognition (ICMPC), Sapporo, Japan (2008)
Todd, N.: The dynamics of dynamics: a model of musical expression. J. Acoust. Soc. Am. 91, 3540–3550 (1992)
Widmer, G.: Machine discoveries: a few simple, robust local expression principles. J. New Music Res. 31(1), 37–50 (2002)
Widmer, G.: Discovering simple rules in complex data: a meta-learning algorithm and some surprising musical discoveries. Artif. Intell. 146(2), 129–148 (2003)
Widmer, G., Goebl, W.: Computational models of expressive music performance: the state of the art. J. New Music Res. 33(3), 203–216 (2004)
Windsor, W.L., Clarke, E.F.: Expressive timing and dynamics in real and artificial musical performances: using an algorithm as an analytical tool. Music Percept. 15(2), 127–152 (1997)
Acknowledgment
This work is supported by European Union Seventh Framework Programme, through the Lrn2Cre8 (FET grant agreement no. 610859) and the PHENICX (grant agreement no. 601166) projects.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Cancino Chacón, C.E., Grachten, M. (2015). An Evaluation of Score Descriptors Combined with Non-linear Models of Expressive Dynamics in Music. In: Japkowicz, N., Matwin, S. (eds) Discovery Science. DS 2015. Lecture Notes in Computer Science(), vol 9356. Springer, Cham. https://doi.org/10.1007/978-3-319-24282-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-24282-8_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-24281-1
Online ISBN: 978-3-319-24282-8
eBook Packages: Computer ScienceComputer Science (R0)