Abstract
In Chap. 5 I discussed how the impact of music on human decision-making can be modeled. Subsequently in Chap. 6 I discussed how this impact can be leveraged by an agent to engender better interaction with a person. However, that is only one facet of person-agent interaction in musical context. The other includes a scenario in which people and machines actively collaborate in music generation. What would such an interaction be like? An important aspect of person-agent interaction, or of agents interacting with multiple people and/or other agents, is that of reasoning about preferences. Particularly in a domain such as music generation, people’s subjective tastes play a pivotal role, and reasoning about them when trying to collaborate is critical. This train of thought leads to a deeper question: how can multiple agents reason with each other in a shared task while also maintaining individual preferences that may be at odds with the shared task and with the preferences of others? Studying the balance between shared tasks and individual preferences in multiagent interaction is a significant step in fulfillment of Contribution 4 of this book, building towards multiagent music interaction as a meaningful step towards person-agent music generation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Relating the scale to actual notes, 0 denotes C, 1 denotes \(C\#\), 2 denotes D and so forth up to 11 = B.
References
L. Busoniu, R. Babuska, B. De Schutter, A comprehensive survey of multiagent reinforcement learning. IEEE Trans. Syst. Man Cybern.-Part C: Appl. Rev. 38(2), 2008 (2008)
S. Barrett, P. Stone, S. Kraus, A. Rosenfeld, Teamwork with limited knowledge of teammates, in Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence (AAAI), July 2013
P. Stone, G.A. Kaminka, S. Kraus, J.S. Rosenschein, et al., Ad hoc autonomous agent teams: collaboration without pre-coordination, AAAI (2010)
R.S. Sutton, A.G. Barto, in Reinforcement Learning: An Introduction (MIT press, Cambridge, 2018)
R.S. Sutton, D.A. McAllester, S.P. Singh, Y. Mansour, Policy gradient methods for reinforcement learning with function approximation, in Advances in Neural Information Processing Systems (2000), pp. 1057–1063
J. Schulman, F. Wolski, P. Dhariwal, A. Radford, O. Klimov, Proximal policy optimization algorithms (2017), arXiv:1707.06347
J. Schulman, S. Levine, P. Abbeel, M. Jordan, P. Moritz, Trust region policy optimization, in Proceedings of the 32nd International Conference on Machine Learning (ICML-15) (2015), pp. 1889–1897
S. Ross, G.J. Gordon, D. Bagnell, A reduction of imitation learning and structured prediction to no-regret online learning, in AISTATS JMLR Proceedings, vol. 15 (2011), pp. 627–635
P. Abbeel, A.Y. Ng, Apprenticeship learning via inverse reinforcement learning. In Proceedings of the Twenty-first International Conference on Machine Learning, ICML ’04 (ACM, New York, 2004), p. 1
J. Ho, S. Ermon, Generative adversarial imitation learning, ed. by D.D. Lee, M. Sugiyama, U.V. Luxburg, I. Guyon, R. Garnett, Advances in Neural Information Processing Systems 29 (Curran Associates, Inc., 2016), pp. 4565–4573
N. Cook, in A Guide to Musical Analysis (Oxford University Press, Oxford, 1994)
H.W. Kuhn, The hungarian method for the assignment problem. Nav. Res. Logist. Q. 2(1–2), 83–97 (1955)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Liebman, E. (2020). Multiagent Collaboration Learning: A Music Generation Test Case. In: Sequential Decision-Making in Musical Intelligence. Studies in Computational Intelligence, vol 857. Springer, Cham. https://doi.org/10.1007/978-3-030-30519-2_7
Download citation
DOI: https://doi.org/10.1007/978-3-030-30519-2_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-30518-5
Online ISBN: 978-3-030-30519-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)