Abstract
A gradient based method for both symmetric and asymmetric multiagent reinforcement learning is introduced in this paper. Symmetric multiagent reinforcement learning addresses the problem with agents involved in the learning task having equal information states. Respectively, in asymmetric multiagent reinforcement learning, the information states are not equal, i.e. some agents (leaders) try to encourage agents with less information (followers) to select actions that lead to improved overall utility value for the leaders. In both cases, there is a huge number of parameters to learn and we thus need to use some parametric function approximation methods to represent the value functions of the agents. The method proposed in this paper is based on the VAPS framework that is extended to utilize the theory of Markov games, i.e. a natural basis of multiagent reinforcement learning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Baird, L., Moore, A.: Gradient descent for general reinforcement learning. In: Kearns, M., Solla, S., Cohn, D. (eds.) Advances in Neural Information Processing Systems, Cambridge, MA, USA, vol. 11. MIT Press, Cambridge (1999)
Cottle, R.W., Stone, R.E., Pang, J.-S.: The Linear Complementarity Problem. Academic Press, New York (1992)
Filar, J.A., Vrieze, K.: Competitive Markov Decision Processes. Springer, New York (1997)
Hu, J., Wellman, M.P.: Multiagent reinforcement learning: Theoretical framework and an algorithm. In: Proceedings of the Fifteenth International Conference on Machine Learning (ICML 1998), Madison, Wisconsin, USA. Morgan Kaufmann Publishers, San Francisco (July 1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Könönen, V. (2003). Gradient Based Method for Symmetric and Asymmetric Multiagent Reinforcement Learning. In: Liu, J., Cheung, Ym., Yin, H. (eds) Intelligent Data Engineering and Automated Learning. IDEAL 2003. Lecture Notes in Computer Science, vol 2690. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-45080-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-540-45080-1_9
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40550-4
Online ISBN: 978-3-540-45080-1
eBook Packages: Springer Book Archive