Abstract
Both self-learning architecture (embedded structure) and explicit/implicit teaching from other agents (environmental design issue) are necessary not only for one behavior learning but more seriously for life-time behavior learning. This paper presents a method for a robot to understand unfamiliar behavior shown by others through the collaboration between behavior acquisition and recognition of observed behavior, where the state value has an important role not simply for behavior acquisition (reinforcement learning) but also for behavior recognition (observation). That is, the state value updates can be accelerated by observation without real trials and errors while the learned values enrich the recognition system since it is based on estimation of the state value of the observed behavior. The validity of the proposed method is shown by applying it to a dynamic environment where two robots play soccer.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Whitehead, S.D.: Complexity and cooperation in q-learning. In: Proceedings Eighth International Workshop on Machine Learning (ML 1991), pp. 363–367 (1991)
Price, B., Boutilier, C.: Accelerating reinforcement learning through implicit imitatione. Journal of Articial Intelligence Research (2003)
Bentivegna, D.C., Atkeson, C.G., Chenga, G.: Learning tasks from observation and practice. Robotics and Autonomous Systems 47, 163–169 (2004)
Takahashi, Y., Kawamata, T., Asada, M., Negrello, M.: Emulation and behavior understanding through shared values. In: Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, October 2007, pp. 3950–3955 (2007)
Sutton, R., Barto, A.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)
Connell, J.H., Mahadevan, S.: ROBOT LEARNING. Kluwer Academic Publishers, Dordrecht (1993)
Takahashi, Y., Kawamata, T., Asada, M.: Learning utility for behavior acquisition and intention inference of other agent. In: Proceedings of the 2006 IEEE/RSJ IROS 2006 Workshop on Multi-objective Robotics, October 2006, pp. 25–31 (2006)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Takahashi, Y., Tamura, Y., Asada, M. (2008). Mutual Development of Behavior Acquisition and Recognition Based on Value System. In: Asada, M., Hallam, J.C.T., Meyer, JA., Tani, J. (eds) From Animals to Animats 10. SAB 2008. Lecture Notes in Computer Science(), vol 5040. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-69134-1_29
Download citation
DOI: https://doi.org/10.1007/978-3-540-69134-1_29
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-69133-4
Online ISBN: 978-3-540-69134-1
eBook Packages: Computer ScienceComputer Science (R0)