Skip to main content

Flexibility of Emulation Learning from Pioneers in Nonstationary Environments

  • Conference paper
  • First Online:
Advances in Artificial Intelligence (JSAI 2019)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1128))

Included in the following conference series:

Abstract

This is an extension from a selected paper from JSAI2019. Social learning is crucial in acquisition of the intelligent behaviors of humans and many kinds of animals, as it makes behavior learning far more efficient than pure trial-and-error. In imitation learning, a representative form of social learning, the agent observes specific action-state pair sequences produced by another agent (expert) and reflect them into its own action. One of its implementations in reinforcement learning is the inverse reinforcement learning. We propose another form of social learning, emulation learning, which requires much less information from another agent (pioneer). In emulation learning, the agent is given only a certain level of achievement by another agent, or a record. In this study, we implement emulation learning in the reinforcement learning setting by applying a model of satisficing action policy. We show that the emulation learning algorithm works well both in stationary and non-stationary reinforcement learning tasks, breaking the often observed trade-off like relationship between efficiency and flexibility.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Whiten, A., McGuigan, N., Marshall-Pescini, S., Hopper, L.M.: Emulation, imitation, over-imitation and the scope of culture for child and chimpanzee. Philos. Trans. R. Soc. B: Biol. Sci. 364(1528), 2417–2428 (2009). https://doi.org/10.1098/rstb.2009.0069

    Article  Google Scholar 

  2. Takahashi, T., Kohno, Y., Uragami, D.: Cognitive satisficing: bounded rationality in reinforcement learning. Trans. Jpn. Soc. Artif. Intell. 31(6), AI30-M\_1–11 (2016). (in Japanese)

    Article  Google Scholar 

  3. Tamatsukuri, A., Takahashi, T.: Guaranteed satisficing and finite regret: analysis of a cognitive satisficing value function. BioSystems 180, 46–53 (2019)

    Article  Google Scholar 

  4. Andrew Maas, J., Bagnell, A., Dey, A.K., Ziebart, B.D.: Maximum entropy inverse reinforcement learning. In: AAAI 2008 (2008)

    Google Scholar 

  5. Levy, K.Y., Shimkin, N.: Unified inter and intra options learning using policy gradient methods. In: EWRL, pp. 153–164 (2011)

    Google Scholar 

  6. Simon, H.A.: Rational choice and the structure of the environment. Psychol. Rev. 63(2), 129–138 (1956)

    Article  Google Scholar 

  7. Ushida, U., Kono, Y., Takahashi, T.: Satisficing reinforcement learning for survival. In: Proceedings of JSAI 2017, 4C2-2in2 (2017). (in Japanese)

    Google Scholar 

  8. Kono, Y., Takahashi, T.: Autonomous optimal exploration through satisficing. In: Proceedings of JSAI 2018, 1Z3-04 (2018). (in Japanese)

    Google Scholar 

Download references

Acknowledgments

This work was supported by JSPS KAKENHI Grant Number 17H04696.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tatsuji Takahashi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Shinriki, M., Wakabayashi, H., Kono, Y., Takahashi, T. (2020). Flexibility of Emulation Learning from Pioneers in Nonstationary Environments. In: Ohsawa, Y., et al. Advances in Artificial Intelligence. JSAI 2019. Advances in Intelligent Systems and Computing, vol 1128. Springer, Cham. https://doi.org/10.1007/978-3-030-39878-1_9

Download citation

Publish with us

Policies and ethics