Skip to main content

Gradient-Based Learning of Compositional Dynamics with Modular RNNs

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 11727))

Abstract

Learning compositional dynamics with recurrent neural networks (RNNs) trained with back-propagation through time (BPTT) is usually a difficult task. Typically RNNs learn the consecutive shape along target sequences from time step to time step, focusing on local temporal correlations. When the challenge is to identify and model independent, unknown data subcomponents, that is, data generating causes on-the-fly during training, however, this local temporal shape-oriented inductive learning bias is obstructive. We propose a modular, compositional RNN architecture and derive simple procedures to automatically infer the source subdynamics that generate the data. We show that the involved error signal separation can be used for both teacher forcing and model-distinct target signal provision in the compositional RNN architecture. As a result, the entire network is able to learn compositional dynamics, developing emergent, flexibly adaptable signal decompositions within the distributed modules. We demonstrate that in this way simple RNNs trained with BPTT can learn sequences that could so far only be solved effectively with reservoir computing approaches. Moreover we show that these RNNs are much more robust against signal noise when compared to traditional BPTT or reservoir computing approaches.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Butz, M.V., Kutter, E.F.: How the Mind Comes Into Being: Introducing Cognitive Science from a Functional and Computational Perspective. Oxford University Press, Oxford (2017)

    Book  Google Scholar 

  2. Gao, Z., et al.: A cortico-cerebellar loop for motor planning. Nature 563(7729), 113 (2018)

    Article  Google Scholar 

  3. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016

    Google Scholar 

  4. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997). https://doi.org/10.1162/neco.1997.9.8.1735

    Article  Google Scholar 

  5. Jaeger, H.: The “echo state” approach to analysing and training recurrent neural networks. Technical report, GMD Report, 148, Fraunhofer Institute for Analysis and Information Systems AIS, Sankt Augustin, Germany (2001)

    Google Scholar 

  6. Kandel, E.R., Schwartz, J.H., Jessell, T.M., Jessell, M.B.T., Siegelbaum, S., Hudspeth, A.: Principles of Neural Science. McGraw-hill, New York (2000)

    Google Scholar 

  7. Kingma, D.P., Ba, J.L.: Adam: a method for stochastic optimization. 3rd International Conference for Learning Representations abs/1412.6980 (2015)

    Google Scholar 

  8. Koryakin, D., Lohmann, J., Butz, M.V.: Balanced echo state networks. Neural Netw. 36, 35–45 (2012)

    Article  Google Scholar 

  9. Otte, S., Krechel, D., Liwicki, M.: JANNLab neural network framework for Java. In: Poster Proceedings of MLDM 2013, pp. 39–46. ibai-publishing, New York (2013)

    Google Scholar 

  10. Otte, S., Butz, M.V., Koryakin, D., Becker, F., Liwicki, M., Zell, A.: Optimizing recurrent reservoirs with neuro-evolution. Neurocomputing 192, 128–138 (2016)

    Article  Google Scholar 

  11. Pathak, J., Hunt, B., Girvan, M., Lu, Z., Ott, E.: Model-free prediction of large spatiotemporally chaotic systems from data: a reservoir computing approach. Phys. Rev. Lett. 120, 024102 (2018)

    Article  Google Scholar 

  12. Schmidhuber, J., Wierstra, D., Gagliolo, M., Gomez, F.: Training recurrent neural networks by evolino. Neural Comput. 19, 757–779 (2007)

    Article  Google Scholar 

  13. Svoboda, K., Li, N.: Neural mechanisms of movement planning: motor cortex and beyond. Curr. Opin. Neurobiol. 49, 33–41 (2018)

    Article  Google Scholar 

  14. Werbos, P.: Backpropagation through time: what it does and how to do it. Proc. IEEE 78(10), 1550–1560 (1990). https://doi.org/10.1109/5.58337

    Article  Google Scholar 

  15. Wu, Y., et al.: Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation. ArXiv e-prints 1609.08144 (2016)

Download references

Acknowledgments

The authors would like to thank Sander Bothé, CWI Amsterdam, for helpful comments and suggestions regarding this work.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sebastian Otte .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Otte, S., Rubisch, P., Butz, M.V. (2019). Gradient-Based Learning of Compositional Dynamics with Modular RNNs. In: Tetko, I., Kůrková, V., Karpov, P., Theis, F. (eds) Artificial Neural Networks and Machine Learning – ICANN 2019: Theoretical Neural Computation. ICANN 2019. Lecture Notes in Computer Science(), vol 11727. Springer, Cham. https://doi.org/10.1007/978-3-030-30487-4_38

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-30487-4_38

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-30486-7

  • Online ISBN: 978-3-030-30487-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics