Abstract
This paper describes the technology behind a performance where human dancers interact with an “artificial” performer projected on a screen. The system learns movement patterns from the human dancers in real time. It can also generate novel movement sequences that go beyond what it has been taught, thereby serving as a source of inspiration for the human dancers, challenging their habits and normal boundaries and enabling a mutual exchange of movement ideas. It is central to the performance concept that the system’s learning process is perceivable for the audience. To this end, an autoencoder neural network is trained in real time with motion data captured live on stage. As training proceeds, a “pose map” emerges that the system explores in a kind of improvisational state. The paper shows how this method is applied in the performance, and shares observations and lessons made in the process.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Schiphorst, T.: A case study of merce cunningham’s use of the lifeforms computer choreographic system in the making of trackers. M.A. thesis, Simon Fraser University (1993)
deLahunta, S.: The choreographic language agent. In: Stock, C. (ed.) Conference Proceedings of the 2008 World Dance Alliance Global Summit (2009)
Rothwell, N.: Programming languages, software thinking and creative process. In: Maragiannis, A. (ed.) Book of Abstracts DRHA 2014 (2014)
Leach, J., deLahunta, S.: Dance becoming knowledge: designing a digital “body”. Leonardo 50(5), 461–467 (2017)
Carlson, K., Schiphorst, T., Pasquier, P.: Scuddle: generating movement catalysts for computer-aided choreography. In: Proceedings of the Second International Conference on Computational Creativity, pp. 123–128 (2011)
Augello, A., Cipolla, E., Infantino, I., Manfre, A., Pilato, G., Vella, F.: Creative robot dance with variational encoder. In: Goel, A., Jordanous, A., Pease, A. (eds.) Proceedings of the Eighth International Conference on Computational Creativity (2017)
Crnkovic-Friis, L., Crnkovic-Friis, L.: Generative choreography using deep learning. In: Proceedings of the Seventh International Conference on Computational Creativity (2016)
Bret, M., Tramus, M.H., Berthoz, A.: Interacting with an intelligent dancing figure: artistic experiments at the crossroads between art and cognitive science. Leonardo 38(1), 47–53 (2005). http://www.jstor.org/stable/1577645
Arnal Romero, G.: Dancing with deep learning. B.Sc. thesis, Universitat Politècnica de Catalunya (2017)
McCormick, J., Vincs, K., Nahavandi, S., Creighton, D.: Learning to dance with a human. In: Proceedings of the 19th International Symposium on Electronic Art (2013)
Berman, A., James, V.: Kinetic imaginations: exploring the possibilities of combining AI and dance. In: Proceedings of the 24th International Conference on Artificial Intelligence, IJCAI 2015, pp. 2431–2437. AAAI Press (2015). http://dl.acm.org/citation.cfm?id=2832581.2832588
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672–2680 (2014)
Jones, K., Bonafilia, D.: Gangogh: creating art with GANs. https://towardsdatascience.com/gangogh-creating-art-with-gans-8d087d8f74a1. Accessed 15 Jan 2018
Sangkloy, P., Lu, J., Fang, C., Yu, F., Hays, J.: Scribbler: controlling deep image synthesis with sketch and color. CoRR abs/1612.00835 (2016). http://arxiv.org/abs/1612.00835
Elwes, J.: Latent space. http://lumenprize.com/artwork/latent-space. Accessed 15 Jan 2018
Broad, T., Grierson, M.: Autoencoding blade runner: reconstructing films with artificial neural networks. Leonardo 50(4), 376–383 (2017)
Elgammal, A., Liu, B., Elhoseiny, M., Mazzone, M.: CAN: creative adversarial networks: generating “art” by learning about styles and deviating from style norms. In: Proceedings of the Seventh International Conference on Computational Creativity (2016)
Baldi, P., Hornik, K.: Neural networks and principal component analysis: learning from examples without local minima. Neural Netw. 2(1), 53–58 (1989). https://doi.org/10.1016/0893-6080(89)90014-2
Zeestraten, M.J.A., Havoutis, I., Silvério, J., Calinon, S., Caldwell, D.G.: An approach for imitation learning on riemannian manifolds. IEEE Robot. Autom. Lett. (RA-L) 2, 1240–1247 (2017)
Acknowledgments
The work presented in this paper was supported by Kulturbryggan/Swedish Arts Grants Committee, European Commission Culture Program, Life Long Burning, National Cultural Fund of Hungary, Trafó House of Contemporary Arts, CAFe Budapest Contemporary Arts Festival and 3:e Våningen.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Berman, A., James, V. (2018). Learning as Performance: Autoencoding and Generating Dance Movements in Real Time. In: Liapis, A., Romero Cardalda, J., Ekárt, A. (eds) Computational Intelligence in Music, Sound, Art and Design. EvoMUSART 2018. Lecture Notes in Computer Science(), vol 10783. Springer, Cham. https://doi.org/10.1007/978-3-319-77583-8_17
Download citation
DOI: https://doi.org/10.1007/978-3-319-77583-8_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-77582-1
Online ISBN: 978-3-319-77583-8
eBook Packages: Computer ScienceComputer Science (R0)