Advertisement

Evaluating Input Devices for Dance Research

  • Mari Romarheim Haugen
  • Kristian Nymoen
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9617)

Abstract

Recording music-related motions in ecologically valid situations can be challenging. We investigate the performance of three devices providing 3D acceleration data, namely Axivity AX3, iPhone 4s and a Wii controller tracking rhythmic motions. The devices are benchmarked against an infrared motion capture system, tested on both simple and complex music-related body motions, and evaluations are presented of the data quality and suitability for tracking music-related motions in real-world situations. The various systems represent different trade-offs with respect to data quality, user interface and physical attributes.

Keywords

Music and motion Dance Samba Motion capture Motion analysis Qualisys AX3 iPhone Wii 

References

  1. 1.
    Amelynck, D., Grachten, M., Van Noorden, L., Leman, M.: Toward e-motion-based music retrieval a study of affective gesture recognition. IEEE Trans. Affect. Comput. 3(2), 250–259 (2012)CrossRefGoogle Scholar
  2. 2.
    Burger, B., Toiviainen, P.: MoCap Toolbox – a matlab toolbox for computational analysis of movement data. In: Bresin, R. (ed.) Proceedings of 10th Sound and Music Computing Conference (SMC), pp. 172–178 (2013)Google Scholar
  3. 3.
    De Bruyn, L., Leman, M., Moelants, D., Demey, M.: Does social interaction activate music listeners? In: Ystad, S., Kronland-Martinet, R., Jensen, K. (eds.) CMMR 2008. LNCS, vol. 5493, pp. 93–106. Springer, Heidelberg (2009)Google Scholar
  4. 4.
    Gerischer, C.: O suingue baiano: rhythmic feeling and microrhythmic phenomena in Brazilian percussion. Ethnomusicology 50(1), 99–119 (2006)Google Scholar
  5. 5.
    Godøy, R.I.: Gestural-sonorous objects: embodied extensions of Schaeffer’s conceptual apparatus. Org. Sound 11(02), 149–157 (2006)CrossRefGoogle Scholar
  6. 6.
    Gouyon, F.: Microtiming in Samba de Roda – preliminary experiments with polyphonic audio. In: SBCM 2007 Proceedings, pp. 197–203 (2007)Google Scholar
  7. 7.
    Khan, A.M.: Recognizing physical activities using the axivity device. In: The 5th International Conference on eHealth, Telemedicine, and Social Medicine (2013)Google Scholar
  8. 8.
    Lartillot, O., Toiviainen, P.: A matlab toolbox for musical feature extraction from audio. In: International Conference on Digital Audio Effects (2007)Google Scholar
  9. 9.
    Naveda, L., Leman, M.: Sonification of Samba dance using periodic pattern analysis. In: 4th International Conference on Digital Arts, ARTECH 2008, pp. 16–26 (2008)Google Scholar
  10. 10.
    Naveda, L., Leman, M.: A cross-modal heuristic for periodic pattern analysis of Samba music and dance. J. New Music Res. 38(3), 255–283 (2009)CrossRefGoogle Scholar
  11. 11.
    Nymoen, K., Volsund, A., Skogstad, S.A., Jensenius, A.R., Torresen, J.: Comparing motion data from an iPod touch to an optical infrared marker-based motion capture system. In: Proceedings of International Conference on New Interfaces for Musical Expression, pp. 88–91 (2012)Google Scholar
  12. 12.
    Schubert, E.: Correlation analysis of continuous emotional response to music: correcting for the effects of serial correlation. Musicae Scientiae 5(1 Suppl.), 213–236 (2002)MathSciNetGoogle Scholar
  13. 13.
    Van Dyck, E., Moelants, D., Demey, M., Deweppe, A., Coussement, P., Leman, M.: The impact of the bass drum on human dance movement. Music Percept.: Interdiscip. J. 30(4), 349–359 (2013)Google Scholar
  14. 14.
    Wang, G.: Designing Smule’s Ocarina: the iPhone’s magic flute. In: Proceedings of International Conference on New Interfaces for Musical Expression, Pittsburgh, PA, pp. 303–307 (2009)Google Scholar
  15. 15.
    Witek, M.: ‘\(\ldots \) and I feel good!’: the relationship between body-movement, pleasure and groove in music. Ph.D. thesis, University of Oxford (2013)Google Scholar

Copyright information

© Springer International Publishing Switzerland 2016

Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 2.5 International License (http://creativecommons.org/licenses/by-nc/2.5/), which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.

The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

Authors and Affiliations

  1. 1.Department of MusicologyUniversity of OsloOsloNorway

Personalised recommendations