Abstract Sounds and Their Applications in Audio and Perception Research

  • Adrien Merer
  • Sølvi Ystad
  • Richard Kronland-Martinet
  • Mitsuko Aramaki
Part of the Lecture Notes in Computer Science book series (LNCS, volume 6684)


Recognition of sound sources and events is an important process in sound perception and has been studied in many research domains. Conversely sounds that cannot be recognized are not often studied except by electroacoustic music composers. Besides, considerations on recognition of sources might help to address the problem of stimulus selection and categorization of sounds in the context of perception research. This paper introduces what we call abstract sounds with the existing musical background and shows their relevance for different applications.


abstract sound stimuli selection acousmatic 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Association, A.P.: The Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-IV). American Psychiatric Association (1994), (last viewed February 2011)
  2. 2.
    Ballas, J.A.: Common factors in the identification of an assortment of brief everyday sounds. Journal of Experimental Psychology: Human Perception and Performance 19, 250–267 (1993)Google Scholar
  3. 3.
    Bentin, S., McCarthy, G., Wood, C.C.: Event-related potentials, lexical decision and semantic priming. Electroencephalogr Clin. Neurophysiol. 60, 343–355 (1985)CrossRefGoogle Scholar
  4. 4.
    Bergman, P., Skold, A., Vastfjall, D., Fransson, N.: Perceptual and emotional categorization of sound. The Journal of the Acoustical Society of America 126, 3156–3167 (2009)CrossRefGoogle Scholar
  5. 5.
    Bloit, J., Rasamimanana, N., Bevilacqua, F.: Towards morphological sound description using segmental models. In: DAFX, Milan, Italie (2009)Google Scholar
  6. 6.
    Bonebright, T.L., Miner, N.E., Goldsmith, T.E., Caudell, T.P.: Data collection and analysis techniques for evaluating the perceptual qualities of auditory stimuli. ACM Trans. Appl. Percept. 2, 505–516 (2005)CrossRefGoogle Scholar
  7. 7.
    Bonebright, T.L., Nees, M.A.: Most earcons do not interfere with spoken passage comprehension. Applied Cognitive Psychology 23, 431–445 (2009)CrossRefGoogle Scholar
  8. 8.
    Bregman, A.S.: Auditory Scene Analysis. The MIT Press, Cambridge (1990)Google Scholar
  9. 9.
    Chion, M.: Audio-vision, Sound on Screen. Columbia University Press, New-York (1993)Google Scholar
  10. 10.
    Cross, L.: Electronic music, 1948-1953. Perspectives of New Music (1968)Google Scholar
  11. 11.
    Dack, J.: Abstract and concrete. Journal of Electroacoustic Music 14 (2002)Google Scholar
  12. 12.
    Dack, J., North, C.: Translating pierre schaeffer: Symbolism, literature and music. In: Proceedings of EMS 2006 Conference, Beijing (2006)Google Scholar
  13. 13.
    Dessein, A., Lemaitre, G.: Free classification of vocal imitations of everyday sounds. In: Sound And Music Computing (SMC 2009), Porto, Portugal, pp. 213–218 (2009)Google Scholar
  14. 14.
    Dubois, D., Guastavino, C., Raimbault, M.: A cognitive approach to urban soundscapes: Using verbal data to access everyday life auditory categories. Acta Acustica United with Acustica 92, 865–874 (2006)Google Scholar
  15. 15.
    Eimert, H.: What is electronic music. Die Reihe 1 (1957)Google Scholar
  16. 16.
    Fastl, H.: Neutralizing the meaning of sound for sound quality evaluations. In: Proc. Int. Congress on Acoustics ICA 2001, Rome, Italy, vol. 4, CD-ROM (2001)Google Scholar
  17. 17.
    Gaver, W.W.: How do we hear in the world? explorations of ecological acoustics. Ecological Psychology 5, 285–313 (1993)CrossRefGoogle Scholar
  18. 18.
    Gaver, W.W.: What in the world do we hear? an ecological approach to auditory source perception. Ecological Psychology 5, 1–29 (1993)CrossRefGoogle Scholar
  19. 19.
    Hermann, T.: Taxonomy and definitions for sonification and auditory display. In: Proceedings of the 14th International Conference on Auditory Display, Paris, France (2008)Google Scholar
  20. 20.
    Hoffman, M., Cook, P.R.: Feature-based synthesis: Mapping acoustic and perceptual features onto synthesis parameters. In: Proceedings of the 2006 International Computer Music Conference (ICMC), New Orleans (2006)Google Scholar
  21. 21.
    Jekosch, U.: 8. Assigning Meaning to Sounds - Semiotics in the Context of Product-Sound Design. J. Blauert, 193–221 (2005)Google Scholar
  22. 22.
    McKay, C., McEnnis, D., Fujinaga, I.: A large publicly accessible prototype audio database for music research (2006)Google Scholar
  23. 23.
    Merer, A., Ystad, S., Kronland-Martinet, R., Aramaki, M.: Semiotics of sounds evoking motions: Categorization and acoustic features. In: Kronland-Martinet, R., Ystad, S., Jensen, K. (eds.) CMMR 2007. LNCS, vol. 4969, pp. 139–158. Springer, Heidelberg (2008)CrossRefGoogle Scholar
  24. 24.
    Micoulaud-Franchi, J.A., Cermolacce, M., Vion-Dury, J.: Bizzare and familiar recognition troubles of auditory perception in patient with schizophrenia (2010) (in preparation) Google Scholar
  25. 25.
    Moore, B.C.J., Tan, C.T.: Perceived naturalness of spectrally distorted speech and music. The Journal of the Acoustical Society of America 114, 408–419 (2003)CrossRefGoogle Scholar
  26. 26.
    Peeters, G., Deruty, E.: Automatic morphological description of sounds. In: Acoustics 2008, Paris, France (2008)Google Scholar
  27. 27.
    Ricard, J., Herrera, P.: Morphological sound description computational model and usability evaluation. In: AES 116th Convention (2004)Google Scholar
  28. 28.
    Risset, J.C., Wessel, D.L.: Exploration of timbre by analysis and synthesis. In: Deutsch, D. (ed.) The psychology of music. Series in Cognition and Perception, pp. 113–169. Academic Press, London (1999)Google Scholar
  29. 29.
    Schaeffer, P.: Traité des objets musicaux. Editions du seuil (1966)Google Scholar
  30. 30.
    Schaeffer, P., Reibel, G.: Solfège de l’objet sonore. INA-GRM (1967)Google Scholar
  31. 31.
    Schlauch, R.S.: 12 - Loudness. In: Ecological Psychoacoustics, pp. 318–341. Elsevier, Amsterdam (2004)Google Scholar
  32. 32.
    Schön, D., Ystad, S., Kronland-Martinet, R., Besson, M.: The evocative power of sounds: Conceptual priming between words and nonverbal sounds. Journal of Cognitive Neuroscience 22, 1026–1035 (2010)CrossRefGoogle Scholar
  33. 33.
    Shafiro, V., Gygi, B.: How to select stimuli for environmental sound research and where to find them. Behavior Research Methods, Instruments, & Computers 36, 590–598 (2004)CrossRefGoogle Scholar
  34. 34.
    Smalley, D.: Defining timbre — refining timbre. Contemporary Music Review 10, 35–48 (1994)CrossRefGoogle Scholar
  35. 35.
    Smalley, D.: Space-form and the acousmatic image. Org. Sound 12, 35–58 (2007)CrossRefGoogle Scholar
  36. 36.
    Tanaka, K., Matsubara, K., Sato, T.: Study of onomatopoeia expressing strange sounds: Cases of impulse sounds and beat sounds. Transactions of the Japan Society of Mechanical Engineers C 61, 4730–4735 (1995)CrossRefGoogle Scholar
  37. 37.
    Thoresen, L., Hedman, A.: Spectromorphological analysis of sound objects: an adaptation of pierre schaeffer’s typomorphology. Organised Sound 12, 129–141 (2007)CrossRefGoogle Scholar
  38. 38.
    Zeitler, A., Ellermeier, W., Fastl, H.: Significance of meaning in sound quality evaluation. Fortschritte der Akustik, CFA/DAGA 4, 781–782 (2004)Google Scholar
  39. 39.
    Zeitler, A., Hellbrueck, J., Ellermeier, W., Fastl, H., Thoma, G., Zeller, P.: Methodological approaches to investigate the effects of meaning, expectations and context in listening experiments. In: INTER-NOISE 2006, Honolulu, Hawaii (2006)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Adrien Merer
    • 1
  • Sølvi Ystad
    • 1
  • Richard Kronland-Martinet
    • 1
  • Mitsuko Aramaki
    • 2
    • 3
  1. 1.CNRS - Laboratoire de Mécanique et d’AcoustiqueMarseilleFrance
  2. 2.CNRS - Institut de Neurosciences Cognitives de la MéditerranéeMarseilleFrance
  3. 3.Université Aix-MarseilleMarseilleFrance

Personalised recommendations