Abstract
On meeting another individual for the first time during the course of a conversation we learn a lot unique and idiosyncratic facts about that person. If that meeting is socially or professionally significant, in the future, on crossing paths with that individual again, we can easily remember their details and the circumstances of the meeting. The person’s age, gender, ethnic or racial background, place of birth, current place of residence, professional and personal interests are all important pieces of data for building up a mental picture of that individual which we notice without going to too much effort. We might also notice the tone of their voice, the prosody with which they speak, and a foreign accent, if present. Interestingly, all of those details can be readily recalled on either seeing their face or just hearing their voice. Many of the chapters in this book are devoted to the importance of the face and voice in the formed percept we have of another individual—assigning an individual their own unique identity. Other chapters focus on how animals decode these important conspecific details. The questions asked in this chapter pertain less to brain mechanisms active in identifying specific individuals and their characteristics, but focus on issues relating to how non-verbal face and voice cues are integrated by the human brain. Early behavioral studies have noted how important non-verbal behaviors are for the interpretation of the actions of others, in terms of presenting important information relating to the social interaction (Campbell & Rushton, 1978; Mehrabian & Ferris, 1967). Yet, this remains a poorly studied area in social neuroscience and is a major focus for our laboratory.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Beauchamp, M. S., Nath, A. R., & Pasalar, S. (2010). fMRI-Guided transcranial magnetic stimulation reveals that the superior temporal sulcus is a cortical locus of the McGurk effect. Journal of Neuroscience, 30(7), 2414–2417.
Bentin, S., Allison, T., Puce, A., Perez, A., & McCarthy, G. (1996). Electrophysiological studies of face perception in humans. Journal of Cognitive Neuroscience, 8, 551–565.
Brefczynski-Lewis, J., Lowitszch, S., Parsons, M., Lemieux, S., & Puce, A. (2009). Audiovisual non-verbal dynamic faces elicit converging fMRI and ERP responses. Brain Topography, 21(3–4), 193–206.
Busch, N. A., Dubois, J., & VanRullen, R. (2009). The phase of ongoing EEG oscillations predicts visual perception. Journal of Neuroscience, 29(24), 7869–7876.
Calvert, G. A. (2001). Crossmodal processing in the human brain: Insights from functional neuroimaging studies. Cerebral Cortex, 11(12), 1110–1123.
Campbell, A., & Rushton, J. P. (1978). Bodily communication and personality. The British Journal of Social and Clinical Psychology, 17(1), 31–36.
Carrick, O. K., Thompson, J. C., Epling, J. A., & Puce, A. (2007). It’s all in the eyes: Neural responses to socially significant gaze shifts. Neuroreport, 18(8), 763–766.
Eggermont, J. J., & Ponton, C. W. (2002). The neurophysiology of auditory perception: From single units to evoked potentials. Audiology & Neuro-Otology, 7(2), 71–99.
Engel, A. K., & Fries, P. (2010). Beta-band oscillations–signalling the status quo? Current Opinion in Neurobiology, 20(2), 156–165.
Engel, A. K., Fries, P., & Singer, W. (2001). Dynamic predictions: Oscillations and synchrony in top-down processing. Nature Reviews. Neuroscience, 2(10), 704–716.
Foxe, J. J., Morocz, I. A., Murray, M. M., Higgins, B. A., Javitt, D. C., & Schroeder, C. E. (2000). Multisensory auditory-somatosensory interactions in early cortical processing revealed by high-density electrical mapping. Brain Research. Cognitive Brain Research, 10(1–2), 77–83.
Ghazanfar, A. A., & Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends in Cognitive Sciences, 10(6), 278–285.
Giard, M. H., & Peronnet, F. (1999). Auditory-visual integration during multimodal object recognition in humans: A behavioral and electrophysiological study. Journal of Cognitive Neuroscience, 11(5), 473–490.
Hagan, C. C., Woods, W., Johnson, S., Calder, A. J., Green, G. G., & Young, A. W. (2009). MEG demonstrates a supra-additive response to facial and vocal emotion in the right superior temporal sulcus. Proceedings of the National Academy of Sciences of the United States of America, 106(47), 20010–20015.
Herrmann, C. S., & Debener, S. (2008). Simultaneous recording of EEG and BOLD responses: A historical perspective. International Journal of Psychophysiology, 67(3), 161–168.
Herrmann, C. S., Frund, I., & Lenz, D. (2010). Human gamma-band activity: A review on cognitive and behavioral correlates and network models. Neuroscience and Biobehavioral Reviews, 34(7), 981–992.
Herrmann, C. S., Munk, M. H., & Engel, A. K. (2004). Cognitive functions of gamma-band activity: Memory match and utilization. Trends in Cognitive Sciences, 8(8), 347–355.
Itier, R. J., & Batty, M. (2009). Neural bases of eye and gaze processing: The core of social cognition. Neuroscience and Biobehavioral Reviews, 33(6), 843–863.
Kayser, C., Petkov, C. I., Augath, M., & Logothetis, N. K. (2007). Functional imaging reveals visual modulation of specific fields in auditory cortex. Journal of Neuroscience, 27(8), 1824–1835.
Kim, S., & James, T. W. (2010). Enhanced effectiveness in visuo-haptic object-selective brain regions with increasing stimulus salience. Human Brain Mapping, 31(5), 678–693.
Kutas, M., & Hillyard, S. A. (1980). Reading senseless sentences: Brain potentials reflect semantic incongruity. Science, 207(4427), 203–205.
Lewis, J. W., Talkington, W. J., Puce, A., Engel, L. R., & Frum, C. (2011). Cortical networks representing object categories and high-level attributes of familiar real-world action sounds. Journal of Cognitive Neuroscience, 23(8), 2079–2101.
Mazaheri, A., & Jensen, O. (2010). Rhythmic pulsing: Linking ongoing brain activity with evoked responses. Frontiers in Human Neuroscience, 4, 177.
McCallum, W. C., Farmer, S. F., & Pocock, P. V. (1984). The effects of physical and semantic incongruities on auditory event-related potentials. Electroencephalography and Clinical Neurophysiology, 59(6), 477–488.
McGurk, H., & MacDonald, J. (1976). Hearing lips and seeing voices. Nature, 264(5588), 746–748.
Mehrabian, A., & Ferris, S. R. (1967). Inference of attitudes from nonverbal communication in two channels. Journal of Consulting Psychology, 31(3), 248–252.
Meredith, M. A., & Stein, B. E. (1986). Visual, auditory, and somatosensory convergence on cells in superior colliculus results in multisensory integration. Journal of Neurophysiology, 56(3), 640–662.
Munhall, K. G., Gribble, P., Sacco, L., & Ward, M. (1996). Temporal constraints on the McGurk effect. Perception & Psychophysics, 58(3), 351–362.
Nath, A. R., & Beauchamp, M. S. (2011a). Dynamic changes in superior temporal sulcus connectivity during perception of noisy audiovisual speech. Journal of Neuroscience, 31(5), 1704–1714.
Nath, A. R., & Beauchamp, M. S. (2011b). A neural basis for interindividual differences in the McGurk effect, a multisensory speech illusion. NeuroImage, 59(1), 781–787.
Naue, N., Rach, S., Struber, D., Huster, R. J., Zaehle, T., Korner, U., et al. (2011). Auditory event-related response in visual cortex modulates subsequent visual responses in humans. Journal of Neuroscience, 31(21), 7729–7736.
Puce, A., Epling, J. A., Thompson, J. C., & Carrick, O. K. (2007). Neural responses elicited to face motion and vocalization pairings. Neuropsychologia, 45(1), 93–106.
Puce, A., & Perrett, D. (2003). Electrophysiology and brain imaging of biological motion. Philosophical Transactions of the Royal Society of London. Series B, Biological Sciences, 358(1431), 435–445.
Puce, A., Smith, A., & Allison, T. (2000). ERPs evoked by viewing facial movements. Cognitive neuropsychology, 17, 221–239.
Reale, R. A., Calvert, G. A., Thesen, T., Jenison, R. L., Kawasaki, H., Oya, H., et al. (2007). Auditory-visual processing represented in the human superior temporal gyrus. Neuroscience, 145(1), 162–184.
Rossion, B., & Jacques, C. (2008). Does physical interstimulus variance account for early electrophysiological face sensitive responses in the human brain? Ten lessons on the N170. NeuroImage, 39(4), 1959–1979.
Sabbagh, M. A., Moulson, M. C., & Harkness, K. L. (2004). Neural correlates of mental state decoding in human adults: An event-related potential study. Journal of Cognitive Neuroscience, 16(3), 415–426.
Sams, M., Aulanko, R., Hamalainen, M., Hari, R., Lounasmaa, O. V., Lu, S. T., et al. (1991). Seeing speech: Visual information from lip movements modifies activity in the human auditory cortex. Neuroscience Letters, 127(1), 141–145.
Stein, B. E., Stanford, T. R., Ramachandran, R., Perrault, T. J., Jr., & Rowland, B. A. (2009). Challenges in quantifying multisensory integration: Alternative criteria, models, and inverse effectiveness. Experimental Brain Research, 198(2–3), 113–126.
Stevenson, R. A., Geoghegan, M. L., & James, T. W. (2007). Superadditive BOLD activation in superior temporal sulcus with threshold non-speech objects. Experimental Brain Research, 179(1), 85–95.
Stevenson, R. A., & James, T. W. (2009). Audiovisual integration in human superior temporal sulcus: Inverse effectiveness and the neural processing of speech and object recognition. NeuroImage, 44(3), 1210–1223.
Stevenson, R. A., Kim, S., & James, T. W. (2009). An additive-factors design to disambiguate neuronal and areal convergence: Measuring multisensory interactions between audio, visual, and haptic sensory streams using fMRI. Experimental Brain Research, 198(2–3), 183–194.
Wheaton, K. J., Pipingas, A., Silberstein, R. B., & Puce, A. (2001). Human neural responses elicited to observing the actions of others. Visual Neuroscience, 18(3), 401–406.
Wright, T. M., Pelphrey, K. A., Allison, T., McKeown, M. J., & McCarthy, G. (2003). Polysensory interactions along lateral temporal regions evoked by audiovisual speech. Cerebral Cortex, 13(10), 1034–1043.
Young, C. K., & Eggermont, J. J. (2009). Coupling of mesoscopic brain oscillations: Recent advances in analytical and theoretical perspectives. Progress in Neurobiology, 89(1), 61–78.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media New York
About this chapter
Cite this chapter
Puce, A. (2013). Neurophysiological Correlates of Face and Voice Integration. In: Belin, P., Campanella, S., Ethofer, T. (eds) Integrating Face and Voice in Person Perception. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-3585-3_9
Download citation
DOI: https://doi.org/10.1007/978-1-4614-3585-3_9
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-3584-6
Online ISBN: 978-1-4614-3585-3
eBook Packages: Biomedical and Life SciencesBiomedical and Life Sciences (R0)