See me, hear me, touch me: multisensory integration in lateral occipital-temporal cortex

https://doi.org/10.1016/j.conb.2005.03.011Get rights and content

Our understanding of multisensory integration has advanced because of recent functional neuroimaging studies of three areas in human lateral occipito-temporal cortex: superior temporal sulcus, area LO and area MT (V5). Superior temporal sulcus is activated strongly in response to meaningful auditory and visual stimuli, but responses to tactile stimuli have not been well studied. Area LO shows strong activation in response to both visual and tactile shape information, but not to auditory representations of objects. Area MT, an important region for processing visual motion, also shows weak activation in response to tactile motion, and a signal that drops below resting baseline in response to auditory motion. Within superior temporal sulcus, a patchy organization of regions is activated in response to auditory, visual and multisensory stimuli. This organization appears similar to that observed in polysensory areas in macaque superior temporal sulcus, suggesting that it is an anatomical substrate for multisensory integration. A patchy organization might also be a neural mechanism for integrating disparate representations within individual sensory modalities, such as representations of visual form and visual motion.

Introduction

In everyday life, perceptual events often occur in multiple sensory modalities at once: we hear someone speaking as we see their mouth move. Most scientific investigations have focused on single modalities (frequently vision) in isolation. Recently, there has been increasing interest in studying integration across sensory modalities. In this review, I discuss progress in studying the brain mechanisms of multisensory integration in human lateral occipital-temporal cortex, especially functional magnetic resonance imaging (fMRI) studies of superior temporal sulcus (STS), area LO and area MT (see glossary for a brief definition of these terms). Links between human neuroimaging studies and studies in non-human primates are made using techniques from computational neuroanatomy that permit alignment of human and monkey brains.

An ongoing discussion concerns the appropriate methods for studying multisensory integration using fMRI [1•, 2•, 3]. One important method is to contrast unisensory stimulation conditions with multisensory conditions. The hallmark of multisensory integration is that unisensory stimuli presented in combination produce an effect different from the linear combination of the unisensory stimuli presented separately. In individual neurons, these differences can be quite dramatic, with multisensory responses that are much greater than the sum of individual unisensory responses (‘super-additivity’). However, because fMRI measurements integrate across thousands or millions of neurons, the super-additivity measure might not be appropriate [2]. Instead, increasingly liberal criteria might be more suitable, such as requiring only that multisensory responses are greater than the maximum or mean of the individual unisensory responses [1].

Another important issue is the high degree of inter-subject and -laboratory variability observed in fMRI studies. STS, LO and MT are attractive targets for a review because there is some consensus on their anatomical location. This is either because they constitute an anatomical structure observed in every normal human hemisphere (such as STS) or because their response properties make it possible to identify them with functional localizers (somewhat ambiguously for LO, unambiguously for MT). By starting out with well-defined regions, a review can sidestep some of the difficulties inherent in deciding if a stereotaxic coordinate reported in one study of multisensory integration corresponds to the same cortical region as a coordinate from a different study.

Although STS, LO and MT are found in relative proximity, within the space of a few centimeters in human lateral occipital temporal cortex, their multisensory response properties are quite different, as is our level of knowledge about their role in multisensory perception. Therefore, this review attempts to compare and contrast the activity in these three areas in response to stimuli in three sensory modalities — visual, auditory and tactile. Figure 1 illustrates the location of STS, LO and MT in folded and inflated versions of a human brain, and their relationship to Brodmann's cytoarchitectonic classification scheme.

Glossary

  • Area MT (V5): A region in extrastriate visual cortex distinguished by its heavy myelination and specialization for processing visual motion. It was first described in the posterior middle temporal cortex of owl monkey [53], leading to the designation MT. In macaque monkeys, this region lies in the posterior bank of the superior temporal sulcus, where some investigators have designated it V5 [54]. A homologous region has been found in many other species, including humans, where it lies near the junction of the inferior temporal sulcus and the lateral occipital sulcus [55].

  • Congruent and incongruent stimuli: Because different sensory modalities can be stimulated independently in an experimental setting, multisensory stimuli can be congruent (such as a picture of a car presented with the sound of a car) or incongruent (such as a picture of a car presented with the sound of a telephone).

  • fMRI (functional magnetic resonance imaging): A non-invasive method for measuring neuronal activity, typically with an indirect measure such as blood-oxygenation level dependent (BOLD) contrast.

  • Localizer: There is only a rough correlation between visible anatomical structures (such as specific sulci or gyri) and the functional areas that comprise the computational organization of the brain. However, in order to make inferences about organization, it is important to compare the same functional area across subjects. A common technique is to use a localizer fMRI scan (for instance, alternating moving and static stimuli) in order to identify a specific region of interest (for instance, area MT). Additional experiments are then performed and the results compared across subjects within this region.

  • Multisensory: Refers to the processing of stimuli presented in multiple sensory modalities at once. Although the term ‘multimodal’ is sometimes used as a synonym for multisensory, it is also used to describe studies that use multiple measurement techniques, such as fMRI and magnetoencephalography (MEG). Therefore, the term multisensory is preferred.

  • Synchronous and asynchronous stimuli: An experimental manipulation that involves artificially changing the temporal offset between stimuli presented in different sensory modalities in order to measure the effect on multisensory integration. For instance, the discomforting sensation when the dialogue in the sound track of a movie is offset from the images.

Section snippets

Multisensory integration in superior temporal sulcus

There is compelling evidence for auditory and visual responses in human STS to a variety of stimuli. (For a review of all regions important for multisensory identification and object recognition, please see Amedi et al. [4]). Because it extends over a large area of cortex, STS certainly contains several functional regions. However, the parcellation of human STS is poorly understood, and in this review STS is used as shorthand for ‘the constellation of cortical areas with multisensory response

Multisensory responses in human and monkey superior temporal sulcus

In macaque, an important multisensory region lies along the fundus of the STS. This region was functionally defined as the superior temporal polysensory (STP) area on the basis of single cell recordings [11] and probably corresponds to the region in macaques that was anatomically defined as temporal–parietal–occipital (TPO) [12]. Although the visual responses of many areas in macaque STS have been characterized, recent neuroimaging studies in macaque demonstrate that complex, behaviorally

Area LO

Area LO was first described as a region of human lateral occipital cortex, just ventral and posterior to area MT, that responded preferentially to images of objects versus those of textured patterns [22]. LO is thought to be important for processing visual shape information [23]. More recently, studies showing that an extended band of visual cortex responds preferentially to images versus patterns [24•, 25] has led to confusion over the location and identity of LO. Figure 2 illustrates the

Area MT

Area MT is recognized as a key locus for visual motion processing in the primate brain (see glossary). In macaque monkeys, MT is located in the lower bank of the STS (Figure 2b), whereas in humans, MT is located in lateral occipital cortex (Figure 2c). This review refers to ‘MT’ as a single area for simplicity, although this region of cortex contains several motion-responsive areas that are grouped together in most imaging studies, often under the rubric MT+ [35].

There are strong

Commonalities between integration across and within modalities

As discussed above, one of the neural substrates for multisensory integration in STS might be a patchy organization, in which neighboring patches respond primarily to unisensory auditory or visual information. Unisensory information might be translated into a common code and integrated in multisensory regions that lie between the unisensory patches. Such an organization might also be amenable to integration of other types of information. Neurons in STS can be selective to both visual form and

The object property model

It is also useful to consider the relationship between multisensory and category-related responses. One of the most surprising findings to arise from recent functional neuroimaging studies is that specific regions of human visual cortex respond preferentially to specific categories of objects. For instance, parts of lateral temporal cortex (middle temporal gyrus and inferior temporal sulcus, including portions of areas MT and LO) respond preferentially to images of man-made graspable objects

Conclusions and future directions

Neuroimaging studies in humans and non-human primates using the same multisensory stimuli will be crucial for forming a link between human neurobiology and the anatomical and physiological insight that can only be obtained from invasive studies. The results of these experiments, combined with advances in neuroimaging methods applicable in humans, such as high-resolution fMRI and MEG, mean that the next few years will surely see further great strides in our understanding of multisensory

References and recommended reading

Papers of particular interest, published within the annual period of review, have been highlighted as:

  • • of special interest

  • •• of outstanding interest

Acknowledgements

D Van Essen, D Hanlon, J Dickson, P Christidis and Z Saad were instrumental in figure preparation. A Martin and A Amedi provided helpful comments on the manuscript. This work was supported by the National Institute of Mental Health Intramural Research Program.

References (55)

  • S. Zeki

    The distribution of wavelength and orientation selective cells in different areas of monkey visual cortex

    Proc R Soc Lond B Biol Sci

    (1983)
  • Beauchamp MS: Statistical criteria in fMRI studies of multisensory integration. Neuroinformatics 2005. In...
  • Laurienti PJ, Perrault TJ Jr, Stanford TR, Wallace MT, Stein BE: On the use of superadditivity as a metric for...
  • G.A. Calvert et al.

    Multisensory integration: methodological approaches and emerging principles in the human brain

    J Physiol (Paris)

    (2004)
  • Amedi A, Kriegstein KV, Van Atteveldt N, Beauchamp MS, Naumer M: Functional imaging of human crossmodal identification...
  • T.M. Wright et al.

    Polysensory interactions along lateral temporal regions evoked by audiovisual speech

    Cereb Cortex

    (2003)
  • E. Macaluso et al.

    Spatial and temporal factors during processing of audiovisual speech: a PET study

    Neuroimage

    (2004)
  • M.S. Beauchamp et al.

    Integration of auditory and visual information about objects in superior temporal sulcus

    Neuron

    (2004)
  • C. Bruce et al.

    Visual properties of neurons in a polysensory area in superior temporal sulcus of the macaque

    J Neurophysiol

    (1981)
  • R. Gil-da-Costa et al.

    Toward an evolutionary perspective on conceptual representation: species-specific calls activate visual and affective processing systems in the macaque

    Proc Natl Acad Sci USA

    (2004)
  • D.C. Van Essen

    Surface-based approaches to spatial localization and registration in primate cerebral cortex

    Neuroimage

    (2004)
  • H. Burton et al.

    Cortical activity to vibrotactile stimulation: an fMRI study in blind and sighted individuals

    Hum Brain Mapp

    (2004)
  • B. Seltzer et al.

    Overlapping and nonoverlapping cortical projections to cortex of the superior temporal sulcus in the rhesus monkey: double anterograde tracer studies

    J Comp Neurol

    (1996)
  • C.G. Cusick et al.

    Chemoarchitectonics and corticocortical terminations within the superior temporal sulcus of the rhesus monkey: evidence for subdivisions of superior temporal polysensory cortex

    J Comp Neurol

    (1995)
  • M.S. Beauchamp et al.

    Unraveling multisensory integration: patchy organization within human STS multisensory cortex

    Nat Neurosci

    (2004)
  • J.A. De Zwart et al.

    Signal to- noise ratio and parallel imaging performance of a 16-channel receive-only brain coil array at 3.0 Tesla

    Magn Reson Med

    (2004)
  • R. Malach et al.

    Object-related activity revealed by functional magnetic resonance imaging in human occipital cortex

    Proc Natl Acad Sci USA

    (1995)
  • Cited by (311)

    View all citing articles on Scopus
    View full text