Multisensory integration is at the core of many cognitive phenomena. It provides a survival advantage because it allows the brain to combine the independent estimates available from different sensory modalities into a single estimate that is more accurate than any single modality in isolation. A key obstacle to progress is our lack of knowledge about how the brain combines different modalities. If sensory modality #1 claims that the environment is """"""""X"""""""" while sensory modality #2 claims that the environment is """"""""Y"""""""", how can the estimates best be combined to guide behavior? An important finding in behavioral studies is that multisensory integration is Bayes-optimal-that is, the reliability of different sensory modalities are taken into account when integrating them. Sensory inputs that are reliable (more informative) receive greater weights, while sensory inputs that are less informative receive less weight. The goal of this proposal is to uncover the neural mechanisms for optimal visual-tactile integration. Our central hypothesis takes the form of a simple model in which the strengths of connections from unisensory to multisensory brain areas are modulated by the reliability of the stimulus in each modality. An unreliable stimulus results in a weak connection, decreasing the effectiveness of that modality in the integration area, while a reliable stimulus results in a strong connection and increased ability to drive behavior. To test our model, we propose four specific aims that will examine two distinct paradigms: a touch delivered to the hand that is both seen and felt;and speech that is both seen and heard. In the first aim, we will determine the brain areas involved in these two types of stimuli using blood oxygen-level dependent functional magnetic resonance imaging (BOLD fMRI). We will test the hypothesis that the intraparietal sulcus (IPS) will respond to visual and tactile touch and that the superior temporal sulcus (STS) will respond to auditory and visual speech. In the second aim, we will show that neural connection strengths are proportional to stimulus reliability. We will test the hypothesis that the effective connectivity between unisensory and multisensory areas will be proportional to the reliability of the stimulus presented in that modality. In the third aim, we will demonstrate a correlation between multisensory brain activity and behavior using multi-voxel pattern analysis (MVPA). In the fourth aim, we will reveal a causal link between brain activity and behavioral multisensory integration. Using fMRI-guided transcranial magnetic stimulation (TMS), we will test the hypothesis that TMS of multisensory areas will eliminate the behavioral advantage of multisensory stimuli and the hypothesis that TMS of unisensory areas will impair behavioral performance proportional to the reliability of the stimulus in that modality.
Multisensory integration is at the core of many cognitive phenomena. We will use functional magnetic resonance imaging (fMRI) and transcranial magnetic stimulation (TMS) in normal human subjects to study the organization and operation of the brain during multisensory integration.
|Magnotti, John F; Beauchamp, Michael S (2015) The noisy encoding of disparity model of the McGurk effect. Psychon Bull Rev 22:701-9|
|Gurler, Demet; Doyle, Nathan; Walker, Edgar et al. (2015) A link between individual differences in multisensory speech perception and eye movements. Atten Percept Psychophys 77:1333-41|
|Magnotti, John F; Basu Mallick, Debshila; Feng, Guo et al. (2015) Similar frequency of the McGurk effect in large samples of native Mandarin Chinese and American English speakers. Exp Brain Res 233:2581-6|
|Mallick, Debshila Basu; Magnotti, John F; Beauchamp, Michael S (2015) Variability and stability in the McGurk effect: contributions of participants, stimuli, time, and response type. Psychon Bull Rev 22:1299-307|
|Schepers, Inga M; Yoshor, Daniel; Beauchamp, Michael S (2015) Electrocorticography Reveals Enhanced Visual Cortex Responses to Visual Speech. Cereb Cortex 25:4103-10|
|Beauchamp, Michael S (2015) The social mysteries of the superior temporal sulcus. Trends Cogn Sci 19:489-90|
|Pollonini, Luca; Olds, Cristen; Abaya, Homer et al. (2014) Auditory cortex activation to natural speech and simulated cochlear implant speech measured with functional near-infrared spectroscopy. Hear Res 309:84-93|
|Baum, Sarah H; Beauchamp, Michael S (2014) Greater BOLD variability in older compared with younger adults during audiovisual speech perception. PLoS One 9:e111121|
|Ro, Tony; Ellmore, Timothy M; Beauchamp, Michael S (2013) A neural link between feeling and hearing. Cereb Cortex 23:1724-30|
|Baum, Sarah H; Martin, Randi C; Hamilton, A Cris et al. (2012) Multisensory speech perception without the left superior temporal sulcus. Neuroimage 62:1825-32|
Showing the most recent 10 out of 20 publications