The representation of speech and other complex auditory signals in the human brain constitutes a major interdisciplinary challenge for cognitive neuroscience. Understanding in a principled manner how acoustic signals are transformed and ultimately recognized as words in a speaker's mental dictionary requires the integration of knowledge across fields ranging from single-cell recording in auditory cortex to linguistic theory. The research program outlined here is focused on two subroutines in speech processing. In the context of the first specific aim, the hypothesis is investigated that speech is analyzed concurrently on two time scales in human auditory cortex, with one corresponding to analysis at the syllabic scale, another at the segmental (phonemic) scale. This multi-time resolution model, which provides an account of hemispheric asymmetry in audition, is tested in a series of behavioral and electrophysiological studies. The goal is to provide a theoretically motivated and neurobiologically sensible answer to how acoustic signals are fractionated in time and how they map to words stored in the brain.
The second aim encompasses both behavioral (often audio- visual) and electrophysiological studies that test how (specifically, how abstractly) speech and words are represented in the human brain. The goal is to test models of the cortical encoding of speech sounds and words. The principal method used in this research program is magnetoencephalography (MEG), typically with parallel behavioral studies performed. Other non-invasive recording modalities are also employed (EEG, fMRI) to validate and extend data from any single approach.
Successfully perceiving speech and recognizing words are processes at the basis of human communication. A mechanistic characterization of the brain structures that mediate these skills is essential to understand the range of disorders associated with problems in speech processing. Health-related phenomena ranging from dyslexia and autism in childhood to aphasia and Alzheimer's disease in the aging population have been repeatedly linked to problems with the auditory analysis of complex signals and the ability to process words appropriately. The development of innovative diagnostic, interventional, and therapeutic approaches critically depends on our enriched knowledge of the brain basis of the processes underlying human speech.
|Scharinger, Mathias; Monahan, Philip J; Idsardi, William J (2016) Linguistic category structure influences early auditory processing: Converging evidence from mismatch responses and cortical oscillations. Neuroimage 128:293-301|
|Steinberg Lowe, Mara; Lewis, Gwyneth A; Poeppel, David (2016) Effects of Part- and Whole-Object Primes on Early MEG Responses to Mooney Faces and Houses. Front Psychol 7:147|
|Tian, Xing; Zarate, Jean Mary; Poeppel, David (2016) Mental imagery of speech implicates two mechanisms of perceptual reactivation. Cortex 77:1-12|
|Almeida, Diogo; Poeppel, David; Corina, David (2016) The Processing of Biologically Plausible and Implausible forms in American Sign Language: Evidence for Perceptual Tuning. Lang Cogn Neurosci 31:361-374|
|Teng, Xiangbin; Tian, Xing; Poeppel, David (2016) Testing multi-scale processing in the auditory system. Sci Rep 6:34390|
|Ding, Nai; Melloni, Lucia; Zhang, Hang et al. (2016) Cortical tracking of hierarchical linguistic structures in connected speech. Nat Neurosci 19:158-64|
|Lewis, Gwyneth A; Poeppel, David; Murphy, Gregory L (2015) The neural bases of taxonomic and thematic conceptual relations: an MEG study. Neuropsychologia 68:176-89|
|Chait, Maria; Greenberg, Steven; Arai, Takayuki et al. (2015) Multi-time resolution analysis of speech: evidence from psychophysics. Front Neurosci 9:214|
|Overath, Tobias; McDermott, Josh H; Zarate, Jean Mary et al. (2015) The cortical analysis of speech-specific temporal structure revealed by responses to sound quilts. Nat Neurosci 18:903-11|
|Rimmele, Johanna M; Zion Golumbic, Elana; SchrÃ¶ger, Erich et al. (2015) The effects of selective attention and speech acoustics on neural speech-tracking in a multi-talker scene. Cortex 68:144-54|
Showing the most recent 10 out of 90 publications