Speech perception is the process by which listeners hear and interpret the sounds of language. It is defined by the unique mapping of a highly variable and complex acoustic signal to a phonetic representation. A convergence of studies has implicated the human posterior superior temporal cortex for the specialized processing of speech sounds. Although localized, the basic neural mechanisms by which linguistic information is extracted are entirely unclear. We propose an innovative methodological approach using customized intracranial high-density electrode arrays to record detailed neural activity directly from the posterior temporal cortex in awake, behaving subjects undergoing clinical evaluation for epilepsy surgery. This method offers the highest possible spatial and temporal resolution, thereby overcoming many limitations of non-invasive imaging approaches. This research is positioned at a critical interface between the fields of auditory neuroscience and linguistics. Our previous results demonstrate that cortical representation of speech sounds manifest important non-linearities that correspond to perceptual boundaries over acoustic parameters (Chang et al, Nature Neuroscience 2010). Building on these findings, we propose experiments to determine: 1) the functional organization of the posterior temporal gyrus for acoustic and phonetic features, 2) the distributed, population-level encoding of emergent phonetic representation, and 3) the correlates of "robust" speech representation in the context of a multi-speaker listening environment. The results of these proposed experiments will have significant impact on the field of neurolinguistics and broader research on sensory perception and cognition.

Public Health Relevance

Alterations in the cortical mechanisms supporting speech perception underlie the pathogenesis of an increasing number of acquired and developmental language disabilities, including aphasia, language learning delay, autism, and dyslexia. Revealing mechanisms by which the brain encodes speech sounds bears tremendous implications for our understanding of these disorders, and more importantly, will guide strategies for their remediation.

Agency
National Institute of Health (NIH)
Institute
National Institute on Deafness and Other Communication Disorders (NIDCD)
Type
Research Project (R01)
Project #
5R01DC012379-03
Application #
8643221
Study Section
Auditory System Study Section (AUD)
Program Officer
Shekim, Lana O
Project Start
2012-04-01
Project End
2017-03-31
Budget Start
2014-04-01
Budget End
2015-03-31
Support Year
3
Fiscal Year
2014
Total Cost
$323,167
Indirect Cost
$110,667
Name
University of California San Francisco
Department
Neurosurgery
Type
Schools of Medicine
DUNS #
094878337
City
San Francisco
State
CA
Country
United States
Zip Code
94143
Leonard, Matthew K; Chang, Edward F (2014) Dynamic speech representations in the human temporal lobe. Trends Cogn Sci 18:472-9
Bouchard, Kristofer E; Chang, Edward F (2014) Control of spoken vowel acoustics and the influence of phonetic context in human speech sensorimotor cortex. J Neurosci 34:12662-77
Conant, David; Bouchard, Kristofer E; Chang, Edward F (2014) Speech map in the human ventral sensory-motor cortex. Curr Opin Neurobiol 24:63-7
Mesgarani, Nima; Cheung, Connie; Johnson, Keith et al. (2014) Phonetic feature encoding in human superior temporal gyrus. Science 343:1006-10
Edwards, Erik; Chang, Edward F (2013) Syllabic (ýýý2-5ýýHz) and fluctuation (ýýý1-10ýýHz) ranges in speech andýýauditory processing. Hear Res 305:113-34
Chang, Edward F; Niziolek, Caroline A; Knight, Robert T et al. (2013) Human cortical sensorimotor network underlying feedback control of vocal pitch. Proc Natl Acad Sci U S A 110:2653-8
Bouchard, Kristofer E; Mesgarani, Nima; Johnson, Keith et al. (2013) Functional organization of human sensorimotor cortex for speech articulation. Nature 495:327-32
Mesgarani, Nima; Chang, Edward F (2012) Selective cortical representation of attended speaker in multi-talker speech perception. Nature 485:233-6