Adults can recognize between 50,000 and 100,000 words, spoken by different talkers in varying acoustic environments. This ability is remarkable because speech unfolds as a series of transient acoustic events extending over a few hundred ms, without reliable cues to word boundaries. Despite the ubiquitous success of spoken language processing among the general population, 5 percent of first graders enter school with some type of speech sound disorder that cannot be accounted for by hearing impairment. In addition, once the language system has been successfully acquired, it is susceptible to insult from injury or stroke (accounting for 1 million adults in the U.S. with some form of aphasia). Spoken word recognition plays a central role in language acquisition and spoken language comprehension, allowing for storage of the rich array of syntactic, semantic and pragmatic knowledge that is linked to lexical representations and rapid access to this information during comprehension. A more complete understanding of the perceptual and computational capacities underlying spoken word recognition is essential to advancing understanding of both normal and deviant language acquisition and processing. Because the speech signal unfolds over time and the acoustic realization of a word varies with its local environment, it is important to evaluate spoken word recognition at a fine temporal grain using words embedded in continuous speech. This project has established """"""""visual world"""""""" eye tracking as a powerful tool for examining spoken word recognition. The methods that we have developed are increasingly being used to address questions about spoken language processing in participant populations across the lifespan from infants to older adults, and in normal and impaired populations. The proposed research has two aims.
The first aim i s to evaluate a """"""""data explanation"""""""" framework in which processing words in continuous speech is modulated by expectations based on context which (a) affect how listeners interpret the input and (b) provide a mechanism for rapid perceptual learning/adaptation. We manipulate speech rate and discourse-based information structure to examine how expectations affect real-time integration of asynchronous cues and how cue-weights are adjusted through perceptual learning.
The second aim focuses on three emerging questions that affect the design, interpretation and analysis of visual world experiments: (1) Are the earliest signal-driven eye movements to pictures (at least partially) mediated by phonological information from displayed pictures or are eye-movements primarily mediated by perceptual representations activated by the spoken word;(2) What is the minimal lag between cues in the speech signal and the first stimulus-driven fixations;and (3) Are fixations affected b state-dependencies, and if so, under what conditions, and how can these effects be modeled within an event-based statistical framework.

Public Health Relevance

Speech unfolds as a series of transient acoustic events without reliable cues to word boundaries. The proposed research uses eye movements to pictures in a display to examine the mechanisms by which listeners rapidly recognize spoken words in continuous speech and adapt to differences among speakers.

Agency
National Institute of Health (NIH)
Institute
Eunice Kennedy Shriver National Institute of Child Health & Human Development (NICHD)
Type
Research Project (R01)
Project #
5R01HD073890-13
Application #
8721997
Study Section
Language and Communication Study Section (LCOM)
Program Officer
Miller, Brett
Project Start
2001-08-01
Project End
2016-06-30
Budget Start
2014-07-01
Budget End
2015-06-30
Support Year
13
Fiscal Year
2014
Total Cost
Indirect Cost
Name
University of Rochester
Department
Other Basic Sciences
Type
Schools of Arts and Sciences
DUNS #
City
Rochester
State
NY
Country
United States
Zip Code
14627
Bergelson, Elika; Swingley, Daniel (2018) Young Infants' Word Comprehension Given An Unfamiliar Talker or Altered Pronunciations. Child Dev 89:1567-1576
Kurumada, Chigusa; Brown, Meredith; Tanenhaus, Michael K (2018) Effects of distributional information on categorization of prosodic contours. Psychon Bull Rev 25:1153-1160
Bergelson, Elika; Aslin, Richard N (2017) Nature and origins of the lexicon in 6-mo-olds. Proc Natl Acad Sci U S A 114:12916-12921
Finley, Sara (2017) Learning metathesis: Evidence for syllable structure constraints. J Mem Lang 92:142-157
Buz, Esteban; Tanenhaus, Michael K; Jaeger, T Florian (2016) Dynamically adapted context-specific hyper-articulation: Feedback from interlocutors affects speakers' subsequent pronunciations. J Mem Lang 89:68-86
Brown, Meredith; Salverda, Anne Pier; Dilley, Laura C et al. (2015) Metrical expectations from preceding prosody influence perception of lexical stress. J Exp Psychol Hum Percept Perform 41:306-23
Brown, Meredith; Salverda, Anne Pier; Gunlogson, Christine et al. (2015) Interpreting prosodic cues in discourse context. Lang Cogn Neurosci 30:149-166
Heeren, Willemijn F L; Bibyk, Sarah A; Gunlogson, Christine et al. (2015) Asking or Telling--Real-time Processing of Prosodically Distinguished Questions and Statements. Lang Speech 58:474-501
Carbary, Kathleen; Brown, Meredith; Gunlogson, Christine et al. (2015) Anticipatory Deaccenting in Language Comprehension. Lang Cogn Neurosci 30:197-211
Aslin, Richard N (2014) Infant learning: Historical, conceptual, and methodological challenges. Infancy 19:2-27

Showing the most recent 10 out of 19 publications