In language processing, our impression is that speech is a sequence of words. This subjective impression conceals a complex coding process that transforms sound into a segmental representation from which word recognition can proceed. The goal of this research is a formal specification of the auditory and phonetic coding of speech. This, in turn, can serve as a model of normal perception for comparison with theories of perceptual disorders and perceptual development. This research may also contribute to the development of machine-based speech perception which could serve as an aid to the deaf. The experimental studies and computer modeling in this research are aimed at answering three interrelated questions. How are the acoustic qualities of speech mapped onto a segmental representation of the sound? What is the precise nature of this segmental representation? How is the auditory and phonetic coding of this segmental representation influenced by higher level processes, including the listener's knowledge of the words of the language? The approach to answering these questions involves converging operations. Different experimental tasks that tap into different aspects of human perception are used to provide a pattern of data that can distinguish among competing theories. The experimental procedures of identification, discrimination, monitoring, naming, lexical decision, and similarity judgment will be used in conjunction with natural speech, synthetic speech and complex nonspeech stimuli. The choice of stimuli is based on models of the auditory and phonetic coding of speech and models of auditory word recognition. These models allow us to manipulate certain qualities of the speech sounds (the shape of the spectrum of a vowel, the number of words a syllable is similar to) while holding other aspects of the sounds constant. The experimental tasks include both on-line measures related to fluent language processing as well as more traditional, psychophysical tasks. The experimental results will be incorporated into computer simulation models of human speech perception designed to assess the adequacy of the underlying theories. The simulation models can also serve as a basis for understanding perceptual disorders and possible methods for rehabilitation.

Agency
National Institute of Health (NIH)
Institute
National Institute on Deafness and Other Communication Disorders (NIDCD)
Type
Research Project (R01)
Project #
2R01DC000219-15A2
Application #
2903499
Study Section
Special Emphasis Panel (ZRG1-CMS (02))
Project Start
1983-07-01
Project End
2003-06-30
Budget Start
1999-07-01
Budget End
2000-06-30
Support Year
15
Fiscal Year
1999
Total Cost
Indirect Cost
Name
State University of New York at Buffalo
Department
Psychology
Type
Schools of Arts and Sciences
DUNS #
038633251
City
Buffalo
State
NY
Country
United States
Zip Code
14260
Newman, Rochelle S; Sawusch, James R (2009) Perceptual normalization for speaking rate III: Effects of the rate of one voice on perception of another. J Phon 37:46-65
Sawusch, J R; Newman, R S (2000) Perceptual normalization for speaking rate. II: Effects of signal discontinuities. Percept Psychophys 62:285-300
Turk, A E; Sawusch, J R (1996) The processing of duration and intensity cues to prominence. J Acoust Soc Am 99:3782-90
Newman, R S; Sawusch, J R (1996) Perceptual normalization for speaking rate: effects of temporal distance. Percept Psychophys 58:540-60
Sawusch, J R; Gagnon, D A (1995) Auditory coding, cues, and coherence in phonetic perception. J Exp Psychol Hum Percept Perform 21:635-52