Award Abstract # 1431118
Collaborative Research: Making words disappear or appear: A neurocognitive and behavioral investigation of effects of speech rate on spoken word perception

NSF Org: BCS
Division of Behavioral and Cognitive Sciences
Recipient: UNIVERSITY OF MASSACHUSETTS
Initial Amendment Date: August 14, 2014
Latest Amendment Date: August 14, 2014
Award Number: 1431118
Award Instrument: Standard Grant
Program Manager: Betty Tuller
btuller@nsf.gov
 (703)292-7238
BCS
 Division of Behavioral and Cognitive Sciences
SBE
 Directorate for Social, Behavioral and Economic Sciences
Start Date: August 15, 2014
End Date: July 31, 2017 (Estimated)
Total Intended Award Amount: $102,214.00
Total Awarded Amount to Date: $102,214.00
Funds Obligated to Date: FY 2014 = $102,214.00
History of Investigator:
  • Lisa Sanders (Principal Investigator)
    lsanders@psych.umass.edu
Recipient Sponsored Research Office: University of Massachusetts Amherst
101 COMMONWEALTH AVE
AMHERST
MA  US  01003-9252
(413)545-0698
Sponsor Congressional District: 02
Primary Place of Performance: University of Massachusetts Amherst
Research Administration Building
Amherst
MA  US  01003-9242
Primary Place of Performance
Congressional District:
02
Unique Entity Identifier (UEI): VGJHK59NMPK9
Parent UEI: VGJHK59NMPK9
NSF Program(s): Perception, Action & Cognition
Primary Program Source: 01001415DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 7252, 7298
Program Element Code(s): 725200
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.075

ABSTRACT

Understanding how humans comprehend speech is an unsolved and challenging problem, in part because factors such as different speakers, dialects, and speaking rates introduce a great deal of temporal and spectral variability into the speech signal. The focus of this research is on the influence of temporal context on perception of segments, syllables, and words. Results of the research may offer insights into treatment of disorders that involve disruption of speech rate (e.g., dysarthria, stuttering, Parkinson's disease, and aphasia), inform approaches to improve speech technology applications (e.g., enhanced automatic speech recognition, more natural sounding computer-generated speech), and lead to new discoveries related to brain mechanisms involved in understanding spoken language. The investigators will also involve students in the research, including those from a primarily undergraduate institution collaborating on the project.

The investigators will test different accounts of temporal phenomena in the perception of speech. They propose two interacting cognitive mechanisms controlling phenomena at lexical and phonetic levels, each driven by a different neural timing mechanism. The hypothesis is that effects of lexical rate primarily stem from top-down, speech-specific temporal expectancies, while phonetic rate effects originate in bottom-up, transient rhythmic expectations that are not specific to speech. This hypothesis will be assessed using psychoacoustic studies, non-invasive measures of brain activity, and theoretical modeling in order to identify the processing characteristics revealed by neural representations of temporal properties of speech.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

Dilley, L., Ireland, Z., Viswanathan, N., Sanders, L., Arjmandi, M., & Wieland, E. "Statistical learning of talker rate influences word perception" 30th CUNY Conference on Human Sentence Processing , 2017

PROJECT OUTCOMES REPORT

Disclaimer

This Project Outcomes Report for the General Public is displayed verbatim as submitted by the Principal Investigator (PI) for this award. Any opinions, findings, and conclusions or recommendations expressed in this Report are those of the PI and do not necessarily reflect the views of the National Science Foundation; NSF has not approved or endorsed its content.

Understanding speech requires that listeners be able to adjust to a wide variety of talkers and listening conditions. As such, there are multiple influences of the speech a listener has already heard on how they process incoming speech signals. One such effect of heard speech on speech processing involves the timing of speech information; the rate of context speech influences whether listeners report hearing acoustically blended words (e.g., “or” in “leisure or time”). Competing hypotheses posited that 1) listeners formulate predictions about the timing of upcoming speech signals based on the rate of the preceding context and 2) speech rate information can influence lexical competition in a way that favors words that are consistent with context speech rate. To differentiate between these accounts, we conducted event-related potential (ERP) research that measured the electrical activity of the brain that is time locked to specific speech events. The neurophysiological response to blended function words in acoustically identical portions of speech streams differed based on context speech rate by 100 ms after onset. Specifically, the same speech sounds that failed to elicit ERPs associated with auditory onsets and function words when presented with fast, casual contexts did elicit these brain responses when presented with slowed contexts. The ERP results were consistent with transcriptions of the phrases. The acoustically blended function words were typically not reported in the faster context condition and typically were reported in the slowed context condition. The timing of the differences in the brain response indicates that distal speech rate affects early perceptual processing of speech sounds. This result is consistent with listeners formulating predictions about the timing of upcoming speech signals, and those predictions influencing perception of incoming speech signals.

 

This research significantly expands our shared knowledge of the mechanism by which context influences speech processing. The finding that context speech rate affects early perceptual processing of speech sounds has important implications for models of speech processing. The results suggest that listeners automatically extract rate information from speech signals under typical listening conditions. Further, they demonstrate that predictions about the timing of speech sounds are typically employed when listening to speech for comprehension. Finally, the results show that factors which influence predictive processes can be functionally equivalent to acoustic modulations.

 

This research involved the contributions of seven undergraduate students and two post-baccalaureate assistants, five of whom are under-represented minorities. The research experience on this project compelled six of those students (three under-represented minorities) to apply to graduate programs; four have already begun their graduate work. One of the students who has started her PhD program is an author on a manuscript describing this research that has been submitted for publication.


Last Modified: 10/30/2017
Modified by: Lisa D Sanders

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page