首页 | 本学科首页   官方微博 | 高级检索  
   检索      


Lexical Mediation Between Sight and Sound in Speechreading
Authors:Bruno H Repp  Ram Frost  Elizabeth Zsiga
Institution:  a Haskins Laboratories, New Haven, Connecticut, U.S.A. b Hebrew University of Jerusalem, Israel c Yale University and Haskins Laboratories, New Haven, Connecticut, U.S.A.
Abstract:In two experiments, we investigated whether simultaneous speech reading can influence the detection of speech in envelope-matched noise. Subjects attempted to detect the presence of a disyllabic utterance in noise while watching a speaker articulate a matching or a non-matching utterance. Speech detection was not facilitated by an audio-visual match, which suggests that listeners relied on low-level auditory cues whose perception was immune to cross-modal top-down influences. However, when the stimuli were words (Experiment 1), there was a (predicted) relative shift in bias, suggesting that the masking noise itself was perceived as more speechlike when its envelope corresponded to the visual information. This bias shift was absent, however, with non-word materials (Experiment 2). These results, which resemble earlier findings obtained with orthographic visual input, indicate that the mapping from sight to sound is lexically mediated even when, as in the case of the articulatory-phonetic correspondence, the cross-modal relationship is non-arbitrary.
Keywords:
本文献已被 InformaWorld 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号