首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Listeners had the task of following a target speech signal heard against two competitors either located at the same spatial position as the target or displaced symmetrically to locations flanking it. When speech was the competitor, there was a significantly higher separation effect (maintained intelligibility with reduced target sound level), as compared with either steady-state or fluctuating noises. Increasing the extent of spatial separation slightly increased the effect, and a substantial contribution of interaural time differences was observed. When same- and opposite-sex voices were used, a hypothesis that the similarity between target and competing speech would explain the role for spatial separation was partly supported. High- and low-pass filtering showed that both parts of an acoustically similar competing signal contribute to the phenomenon. We conclude that, in parsing the auditory array, attention to spatial cues is heightened when the components of the array are confusable on other acoustic grounds.  相似文献   

2.
When speech is rapidly alternated between the two ears, intelligibility declines as the rate of alternation approaches 3 to 5 switching cycles per second, and then, paradoxically, returns to a good level beyond that point. We tested intelligibility when shadowing was used as a response measure (Experiment 1), when recall was used as a response measure (Experiment 2), and when time-compression was used to vary the speech rate of the presented materials (Experiment 3). In spite of claims that older adults are generally slower in switching attention, younger and older adults did not differ in the critical alternation rates producing minimal intelligibility. We suggest that the point of minimal intelligibility in alternated speech reflects an interaction between (1) the rate of disruption induced by breaking the speech stream between two sound sources, (2) the amount of contextual information per ear, and (3) the size of the silent gaps separating the speech elements that must be perceptually bridged.  相似文献   

3.
Classical views of speech perception argue that the static and dynamic characteristics of spectral energy peaks (formants) are the acoustic features that underpin phoneme recognition. Here we use representations where the amplitude modulations of sub-band filtered speech are described, precisely, in terms of co-sinusoidal pulses. These pulses are parameterised in terms of their amplitude, duration and position in time across a large number of spectral channels. Coherent sweeps of energy across this parameter space are identified and the local transitions of pulse features across spectral channels are extracted. Synthesised speech based on manipulations of these local amplitude modulation features was used to explore the basis of intelligibility. The results show that removing changes in amplitude across channels has a much greater impact on intelligibility than differences in sweep transition or duration across channels. This finding has severe implications for future experimental design in the fields of psychophysics, electrophysiology and neuroimaging.  相似文献   

4.
20 female students in speech-language pathology provided magnitude estimation scaling responses for the speech intelligibility and acceptability of audio-taped speech samples varying systematically the number of consonant sounds produced correctly. Analysis indicated no significant over-all differences between listeners' judgments of intelligibility and acceptability; however, listeners tended to judge samples with fewer than 50% of the consonants correct as more acceptable than intelligible, and they judged samples with more than 50% consonants correct as less acceptable than intelligible.  相似文献   

5.
It is not unusual that people have to write in an environment where background speech is present. Background speech can vary in both speech intelligibility and location of the sound source. Earlier research has shown disruptive effects of background speech on writing performance. To expand and reinforce this knowledge, the present study investigated the role of number of voices and sound source location in the relation between background speech and writing performance. Participants wrote texts in quiet or in background speech consisting of one or seven voices talking simultaneously located in front of or behind them. Overall, one voice was more disruptive than seven voices talking simultaneously. Self‐reports showed that sound from the front was more disruptive compared with sound from behind. Results are in line with theory of interference‐by‐process, attentional capture, and the cross‐modal theory of attention. The relevance of the results for open‐office environments is discussed.  相似文献   

6.
Outside of the laboratory, listening conditions are often less than ideal, and when attending to sounds from a particular source, portions are often obliterated by extraneous noises. However, listeners possess rather elegant reconstructive mechanisms. Restoration can be complete, so that missing segments are indistinguishable from those actually present and the listener is unaware that the signal is fragmented. This phenomenon, called temporal induction (TI), has been studied extensively with nonverbal signals and to a lesser extent with speech. Earlier studies have demonstrated that TI can produce illusory continuity spanning gaps of a few hundred milliseconds when portions of a signal are replaced by a louder sound capable of masking the signal were it actually present. The present study employed various types of speech signals with periodic gaps and measured the effects upon intelligibility produced by filling these gaps with noises. Enhancement of intelligibility through multiple phonemic restoration occurred when the acoustic requirements for TI were met and when sufficient contextual information was available in the remaining speech fragments. It appears that phonemic restoration is a specialized form of TI that uses linguistic skills for the reconstruction of obliterated speech.  相似文献   

7.
Outside of the laboratory, listening conditions are often less than ideal, and when attending to sounds from a particular source, portions are often obliterated by extraneous noises. However, listeners possess rather elegant reconstructive mechanisms. Restoration can be complete, so that missing segments are indistinguishable from those actually present and the listener is unaware that the signal is fragmented. This phenomenon, called temporal induction (TI), has been studied extensively with nonverbal signals and to a lesser extent with speech. Earlier studies have demonstrated that TI can produce illusory continuity spanning gaps of a few hundred milliseconds when portions of a signal are replaced by a louder sound capable of masking the signal were it actually present. The present study employed various types of speech signals with periodic gaps and measured the effects upon intelligibility produced by filling these gaps with noises. Enhancement of intelligibility through multiple phonemic restoration occurred when the acoustic requirements for TI were met and when sufficient contextual information was available in the remaining speech fragments. It appears that phonemic restoration is a specialized form of TI that uses linguistic skills for the reconstruction of obliterated speech.  相似文献   

8.
This study assessed intelligibility in a dysarthric patient with Parkinson's disease (PD) across five speech production tasks: spontaneous speech, repetition, reading, repeated singing, and spontaneous singing, using the same phrases for all but spontaneous singing. The results show that this speaker was significantly less intelligible when speaking spontaneously than in the other tasks. Acoustic analysis suggested that relative intensity and word duration were not independently linked to intelligibility, but dysfluencies (from perceptual analysis) and articulatory/resonance patterns (from acoustic records) were related to intelligibility in predictable ways. These data indicate that speech production task may be an important variable to consider during the evaluation of dysarthria. As speech production efficiency was found to vary with task in a patient with Parkinson's disease, these results can be related to recent models of basal ganglia function in motor performance.  相似文献   

9.
The purpose of this investigation was to judge whether the Lombard effect, a characteristic change in the acoustical properties of speech produced in noise, existed in adductor spasmodic dysphonia speech, and if so, whether the effect added to or detracted from speaker intelligibility. Intelligibility, as described by Duffy, is the extent to which the acoustic signal produced by a speaker is understood by a listener based on the auditory signal alone. Four speakers with adductor spasmodic dysphonia provided speech samples consisting of low probability sentences from the Speech Perception in Noise test to use as stimuli. The speakers were first tape-recorded as they read the sentences in a quiet speaking condition and were later tape-recorded as they read the same sentences while exposed to background noise. The listeners used as subjects in this study were 50 undergraduate university students. The results of the statistical analysis indicated a significant difference between the intelligibility of the speech recorded in the quiet versus noise conditions (F(1,49) = 57.80, p < or = .001). It was concluded that a deleterious Lombard effect existed for the adductor spasmodic dysphonia speaker group, with the premise that the activation of a Lombard effect in such patients may detract from their overall speech intelligibility.  相似文献   

10.
We studied speech intelligibility and memory performance for speech material heard under different signal‐to‐noise (S/N) ratios. Pre‐experimental measures of working memory capacity (WMC) were taken to explore individual susceptibility to the disruptive effects of noise. Thirty‐five participants first completed a WMC‐operation span task in quiet and later listened to spoken word lists containing 11 one‐syllable phonetically balanced words presented at four different S/N ratios (+12, +9, +6, and +3). Participants repeated each word aloud immediately after its presentation, to establish speech intelligibility and later on performed a free recall task for those words. The speech intelligibility function decreased linearly with increasing S/N levels for both the high‐WMC and low‐WMC groups. However, only the low‐WMC group had decreasing memory performance with increasing S/N levels. The memory of the high‐WMC individuals was not affected by increased S/N levels. Our results suggest that individual differences in WMC counteract some of the negative effects of speech noise. Copyright © 2012 John Wiley & Sons, Ltd.  相似文献   

11.
When deleted segments of speech are replaced by extraneous sounds rather than silence, the missing speech fragments may be perceptually restored and intelligibility improved. This phonemic restoration (PhR) effect has been used to measure various aspects of speech processing, with deleted portions of speech typically being replaced by stochastic noise. However, several recent studies of PhR have used speech-modulated noise, which may provide amplitude-envelope cues concerning the replaced speech. The present study compared the effects upon intelligibility of replacing regularly spaced portions of speech with stochastic (white) noise versus speech-modulated noise. In Experiment 1, filling periodic gaps in sentences with noise modulated by the amplitude envelope of the deleted speech fragments produced twice the intelligibility increase obtained with interpolated stochastic noise. Moreover, when lists of isolated monosyllables were interrupted in Experiment 2, interpolation of speech-modulated noise increased intelligibility whereas stochastic noise reduced intelligibility. The augmentation of PhR produced by modulated noise appeared without practice, suggesting that speech processing normally involves not only a narrowband analysis of spectral information but also a wideband integration of amplitude levels across critical bands. This is of considerable theoretical interest, but it also suggests that since PhRs produced by speech-modulated noise utilize potent bottom-up cues provided by the noise, they differ from the PhRs produced by extraneous sounds, such as coughs and stochastic noise.  相似文献   

12.
Kim J  Sironic A  Davis C 《Perception》2011,40(7):853-862
Seeing the talker improves the intelligibility of speech degraded by noise (a visual speech benefit). Given that talkers exaggerate spoken articulation in noise, this set of two experiments examined whether the visual speech benefit was greater for speech produced in noise than in quiet. We first examined the extent to which spoken articulation was exaggerated in noise by measuring the motion of face markers as four people uttered 10 sentences either in quiet or in babble-speech noise (these renditions were also filmed). The tracking results showed that articulated motion in speech produced in noise was greater than that produced in quiet and was more highly correlated with speech acoustics. Speech intelligibility was tested in a second experiment using a speech-perception-in-noise task under auditory-visual and auditory-only conditions. The results showed that the visual speech benefit was greater for speech recorded in noise than for speech recorded in quiet. Furthermore, the amount of articulatory movement was related to performance on the perception task, indicating that the enhanced gestures made when speaking in noise function to make speech more intelligible.  相似文献   

13.
In this report, we have analyzed the effects of left and right STN stimulation separately on different aspects of speech. Significant differences were found between left and right stimulation. It appears that selective left-sided stimulation has a profoundly negative effect on prosody, articulation and hence intelligibility. Right-sided stimulation does not display this side-effect. There is no significant difference in speech characteristics between bilateral stimulation on and off. We suggest that a balanced tuning of bilateral basal ganglia networks is necessary for speech, and that the left circuit is probably dominant.  相似文献   

14.
An automated threshold method has been developed for determining the maximum rate of speech understood by individual listeners. Two experiments were undertaken to determine whether the threshold was related to the comprehension of speech or to speech intelligibility. The first experiment compared thresholds of two types of rapid speech reportedly different in intelligibility: simple speeded speech and speech compressed by the sampling method. The second experiment sought to determine the relationship of the threshold to traditional comprehension measures. The results are discussed in terms of the intelligibility and comprehensibility of speech.  相似文献   

15.
Understanding low-intelligibility speech is effortful. In three experiments, we examined the effects of intelligibility on working memory (WM) demands imposed by perception of synthetic speech. In all three experiments, a primary speeded word recognition task was paired with a secondary WM-load task designed to vary the availability of WM capacity during speech perception. Speech intelligibility was varied either by training listeners to use available acoustic cues in a more diagnostic manner (as in Experiment 1) or by providing listeners with more informative acoustic cues (i.e., better speech quality, as in Experiments 2 and 3). In the first experiment, training significantly improved intelligibility and recognition speed; increasing WM load significantly slowed recognition. A significant interaction between training and load indicated that the benefit of training on recognition speed was observed only under low memory load. In subsequent experiments, listeners received no training; intelligibility was manipulated by changing synthesizers. Improving intelligibility without training improved recognition accuracy, and increasing memory load still decreased it, but more intelligible speech did not produce more efficient use of available WM capacity. This suggests that perceptual learning modifies the way available capacity is used, perhaps by increasing the use of more phonetically informative features and/or by decreasing use of less informative ones.  相似文献   

16.
Microswitches are technical devices designed to enable persons with multiple disabilities to control environmental events, generally preferred stimuli such as music and lights, using simple/feasible responses. For example, a sound-detecting microswitch may allow a person to turn on brief periods of preferred stimulation through simple vocal responses. This study compared the accuracy of two versions of a sound-detecting microswitch, i.e., one using a throat microphone and the other both throat and airborne microphones, for a man with multiple disabilities. Analysis showed that the latter microswitch version radically reduced the false activations present with the former microswitch version. Thus, the latter version seemed to have a clear accuracy advantage over the former with important practical implications.  相似文献   

17.
18.
The role of rhythm in the speech intelligibility of 18 hearing-impaired children, aged 15 years with hearing losses from 40 to 108 db, was investigated. Their perceptual judgement of visual rhythm sequences was superior to that of the hearing controls, but their scores were not correlated with their speech intelligibility.  相似文献   

19.
In order to function effectively as a means of communication, speech must be intelligible under the noisy conditions encountered in everyday life. Two types of perceptual synthesis have been reported that can reduce or cancel the effects of masking by extraneous sounds: Phonemic restoration can enhance intelligibility when segments are replaced or masked by noise, and contralateral induction can prevent mislateralization by effectively restoring speech masked at one ear when it is heard in the other. The present study reports a third type of perceptual synthesis induced by noise: enhancement of intelligibility produced by adding noise to spectral gaps. In most of the experiments, the speech stimuli consisted of two widely separated narrow bands of speech (center frequencies of 370 and 6000 Hz, each band having high-pass and low-pass slopes of 115 dB/octave meeting at the center frequency). These very narrow bands effectively reduced the available information to frequency-limited patterns of amplitude fluctuation lacking information concerning formant structure and frequency transitions. When stochastic noise was introduced into the gap separating the two speech bands, intelligibility increased for “everyday” sentences, for sentences that varied in the transitional probability of keywords, and for monosyllabic word lists. Effects produced by systematically varying noise amplitude and noise bandwidth are reported, and the implications of some of the novel effects observed are discussed.  相似文献   

20.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号