首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 250 毫秒
1.
It has been demonstrated using the "silent-center" (SC) syllable paradigm that there is sufficient information in syllable onsets and offsets, taken together, to support accurate identification of vowels spoken in both citation-form syllables and syllables spoken in sentence context. Using edited natural speech stimuli, the present study examined the identification of American English vowels when increasing amounts of syllable onsets alone or syllable offsets alone were presented in their original sentence context. The stimuli were /d/-vowel-/d/ syllables spoken in a short carrier sentence by a male speaker. Listeners attempted to identify the vowels in experimental conditions that differed in the number of pitch periods presented and whether the pitch periods were from syllable onsets or syllable offsets. In general, syllable onsets were more informative than syllable offsets, although neither onsets nor offsets alone specified vowel identity as well as onsets and offsets together (SC syllables). Vowels differed widely in ease of identification; the diphthongized long vowels /e/, /ae/, /o/ were especially difficult to identify from syllable offsets. Identification of vowels as "front" or "back" was accurate, even from short samples of the syllable; however, vowel "height" was quite difficult to determine, again, especially from syllable offsets. The results emphasize the perceptual importance of time-varying acoustic parameters, which are the direct consequence of the articulatory dynamics involved in producing syllables.  相似文献   

2.
Traditionally, it has been held that the primary information for vowel identification is provided by formant frequencies in the quasi-steady-state portion of the spoken syllable. Recent search has advanced an alternative view that emphasizes the role of temporal factors and dynamic (time-varying) spectral information in determining the perception of vowels. Nine vowels spoken in /b/ + vowel + /b/ syllables were recorded. The syllables were modified electronically in several ways to suppress various sources of spectral and durational information. Two vowel-perception experiments were performed, testing subjects’ ability to identify vowels in these modified syllables. Results of both experiments revealed the importance of dynamic spectral information at syllable onset and offset (in its proper temporal relation) in permitting vowel identification. On the other hand, steady-state spectral information, deprived of its durational variation, was a poor basis for identification. Results constitute a challenge to traditional accounts of vowel perception and point toward important sources of dynamic information.  相似文献   

3.
We explore how listeners perceive distinct pieces of phonetic information that are conveyed in parallel by the fundamental frequency (f0) contour of spoken and sung vowels. In a first experiment, we measured differences inf0 of /i/ and /a/ vowels spoken and sung by unselected undergraduate participants. Differences in “intrinsicf0” (withf0 of /i/ higher than of /a/) were present in spoken and sung vowels; however, differences in sung vowels were smaller than those in spoken vowels. Four experiments tested a hypothesis that listeners would not hear the intrinsicf0 differences as differences in pitch on the vowel, because they provide information, instead, for production of a closed or open vowel. The experiments provide clear evidence of “parsing” of intrinsicf0 from thef0 that contributes to perceived vowel pitch. However, only some conditions led to an estimate of the magnitude of parsing that closely matched the magnitude of produced intrinsicf0 differences.  相似文献   

4.
Three experiments were performed to examine listeners’ thresholds for identifying stimuli whose spectra were modeled after the vowels /i/ and /ε/, with the differences between these stimuli restricted to the frequency of the first formant. The stimuli were presented in a low-pass masking noise that spectrally overlapped the first formant but not the higher formants. Identification thresholds were lower when the higher formants were present than when they were not, even though the first formant contained the only distinctive information for stimulus identification. This indicates that listeners were more sensitive in identifying the first formant energy through its contribution to the vowel than as an independent percept; this effect is given the namecoherence masking protection. The first experiment showed this effect for synthetic vowels in which the distinctive first formant was supported by a series of harmonics that progressed through the higher formants. In the second two experiments, the harmonics in the first formant region were removed, and the first formant was simulated by a narrow band of noise. This was done so that harmonic relations did not provide a basis for grouping the lower formant with the higher formants; coherence masking protection was still observed. However, when the temporal alignment of the onsets and offsets of the higher and lower formants was disrupted, the effect was eliminated, although the stimuli were still perceived as vowels. These results are interpreted as indicating that general principles of auditory grouping that can exploit regularities in temporal patterns cause acoustic energy belonging to a coherent speech sound to stand out in the auditory scene.  相似文献   

5.
Recent studies in alphabetic writing systems have investigated whether the status of letters as consonants or vowels influences the perception and processing of written words. Here, we examined to what extent the organisation of consonants and vowels within words affects performance in a syllable counting task in English. Participants were asked to judge the number of syllables in written words that were matched for the number of spoken syllables but comprised either 1 orthographic vowel cluster less than the number of syllables (hiatus words, e.g., triumph) or as many vowel clusters as syllables (e.g., pudding). In 3 experiments, we found that readers were slower and less accurate on hiatus than control words, even when phonological complexity (Experiment 1), number of reduced vowels (Experiment 2), and number of letters (Experiment 3) were taken into account. Interestingly, for words with or without the same number of vowel clusters and syllables, participants’ errors were more likely to underestimate the number of syllables than to overestimate it. Results are discussed in a cross-linguistic perspective.  相似文献   

6.
Two dichotic listening experiments assess the lateralization of speaker identification in right-handed native English speakers. Stimuli were tokens of /ba/, /da/, /pa/, and /ta/ pronounced by two male and two female speakers. In Experiment 1, subjects identified either the two consonants in dichotic stimuli spoken by the same person, or identified two speakers in dichotic tokens of the same syllable. In Experiment 2 new subjects identified the two consonants or the two speakers in pairs in which both consonant and speaker distinguished the pair members. Both experiments yielded significant right-ear advantages for consonant identification and nonsignificant ear differences for speaker identification. Fewer errors were made for speaker judgments than for consonant judgments, and for speaker judgments for pairs in which the speakers were of the same sex than for pairs in which speaker sex differed. It is concluded that, as in vowel identification, neither hemisphere clearly dominates in dichotic speaker identification, perhaps because of minor information loss in the ipsilateral pathways.  相似文献   

7.
Ss either saw or heard lists of three syllables which differed by their initial consonant phoneme or their final vowel phoneme. After 5 or 15 sec of mental arithmetic, Ss were required to recall the syllables. Following auditory presentation, vowels were recalled more accurately than consonants in all serial positions and at both delays. In addition, spoken consonants and vowels showed primacy and recency effects. Following visual presentation, consonants and vowels were recalled with equal accuracy at both delays, and no recency effects were observed. These data suggest that superior recall of vowels over consonants results from differential decay of these stimuli in an acoustic storage. These data are consistent with previous experiments showing that, during serial recall, the final vowels in a sequence are recalled more accurately than the final consonants.  相似文献   

8.
The ability of 3-year-old children to perceive the identity of vowels in full-vowel and silentcenter, consonant-vowel-consonant (CVC) syllables was investigated using a two-alternative pointing procedure. Silence replaced the middle 10%, 35%, 65%, or 90% of the steady-state formants of synthetic “bad” and “bud” syllables. Identification of the two full-vowel syllables was 87% correct, whereas performance for the silent-center syllables was somewhat lower (72%, 70%, 67%, and 66% correct for the 10%, 35%, 65%, and 90% deletion stimuli, respectively). The performance of individual children fell into two subgroups: (1) those who performed like adults by maintaining correct vowel identification for all of the silent-center syllables, and (2) those who identified the full-vowel syllables correctly but performed at chance for all of the silent-center syllables. Three additional experiments showed that none of the children performed poorly when noise replaced the gap in the silent-center syllables. These results demonstrate that many 3-year-olds can identify vowels correctly in CVC syllables in the absence of the full spectral properties of steady-state formants.  相似文献   

9.
Research has shown that speaking rate provides an important context for the perception of certain acoustic properties of speech. For example, syllable duration, which varies as a function of speaking rate, has been shown to influence the perception of voice onset time (VOT) for syllableinitial stop consonants. The purpose of the present experiments was to examine the influence of syllable duration when the initial portion of the syllable was produced by one talker and the remainder of the syllable was produced by a different talker. A short-duration and a long-duration /bi/-/pi/ continuum were synthesized with pitch and formant values appropriate to a female talker. When presented to listeners for identification, these stimuli demonstrated the typical effect of syllable duration on the voicing boundary: a shorter VOT boundary for the short stimuli than for the long stimuli. An /i/ vowel, synthesized with pitch and formant values appropriate to a male talker, was added to the end of each of the short tokens, producing a new hybrid continuum. Although the overall syllable duration of the hybrid stimuli equaled the original long stimuli, they produced a VOT boundary similar to that for the short stimuli. In a second experiment, two new /i/ vowels were synthesized. One had a pitch appropriate to a female talker with formant values appropriate to a male talker; the other had a pitch appropriate to a male talker and formants appropriate to a female talker. These vowels were used to create two new hybrid continua. In a third experiment, new hybrid continua were created by using more extreme male formant values. The results of both experiments demonstrated that the hybrid tokens with a change in pitch acted like the short stimuli, whereas the tokens with a change in formants acted like the long stimuli. A fourth experiment demonstrated that listeners could hear a change in talker with both sets of hybrid tokens. These results indicate that continuity of pitch but not formant structure appears to be the critical factor in the calculation of speaking rate within a syllable.  相似文献   

10.
Anticipatory coarticulation in a patient with apraxia of speech   总被引:1,自引:1,他引:0  
Articulatory anticipation of vowel gestures was assessed in an apraxic patient, a dysarthric patient, and three normal speakers. The technique of assessment included perceptual identification of gated speech stimuli. The speech material consisted of /getVte/ utterances with the target vowels /i/, /y/, and /u/. In the case of the apraxic patient the gated vowels were identified at a later instant relative to the preceding plosion burst than in the normal speakers. This result was interpreted as reflecting a delayed onset of coarticulatory gestures, in particular lip rounding. The identification rates for the dysarthric's vowels rather reflected the general reduction of this patient's vowel space.  相似文献   

11.
Speech errors follow the phonotactics of the language being spoken. For example, in English, if [n] is mispronounced as [n], the [n] will always appear in a syllable coda. The authors created an analogue to this phenomenon by having participants recite lists of consonant-vowel-consonant syllables in 4 sessions on different days. In the first 2 experiments, some consonants were always onsets, some were always codas, and some could be both. In a third experiment, the set of possible onsets and codas depended on vowel identity. In all 3 studies, the production errors that occurred respected the "phonotactics" of the experiment. The results illustrate the implicit learning of the sequential constraints present in the stimuli and show that the language production system adapts to recent experience.  相似文献   

12.
On each trial, subjects were played a dichotic pair of syllables differing in the consonant (/ba/, /da/, /ga/) or in the vowel (/ba/, /b?/, /bi/). The pair of syllables was preceded by a melody, or a sentence, and followed by the same or a different melody, or sentence. Subjects either had to retain the first piece of additional material or were free to ignore it. The different combinations of phonemic contrast, additional material, and instruction concerning the additional material were used in different sessions. In each case, the main task of the subjects was to respond to the presence or the absence of the target /ba/ on the ear previously indicated. There was no effect of context on relative ear accuracy, but the right-ear advantage observed for consonants in response latency when subjects retained a sentence gave way to a small nonsignificant left-ear advantage when subjects retained a melody. Right-ear advantage in response latencies was also observed for vowels in the verbal context, but the contextual effect, although in the same direction as for consonants, was very slight. The implications of contextual effects for a theory of the determinants of the auditory laterality effects are discussed.  相似文献   

13.
Blind people can learn to understand speech at ultra-high syllable rates (ca. 20 syllables/s), a capability associated with hemodynamic activation of the central-visual system. To further elucidate the neural mechanisms underlying this skill, magnetoencephalographic (MEG) measurements during listening to sentence utterances were cross-correlated with time courses derived from the speech signal (envelope, syllable onsets and pitch periodicity) to capture phase-locked MEG components (14 blind, 12 sighted subjects; speech rate = 8 or 16 syllables/s, pre-defined source regions: auditory and visual cortex, inferior frontal gyrus). Blind individuals showed stronger phase locking in auditory cortex than sighted controls, and right-hemisphere visual cortex activity correlated with syllable onsets in case of ultra-fast speech. Furthermore, inferior-frontal MEG components time-locked to pitch periodicity displayed opposite lateralization effects in sighted (towards right hemisphere) and blind subjects (left). Thus, ultra-fast speech comprehension in blind individuals appears associated with changes in early signal-related processing mechanisms both within and outside the central-auditory terrain.  相似文献   

14.
Subjects monitored for the syllable-initial phonemes /b/ and /s/, as well as for the syllables containing those phonemes, in lists of nonsense syllables. Time to detect /b/ was a function of the amount of uncertainty as to the identity of the vowel following the target consonant; when uncertainty was low, no difference existed between phoneme and syllable monitoring latencies, but when uncertainty was high, syllables were detected faster than phonemes. Time to detect /s/ was independent of uncertainty concerning the accompanying vowel and was always slower than syllable detection. The role of knowledge of contexts in a phoneme-monitoring task as well as the relative availability of phonemic information to the listener in this task are discussed.  相似文献   

15.
Two experiments investigating the selective adaptation of vowels examined changes in listeners’ identification functions for the vowel continuum [i-I-∈] as a function of the adapting stimulus. In Experiment I, the adapting stimuli were [i], [I], and [∈]. Both the [i] and [∈] stimuli produced significant shifts in the neighboringand distant phonetic boundaries, whereas [I] did not result in any adaptation effects. In order to explore the phonetic nature of feature adaptation in vowels, a second experiment was conducted using the adapting stimuli [gig] and [g ∈ g], which differed acoustically from the [i] and [∈] vowels on the identification continuum. Only [gig] yielded reliable adaptation effects. The results of these experiments were interpreted as suggesting arelative rather than a stableauditory mode of feature analysis in vowels and a possibly more complex auditory feature analysis for the vowel [i].  相似文献   

16.
This study explored the extent to which rapid temporal processing and duration contribute to the right-ear advantage (REA) and presumably left-hemisphere processing for stop consonants and the lack of clear-cut laterality effects for vowels. Three sets of synthetic stimuli were constructed: consonant vowel stimuli [ba da ga bi di gi bu du gu] of 300 msec duration (full stimuli) and two shortened stimuli consisting either of a noise burst and 40-msec transitions (40-msec stimuli), or a noise burst and 20-msec transitions (20-msec stimuli). Stimuli were presented dichotically for consonant, vowel, and syllable identification. Results indicated a significant REA for consonants in the full and 40-msec conditions and a non-significant REA in the 20-msec condition. Nevertheless, the magnitude of laterality did not change across the three conditions. These results suggest that although transition information including duration contributes to lateralization for stop consonants, it is the presence of abrupt onsets which crucially determines lateralized processing. For vowels, there was a significant REA only in the full stimulus condition, and a significant decrement in the magnitude of the laterality effect in the two shortened stimulus conditions. These results suggest that for vowel perception, it is the nature of the acoustic cue used for phonetic identification and not duration that seems to be the critical determinant of lateralization effects.  相似文献   

17.
Onsets and rimes as units of spoken syllables: evidence from children   总被引:6,自引:0,他引:6  
The effects of syllable structure on the development of phonemic analysis and reading skills were examined in four experiments. The experiments were motivated by theories that syllables consist of an onset (initial consonant or cluster) and a rime (vowel and any following consonants). Experiment 1 provided behavioral support for the syllable structure model by showing that 8-year-olds more easily learned word games that treated onsets and rimes as units than games that did not. Further support for the cohesiveness of the onset came from Experiments 2 and 3, which found that 4- and 5-year-olds less easily recognized a spoken or printed consonant target when it was the first phoneme of a cluster than when it was a singleton. Experiment 4 extended these results to printed words by showing that consonant-consonant-vowel nonsense syllables were more difficult for beginning readers to decode than consonant-vowel-consonant syllables.  相似文献   

18.
The “McGurk effect” demonstrates that visual (lip-read) information is used during speech perception even when it is discrepant with auditory information. While this has been established as a robust effect in subjects from Western cultures, our own earlier results had suggested that Japanese subjects use visual information much less than American subjects do (Sekiyama & Tohkura, 1993). The present study examined whether Chinese subjects would also show a reduced McGurk effect due to their cultural similarities with the Japanese. The subjects were 14 native speakers of Chinese living in Japan. Stimuli consisted of 10 syllables (/ba/, /pa/, /ma/, /wa/, /da/, /ta/, /na/, /ga/, /ka/, /ra/ ) pronounced by two speakers, one Japanese and one American. Each auditory syllable was dubbed onto every visual syllable within one speaker, resulting in 100 audiovisual stimuli in each language. The subjects’ main task was to report what they thought they had heard while looking at and listening to the speaker while the stimuli were being uttered. Compared with previous results obtained with American subjects, the Chinese subjects showed a weaker McGurk effect. The results also showed that the magnitude of the McGurk effect depends on the length of time the Chinese subjects had lived in Japan. Factors that foster and alter the Chinese subjects’ reliance on auditory information are discussed.  相似文献   

19.
Geudens and Sandra, in their 2003 study, investigated the special role of onsets and rimes in Dutch-speaking children's explicit phonological awareness. In the current study, we tapped implicit phonological knowledge using forced-choice similarity judgment (Experiment 1) and recall of syllable lists (Experiment 2). In Experiment 1, Dutch-speaking prereaders judged rime-sharing pseudowords (/fas/-/mas/) to sound more similar than pseudowords sharing an equally sized nonrime unit (/fas/-/fak/). However, in a syllable recall task (/tepsilonf/, /ris/, /nal/), Dutch-speaking prereaders were as likely to produce recombination errors that broke up the rime (/tepsilons/) as to produce errors that retained the rime (/repsilonf/). Thus, a rime effect was obtained in a task that highlighted the phonological similarity between items sharing their rimes, but this effect disappeared in tasks without repetition of rime units. We conclude that children's sensitivity to rimes depends on similarity relations and might not reflect a fixed perceived structure of spoken syllables.  相似文献   

20.
赵荣  王小娟  杨剑峰 《心理学报》2016,48(8):915-923
探讨超音段(如声调)与音段信息的共同作用机制, 对口语词汇识别研究具有重要的理论意义。有研究探讨了声调在口语词汇语义通达阶段的作用, 但在相对早期的音节感知阶段, 声调与声母、韵母的共同作用机制还缺乏系统的认识。本研究采用oddball实验范式, 通过两个行为实验考察了声调在汉语音节感知中的作用。实验1发现检测声调和声母变化的时间没有差异, 均比检测韵母变化的时间长, 表明在汉语音节感知中对声调的敏感性不及韵母。实验2发现声母和韵母的组合并没有明显优于对韵母的觉察, 但声调与声母或韵母的同时变化都促进了被试对偏差刺激的觉察, 表明声调通过与声母、韵母的结合来共同影响汉语音节的感知加工。研究结果在认知行为层面为声调在音节感知中的作用机制提供了直接的实验证据, 为进一步探讨超音段与音段信息共同作用的认知神经机制提供了基础。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号