首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Auditory evoked responses (AER) to series of consonant—vowel syllables were recorded from temporal and parietal scalp locations from 20 right-handed female college students. Averaged AERs were submitted to principal components analysis and analysis of variance. Seven components of the group's AERs were found to reflect various aspects of the stimulus parameters. One component reflected changes over only the left hemisphere to different consonants independent of the following vowel sound. A second component changed systematically over both hemispheres in response to only consonant changes. A third component systematically changed for the different consonants depending on the following vowel.  相似文献   

2.
In 10 right-handed Ss, auditory evoked responses (AERs) were recorded from left and right temporal and parietal scalp regions during simple discrimination responses to binaurally presented pairs of synthetic speech sounds ranging perceptually from /ba/ to /da/. A late positive component (P3) in the AER was found to reflect the categorical or phonetic analysis of the stop consonants, with only left scalp sites averaging significantly different responses between acoustic and phonetic comparisons. The result is interpreted as evidence of hemispheric differences in the processing of speech in respect of the level of processing accessed by the particular information processing task.  相似文献   

3.
采用启动范式,在三个实验中通过操纵启动音和目标音的频谱相似度和时间间隔,考察了汉语听者元音范畴知觉中特征分析和整合的时间进程。结果发现随着启动音(从纯音、复合音到目标元音本身)和目标元音频谱相似度的增加,启动效应延续的时间越来越长。实验结果支持语音范畴知觉存在早期的声学特征分析和整合到后期的范畴知觉阶段,并为这些加工阶段的时间进程提供了初步的证据。  相似文献   

4.
Working memory uses central sound representations as an informational basis. The central sound representation is the temporally and feature-integrated mental representation that corresponds to phenomenal perception. It is used in (higher-order) mental operations and stored in long-term memory. In the bottom-up processing path, the central sound representation can be probed at the level of auditory sensory memory with the mismatch negativity (MMN) of the event-related potential. The present paper reviews a newly developed MMN paradigm to tap into the processing of speech sound representations. Preattentive vowel categorization based on F1-F2 formant information occurs in speech sounds and complex tones even under conditions of high variability of the auditory input. However, an additional experiment demonstrated the limits of the preattentive categorization of language-relevant information. It tested whether the system categorizes complex tones containing the F1 and F2 formant components of the vowel /a/ differently than six sounds with nonlanguage-like F1-F2 combinations. From the absence of an MMN in this experiment, it is concluded that no adequate vowel representation was constructed. This shows limitations of the capability of preattentive vowel categorization.  相似文献   

5.
Auditory evoked responses (AER) were recorded from frontal, temporal, and parietal scalp regions to a series of consonant-vowel syllables which varied in the duration of the consonant transition. Multivariate analyses of the AER waveforms identified one component of the AERs occurring only over right hemisphere regions which discriminated between differences in transition durations. A second component detected over only left hemisphere areas discriminated differences in place of articulation. These data are consistent with previous behavioral and electrophysiological reports that the right hemisphere is sensitive to temporal discriminations.  相似文献   

6.
A group of five anterior and seven posterior aphasic patients were recorded for their vowel productions of the nine nondipthong vowels of American English and compared to a group of seven normal speakers. All phonemic substitutions were eliminated from the data base. A Linear Predictive Coding (LPC) computer program was used to extract the first and the second formant frequencies at the midpoint of the vowel for each of the remaining repetitions of the nine vowels. The vowel duration and the fundamental frequency of phonation were also measured. Although there were no significant differences in the formant frequency means across groups, there were significantly larger standard deviations for the aphasic groups compared to normals. Anterior aphasics were not significantly different from posterior aphasics with respect to this greater formant variability. There was a main effect for vowel duration means, but no individual group was significantly different from the other. Standard deviations of duration were significantly greater for the anterior aphasics compared to normal speakers, but not significantly different from posterior aphasics. Posterior aphasics did not have significantly greater standard deviations of duration than did normal subjects. Greater acoustic variability was considered to evidence a phonetic production deficit on the part of both groups of aphasic speakers, in the context of fairly well-preserved phonemic organization for vowels.  相似文献   

7.
刘文理  祁志强 《心理科学》2016,39(2):291-298
采用启动范式,在两个实验中分别考察了辅音范畴和元音范畴知觉中的启动效应。启动音是纯音和目标范畴本身,目标音是辅音范畴和元音范畴连续体。结果发现辅音范畴连续体知觉的范畴反应百分比受到纯音和言语启动音影响,辅音范畴知觉的反应时只受言语启动音影响;元音范畴连续体知觉的范畴反应百分比不受两种启动音影响,但元音范畴知觉的反应时受到言语启动音影响。实验结果表明辅音范畴和元音范畴知觉中的启动效应存在差异,这为辅音和元音范畴内在加工机制的差异提供了新证据。  相似文献   

8.
To learn to produce speech, infants must effectively monitor and assess their own speech output. Yet very little is known about how infants perceive speech produced by an infant, which has higher voice pitch and formant frequencies compared to adult or child speech. Here, we tested whether pre‐babbling infants (at 4–6 months) prefer listening to vowel sounds with infant vocal properties over vowel sounds with adult vocal properties. A listening preference favoring infant vowels may derive from their higher voice pitch, which has been shown to attract infant attention in infant‐directed speech (IDS). In addition, infants' nascent articulatory abilities may induce a bias favoring infant speech given that 4‐ to 6‐month‐olds are beginning to produce vowel sounds. We created infant and adult /i/ (‘ee’) vowels using a production‐based synthesizer that simulates the act of speaking in talkers at different ages and then tested infants across four experiments using a sequential preferential listening task. The findings provide the first evidence that infants preferentially attend to vowel sounds with infant voice pitch and/or formants over vowel sounds with no infant‐like vocal properties, supporting the view that infants' production abilities influence how they process infant speech. The findings with respect to voice pitch also reveal parallels between IDS and infant speech, raising new questions about the role of this speech register in infant development. Research exploring the underpinnings and impact of this perceptual bias can expand our understanding of infant language development.  相似文献   

9.
These studies examined the perceptual role of various components of naturally produced stop consonants (/b, d, g, p, t, k/) in CV syllables. In the first experiment, the context-sensitive voiced formant transitions were removed with a computer-splicing technique. Identification accuracy was 84% when the consonant was presented with the same vowel as had been used to produce it. Performance fell to 66% when the consonant was juxtaposed with a different vowel. The second experiment not only deleted the voiced formant transition, but also replaced the aspiration with silence. Here, identification accuracy dropped substantially, especially for voiceless stops, which had contained devoiced formant transitions in the replaced interval. The pattern of errors suggested that listeners try to extract the missing locus of the consonant from the vowel transition, and in the absence of a vowel transition, they try to extrapolate it from the second formant of the steady-state vowel.  相似文献   

10.
According to the formant centre of gravity (FCOG) hypothesis, two vowel formants in close proximity are merged during perceptual analysis, and their contribution to vowel quality depends on the centre of gravity of the formant cluster. Findings consistent with this hypothesis are that two formants can be replaced by a single formant of intermediate centre frequency, provided their separation is less than 3-3.5 Bark; and that changes in their relative amplitudes produce systematic shifts in vowel quality. In Experiment 1, listeners adjusted the frequencies of F1 and F2 in a synthesized 6-formant vowel (with the F1-F2 separation fixed at 250 Hz, i.e. less than 3 Bark) to find the best phonetic match to a reference vowel with modified formant amplitudes. Contrary to FCOG predictions, F2 attenuation did not produce lower frequency matches. Raising the amplitude of F2 led to predicted upward shifts in formant frequencies of the matched vowel, but with increased variability of matches for some stimuli. In Experiment 2, listeners identified synthesized vowels with a range of separations of F1 and F2. Formant amplitude manipulations had no effect on listeners' judgements when the fundamental frequency was low (125 Hz). Small shifts in vowel quality appeared for stimuli with a high fundamental (250 Hz), but the shifts were significantly larger for F1-F2 separations greater than 3.5 Bark. These effects of formant amplitude are qualitatively different from those observed with single-formant vowels and are generally incompatible with a formant-averaging mechanism.  相似文献   

11.
This study examines the acoustic characteristics of individual vowels and those produced in sequences of three or more during vocal play by full term and preterm at age 6 months. Laboratory and home audiotape recordings of infant vowel sounds were made and digitized for acoustic analysis. Of interest was whether, during production of vowel sequences compared to those produced singly, infants explored the relation between tongue height and tongue advancement, as measured acoustically by the first formant (F1) and second formant (F2) frequencies, respectively. In both groups of infants, the region of F1-F2 space for individually produced vowels was significantly greater than for vowels produced in sequence. High correlation coefficients for F1 and F2 during exploration of vowels produced in sequence were apparent in full term, but not preterm infants. The data support the claim that vowels produced in sequence occupy a more limited region of the vocal tract than those produced singly, and that infants explore the characteristics of their vocalizations within sequences.  相似文献   

12.
Memory for speech sounds is a key component of models of verbal working memory (WM). But how good is verbal WM? Most investigations assess this using binary report measures to derive a fixed number of items that can be stored. However, recent findings in visual WM have challenged such “quantized” views by employing measures of recall precision with an analogue response scale. WM for speech sounds might rely on both continuous and categorical storage mechanisms. Using a novel speech matching paradigm, we measured WM recall precision for phonemes. Vowel qualities were sampled from a formant space continuum. A probe vowel had to be adjusted to match the vowel quality of a target on a continuous, analogue response scale. Crucially, this provided an index of the variability of a memory representation around its true value and thus allowed us to estimate how memories were distorted from the original sounds. Memory load affected the quality of speech sound recall in two ways. First, there was a gradual decline in recall precision with increasing number of items, consistent with the view that WM representations of speech sounds become noisier with an increase in the number of items held in memory, just as for vision. Based on multidimensional scaling (MDS), the level of noise appeared to be reflected in distortions of the formant space. Second, as memory load increased, there was evidence of greater clustering of participants' responses around particular vowels. A mixture model captured both continuous and categorical responses, demonstrating a shift from continuous to categorical memory with increasing WM load. This suggests that direct acoustic storage can be used for single items, but when more items must be stored, categorical representations must be used.  相似文献   

13.
Discrimination of natural, sustained vowels was studied in 5 budgerigars. The birds were trained using operant conditioning procedures on a same-different task, which was structured so that response latencies would provide a measure of stimulus similarity. These response latencies were used to construct similarity matrices, which were then analyzed by multidimensional scaling (MDS) procedures. MDS produced spatial maps of these speech sounds where perceptual similarity was represented by spatial proximity. The results of the three experiments suggest that budgerigars perceive natural, spoken vowels according to phonetic categories, find the acoustic differences among different talkers less salient than the acoustic differences among vowel categories, and use formant frequencies in making these complex discriminations.  相似文献   

14.
Exaggeration of the vowel space in infant-directed speech (IDS) is well documented for English, but not consistently replicated in other languages or for other speech-sound contrasts. A second attested, but less discussed, pattern of change in IDS is an overall rise of the formant frequencies, which may reflect an affective speaking style. The present study investigates longitudinally how Dutch mothers change their corner vowels, voiceless fricatives, and pitch when speaking to their infant at 11 and 15 months of age. In comparison to adult-directed speech (ADS), Dutch IDS has a smaller vowel space, higher second and third formant frequencies in the vowels, and a higher spectral frequency in the fricatives. The formants of the vowels and spectral frequency of the fricatives are raised more strongly for infants at 11 than at 15 months, while the pitch is more extreme in IDS to 15-month olds. These results show that enhanced positive affect is the main factor influencing Dutch mothers’ realisation of speech sounds in IDS, especially to younger infants. This study provides evidence that mothers’ expression of emotion in IDS can influence the realisation of speech sounds, and that the loss or gain of speech clarity may be secondary effects of affect.  相似文献   

15.
Do vowel-contingent selective adaptation effects for place of articulation depend on vowel identity, or on the particular formant frequencies used? An experiment is reported here which tested the adaptation effects of consonants with exactly the same formant transitions before different diphthongs. In this experiment, the phonetic identity of the vowel and the formant frequencies of the consonant are not confounded as they have been in previous studies. In the contingent adaptation condition, no phoneme boundary shifts were observed, and this is interpreted along with previous evidencefor such phoneme boundary shifts when phonetic identity of the vowel and formant frequencies are confounded as indicating that adaptation operates on the spectral representation of the stimulus. Other evidence consistent with this conclusion is that adaptation with alternating adaptors from each end of the test series produced negligible shifts, and that a single adaptor from a diphthong series different from that of the test series produced phoneme boundary shifts as large as those with a single adaptor from the same test series.  相似文献   

16.
Auditory evoked responses (AER) were recorded from scalp locations over the left and right temporal regions in response to CVC words and nonsense syllables. Various components of the AER were found to vary systematically with changes in stimulus meaning. One such component reflected subcortical involvement in semantic processing. Other components reflected changes in voicing and place of articulation as well as hemisphere differences.  相似文献   

17.
Subjects (average age 21 years, recruited by personal contact and through a school) were presented with a spoken sentence on tape and then heard six speakers of the same sex, including the original speaker, say the same sentence. They were required to indicate which was the original speaker. The task was repeated with seven different sentences and sets of speakers. One group of subjects heard short sentences containing an average of 2.14 different vowel sounds and 6.28 syllables, another group heard short sentences containing an average of 6.14 vowel sounds (7.28 syllables) and a third group heard longer sentences containing an average of 6.28 vowel sounds (11.00 syllables). Accuracy of speaker identification improved significantly when more vowel sounds were heard, but increased sentence length had no significant effect on performance. Performance was significantly better when the listener was the same sex as the speaker than when the listener was of the other sex.  相似文献   

18.
Two experiments were performed employing acoustic continua which change from speech to nonspeech. The members of one continuum, synthesized on the Pattern Playback, varied in the bandwidths of the first three formants in equal steps of change, from the vowel /α/ to a nonspeech buzz. The other continuum, achieved through digital synthesis, varied in the bandwidths of the first five formants, from the vowel /æ/ to a buzz. Identification and discrimination tests were carried out to establish that these continua were perceived categorically. Perceptual adaptation of these continua revealed shifts in the category boundaries comparable to those previously reported for speech sounds. The results were interpreted as suggesting that neither phonetic nor auditory feature detectors are responsible for perceptual adaptation of speech sounds, and that feature detector accounts of speech perception should therefore be reconsidered.  相似文献   

19.
Adults and infants were tested for the capacity to detect correspondences between nonspeech sounds and real vowels. The /i/ and /a/ vowels were presented in 3 different ways: auditory speech, silent visual faces articulating the vowels, or mentally imagined vowels. The nonspeech sounds were either pure tones or 3-tone complexes that isolated a single feature of the vowel without allowing the vowel to be identified. Adults perceived an orderly relation between the nonspeech sounds and vowels. They matched high-pitched nonspeech sounds to /i/ vowels and low-pitched nonspeech sounds to /a/ vowels. In contrast, infants could not match nonspeech sounds to the visually presented vowels. Infants' detection of correspondence between auditory and visual speech appears to require the whole speech signal; with development, an isolated feature of the vowel is sufficient for detection of the cross-modal correspondence.  相似文献   

20.
Traditionally, it has been held that the primary information for vowel identification is provided by formant frequencies in the quasi-steady-state portion of the spoken syllable. Recent search has advanced an alternative view that emphasizes the role of temporal factors and dynamic (time-varying) spectral information in determining the perception of vowels. Nine vowels spoken in /b/ + vowel + /b/ syllables were recorded. The syllables were modified electronically in several ways to suppress various sources of spectral and durational information. Two vowel-perception experiments were performed, testing subjects’ ability to identify vowels in these modified syllables. Results of both experiments revealed the importance of dynamic spectral information at syllable onset and offset (in its proper temporal relation) in permitting vowel identification. On the other hand, steady-state spectral information, deprived of its durational variation, was a poor basis for identification. Results constitute a challenge to traditional accounts of vowel perception and point toward important sources of dynamic information.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号