首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
A dichotic listening experiment was conducted to determine if vowel perception is based on phonetic feature extraction as is consonant perception. Twenty normal right-handed subjects were given dichotic CV syllables contrasting in final vowels. It was found that, unlike consonants, the perception of dichotic vowels was not significantly lateralized, that the dichotic perception of vowels was not significantly enhanced by the number of phonetic features shared, and that the occurrence of double-blend errors was not greater than chance. However, there was strong evidence for the use of phonetic features at the level of response organization. It is suggested that the differences between vowel and consonant perception reflect the differential availability of the underlying acoustic information from auditory store, rather than differences in processing mechanisms.  相似文献   

2.
This study explored a number of temporal (durational) parameters of consonant and vowel production in order to determine whether the speech production impairments of aphasics are the result of the same or different underlying mechanisms and in particular whether they implicate deficits that are primarily phonetic or phonological in nature. Detailed analyses of CT scan lesion data were also conducted to explore whether more specific neuroanatomical correlations could be made with speech production deficits. A series of acoustic analyses were conducted including voice-onset time, intrinsic and contrastive fricative duration, and intrinsic and contrastive vowel duration as produced by Broca's aphasics with anterior lesions (A patients), nonfluent aphasics with anterior and posterior lesions (AP patients), and fluent aphasics with posterior lesions (P patients). The constellation of impairments for the anterior aphasics including both the A and AP patients suggests that their disorder primarily reflects an inability to implement particular types of articulatory gestures or articulatory parameters rather than an inability to implement particular phonetic features. They display impairments in the implementation of laryngeal gestures for both consonant and vowel production. These patterns seem to relate to particular anatomical sites involving Broca's area, the anterior limb of the internal capsule, and the lowest motor cortex areas for larynx and tongue. The posterior patients also show evidence of subtle phonetic impairments suggesting that the neural instantiation of speech may require more extensive involvement, including the perisylvian area, than previously suggested.  相似文献   

3.
Coarticulatory acoustic variation is presumed to be caused by temporally overlapping linguistically significant gestures of the vocal tract. The complex acoustic consequences of such gestures can be hypothesized to specify them without recourse to context-sensitive representations of phonetic segments. When the consequences of separate gestures converge on a common acoustic dimension (e.g., fundamental frequency), perceptual parsing of the acoustic consequences of overlapping spoken gestures, rather than associations of acoustic features, is required to resolve the distinct gestural events. Direct tests of this theory were conducted. These tests revealed mutual influences of (1) fundamental frequency during a vowel on prior consonant perception, and (2) consonant identity on following vowel stress and pitch perception. The results of these converging tests lead to the conclusion that speech perception involves a process in which acoustic information for coarticulated gestures is parsed from the stream of speech.  相似文献   

4.
How do acoustic attributes of the speech signal contribute to feature-processing interactions that occur in phonetic classification? In a series of five experiments addressed to this question, listeners performed speeded classification tasks that explicitly required a phonetic decision for each response. Stimuli were natural consonant-vowel syllables differing by multiple phonetic features, although classification responses were based on a single target feature. In control tasks, no variations in nontarget features occurred, whereas in orthogonal tasks nonrelevant feature variations occurred but had to be ignored. Comparison of classification times demonstrated that feature information may either be processed separately as independent cues for each feature or as a single integral segment that jointly specifies several features. The observed form on processing depended on the acoustic manifestations of feature variation in the signal. Stop-consonant place of articulation and voicing cues, conveyed independently by the pattern and excitation source of the initial formant transitions, may be processed separately. However, information for consonant place of articulation and vowel quality, features that interactively affect the shape of initial formant transitions, are processed as an integral segment. Articulatory correlates of each type of processing are discussed in terms of the distinction between source features that vary discretely in speech production and resonance features that can change smoothly and continuously. Implications for perceptual models that include initial segmentation of an input utterance into a phonetic feature representation are also considered.  相似文献   

5.
We propose a psycholinguistic model of lexical processing which incorporates both process and representation. The view of lexical access and selection that we advocate claims that these processes are conducted with respect to abstract underspecified phonological representations of lexical form. The abstract form of a given item in the recognition lexicon is an integrated segmental-featural representation, where all predictable and non-distinctive information is withheld. This means that listeners do not have available to them, as they process the speech input, a representation of the surface phonetic realisation of a given word-form. What determines performance is the abstract, underspecified representation with respect to which this surface string is being interpreted. These claims were tested by studying the interpretation of the same phonological feature, vowel nasality, in two languages, English and Bengali. The underlying status of this feature differs in the two languages; nasality is distinctive only in consonants in English, while both vowels and consonants contrast in nasality in Bengali. Both languages have an assimilation process which spreads nasality from a nasal consonant to the preceding vowel. A cross-linguistic gating study was conducted to investigate whether listeners would interpret nasal and oral vowels differently in two languages. The results show that surface phonetic nasality in the vowel in VN sequences is used by English listeners to anticipate the upcoming nasal consonant. In Bengali, however, nasality is initially interpreted as an underlying nasal vowel. Bengali listeners respond to CVN stimuli with words containing a nasal vowel, until they get information about the nasal consonant. In contrast, oral vowels in both languages are unspecified for nasality and are interpreted accordingly. Listeners in both languages respond with CVN words (which have phonetic nasality on the surface) as well as with CVC words while hearing an oral vowel. The results of this cross-linguistic study support, in detail, the hypothesis that the listener's interpretation of the speech input is in terms of an abstract underspecified representation of lexical form.  相似文献   

6.
刘文理  祁志强 《心理科学》2016,39(2):291-298
采用启动范式,在两个实验中分别考察了辅音范畴和元音范畴知觉中的启动效应。启动音是纯音和目标范畴本身,目标音是辅音范畴和元音范畴连续体。结果发现辅音范畴连续体知觉的范畴反应百分比受到纯音和言语启动音影响,辅音范畴知觉的反应时只受言语启动音影响;元音范畴连续体知觉的范畴反应百分比不受两种启动音影响,但元音范畴知觉的反应时受到言语启动音影响。实验结果表明辅音范畴和元音范畴知觉中的启动效应存在差异,这为辅音和元音范畴内在加工机制的差异提供了新证据。  相似文献   

7.
This study investigated whether consonant phonetic features or consonant acoustic properties more appropriately describe perceptual confusions among speech stimuli in multitalker babble backgrounds. Ten normal-hearing subjects identified 19 consonants, each paired with /a/, 1–19 and lui in a CV format. The stimuli were presented in quiet and in three levels of babble. Multidimensional scaling analyses of the confusion data retrieved stimulus dimensions corresponding to consonant acoustic parameters. The acoustic dimensions identified were: periodicity/burst onset, friction duration, consonant-vowel ratio, second formant transition slope, and first formant transition onset. These findings are comparable to previous reports of acoustic effects observed in white-noise conditions, and support the theory that acoustic characteristics are the relevant perceptual properties of speech in noise conditions. Perceptual effects of vowel context and level of the babble also were observed. These condition effects contrast with those previously reported for white-noise interference, and are attributed to direct masking of the low-frequency acoustic cues in the nonsense syllables by the low-frequency spectrum of the babble.  相似文献   

8.
Speech perception can be viewed in terms of the listener’s integration of two sources of information: the acoustic features transduced by the auditory receptor system and the context of the linguistic message. The present research asked how these sources were evaluated and integrated in the identification of synthetic speech. A speech continuum between the glide-vowel syllables /ri/ and /li/ was generated by varying the onset frequency of the third formant. Each sound along the continuum was placed in a consonant-cluster vowel syllable after an initial consonant /p/, /t/, /s/, and /v/. In English, both /r/ and /l/ are phonologically admissible following /p/ but are not admissible following /v/. Only /l/ is admissible following /s/ and only /r/ is admissible following /t/. A third experiment used synthetic consonant-cluster vowel syllables in which the first consonant varied between /b/ and /d and the second consonant varied between /l/ and /r/. Identification of synthetic speech varying in both acoustic featural information and phonological context allowed quantitative tests of various models of how these two sources of information are evaluated and integrated in speech perception.  相似文献   

9.
Ramus F  Nespor M  Mehler J 《Cognition》1999,73(3):265-292
Spoken languages have been classified by linguists according to their rhythmic properties, and psycholinguists have relied on this classification to account for infants' capacity to discriminate languages. Although researchers have measured many speech signal properties, they have failed to identify reliable acoustic characteristics for language classes. This paper presents instrumental measurements based on a consonant/vowel segmentation for eight languages. The measurements suggest that intuitive rhythm types reflect specific phonological properties, which in turn are signaled by the acoustic/phonetic properties of speech. The data support the notion of rhythm classes and also allow the simulation of infant language discrimination, consistent with the hypothesis that newborns rely on a coarse segmentation of speech. A hypothesis is proposed regarding the role of rhythm perception in language acquisition.  相似文献   

10.
Ramus F  Nespor M  Mehler J 《Cognition》2000,75(1):AD3-AD30
Spoken languages have been classified by linguists according to their rhythmic properties, and psycholinguists have relied on this classification to account for infants' capacity to discriminate languages. Although researchers have measured many speech signal properties, they have failed to identify reliable acoustic characteristics for language classes. This paper presents instrumental measurements based on a consonant/vowel segmentation for eight languages. The measurements suggest that intuitive rhythm types reflect specific phonological properties, which in turn are signaled by the acoustic/phonetic properties of speech. The data support the notion of rhythm classes and also allow the simulation of infant language discrimination, consistent with the hypothesis that newborns rely on a coarse segmentation of speech. A hypothesis is proposed regarding the role of rhythm perception in language acquisition.  相似文献   

11.
Three experiments were designed to investigate how listeners to coarticulated speech use the acoustic speech signal during a vowel to extract information about a forthcoming oral or nasal consonant. A first experiment showed that listeners use evidence of nasalization in a vowel as information for a forthcoming nasal consonant. A second and third experiment attempted to distinguish two accounts of their ability to do so. According to one account, listeners hear nasalization in the vowel as such and use it to predict that a forthcoming nasal consonant is nasal. According to a second, they perceive speech gestures and hear nasalization in the acoustic domain of a vowel as the onset of a nasal consonant. Therefore, they parse nasal information from a vowel and hear the vowel as oral. In Experiment 2, evidence in favor of the parsing hypothesis was found. Experiment 3 showed, however, that parsing is incomplete.  相似文献   

12.
Although copious research has investigated the role of phonology in reading, little research has investigated the precise nature of the entailed speech representations. The present study examined the similarity of “inner speech” in reading to overt speech. Two lexical decision experiments (in which participants gave speeded word/nonword classifications to letter strings) assessed the effects of implicit variations in vowel and word-initial consonant length. Responses were generally slower for phonetically long stimuli than for phonetically short stimuli, despite equal orthographic lengths. Moreover, the phonetic length effects displayed principled interactions with common factors known to affect lexical decisions, such as word frequency and the similarity of words to nonwords. Both phonetic length effects were stronger among slower readers. The data suggest that acoustic representations activated in silent reading are best characterized as inner speech rather than as abstract phonological codes.  相似文献   

13.
Morton, Marcus, and Frankish (1976) defined “perceptual center,” or “P-center,” as a neutral term to describe that which is regular in a perceptually regular sequence of speech sounds. This paper describes a paradigm for the determination of P-center location and the effect of various acoustic parameters on empirically determined P-center locations. It is shown that P-center location is affected by both initial consonant duration and, secondarily, subsequent vowel and consonant duration. A simple two-parameter model involving the duration of the whole stimulus is developed and gives good performance in predicting P-center location. The application of this model to continuous speech is demonstrated. It is suggested that there is little value in attempting to determine any single acoustic or articulatory correlate of P-center location, or in attempting to define P-center location absolutely in time. Rather, these results indicate that P-centers are a property of the whole stimulus and reflect properties of both the production and perception of speech.  相似文献   

14.
《Ecological Psychology》2013,25(2):161-180
An analysis of the auditory enhancement theory of speech perception reveals its liabilities despite its plausibility. Four points are featured in a critique of its main premise: that the objects of speech perception are sounds, and, therefore, the perception of speech relies primarily on the categorization of auditory qualities. First, this critique notes that auditory enhancement theory is conventional, despite professed differences from the accounts which it disputes, in adopting consonant and vowel inventories as the objects of perception. Second, the critique argues that auditory enhancement theory fails three empirical tests of this equa- tion of phonetic attributes and auditory qualities. Third, the critique alleges that the auditory enhancement theory seeks support from equivocal experiments on animal training, in which the limited similarity of human and nonhuman test performance remains unexplored empirically and unexplained. Last, the critique concludes that the theory of speech perception by auditory enhancement relies on parsimony and distinctiveness as organizing principles for phonetic sound pat- terns, both of which are clearly inconsistent with central findings of linguistic and psychological research on speech perception and production.  相似文献   

15.

An experiment was performed to determine the effect of selective adaptation on the identification of synthetic speech sounds which varied along the phonetic dimensionplace of articulation. Adaptation with a stimulus of a particular place value led to a reduction in the number of test stimuli identified as having that place value. An identification shift was obtained even when the acoustic information specifying place value for the adapting stimulus had virtually nothing in common with the information specifying place value for any of the test stimuli. Removing the vowel portion of an adapting stimulus eliminated identification shift only when the resulting stimulus was no longer perceived as speech-like. The results indicate that at least part of the adaptation effect occurs at a site of phonetic, not merely acoustic, feature analysis.

  相似文献   

16.
Previous research has demonstrated that infants under 17 months have difficulty learning novel words in the laboratory when the words differ by only one consonant sound, irrespective of the magnitude of that difference. The current study explored whether 15‐month‐old infants can learn novel words that differ in only one vowel sound. The rich acoustic/phonetic properties of vowels allow for a detailed analysis of the contribution of acoustic/phonetic cues to infants’ performance with similar‐sounding words. Infants succeeded with the vowel pair /i/–/I/, but failed with vowel pairs /i/–/u/ and /I/–/u/. These results suggest that infants initially use the most salient acoustic cues for vowels and that this staged use of acoustic cues both predicts and explains why infants can learn some words that differ in only a single vowel.  相似文献   

17.
The stop consonants /b, d, g, p, t, k/were recorded before/i/,/a/,/u/. The energy spectrum for each stop consonant was removed from its original vowel and spliced onto a different steady-state vowel. Results of a recognition test revealed that consonants were accurately recognized in all cases except when /k/ or/g[ was spliced from/i/to/u/. Further demonstrations suggested that/k/ and /g/ do have invariant characteristics before/i/, /a/, and /u/. These results support the general notion that stop consonants may be recognized before different vowels in normal speech in terms of invariant acoustic features.  相似文献   

18.
Speech sounds can be classified on the basis of their underlying articulators or on the basis of the acoustic characteristics resulting from particular articulatory positions. Research in speech perception suggests that distinctive features are based on both articulatory and acoustic information. In recent years, neuroelectric and neuromagnetic investigations provided evidence for the brain's early sensitivity to distinctive features and their acoustic consequences, particularly for place of articulation distinctions. Here, we compare English consonants in a Mismatch Field design across two broad and distinct places of articulation - labial and coronal - and provide further evidence that early evoked auditory responses are sensitive to these features. We further add to the findings of asymmetric consonant processing, although we do not find support for coronal underspecification. Labial glides (Experiment 1) and fricatives (Experiment 2) elicited larger Mismatch responses than their coronal counterparts. Interestingly, their M100 dipoles differed along the anterior/posterior dimension in the auditory cortex that has previously been found to spatially reflect place of articulation differences. Our results are discussed with respect to acoustic and articulatory bases of featural speech sound classifications and with respect to a model that maps distinctive phonetic features onto long-term representations of speech sounds.  相似文献   

19.
Laboratoire de Sciences Cognitives et Psycholinguistique, Paris, France This study introduces a new paradigm for investigating lexical processing. First, an analysis of data from a series of word-spotting experiments is presented suggesting that listeners treat vowels as more mutable than consonants in auditory word recognition in English. In order to assess this hypothesis, a word reconstruction task was devised in which listeners were required to turn word-like nonwords into words by adapting the identity of either one vowel or one consonant. Listeners modified vowel identity more readily than consonant identity. Furthermore, incorrect responses more often involved a vowel change than a consonant change. These findings are compatible with the proposal that English listeners are equipped to deal with vowel variability by assuming that vowel identity is comparatively underdefined. The results are discussed in the light of theoretical accounts of speech processing.  相似文献   

20.
The influence of motor knowledge on speech perception is well established, but the functional role of the motor system is still poorly understood. The present study explores the hypothesis that speech production abilities may help infants discover phonetic categories in the speech stream, in spite of coarticulation effects. To this aim, we examined the influence of babbling abilities on consonant categorization in 6‐ and 9‐month‐old infants. Using an intersensory matching procedure, we investigated the infants’ capacity to associate auditory information about a consonant in various vowel contexts with visual information about the same consonant, and to map auditory and visual information onto a common phoneme representation. Moreover, a parental questionnaire evaluated the infants’ consonantal repertoire. In a first experiment using /b/–/d/ consonants, we found that infants who displayed babbling abilities and produced the /b/ and/or the /d/ consonants in repetitive sequences were able to correctly perform intersensory matching, while non‐babblers were not. In a second experiment using the /v/–/z/ pair, which is as visually contrasted as the /b/–/d/ pair but which is usually not produced at the tested ages, no significant matching was observed, for any group of infants, babbling or not. These results demonstrate, for the first time, that the emergence of babbling could play a role in the extraction of vowel‐independent representations for consonant place of articulation. They have important implications for speech perception theories, as they highlight the role of sensorimotor interactions in the development of phoneme representations during the first year of life.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号