首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 0 毫秒
1.
Research has shown that speaking rate provides an important context for the perception of certain acoustic properties of speech. For example, syllable duration, which varies as a function of speaking rate, has been shown to influence the perception of voice onset time (VOT) for syllableinitial stop consonants. The purpose of the present experiments was to examine the influence of syllable duration when the initial portion of the syllable was produced by one talker and the remainder of the syllable was produced by a different talker. A short-duration and a long-duration /bi/-/pi/ continuum were synthesized with pitch and formant values appropriate to a female talker. When presented to listeners for identification, these stimuli demonstrated the typical effect of syllable duration on the voicing boundary: a shorter VOT boundary for the short stimuli than for the long stimuli. An /i/ vowel, synthesized with pitch and formant values appropriate to a male talker, was added to the end of each of the short tokens, producing a new hybrid continuum. Although the overall syllable duration of the hybrid stimuli equaled the original long stimuli, they produced a VOT boundary similar to that for the short stimuli. In a second experiment, two new /i/ vowels were synthesized. One had a pitch appropriate to a female talker with formant values appropriate to a male talker; the other had a pitch appropriate to a male talker and formants appropriate to a female talker. These vowels were used to create two new hybrid continua. In a third experiment, new hybrid continua were created by using more extreme male formant values. The results of both experiments demonstrated that the hybrid tokens with a change in pitch acted like the short stimuli, whereas the tokens with a change in formants acted like the long stimuli. A fourth experiment demonstrated that listeners could hear a change in talker with both sets of hybrid tokens. These results indicate that continuity of pitch but not formant structure appears to be the critical factor in the calculation of speaking rate within a syllable.  相似文献   

2.
3.
Results of auditory speech experiments show that reaction times (RTs) for place classification in a test condition in which stimuli vary along the dimensions of both place and voicing are longer than RTs in a control condition in which stimuli vary only in place. Similar results are obtained when subjects are asked to classify the stimuli along the voicing dimension. By taking advantage of the "McGurk" effect (McGurk & MacDonald, 1976), the present study investigated whether a similar pattern of interference extends to situations in which variation along the place dimension occurs in the visual modality. The results showed that RTs for classifying phonetic features in the test condition were significantly longer than in the control condition for the place and voicing dimensions. These results indicate a mutual and symmetric interference exists in the classification of the two dimensions, even when the variation along the dimensions occurs in separate modalities.  相似文献   

4.
Effects of speaking rate and lexical status on phonetic perception   总被引:1,自引:0,他引:1  
Among the contextual factors known to play a role in segmental perception are the rate at which the speech was produced and the lexical status of the item, that is, whether it is a meaningful word of the language. In a series of experiments on the word-initial /b/-/p/ voicing distinction, we investigated the conditions under which these factors operate during speech processing. The results indicated that under instructions of speeded responding, listeners could, on some trials, ignore some later occurring contextual information within the word that specified rate and lexical status. Importantly, however, they could not ignore speaking rate entirely. Although they could base their decision on only the early portion of the word, when doing so they treated the word as if it were physically short--that is to say, as if there were no later occurring information specifying a slower rate. This suggests that listeners always take account of rate when identifying the voicing value of a consonant, but precisely which information within the word is used to specify rate can vary with task demands.  相似文献   

5.
6.
In two studies we investigated the way in which the components of speaking rate, articulation rate and pause rate, combine to influence processing of the silence-duration cue for the voicing distinction in medial stop consonants. First, we replicated the finding that the articulation rate of a carrier sentence, that is, the rate at which the speech itself is produced, influences how the duration information is used to assign voicing values. Second, and more importantly, the assignment of voicing values was also influenced by the pause rate of the sentence. Thus, the listener adjusts for both articulation rate and pause rate when processing the phonetically relevant information. Finally, the two rate components did not function in an equivalent manner, since changes in articulation rate had considerably more effect on phonetic judgments than did changes in pause rate. Alternative explanations fo the relative weighting of the two variables are discussed.  相似文献   

7.
Understanding the circumstances under which talker (and other types of) variability affects language perception represents an important area of research in the field of spoken word recognition. Previous work has demonstrated that talker effects are more likely when processing is relatively slow (McLennan & Luce, 2005). Given that listeners may take longer to process foreign-accented speech than native-accented speech (Munro & Derwing, Language and Speech, 38, 289?C306 1995), talker effects should be more likely when listeners are presented with words spoken in a foreign accent than when they are presented with those same words spoken in a native accent. The results of two experiments, conducted in two different countries and in two different languages, are consistent with this prediction.  相似文献   

8.
9.
When third-formant transitions are appropriately incorporated into an acoustic syllable, they provide critical support for the phonetic percepts we call [d] and [g], but when presented in isolation they are perceived as time-varying ‘chirps’. In the present experiment, both modes of perception were made available simultaneously by presenting the third-format transitions to one ear and the remainder of the acoustic syllable to the other. On the speech side of this duplex percept, where the transitions supported the perception of stop-vowel syllables, perception was categorical and influenced by the presence of a preposed [al] or [ar]. On the nonspeech side, where the same transitions were heard as ‘chirps’, perception was continuous and free of influence from the preposed syllables. As both differences occurred under conditions in which the acoustic input was constant, we should suppose that they reflect the different properties of auditory and phonetic modes of perception.  相似文献   

10.
Use of the selective adaptation procedure with speech stimuli has led to a number of theoretical positions with regard to the level or levels of processing affected by adaptation. Recent experiments (i.e., Sawusch & Jusczyk, 1981) have, however, yielded strong evidence that only auditory coding processes are affected by selective adaptation. In the present experiment, a test series that varied along the phonetic dimension of place of articulation for stops ([da]-[ga]) was used in conjunction with a [ska] syllable that shared the phonetic value of velar with the [ga] end of the test series but had a spectral structure that closely matched a stimulus from the [da] end of the series. As an adaptor, the [ska] and Ida] stimuli produced identical effects, whereas in a paired-comparison procedure, the [ska] produced effects consistent with its phonetic label. These results offer further support for the contention that selective adaptation affects only the auditory coding of speech, whereas the paired-comparison procedure affects only the phonetic coding of speech. On the basis of these results and previous place-adaptation results, a process model of speech perception is described.  相似文献   

11.
An experiment was conducted to study the dimensions of encoding of verbal material in short-term memory as a function of other demands placed on Ss attention. Eighty Ss shadowed one of two simultaneous lists, under instructions to remember the shadowed or the nonshadowed list, A single recognition probe followed each shadowing trial. Ss judged whether the probed item was (1) identical to, (2) a rhyme of, or (3) a synonym of one of the to be remembered items. Results indicated that acoustic information was encoded from all inputs, regardless of the focus of attention. Evidence for semantic encoding, however, was limited to those conditions in which retention of shadowed material was required. The data were interpreted as contradicting “late selection” theories of attention, which propose that all inputs are analyzed for meaning prior to the focusing of attention.  相似文献   

12.
Categorical effects are found across speech sound categories, with the degree of these effects ranging from extremely strong categorical perception in consonants to nearly continuous perception in vowels. We show that both strong and weak categorical effects can be captured by a unified model. We treat speech perception as a statistical inference problem, assuming that listeners use their knowledge of categories as well as the acoustics of the signal to infer the intended productions of the speaker. Simulations show that the model provides close fits to empirical data, unifying past findings of categorical effects in consonants and vowels and capturing differences in the degree of categorical effects through a single parameter.  相似文献   

13.
In each of four experiments, complex visual stimuli--pictures and digit arrays--were remembered better when shown at high luminance than when shown at low luminance. Why does this occur? Two possibilities were considered: first that lowering luminance reduces the amount of available information in the stimulus, and second that lowering luminance reduces the rate at which the information is extracted from the stimulus. Evidence was found for both possibilities. When stimuli were presented at durations short enough to permit only a single eye fixation, luminance affected only the rate at which information is extracted: decreasing luminance by a factor of 100 caused information to be extracted more slowly by a factor that ranged, over experiments, from 1.4 to 2.0. When pictures were presented at durations long enough to permit multiple fixations, however, luminance affected the total amount of extractable information. In a fifth experiment, converging evidence was sought for the proposition that within the first eye fixation on a picture, luminance affects the rate of information extraction. If this proposition is correct and, in addition, the first eye fixation lasts until some criterion amount of information is extracted, then fixation duration should increase with decreasing luminance. This prediction was confirmed.  相似文献   

14.
A complete understanding of visual phonetic perception (lipreading) requires linking perceptual effects to physical stimulus properties. However, the talking face is a highly complex stimulus, affording innumerable possible physical measurements. In the search for isomorphism between stimulus properties and phoneticeffects, second-order isomorphism was examined between theperceptual similarities of video-recorded perceptually identified speech syllables and the physical similarities among the stimuli. Four talkers produced the stimulus syllables comprising 23 initial consonants followed by one of three vowels. Six normal-hearing participants identified the syllables in a visual-only condition. Perceptual stimulus dissimilarity was quantified using the Euclidean distances between stimuli in perceptual spaces obtained via multidimensional scaling. Physical stimulus dissimilarity was quantified using face points recorded in three dimensions by an optical motion capture system. The variance accounted for in the relationship between the perceptual and the physical dissimilarities was evaluated using both the raw dissimilarities and the weighted dissimilarities. With weighting and the full set of 3-D optical data, the variance accounted for ranged between 46% and 66% across talkers and between 49% and 64% across vowels. The robust second-order relationship between the sparse 3-D point representation of visible speech and the perceptual effects suggests that the 3-D point representation is a viable basis for controlled studies of first-order relationships between visual phonetic perception and physical stimulus attributes.  相似文献   

15.
During the first year of life, infants show decreased sensitivity to phonetic differences not used in their native language and increased sensitivity to the differences that are used. It has been shown that this change in speech perception is a function of the distributional properties of the input. The present study explores whether the mechanism responsible for the developmental changes regarding the organization of phonetic categories is a general mechanism shared with other animals. The results demonstrate that the distributional exposure to a phonetic continuum affects the subsequent discrimination of these phonemes in rats, indicating that the ability to use distributional cues to change the phonetic category structure extends beyond humans.  相似文献   

16.
Letter perception has been traditionally viewed as a process in which individual features are accumulated over time. In order to test this notion, a special stimulus set was created having little or no featural redundancy. Using a masking paradigm, confusion matrices were generated at each of eight interstimulus intervals. Few, if any, of the predictions made by the feature accumulation models were upheld. Instead, it is suggested that letter perception is better thought of as a global-to-local process. When a letter is presented, an observer initially perceives a large array of perceptual data. Over time, a clearer view of the stimulus emerges as the perceptual system brings the letter into focus. Thus, global information about the letter is available quite early in processing, while the letter’s more local aspects become available only after relatively extensive perceptual processing.  相似文献   

17.
18.
Recent experiments using a variety of techniques have suggested that speech perception involves separate auditory and phonetic levels of processing. Two models of auditory and phonetic processing appear to be consistent with existing data: (a) a strictserial model in which auditory information would be processed at one level, followed by the processing of phonetic information at a subsequent level; and (b) aparallel model in which auditory and phonetic processing could proceed simultaneously. The present experiment attempted to distinguish empirically between these two models. Ss identified either an auditory dimension (fundamental frequency) or a phonetic dimension (place of articulation of the consonant) of synthetic consonant-vowel syllables. When the two dimensions varied in a completely correlated manner, reaction times were significantly shorter than when either dimension varied alone. This “redundancy gain” could not be attributed to speed-accuracy trades, selective serial processing, or differential transfer between conditions. These results allow rejection of a completely serial model, suggesting instead that at least some portion of auditory and phonetic processing can occur in parallel.  相似文献   

19.
6 subjects were each auditorily presented six lists of 7-digit numbers for retention intervals of 0, 5, and 10 sec. Pupil size was recorded during stimulus presentation, retention, interval, and recall of items. Results indicated that pupil dilation occurred during encoding and retrieval of stimulus items. Pupillary constriction was found during the retention interval when rehearsal was presumed to occur.  相似文献   

20.
The methods of magnitude production and estimation were used to scale the perception of signing rate by signers and observers. As in the case of voice level and speech rate, the autophonic scale of signing rate has a slope greater than unity and is steeper than the corresponding extraphonic scale; the obtained exponents of the two power functions for signing are, respectively, 2.66 and 1.56. When English-speaking subjects estimated their own rate of reading of the translated version of the signed passage, they produced an autophonic reading scale quite similar to that for signing (exponent of 2.51), but when they made magnitude estimations of English rates covering the same range of rates as the signed passage, the exponent of the extraphonic reading scale was significantly larger (1.89). This was also the case when French subjects estimated French reading rates. The difference between extraphonic signing and reading scales was confirmed by subjects who knew no Sign Language or French; their results appear to indicate, in addition, that the processes involved in extraphonic perception of rate are purely acoustic (speech) or visual (sign) and do not require, as one could have thought, deeper linguistic operations.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号