首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Comparisons between infant-directed and adult-directed speech were conducted to determine whether word-final syllables are highlighted in infant-directed speech. Samples of adult-directed and infant-directed speech were collected from 8 mothers of 6-month-old and 8 mothers of 9- month-old infants. Mothers were asked to label seven objects both to an experimenter and to their infant. Duration, pitch, and amplitude were measured for whole words and for each of the target word syllables. As in prior research, the infant-directed targets were higher pitched and longer than adult-directed targets. The results also extend beyond previous results in showing that lengthening of final syllables in infant-directed speech is particularly exaggerated. Results of analyses comparing word-final versus nonfinal unstressed syllables in utterance-medial position in infant-directed speech showed that lengthening of unstressed word-final syllables occurs even in utterance-internal positions. These results could suggest a mechanism for proposals that word-final syllables are perceptually salient to young children.  相似文献   

2.
This study investigates whether or not the representation of lexical stress information can be primed during speech production. In four experiments, we attempted to prime the stress position of bisyllabic target nouns (picture names) having initial and final stress with auditory prime words having either the same or different stress as the target (e.g., WORtel-MOtor vs. koSTUUM-MOtor; capital letters indicate stressed syllables in prime-target pairs). Furthermore, half of the prime words were semantically related, the other half unrelated. Overall, picture names were not produced faster when the prime word had the same stress as the target than when the prime had different stress, i.e., there was no stress-priming effect in any experiment. This result would not be expected if stress were stored in the lexicon. However, targets with initial stress were responded to faster than final-stress targets. The reason for this effect was neither the quality of the pictures nor frequency of occurrence or voice-key characteristics. We hypothesize here that this stress effect is a genuine encoding effect, i.e., words with stress on the second syllable take longer to be encoded because their stress pattern is irregular with respect to the lexical distribution of bisyllabic stress patterns, even though it can be regular with respect to metrical stress rules in Dutch. The results of the experiments are discussed in the framework of models of phonological encoding.  相似文献   

3.
The use of rhythm in attending to speech   总被引:1,自引:0,他引:1  
Three experiments examined attentional allocation during speech processing to determine whether listeners capitalize on the rhythmic nature of speech and attend more closely to stressed than to unstressed syllables. Ss performed a phoneme monitoring task in which the target phoneme occurred on a syllable that was either predicted to be stressed or unstressed by the context preceding the target word. Stimuli were digitally edited to eliminate the local acoustic correlates of stress. A sentential context and a context composed of word lists, in which all the words had the same stress pattern, were used. In both cases, the results suggest that attention may be preferentially allocated to stressed syllables during speech processing. However, a normal sentence context may not provide strong predictive cues to lexical stress, limiting the use of the attentional focus.  相似文献   

4.
In four experiments, listeners’ response times to detect vowel targets in spoken input were measured. The first three experiments were conducted in English. In two, one using real words and the other, nonwords, detection accuracy was low, targets in initial syllables were detected more slowly than targets in final syllables, and both response time and missed-response rate were inversely correlated with vowel duration. In a third experiment, the speech context for some subjects included all English vowels, while for others, only five relatively distinct vowels occurred. This manipulation had essentially no effect, and the same response pattern was again observed. A fourth experiment, conducted in Spanish, replicated the results in the first three experiments, except that miss rate was here unrelated to vowel duration. We propose that listeners’ responses to vowel targets in naturally spoken input are effectively cautious, reflecting realistic appreciation of vowel variability in natural context.  相似文献   

5.
Previous research comparing detection times for syllables and for phonemes has consistently found that syllables are responded to faster than phonemes. This finding poses theoretical problems for strictly hierarchical models of speech recognition, in which smaller units should be able to be identified faster than larger units. However, inspection of the characteristics of previous experiments’ stimuli reveals that subjects have been able to respond to syllables on the basis of only a partial analysis of the stimulus. In the present experiment, five groups of subjects listened to identical stimulus material. Phoneme and syllable monitoring under standard conditions was compared with monitoring under conditions in which near matches of target and stimulus occurred on no-response trials. In the latter case, when subjects were forced to analyze each stimulus fully, phonemes were detected faster than syllables.  相似文献   

6.
Navon's (1977) global precedence hypothesis was based primarily on the joint occurrence of two effects: a response time (RT) advantage for identifying global targets, and interference by global distractors on responding to local targets. Although the hypothesis has been questioned on the basis of experiments in which it has been shown that a local RT advantage and local interference can occur, it is still frequently assumed that these two effects are a valid measure of the order in which local and global levels of structure are processed. In the present experiment, this assumption was examined. Subjects identified target letters that occurred randomly at the global or local level in a divided-attention task. The visual angle subtended by the stimulus pattern was varied, a manipulation known to affect the relative speed of response to local- or global-level information. Local targets were identified faster than global targets at the larger visual angles, but there was no difference in RT at the smallest visual angle. Despite this change in RT advantage, the interference effect did not change as a function of the visual angle of the stimulus pattern. Moreover, global distractors interfered with responding to local targets but local targets had no effect on responding to global targets, which is exactly the opposite of the finding one would expect if RT advantage and interference reflected order of processing. These findings are not consistent with the assumption that RT advantage and interference reflect order of processing in a simple way.  相似文献   

7.
The 'audio-phonatoric coupling' (APC) was investigated in two independent experiments. Slightly delayed auditory feedback (delay time 40 ms) of the subjects' own speech was used as experimental method. The first experiment was conducted to examine whether the strength of the APC depends on the speech rate. In this experiment 16 male Subjects (Ss) were required to utter the testword/tatatas/either with stress placing on the first or second syllable at two different speech rates (fast and slow). In 16% of the randomly chosen speech trials, the delayed auditory feedback (DAF; 40 ms delay) was introduced. It could be shown that the stressed phonation was significantly lengthened under the DAF condition. This lengthening was greater when Ss spoke slowly. The unstressed phonations were not influenced by the DAF condition. The second experiment was conducted to examine whether or not speech intensity effects APC. Nine male Ss were required to utter the testword/tatatas/either with stress placing on the first or second syllable using three different speech intensities (30 dB, 50 dB and 70 dB). In 16% of the randomly chosen speech trials DAF condition was introduced. It could be shown that speech intensity does not influence the DAF effect (lengthening of stressed phonation). These findings were taken as evidence that the auditory feedback of the subjects' own speech can be incorporated into speech control during ongoing speech. Obviously, this feedback information is efficient only during the production of stressed syllables, and varies as a function of speech rate. In addition, the significance of stressed syllables for the structuring of speech is discussed.  相似文献   

8.
Memory for targets moving in depth and for stationary targets was examined in five experiments. Memory for targets moving in depth was displaced behind the target with slower target velocities (longer ISIS and retention intervals) and beyond the target with faster target velocities (shorter ISIS and retention intervals), and the overall magnitude of forward displacement for motion in depth was less than the overall magnitude of forward displacement for motion in the picture plane. Memory for stationary targets was initially displaced away from the observer, but memory for smaller stationary targets was subsequently displaced toward the observer and memory for larger stationary targets was subsequently displaced away from the observer; memory for the top or bottom edge of a stationary target was displaced inside the target perimeter. The data are consistent with Freyd and Johnson's (1987) two-component model of the time course of representational momentum and with Intraub et al.'s (1992) two-component model of boundary extension.  相似文献   

9.
Listeners must adapt to differences in speech rate across talkers and situations. Speech rate adaptation effects are strong for adjacent syllables (i.e., proximal syllables). For studies that have assessed adaptation effects on speech rate information more than one syllable removed from a point of ambiguity in speech (i.e., distal syllables), the difference in strength between different types of ambiguity is stark. Studies of word segmentation have shown large shifts in perception as a result of distal rate manipulations, while studies of segmental perception have shown only weak, or even nonexistent, effects. However, no study has standardized methods and materials to study context effects for both types of ambiguity simultaneously. Here, a set of sentences was created that differed as minimally as possible except for whether the sentences were ambiguous to the voicing of a consonant or ambiguous to the location of a word boundary. The sentences were then rate-modified to slow down the distal context speech rate to various extents, dependent on three different definitions of distal context that were adapted from previous experiments, along with a manipulation of proximal context to assess whether proximal effects were comparable across ambiguity types. The results indicate that the definition of distal influenced the extent of distal rate effects strongly for both segments and segmentation. They also establish the presence of distal rate effects on word-final segments for the first time. These results were replicated, with some caveats regarding the perception of individual segments, in an Internet-based sample recruited from Mechanical Turk.  相似文献   

10.
Ishihara M  Imanaka K 《Perception》2007,36(9):1375-1390
We conducted two experiments to investigate whether the motor preparation of manual aiming to a visual target is affected by either the physical characteristics (size or luminance contrast) or spatial characteristics (location) of the target. Reaction time (RT) of both finger lifting (ie stimulus-detection time) and manual aiming (ie movement-triggering time) to the onset of the target was measured. The difference of RT (DRT) between two tasks (ie the difference of task complexity) was examined to clarify the temporal characteristics of manual aiming per se during visuomotor integration. Results show classical characteristics: RT decreased as either the target size or luminance contrast increased. Furthermore, the task-complexity and target-location factors significantly interacted with each other, where the aiming RT was longer than the finger-lifting RT and the effects of target location on RT differed for each task. However, the task factor did not interact with either the size or luminance-contrast factor, implying that the motor preparation of manual aiming is associated with the spatial characteristics rather than the physical characteristics of the target. Inspection of DRT revealed that the time needed for motor preparation for an ipsilateral target was significantly shorter than that for a contralateral target. This was the case both for the left and for the right hand. Foveal targets required longer processing time, implying a disadvantageous function of motor preparation for the gazed target. The left-hand superiority for the target appearing in the left visual field was also observed. Such lateralised effect and left-hand advantage to the left visual field in manual aiming suggest that visuospatial information processing is activated during the preparation of aiming action, with faster processing in the right hemisphere.  相似文献   

11.
Pairs of consonant-vowel (CV) syllables were presented dichotically to Ss who were instructed to monitor for the presence of a target CV which could occur in either ear. Ss responded by depressing a response button ; reaction time (RT) was also recorded. Right ear targets were detected 6.2% more frequently, on the average, than left ear targets and had an average RT 50 msec quicker than their left ear counterparts. These results demonstrate the existence of a right ear superiority in dichotic listening when a nonverbal motor response measure is used, supporting the contention that the ear asymmetry phenomenon is truly perceptual in nature and not merely due to the lateralization of verbal output. Two alternative explanations of the RT difference between left and right ear targets are offered. One attributes this difference to the time necessary for intercortical transfer of right hemisphere information, while the second holds that it is due to the longer times needed by the right hemisphere to process information projected to it.  相似文献   

12.
We investigated the role of syllables during speech planning in English by measuring syllable-frequency effects. So far, syllable-frequency effects in English have not been reported. English has poorly defined syllable boundaries, and thus the syllable might not function as a prominent unit in English speech production. Speakers produced either monosyllabic (Experiment 1) or disyllabic (Experiment 2-4) pseudowords as quickly as possible in response to symbolic cues. Monosyllabic targets consisted of either high- or low-frequency syllables, whereas disyllabic items contained either a 1st or 2nd syllable that was frequency-manipulated. Significant syllable-frequency effects were found in all experiments. Whereas previous findings for disyllables in Dutch and Spanish-languages with relatively clear syllable boundaries-showed effects of a frequency manipulation on 1st but not 2nd syllables, in our study English speakers were sensitive to the frequency of both syllables. We interpret this sensitivity as an indication that the production of English has more extensive planning scopes at the interface of phonetic encoding and articulation. (PsycINFO Database Record (c) 2010 APA, all rights reserved).  相似文献   

13.
This study investigates the influence of stress grouping on verbal short-term memory (STM). English speakers show a preference to combine syllables into trochaic groups, both lexically and in continuous speech. In two serial recall experiments, auditory lists of nonsense syllables were presented with either trochaic (STRONG–weak) or iambic (weak–STRONG) stress patterns, or in monotone. The acoustic correlates that carry stress were also manipulated in order to examine the relationship between input and output processes during recall. In Experiment 1, stressed and unstressed syllables differed in intensity and pitch but were matched for spoken duration. Significantly more syllables were recalled in the trochaic stress pattern condition than in the iambic and monotone conditions, which did not differ. In Experiment 2, spoken duration and pitch were manipulated but intensity was held constant. No effects of stress grouping were observed, suggesting that intensity is a critical acoustic factor for trochaic grouping. Acoustic analyses demonstrated that speech output was not identical to the auditory input, but that participants generated correct stress patterns by manipulating acoustic correlates in the same way in both experiments. These data challenge the idea of a language-independent STM store and support the notion of separable phonological input and output processes.  相似文献   

14.
Some reaction time experiments are reported on the relation between the perception and production of phonetic features in speech. Subjects had to produce spoken consonant-vowel syllables rapidly in response to other consonant-vowel stimulus syllables. The stimulus syllables were presented auditorily in one condition and visually in another. Reaction time was measured as a function of the phonetic features shared by the consonants of the stimulus and response syllables. Responses to auditory stimulus syllables were faster when the response syllables started with consonants that had the same voicing feature as those of the stimulus syllables. A shared place-of-articulation feature did not affect the speed of responses to auditory stimulus syllables, even though the place feature was highly salient. For visual stimulus syllables, performance was independent of whether the consonants of the response syllables had the same voicing, same place of articulation, or no shared features. This pattern of results occurred in cases where the syllables contained stop consonants and where they contained fricatives. It held for natural auditory stimuli as well as artificially synthesized ones. The overall data reveal a close relation between the perception and production of voicing features in speech. It does not appear that such a relation exists between perceiving and producing places of articulation. The experiments are relevant to the motor theory of speech perception and to other models of perceptual-motor interactions.  相似文献   

15.
Cholin J  Levelt WJ  Schiller NO 《Cognition》2006,99(2):205-235
In the speech production model proposed by [Levelt, W. J. M., Roelofs, A., Meyer, A. S. (1999). A theory of lexical access in speech production. Behavioral and Brain Sciences, 22, pp. 1-75.], syllables play a crucial role at the interface of phonological and phonetic encoding. At this interface, abstract phonological syllables are translated into phonetic syllables. It is assumed that this translation process is mediated by a so-called Mental Syllabary. Rather than constructing the motor programs for each syllable on-line, the mental syllabary is hypothesized to provide pre-compiled gestural scores for the articulators. In order to find evidence for such a repository, we investigated syllable-frequency effects: If the mental syllabary consists of retrievable representations corresponding to syllables, then the retrieval process should be sensitive to frequency differences. In a series of experiments using a symbol-position association learning task, we tested whether high-frequency syllables are retrieved and produced faster compared to low-frequency syllables. We found significant syllable frequency effects with monosyllabic pseudo-words and disyllabic pseudo-words in which the first syllable bore the frequency manipulation; no effect was found when the frequency manipulation was on the second syllable. The implications of these results for the theory of word form encoding at the interface of phonological and phonetic encoding; especially with respect to the access mechanisms to the mental syllabary in the speech production model by (Levelt et al.) are discussed.  相似文献   

16.
Articulatory and acoustic studies of speech production have shown that the effects of anticipatory coarticulation may extend across several segments of an utterance. The present experiments show that such effects have perceptual significance. In two experiments, a talker produced consonant (C) and vowel (V) sequences in a sentence frame (e.g., "I say pookee") of the form "I say / C V1 C V2/" in which V1 was /u, ae/ and V2 was /i, a/. Each /i, a/ sentence pair was cross-spliced by exchanging the final syllable /C V2/ so that coarticulatory information prior to the crosspoint was inappropriate for te final vowel (V2) in crossed sentences. Recognition time (RT) for V2 in crossed and intact (as spoken) sentences was obtained from practiced listeners. In both experiments RT was slower in crossed sentences; crossed sentences also attracted more false alarms. The pattern of perceptual results was mirrored in the pattern of precross acoustic differences in experimental sentences (e.g., formants F2 and F3 were higher preceding /i/ than preceding /a/). Pretarget variation in the formants jointly predicted amount of RT interference in crossed sentences. A third experiment found interference (slower RT) and also facilitation (faster RT) from exchanges of pretarget coarticulatory information in sentences. Two final experiments showed that previous results were not dependent on the use of practiced listeners.  相似文献   

17.
Identification time was measured for targets in speech in terms of the number of phonemes occurring between the target and response. Auditory targets in the mam experiment were at three levels of phonetic complexity; (1) monosyllabic words or nonsense syllables, each containing four phonemes; (2) clusters (vowel + consonant) consisting of the second and third phonemes within these syllables; (3) individual phonemes within these clusters. Identification time was always shorter for (1) than for (2) or (3). The effect of semantic and grammatical context upon identification was small. Plosive consonants which did not begin a syllable generally could not be identified as isolated phonemic targets, although they could be identified readily as part of a cluster. Letter targets corresponding to the spelling of the auditory stimulus permitted identification of plosive consonants, and in general followed different rules than the phonemic targets. Additional observations concerning individual targets were made.  相似文献   

18.
Two visual search experiments investigated the detection of odd-one-out feature targets redundantly defined on multiple dimensions. Targets differed from the distractors in either orientation or colour or both (redundant targets). In Experiment 1, the three types of target were presented either in separate trial blocks or randomized within blocks, and the task involved either a simple target detection response or a “compound” response based on the position of dials inside the target. Mean reaction times (RTs) were faster to redundant targets than to singly defined targets, with greater gains in simple detection than in compound tasks. Further, simple detection RTs to redundant targets were faster than the fastest RTs to singly defined targets, violating Miller's (1982) “race model inequality” (RMI). Experiment 2 showed that, with compound tasks, mean RT redundancy gains (and violations of the RMI) depend on practice. The results suggest that separate colour and orientation feature contrast signals coactivate perceptual mechanisms involved in target detection.  相似文献   

19.
English exhibits compensatory shortening, whereby a stressed syllable followed by an unstressed syllable is measured to be shorter than the same stressed syllable alone. This anticipatory shortening is much greater than backward shortening, whereby an unstressed syllable is measured to shorten a following stressed syllable. We speculated that measured shortening reflects not true shortening, but coarticulatory hiding. Hence, we asked whether listeners are sensitive to parts of stressed syllables hidden by following or preceding unstressed syllables. In two experiments (Experiments 1A and 1B), we found the point of subjective equality—that is, the durational difference between a stressed syllable in isolation and one followed by an unstressed syllable—at which listeners cannot tell which is longer. In a third experiment (Experiment 2), we found the point of subjective equality for stressed monosyllables and disyllables with a weak-strong stress pattern. In all of the experiments, the points of subjective equality occurred when stressed syllables in disyllables were measured to be shorter than those in monosyllables, as if the listeners heard the coarticulatory onset or the continuation of a stressed syllable within unstressed syllables.  相似文献   

20.
In three experiments, listeners detected vowel or consonant targets in lists of CV syllables constructed from five vowels and five consonants. Responses were faster in a predictable context (e.g., listening for a vowel target in a list of syllables all beginning with the same consonant) than in an unpredictable context (e.g., listening for a vowel target in a list of syllables beginning with different consonants). In Experiment 1, the listeners’ native language was Dutch, in which vowel and consonant repertoires are similar in size. The difference between predictable and unpredictable contexts was comparable for vowel and consonant targets. In Experiments 2 and 3, the listeners’ native language was Spanish, which has four times as many consonants as vowels; here effects of an unpredictable consonant context on vowel detection were significantly greater than effects of an unpredictable vowel context on consonant detection. This finding suggests that listeners’ processing of phonemes takes into account the constitution of their language’s phonemic repertoire and the implications that this has for contextual variability.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号