首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The current study assessed the extent to which the use of referential prosody varies with communicative demand. Speaker–listener dyads completed a referential communication task during which speakers attempted to indicate one of two color swatches (one bright, one dark) to listeners. Speakers' bright sentences were reliably higher pitched than dark sentences for ambiguous (e.g., bright red versus dark red) but not unambiguous (e.g., bright red versus dark purple) trials, suggesting that speakers produced meaningful acoustic cues to brightness when the accompanying linguistic content was underspecified (e.g., “Can you get the red one?”). Listening partners reliably chose the correct corresponding swatch for ambiguous trials when lexical information was insufficient to identify the target, suggesting that listeners recruited prosody to resolve lexical ambiguity. Prosody can thus be conceptualized as a type of vocal gesture that can be recruited to resolve referential ambiguity when there is communicative demand to do so.  相似文献   

2.
The minimal unit of phonological encoding: prosodic or lexical word   总被引:1,自引:0,他引:1  
Wheeldon LR  Lahiri A 《Cognition》2002,85(2):B31-B41
Wheeldon and Lahiri (Journal of Memory and Language 37 (1997) 356) used a prepared speech production task (Sternberg, S., Monsell, S., Knoll, R. L., & Wright, C. E. (1978). The latency and duration of rapid movement sequences: comparisons of speech and typewriting. In G. E. Stelmach (Ed.), Information processing in motor control and learning (pp. 117-152). New York: Academic Press; Sternberg, S., Wright, C. E., Knoll, R. L., & Monsell, S. (1980). Motor programs in rapid speech: additional evidence. In R. A. Cole (Ed.), The perception and production of fluent speech (pp. 507-534). Hillsdale, NJ: Erlbaum) to demonstrate that the latency to articulate a sentence is a function of the number of phonological words it comprises. Latencies for the sentence [Ik zoek het] [water] 'I seek the water' were shorter than latencies for sentences like [Ik zoek] [vers] [water] 'I seek fresh water'. We extend this research by examining the prepared production of utterances containing phonological words that are less than a lexical word in length. Dutch compounds (e.g. ooglid 'eyelid') form a single morphosyntactic word and a phonological word, which in turn includes two phonological words. We compare their prepared production latencies to those syntactic phrases consisting of an adjective and a noun (e.g. oud lid 'old member') which comprise two morphosyntactic and two phonological words, and to morphologically simple words (e.g. orgel 'organ') which comprise one morphosyntactic and one phonological word. Our findings demonstrate that the effect is limited to phrasal level phonological words, suggesting that production models need to make a distinction between lexical and phrasal phonology.  相似文献   

3.
Two sentence processing experiments on a dative NP ambiguity in Korean demonstrate effects of phrase length on overt and implicit prosody. Both experiments controlled non-prosodic length factors by using long versus short proper names that occurred before the syntactically critical material. Experiment 1 found that long phrases induce different prosodic phrasing than short phrases in a read-aloud task and change the preferred interpretation of globally ambiguous sentences. It also showed that speakers who have been told of the ambiguity can provide significantly different prosody for the two interpretations, for both lengths. Experiment 2 verified that prosodic patterns found in first-pass pronunciations predict self-paced reading patterns for silent reading. The results extend the coverage of the Implicit Prosody Hypothesis [Fodor, J Psycholinguist Res 27:285–319, 1998; Prosodic disambiguation in silent reading. In M. Hirotani (Ed.), NELS 32 (pp. 113–132). Amherst, MA: GLSA Publications, 2002] to another construction and to Korean. They further indicate that strong syntactic biases can have rapid effects on the formulation of implicit prosody.  相似文献   

4.
Rhythmic structure in speech is characterized by sequences of stressed and unstressed syllables. A large body of literature suggests that speakers of English attempt to achieve rhythmic harmony by evenly distributing stressed syllables throughout prosodic phrases. The question remains as to how speakers plan metrical structure during speech production and whether it is planned independently of phonemes. To examine this, we designed a tongue twister task consisting of disyllabic word pairs with overlapping phonological segments and either matching or non-matching metrical structure. Results showed that speakers had more difficulty producing metrically regular word pairs, compared to irregular pairs; that is, word pairs with irregular meter had faster productions and fewer speech errors in this production task. This finding of metrical regularity inhibiting production is inconsistent with an abstract metrical structure that is planned independently of phonemes at the point of phonological encoding.  相似文献   

5.
Recent research with cotton-top tamarin monkeys has revealed language discrimination abilities similar to those found in human infants, demonstrating that these perceptual abilities are not unique to humans but are also present in non-human primates. Specifically, tamarins could discriminate forward but not backward sentences of Dutch from Japanese, using both natural and synthesized utterances. The present study was designed as a conceptual replication of the work on tamarins. Results show that rats trained in a discrimination learning task readily discriminate forward, but not backward sentences of Dutch from Japanese; the results are particularly robust for synthetic utterances, a pattern that shows greater parallels with newborns than with tamarins. Our results extend the claims made in the research with tamarins that the capacity to discriminate languages from different rhythmic classes depends on general perceptual abilities that evolved at least as far back as the rodents. Electronic Publication  相似文献   

6.
Emotional inferences from speech require the integration of verbal and vocal emotional expressions. We asked whether this integration is comparable when listeners are exposed to their native language and when they listen to a language learned later in life. To this end, we presented native and non-native listeners with positive, neutral and negative words that were spoken with a happy, neutral or sad tone of voice. In two separate tasks, participants judged word valence and ignored tone of voice or judged emotional tone of voice and ignored word valence. While native listeners outperformed non-native listeners in the word valence task, performance was comparable in the voice task. More importantly, both native and non-native listeners responded faster and more accurately when verbal and vocal emotional expressions were congruent as compared to when they were incongruent. Given that the size of the latter effect did not differ as a function of language proficiency, one can conclude that the integration of verbal and vocal emotional expressions occurs as readily in one's second language as it does in one's native language.  相似文献   

7.
The purpose of the present investigation was to determine whether mothers use discernible tunes (i.e., specific interval sequences) in their speech to infants and whether such tunes are individually distinctive. Mothers were recorded speaking with their infants on two occasions separated by 1 week or more. Examination of the tunes of each mother revealed discernible tunes and frequent repetitions of tunes within and across sessions. Comparisons of utterances with the most common pitch contour (i.e., rising), both within and across mothers, revealed interval patterns that were individually distinctive, or unique. The findings confirm the prominence of tunes and the presence of signature tunes in maternal speech to infants.  相似文献   

8.
To successfully infer a speaker's emotional state, diverse sources of emotional information need to be decoded. The present study explored to what extent emotional speech recognition of 'basic' emotions (anger, disgust, fear, happiness, pleasant surprise, sadness) differs between different sex (male/female) and age (young/middle-aged) groups in a behavioural experiment. Participants were asked to identify the emotional prosody of a sentence as accurately as possible. As a secondary goal, the perceptual findings were examined in relation to acoustic properties of the sentences presented. Findings indicate that emotion recognition rates differ between the different categories tested and that these patterns varied significantly as a function of age, but not of sex.  相似文献   

9.
言语产生中的韵律生成   总被引:1,自引:1,他引:0  
首先描述了韵律生成的相关模型。然后以研究的问题为着眼点,回顾了言语产生中韵律生成的相关研究。发现大多数研究结果都证明,在单词产生过程中,存在一个独立于音段内容的抽象韵律结构,其中包含了词的重音模式和音节数量等信息;在短语和句子产生过程中,韵律结构和停顿模式都不是完全决定于句法结构,而具有相对的独立性;句子韵律产生的最佳单位是韵律词。语言学韵律产生的脑区激活呈左侧化趋势。  相似文献   

10.
From birth, newborns show a preference for faces talking a native language compared to silent faces. The present study addresses two questions that remained unanswered by previous research: (a) Does the familiarity with the language play a role in this process and (b) Are all the linguistic and paralinguistic cues necessary in this case? Experiment 1 extended newborns’ preference for native speakers to non-native ones. Given that fetuses and newborns are sensitive to the prosodic characteristics of speech, Experiments 2 and 3 presented faces talking native and nonnative languages with the speech stream being low-pass filtered. Results showed that newborns preferred looking at a person who talked to them even when only the prosodic cues were provided for both languages. Nonetheless, a familiarity preference for the previously talking face is observed in the “normal speech” condition (i.e., Experiment 1) and a novelty preference in the “filtered speech” condition (Experiments 2 and 3). This asymmetry reveals that newborns process these two types of stimuli differently and that they may already be sensitive to a mismatch between the articulatory movements of the face and the corresponding speech sounds.  相似文献   

11.
We examined whether children's ability to integrate speech and gesture follows the pattern of a broader developmental shift between 3‐ and 5‐year‐old children (Ramscar & Gitcho, 2007) regarding the ability to process two pieces of information simultaneously. In Experiment 1, 3‐year‐olds, 5‐year‐olds, and adults were presented with either an iconic gesture or a spoken sentence or a combination of the two on a computer screen, and they were instructed to select a photograph that best matched the message. The 3‐year‐olds did not integrate information in speech and gesture, but 5‐year‐olds and adults did. In Experiment 2, 3‐year‐old children were presented with the same speech and gesture as in Experiment 1 that were produced live by an experimenter. When presented live, 3‐year‐olds could integrate speech and gesture. We concluded that development of the integration ability is a part of the broader developmental shift; however, live‐presentation facilitates the nascent integration ability in 3‐year‐olds.  相似文献   

12.
Mitterer H  Ernestus M 《Cognition》2008,109(1):168-173
This study reports a shadowing experiment, in which one has to repeat a speech stimulus as fast as possible. We tested claims about a direct link between perception and production based on speech gestures, and obtained two types of counterevidence. First, shadowing is not slowed down by a gestural mismatch between stimulus and response. Second, phonetic detail is more likely to be imitated in a shadowing task if it is phonologically relevant. This is consistent with the idea that speech perception and speech production are only loosely coupled, on an abstract phonological level.  相似文献   

13.
Using cross-modal form priming, we compared the use of stress and lexicality in the segmentation of spoken English by native English speakers (L1) and by native Hungarian speakers of second-language English (L2). For both language groups, lexicality was found to be an effective segmentation cue. That is, spoken disyllabic word fragments were stronger primes in a subsequent visual word recognition task when preceded by meaningful words than when preceded by nonwords: For example, the first two syllables of corridor were a more effective prime for visually presented corridor when heard in the phrase anythingcorri than in imoshingcorri. The stress pattern of the prime (strong–weak vs. weak–strong) did not affect the degree of priming. For L1 speakers, this supports previous findings about the preferential use of high-level segmentation strategies in clear speech. For L2 speakers, the lexical strategy was employed regardless of L2 proficiency level and instead of exploiting the consistent stress pattern of their native language. This is clear evidence for the primacy and robustness of segmentation by lexical subtraction even in individuals whose lexical knowledge is limited.  相似文献   

14.
Sensitivity to prosodic cues might be used to constrain lexical search. Indeed, the prosodic organization of speech is such that words are invariably aligned with phrasal prosodic edges, providing a cue to segmentation. In this paper we devise an experimental paradigm that allows us to investigate the interaction between statistical and prosodic cues to extract words from a speech stream. We provide evidence that statistics over the syllables are computed independently of prosody. However, we also show that trisyllabic sequences with high transition probabilities that straddle two prosodic constituents appear not to be recognized. Taken together, our findings suggest that prosody acts as a filter, suppressing possible word-like sequences that span prosodic constituents.  相似文献   

15.
Listeners’ comprehension of phrase final rising pitch on declarative utterances, or uptalk, was examined to test the hypothesis that prolongations might differentiate conflicting functions of rising pitch. In Experiment 1 we found that listeners rated prolongations as indicating more speaker uncertainty, but that rising pitch was unrelated to ratings. In Experiment 2 we found that prolongations interacted with rising pitch when listeners monitored for words in the subsequent utterance. Words preceded by prolonged uptalk were monitored faster than words preceded by non-prolonged uptalk. In Experiment 3 we found that the interaction between rising pitch and prolongations depended on listeners’ beliefs about speakers’ mental states. Results support the theory that temporal and situational context are important in determining intonational meaning.  相似文献   

16.
Ozdemir R  Roelofs A  Levelt WJ 《Cognition》2007,105(2):457-465
Disagreement exists about how speakers monitor their internal speech. Production-based accounts assume that self-monitoring mechanisms exist within the production system, whereas comprehension-based accounts assume that monitoring is achieved through the speech comprehension system. Comprehension-based accounts predict perception-specific effects, like the perceptual uniqueness-point effect, in the monitoring of internal speech. We ran an extensive experiment testing this prediction using internal phoneme monitoring and picture naming tasks. Our results show an effect of the perceptual uniqueness point of a word in internal phoneme monitoring in the absence of such an effect in picture naming. These results support comprehension-based accounts of the monitoring of internal speech.  相似文献   

17.
Young and old adults’ ability to recognize emotions from vocal expressions and music performances was compared. The stimuli consisted of (a) acted speech (anger, disgust, fear, happiness, and sadness; each posed with both weak and strong emotion intensity), (b) synthesized speech (anger, fear, happiness, and sadness), and (c) short melodies played on the electric guitar (anger, fear, happiness, and sadness; each played with both weak and strong emotion intensity). The listeners’ recognition of discrete emotions and emotion intensity was assessed and the recognition rates were controlled for various response biases. Results showed emotion-specific age-related differences in recognition accuracy. Old adults consistently received significantly lower recognition rates for negative, but not for positive, emotions for both speech and music stimuli. Some age-related differences were also evident in the listeners’ ratings of emotion intensity. The results show the importance of considering individual emotions in studies on age-related differences in emotion recognition.  相似文献   

18.
Different kinds of speech sounds are used to signify possible word forms in every language. For example, lexical stress is used in Spanish (/‘be.be/, ‘he/she drinks’ versus /be.’be/, ‘baby’), but not in French (/‘be.be/ and /be.’be/ both mean ‘baby’). Infants learn many such native language phonetic contrasts in their first year of life, likely using a number of cues from parental speech input. One such cue could be parents’ object labeling, which can explicitly highlight relevant contrasts. Here we ask whether phonetic learning from object labeling is abstract—that is, if learning can generalize to new phonetic contexts. We investigate this issue in the prosodic domain, as the abstraction of prosodic cues (like lexical stress) has been shown to be particularly difficult. One group of 10-month-old French-learners was given consistent word labels that contrasted on lexical stress (e.g., Object A was labeled /‘ma.bu/, and Object B was labeled /ma.’bu/). Another group of 10-month-olds was given inconsistent word labels (i.e., mixed pairings), and stress discrimination in both groups was measured in a test phase with words made up of new syllables. Infants trained with consistently contrastive labels showed an earlier effect of discrimination compared to infants trained with inconsistent labels. Results indicate that phonetic learning from object labeling can indeed generalize, and suggest one way infants may learn the sound properties of their native language(s).  相似文献   

19.
We hypothesized that chimpanzees could learn to produce attention-getting (AG) sounds via positive reinforcement. We conducted a vocal assessment in 76 captive chimpanzees for their use of AG sounds to acquire the attention of an otherwise inattentive human. Fourteen individuals that did not produce AG sounds during the vocal assessment were evaluated for their ability to acquire the use of an AG sound through operant conditioning and to employ these sounds in an attention-getting context. Nine of the 14 chimpanzees were successfully shaped using positive reinforcement to produce an AG sound. In a post-training vocal assessment, eight of the nine individuals that were successfully trained to produce AG sounds generalized the use of these newly acquired signals to communicatively relevant situations. Chimpanzees possess the ability to acquire the use of a communicative signal via operant conditioning and can generalize the use of this newly acquired signal to appropriate communicative contexts.  相似文献   

20.
This study examined the relationships among prosodic sensitivity, morphological awareness, and reading ability in a sample of 104 8- to 13-year-olds. Using a task adapted from Carlisle (Applied Psycholinguistics, 9 (1988) 247-266), we measured children’s ability to produce morphological derivations with differing levels of phonological complexity between stem and derivation: No Change, Phonemic Change, Stress Change, and Both Phonemic and Stress Change. A 3 (Grade) × 4 (Derivation Type) analysis of variance showed that children perform significantly more poorly on both types of derivations that involve stress changes than on phonemic change and no change derivations. Regression analyses showed that both prosodic sensitivity and morphological awareness, especially in derivations that require manipulation of stress, are significant predictors of reading ability after controlling for age, verbal and nonverbal abilities, and phonological awareness.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号