全文获取类型
收费全文 | 453篇 |
免费 | 42篇 |
国内免费 | 55篇 |
出版年
2024年 | 1篇 |
2023年 | 24篇 |
2022年 | 4篇 |
2021年 | 13篇 |
2020年 | 25篇 |
2019年 | 27篇 |
2018年 | 20篇 |
2017年 | 31篇 |
2016年 | 22篇 |
2015年 | 19篇 |
2014年 | 14篇 |
2013年 | 39篇 |
2012年 | 11篇 |
2011年 | 20篇 |
2010年 | 12篇 |
2009年 | 21篇 |
2008年 | 24篇 |
2007年 | 25篇 |
2006年 | 21篇 |
2005年 | 18篇 |
2004年 | 24篇 |
2003年 | 23篇 |
2002年 | 20篇 |
2001年 | 19篇 |
2000年 | 4篇 |
1999年 | 5篇 |
1998年 | 7篇 |
1997年 | 3篇 |
1996年 | 7篇 |
1995年 | 1篇 |
1994年 | 2篇 |
1993年 | 2篇 |
1992年 | 7篇 |
1991年 | 3篇 |
1990年 | 4篇 |
1989年 | 7篇 |
1988年 | 6篇 |
1987年 | 1篇 |
1986年 | 1篇 |
1985年 | 2篇 |
1984年 | 2篇 |
1983年 | 2篇 |
1981年 | 1篇 |
1979年 | 1篇 |
1978年 | 1篇 |
1976年 | 2篇 |
1975年 | 2篇 |
排序方式: 共有550条查询结果,搜索用时 78 毫秒
281.
Joanne B. Waugh 《Metaphilosophy》2005,36(5):578-612
Abstract: In Historied Thought, Constructed World , Joseph Margolis identifies the philosophical themes that will dominate philosophical discussions in the twenty-first century, given the recognition of the historicity of philosophical thought in the twentieth century. In what follows I examine these themes, especially cognitive intransparency, and the arguments presented in favor of them, noting the extent to which they rest on a view of language that takes a written text, and not speech, as the paradigm of language. I suggest if one takes speech as a mutual embodied action in a shared space as a model for language, the theme of cognitive intransparency—and the problems it brings in its wake—does not loom so large for those of us working in the history of philosophy. I conclude by showing that if we adopt this suggestion in relation to early Greek philosophy, that is, the period in the history of historied thought in which philosophy itself emerges as a linguistic and intellectual activity, we can better understand how and why philosophy emerged as it did—in the form of dialogues by Plato. 相似文献
282.
Maternal speech to infants at 1 and 3 months of age 总被引:1,自引:0,他引:1
The goal of this study was to assess maternal speech and in relation to changes in infant social behavior occurring around the second month post birth. Sixty infants interacted with their mother at 1 and 3 months of age in a face-to-face context. At 3 months, infants gazed, smiled, and positively vocalized significantly more than at 1 month. These findings point to a transition in infant social behavior at around the second month post birth. In addition, maternal speech to infants increased between these times in both amount and complexity, possibly in response to an increase in infant social behavior. Maternal speech was related to infant positive vocalizing at 3 months, suggesting mothers especially monitored infant vocalizing at 3 months. Individual differences in maternal speech were stable across visits. 相似文献
283.
口语感知是当前心理语言学的研究热点, 然而以往研究大多以婴儿和成人为被试, 缺乏对幼儿口语感知的研究。此外, 现有口语感知模型主要是基于非声调语言研究建立起来的, 对汉语并不完全适用。汉语是一种声调语言, 在语音构成上不同于非声调语言。本项目将立足于汉语口语特点, 以3~5岁幼儿为研究对象, 考察幼儿汉语口语感知特点及神经机制。综合使用眼动方法、ERP方法和LORETA源定位技术探讨以下问题:(1)幼儿在前注意阶段和注意阶段的听觉语音辨别特点; (2)幼儿汉语口语词汇识别过程中音段信息和超音段信息的作用; (3)幼儿汉语口语感知的神经机制。本项目研究结果将揭示幼儿汉语口语感知特点, 为完善现有的口语感知模型提供新的实验证据。 相似文献
284.
言语产生是指个体通过口头言语表达想法、感受,而通达过程是言语产生的关键过程。本文以言语产生的不同语言水平为线索,综合分析双语经验对言语产生过程中通达能力的影响,结果发现双语经验降低了言语产生中的通达效率。在此基础上,从跨语言干扰、语言使用频率和词汇量三个角度归纳双语通达劣势的原因。未来的研究应该考察双语经验对言语加工中不同方面的影响,并且对相关解释进行整合。 相似文献
285.
ABSTRACTThe perceptual brain is designed around multisensory input. Areas once thought dedicated to a single sense are now known to work with multiple senses. It has been argued that the multisensory nature of the brain reflects a cortical architecture for which task, rather than sensory system, is the primary design principle. This supramodal thesis is supported by recent research on human echolocation and multisensory speech perception. In this review, we discuss the behavioural implications of a supramodal architecture, especially as they pertain to auditory perception. We suggest that the architecture implies a degree of perceptual parity between the senses and that cross-sensory integration occurs early and completely. We also argue that a supramodal architecture implies that perceptual experience can be shared across modalities and that this sharing should occur even without bimodal experience. We finish by briefly suggesting areas of future research. 相似文献
286.
As research on the construction of a mental representation of referent gender in speech comprehension is scarce, this study examined whether factors identified in reading comprehension exert similar influence in speech comprehension. Conceptually replicating previous research, a sentence continuation evaluation task was set up in two modalities, as a listening task and as a time-confined reading task (i.e. to correspond to the time constraint when listening). In line with previous findings from self-paced reading paradigms we found gender representations in language comprehension to be grounded in the interaction between textual (grammatical) and background (stereotypical) information. Extending previous research, the effect of stereotypical information was however modulated by presentation modality. In all, although speech and reading comprehension share higher-level processes of comprehension, this study provides first evidence that differences in comprehension might occur due to differences such as orthographic access or attention allocation. 相似文献
287.
Perception of intersensory temporal order is particularly difficult for (continuous) audiovisual speech, as perceivers may find it difficult to notice substantial timing differences between speech sounds and lip movements. Here we tested whether this occurs because audiovisual speech is strongly paired (“unity assumption”). Participants made temporal order judgments (TOJ) and simultaneity judgments (SJ) about sine-wave speech (SWS) replicas of pseudowords and the corresponding video of the face. Listeners in speech and non-speech mode were equally sensitive judging audiovisual temporal order. Yet, using the McGurk effect, we could demonstrate that the sound was more likely integrated with lipread speech if heard as speech than non-speech. Judging temporal order in audiovisual speech is thus unaffected by whether the auditory and visual streams are paired. Conceivably, previously found differences between speech and non-speech stimuli are not due to the putative “special” nature of speech, but rather reflect low-level stimulus differences. 相似文献
288.
We report four experiments that examined whether bilinguals’ production of one language is affected by the syntactic properties of their other language. Greek–English and English–Greek highly proficient fluent bilinguals produced sentence completions following subject nouns whose translation had either the same or different number. We manipulated whether participants produced completions in the same language as the subject (the source language; one-language production) or the other language (the non-source language; two-language production), and whether they used only one language or both languages within the experimental session. The results demonstrated that the grammar systems of both languages were activated during both one-language and two-language production. The effects of the non-source language were particularly enhanced in two-language utterances, when both languages were used in the experiment, and when it was the bilinguals’ native language. We interpret our results in terms of a model of bilingual sentence production. 相似文献
289.
In human communication, direct speech (e.g., Mary said: “I’m hungry”) is perceived to be more vivid than indirect speech (e.g., Mary said [that] she was hungry). However, the processing consequences of this distinction are largely unclear. In two experiments, participants were asked to either orally (Experiment 1) or silently (Experiment 2, eye-tracking) read written stories that contained either a direct speech or an indirect speech quotation. The context preceding those quotations described a situation that implied either a fast-speaking or a slow-speaking quoted protagonist. It was found that this context manipulation affected reading rates (in both oral and silent reading) for direct speech quotations, but not for indirect speech quotations. This suggests that readers are more likely to engage in perceptual simulations of the reported speech act when reading direct speech as opposed to meaning-equivalent indirect speech quotations, as part of a more vivid representation of the former. 相似文献
290.
The effectiveness of infant-directed speech (IDS) produced by non-depressed mothers for promoting the acquisition of voice-face associations was investigated in 1-year-old children of depressed mothers in a conditioned-attention paradigm. Prior research suggested that infants of mothers with comparatively longer-duration depressive episodes exhibit poorer learning in response to non-depressed mothers’ IDS, but duration of depression was confounded with infant age. In the current study, 1-year-old infants of currently depressed mothers with relatively longer-duration depressive episodes (i.e., perinatal onset) showed significantly poorer learning than 1-year-olds of currently depressed mothers with relatively shorter duration depressive episodes (non-perinatal onset). This was true despite the fact that there were no measurable differences in the severity of depression, level of social functioning, or antidepressant medication use between the two groups. These findings add support to the hypothesis that there is an experience-based change in responsiveness to female IDS in infants of depressed mothers during the first year of life. 相似文献