首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   2767篇
  免费   125篇
  国内免费   1篇
  2023年   17篇
  2022年   15篇
  2021年   29篇
  2020年   43篇
  2019年   55篇
  2018年   73篇
  2017年   80篇
  2016年   114篇
  2015年   57篇
  2014年   89篇
  2013年   333篇
  2012年   142篇
  2011年   146篇
  2010年   89篇
  2009年   82篇
  2008年   131篇
  2007年   113篇
  2006年   112篇
  2005年   87篇
  2004年   99篇
  2003年   90篇
  2002年   62篇
  2001年   63篇
  2000年   54篇
  1999年   42篇
  1998年   40篇
  1997年   38篇
  1996年   35篇
  1995年   41篇
  1994年   27篇
  1993年   31篇
  1992年   37篇
  1991年   27篇
  1990年   32篇
  1989年   20篇
  1988年   32篇
  1987年   17篇
  1986年   15篇
  1985年   22篇
  1984年   23篇
  1983年   17篇
  1982年   23篇
  1981年   11篇
  1979年   20篇
  1978年   12篇
  1977年   16篇
  1974年   17篇
  1972年   13篇
  1970年   11篇
  1968年   11篇
排序方式: 共有2893条查询结果,搜索用时 15 毫秒
211.
Several methods are available to estimate the total and residual amount of heterogeneity in meta‐analysis, leading to different alternatives when estimating the predictive power in mixed‐effects meta‐regression models using the formula proposed by Raudenbush (1994, 2009). In this paper, a simulation study was conducted to compare the performance of seven estimators of these parameters under various realistic scenarios in psychology and related fields. Our results suggest that the number of studies (k) exerts the most important influence on the accuracy of the results, and that precise estimates of the heterogeneity variances and the model predictive power can only be expected with at least 20 and 40 studies, respectively. Increases in the average within‐study sample size () also improved the results for all estimators. Some differences among the accuracy of the estimators were observed, especially under adverse (small k and ) conditions, while the results for the different methods tended to convergence for more optimal scenarios.  相似文献   
212.
Accounts of comprehension failure, whether in the case of readers with poor skill or when syntactic complexity is high, have overwhelmingly implicated working memory capacity as the key causal factor. However, extant research suggests that this position is not well supported by evidence on the span of active memory during online sentence processing, nor is it well motivated by models that make explicit claims about the memory mechanisms that support language processing. The current study suggests that sensitivity to interference from similar items in memory may provide a better explanation of comprehension failure. Through administration of a comprehensive skill battery, we found that the previously observed association of working memory with comprehension is likely due to the collinearity of working memory with many other reading-related skills, especially IQ. In analyses which removed variance shared with IQ, we found that receptive vocabulary knowledge was the only significant predictor of comprehension performance in our task out of a battery of 24 skill measures. In addition, receptive vocabulary and non-verbal memory for serial order—but not simple verbal memory or working memory—were the only predictors of reading times in the region where interference had its primary affect. We interpret these results in light of a model that emphasizes retrieval interference and the quality of lexical representations as key determinants of successful comprehension.  相似文献   
213.
Research with White participants has demonstrated religious intergroup bias; however, religious identity may be different for Black Americans. Only religiously conscious Black Christians demonstrated a preference for Christian targets over Muslim and Atheist targets. Future research should consider what factors result in a person becoming conscious of other's religion.  相似文献   
214.
School System Evaluation by Value Added Analysis Under Endogeneity   总被引:1,自引:0,他引:1  
Value added is a common tool in educational research on effectiveness. It is often modeled as a (prediction of a) random effect in a specific hierarchical linear model. This paper shows that this modeling strategy is not valid when endogeneity is present. Endogeneity stems, for instance, from a correlation between the random effect in the hierarchical model and some of its covariates. This paper shows that this phenomenon is far from exceptional and can even be a generic problem when the covariates contain the prior score attainments, a typical situation in value added modeling. Starting from a general, model-free definition of value added, the paper derives an explicit expression of the value added in an endogeneous hierarchical linear Gaussian model. Inference on value added is proposed using an instrumental variable approach. The impact of endogeneity on the value added and the estimated value added is calculated accurately. This is also illustrated on a large data set of individual scores of about 200,000 students in Chile.  相似文献   
215.
Approximate Bayesian computation (ABC) is a powerful technique for estimating the posterior distribution of a model’s parameters. It is especially important when the model to be fit has no explicit likelihood function, which happens for computational (or simulation-based) models such as those that are popular in cognitive neuroscience and other areas in psychology. However, ABC is usually applied only to models with few parameters. Extending ABC to hierarchical models has been difficult because high-dimensional hierarchical models add computational complexity that conventional ABC cannot accommodate. In this paper, we summarize some current approaches for performing hierarchical ABC and introduce a new algorithm called Gibbs ABC. This new algorithm incorporates well-known Bayesian techniques to improve the accuracy and efficiency of the ABC approach for estimation of hierarchical models. We then use the Gibbs ABC algorithm to estimate the parameters of two models of signal detection, one with and one without a tractable likelihood function.  相似文献   
216.
We examined the impact of eccentricity on the evaluation of artistic skills and the quality of artworks. Based on the notion that artists are typically perceived as eccentric, creative and skilled, we tested the hypothesis that eccentricity increases perceptions of artistic quality. In Study 1, Van Gogh's Sunflowers painting was evaluated more positively when he was said to have cut off his left ear lobe than when this information was not presented. In Study 2, participants liked art more when the artist was eccentric. In Study 3, the evaluation of fictitious art increased because of the artist's eccentric appearance. Study 4 established that the eccentricity effect was specific to unconventional as opposed to conventional art. In Study 5, Lady Gaga's music was more appreciated when she was displayed as highly eccentric; however, the eccentricity effect emerged only when the display seemed authentic. These novel findings indicate that art evaluations are partly rooted in perceptions of artists' eccentricity and evidence the importance of perceived authenticity and skills for these attributions. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   
217.
Three experiments demonstrate that in the context of U.S. foreign policy decision making, people infer informational quality from secrecy. In Experiment 1, people weighed secret information more heavily than public information when making recommendations about foreign political candidates. In Experiment 2, people judged information presented in documents ostensibly produced by the Department of State and the National Security Council as being of relatively higher quality when those documents were secret rather than public. Finally, in Experiment 3, people judged a National Security Council document as being of higher quality when presented as a secret document rather than a public document and evaluated others' decisions more favorably when those decisions were based on secret information. Discussion centers on the mediators, moderators, and broader implications of this secrecy heuristic in foreign policy contexts.  相似文献   
218.
James A. Van Slyke 《Zygon》2014,49(3):696-707
Robert N. McCauley's new book Why Religion Is Natural and Science Is Not (2011) presents a new paradigm for investigating the relationship between science and religion by exploring the cognitive foundations of religious belief and scientific knowledge. McCauley's contention is that many of the differences and disagreements regarding religion and science are the product of distinct features of human cognition that process these two domains of knowledge very differently. McCauley's thesis provides valuable insights into this relationship while not necessarily leading to a dismissive view of theology or religious belief. His paradigm allows the research lens to focus on cognitive differences in processing scientific versus religious information and the important role of automatic, unconscious, and intuitive cognitive processes in understanding both the natural and supernatural worlds.  相似文献   
219.
The present study explored the effectiveness of a unique intervention for couple therapy, the Canned Attitude Method (a technique developed by Brent Atkinson), which utilized a digital recorder in and between therapy sessions. Participant dyads were randomly assigned to the intervention group utilizing the digital recorders or the control group, in which Pragmatic/Experiential Therapy for Couples (Atkinson in J Syst Ther 17:18–35, 1998) was conducted without the use of digital recorders. Both the intervention and the control groups were composed of five couples each for a total of 20 participants. This study hypothesized that the use of the Canned Attitude Method in couple therapy would produce a significant change in marital satisfaction, emotional flooding, and positive psychological tendencies associated with intimate relationships. Findings approached significance, demonstrating decreases in the experience of emotional flooding, gains in relationship satisfaction, and a shift towards positive attributes within the couples’ relationships.  相似文献   
220.
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号