首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   120篇
  免费   15篇
  国内免费   3篇
  2023年   1篇
  2022年   1篇
  2021年   2篇
  2020年   5篇
  2019年   6篇
  2018年   2篇
  2017年   10篇
  2016年   2篇
  2015年   3篇
  2014年   2篇
  2013年   9篇
  2012年   1篇
  2010年   1篇
  2009年   3篇
  2008年   3篇
  2007年   9篇
  2006年   9篇
  2005年   8篇
  2004年   2篇
  2003年   1篇
  2002年   5篇
  2001年   3篇
  2000年   2篇
  1999年   3篇
  1998年   2篇
  1997年   2篇
  1996年   6篇
  1995年   1篇
  1994年   3篇
  1993年   2篇
  1992年   5篇
  1991年   1篇
  1990年   4篇
  1989年   2篇
  1988年   1篇
  1987年   2篇
  1986年   1篇
  1984年   2篇
  1983年   1篇
  1982年   2篇
  1981年   3篇
  1980年   2篇
  1979年   1篇
  1978年   2篇
排序方式: 共有138条查询结果,搜索用时 15 毫秒
51.
评价神经导航系统在颅脑手术的应用价值.充分利用神经导航系统定位精确和实时导航功能,达到颅脑手术疗效的最优化.对神经导航在颅脑手术最优化方面的几点误区进行分析,对颅脑手术"适度医疗"的问题提出一些建议.  相似文献   
52.
Extreme learning machine (ELM) for random single-hidden-layer feedforward neural networks (RSLFN) has been widely applied in many fields in the past ten years because of its fast learning speed and good generalization performance. But because traditional ELM randomly selects the input weights and hidden biases, it typically requires high number of hidden neurons and thus decreases its convergence performance. It is necessary to select optimal input weights and hidden biases to improve the convergence performance of the traditional ELM. Generally, the single-hidden-layer feedforward neural networks (SLFN) with low input-to-output sensitivity will cause good robustness of the network, which may further lead into good generalization performance. Moreover, particle swarm optimization (PSO) has no complicated evolutionary operators and fewer parameters need to adjust, and is easy to implement. In this study, an improved ELM based on PSO and input-to-output sensitivity information is proposed to improve RSLFN’s convergence performance. In the improved ELM, PSO encoding the input to output sensitivity information of the SLFN is used to optimize the input weights and hidden biases. The improved ELM could obtain better generalization performance as well as improve the conditioning of the SLFN by decreasing the input-to-output sensitivity of the network. Finally, experiment results on the regression and classification problems verify the improved performance the proposed ELM.  相似文献   
53.
The purpose of this study was to explore the level of inter- and intra-individual variability in the kinematic profiles of the back squat movement among skilled weightlifters. Ten competitive weightlifters volunteered for participation in this study. Barbell velocity (VBarbell) and angular velocity of the ankle (ωAnkle), knee (ωKnee) and hip joint (ωHip) were obtained by kinematic recording of six trials at 90% of 1RM in the back squat. Inter-individual variability was assessed by analysing inter-individual differences in the velocity curves through the statistical parametric mapping method. Intra-individual variability was assessed through a correlation analysis between the barbell velocity curves of each trial for each participant. Partial least squares regression analysis, was performed to relate changes in intra-individual variability to movement and anthropometric characteristics. Significant inter- and intra-individual differences were identified in VBarbell, ωAnkle, ωKnee, and ωHip (p ≤ 0.05). Having a short trunk and thigh, and a long shin in combination with greater anterior-posterior displacement of the barbell and slower velocities during the acceleration phase increased intra-individual movement consistency over movement variability. The results of the present study clearly demonstrate that skilled weightlifters display both significant inter- and intra-individual variability in the successful execution of the back squat.  相似文献   
54.
This paper is concerned with a problem where K (n×n) proximity matrices are available for a set of n objects. The goal is to identify a single permutation of the n objects that provides an adequate structural fit, as measured by an appropriate index, for each of the K matrices. A multiobjective programming approach for this problem, which seeks to optimize a weighted function of the K indices, is proposed, and illustrative examples are provided using a set of proximity matrices from the psychological literature. These examples show that, by solving the multiobjective programming model under different weighting schemes, the quantitative analyst can uncover information about the relationships among the matrices and often identify one or more permutations that provide good to excellent index values for all matrices.  相似文献   
55.
Optimization‐based computer systems are used by many airlines to solve crew planning problems by constructing minimal cost tours of duty. However, today airlines do not only require cost effective solutions, but are also very interested in robust solutions. A more robust solution is understood to be one where disruptions in the schedule (due to delays) are less likely to be propagated into the future, causing delays of subsequent flights. Current scheduling systems based solely on cost do not automatically provide robust solutions. These considerations lead to a multiobjective framework, as the maximization of robustness will be in conflict with the minimization of cost. For example crew changing aircraft within a duty period is discouraged if inadequate ground time is provided. We develop a bicriteria optimization framework to generate Pareto optimal schedules for the domestic airline. A Pareto optimal schedule is one which does not allow an improvement in cost and robustness at the same time. We developed a method to solve the bicriteria problem, implemented it and tested it with actual airline data. Our results show that considerable gain in robustness can be achieved with a small increase in cost. The additional cost is mainly due to an increase in overnights, which allows for a reduction of the number of aircraft changes. Copyright © 2003 John Wiley & Sons, Ltd.  相似文献   
56.
Cluster differences scaling is a method for partitioning a set of objects into classes and simultaneously finding a low-dimensional spatial representation ofK cluster points, to model a given square table of dissimilarities amongn stimuli or objects. The least squares loss function of cluster differences scaling, originally defined only on the residuals of pairs of objects that are allocated to different clusters, is extended with a loss component for pairs that are allocated to the same cluster. It is shown that this extension makes the method equivalent to multidimensional scaling with cluster constraints on the coordinates. A decomposition of the sum of squared dissimilarities into contributions from several sources of variation is described, including the appropriate degrees of freedom for each source. After developing a convergent algorithm for fitting the cluster differences model, it is argued that the individual objects and the cluster locations can be jointly displayed in a configuration obtained as a by-product of the optimization. Finally, the paper introduces a fuzzy version of the loss function, which can be used in a successive approximation strategy for avoiding local minima. A simulation study demonstrates that this strategy significantly outperforms two other well-known initialization strategies, and that it has a success rate of 92 out of 100 in attaining the global minimum.  相似文献   
57.
College undergraduates were given repeated opportunities to choose between a fixed-ratio and a progressive-ratio schedule of reinforcement. Completions of a progressive-ratio schedule produced points (exchangeable for money) and incremented that response requirement by 20 responses with each consecutive choice. In the reset condition, completion of a fixed ratio produced the same number of points and also reset the progressive ratio back to its initial value. In the no-reset condition, the progressive ratio continued to increase by increments of 20 throughout the session with each successive selection of this schedule, irrespective of fixed-ratio choices. Subjects' schedule choices were sensitive to parametric manipulations of the size of the fixed-ratio schedule and were consistent with predictions made on the basis of minimizing the number of responses emitted per point earned, which is a principle of most optimality theories. Also, the present results suggest that if data from human performances are to be compared with results for other species, humans should be exposed to schedules of reinforcement for long periods of time, as is commonly done with nonhuman subjects.  相似文献   
58.
A procedure for maximizing the coefficient of generalizability under the constraint of limited resources is presented. The procedure uses optimization techniques that offer an investigator or test constructor the possibility of employing practical constraints. The procedure is illustrated for the two-facet random-model crossed design.  相似文献   
59.
Matching models in the analysis of cross-classifications   总被引:1,自引:0,他引:1  
Inference models motivated by the combinatorial chance literature and the concept of object matching may be used in the analysis of a contingency table if the conditional assumption of fixed row and column totals is imposed. More specifically, by developing a matching reinterpretation for several problems of interest in the prediction analysis of cross-classifications—as defined by Hildebrand, Laing and Rosenthal, appropriate significance tests can be given that may differ from those justified by the more common multinomial models. In the course of the paper the distinction between a degree-1 statistic (based on the relationship between single objects) and a degree-2 statistic (based on the relationship between object pairs) is reviewed in some detail. Also, several specializations are presented to topics of current methodological importance in psychology; for instance, a number of references are made to the measurement of nominal scale response agreement between two raters.Partial support for this research was provided by the National Science Foundation through GSOC-77-28227.  相似文献   
60.
Learning hierarchy research has been characterized by the use of ad hoc statistical procedures to determine the validity of postulated hierarchical connections. The two most substantial attempts to legitimize the procedure are due to White & Clark and Dayton & Macready, although both of these methods suffer from serious inadequacies. Data from a number of sources is analyzed using a restricted maximum likelihood estimation procedure and the results are compared with those obtained using the method suggested by Dayton and Macready. Improved estimates are evidenced by an increase in the computed value of the log likelihood function.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号