首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   99924篇
  免费   2172篇
  国内免费   455篇
电工技术   976篇
综合类   2344篇
化学工业   15368篇
金属工艺   5189篇
机械仪表   3535篇
建筑科学   2677篇
矿业工程   586篇
能源动力   1779篇
轻工业   5325篇
水利工程   1338篇
石油天然气   382篇
武器工业   2篇
无线电   11365篇
一般工业技术   19377篇
冶金工业   4356篇
原子能技术   457篇
自动化技术   27495篇
  2023年   165篇
  2022年   314篇
  2021年   483篇
  2020年   358篇
  2019年   428篇
  2018年   14914篇
  2017年   13782篇
  2016年   10506篇
  2015年   1138篇
  2014年   950篇
  2013年   1404篇
  2012年   4069篇
  2011年   10567篇
  2010年   9115篇
  2009年   6413篇
  2008年   7616篇
  2007年   8401篇
  2006年   772篇
  2005年   1777篇
  2004年   1605篇
  2003年   1629篇
  2002年   951篇
  2001年   420篇
  2000年   474篇
  1999年   342篇
  1998年   617篇
  1997年   455篇
  1996年   370篇
  1995年   226篇
  1994年   213篇
  1993年   183篇
  1992年   111篇
  1991年   113篇
  1990年   91篇
  1989年   86篇
  1988年   73篇
  1987年   81篇
  1986年   75篇
  1985年   68篇
  1984年   52篇
  1983年   52篇
  1981年   51篇
  1977年   45篇
  1976年   75篇
  1968年   51篇
  1966年   48篇
  1965年   49篇
  1956年   40篇
  1955年   67篇
  1954年   71篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
911.
Software Engineering activities are information intensive. Research proposes Information Retrieval (IR) techniques to support engineers in their daily tasks, such as establishing and maintaining traceability links, fault identification, and software maintenance. We describe an engineering task, test case selection, and illustrate our problem analysis and solution discovery process. The objective of the study is to gain an understanding of to what extent IR techniques (one potential solution) can be applied to test case selection and provide decision support in a large-scale, industrial setting. We analyze, in the context of the studied company, how test case selection is performed and design a series of experiments evaluating the performance of different IR techniques. Each experiment provides lessons learned from implementation, execution, and results, feeding to its successor. The three experiments led to the following observations: 1) there is a lack of research on scalable parameter optimization of IR techniques for software engineering problems; 2) scaling IR techniques to industry data is challenging, in particular for latent semantic analysis; 3) the IR context poses constraints on the empirical evaluation of IR techniques, requiring more research on developing valid statistical approaches. We believe that our experiences in conducting a series of IR experiments with industry grade data are valuable for peer researchers so that they can avoid the pitfalls that we have encountered. Furthermore, we identified challenges that need to be addressed in order to bridge the gap between laboratory IR experiments and real applications of IR in the industry.  相似文献   
912.
This paper proposes a novel gait generation method for surely achieving constraint on impact posture in limit cycle walking. First, we introduce an underactuated rimless wheel model without ankle-joint actuation and formulate a state-space realization of the control output using the stance-leg angle as a time parameter through an input–output linearization. Second, we determine a control input that moves the control output to a terminal value at a target stance-leg angle during the single-support phase. Third, we conduct numerical simulations to observe the fundamental gait properties and discuss the relationship between the gait symmetry and mechanical energy restoration. Furthermore, we mathematically prove the asymptotic stability of the generated walking gait by analytically deriving the restored mechanical energy.  相似文献   
913.
Finding dense subgraphs is an important problem in graph mining and has many practical applications. At the same time, while large real-world networks are known to have many communities that are not well-separated, the majority of the existing work focuses on the problem of finding a single densest subgraph. Hence, it is natural to consider the question of finding the top-k densest subgraphs. One major challenge in addressing this question is how to handle overlaps: eliminating overlaps completely is one option, but this may lead to extracting subgraphs not as dense as it would be possible by allowing a limited amount of overlap. Furthermore, overlaps are desirable as in most real-world graphs there are vertices that belong to more than one community, and thus, to more than one densest subgraph. In this paper we study the problem of finding top-k overlapping densest subgraphs, and we present a new approach that improves over the existing techniques, both in theory and practice. First, we reformulate the problem definition in a way that we are able to obtain an algorithm with constant-factor approximation guarantee. Our approach relies on using techniques for solving the max-sum diversification problem, which however, we need to extend in order to make them applicable to our setting. Second, we evaluate our algorithm on a collection of benchmark datasets and show that it convincingly outperforms the previous methods, both in terms of quality and efficiency.  相似文献   
914.
One issue in the dynamic simulation of flexible multibody system is poor computation efficiency, which is due to high frequency components in the solution associated with a deformable body. Standard explicit numerical methods should take very small time steps in order to satisfy the absolute stability condition for the high frequency components and, in turn, the computational efficiency deteriorates. In this study, a hybrid integration scheme is applied to solve the equations of motion of a flexible multibody system for achieving better computational efficiency. The computation times and simulation results are compared between the hybrid scheme and conventional methods. The results demonstrate that the efficiency of a flexible multibody simulation can be improved by using the hybrid scheme.  相似文献   
915.
Dimensional scaling approaches are widely used to develop multi-body human models in injury biomechanics research. Given the limited experimental data for any particular anthropometry, a validated model can be scaled to different sizes to reflect the biological variance of population and used to characterize the human response. This paper compares two scaling approaches at the whole-body level: one is the conventional mass-based scaling approach which assumes geometric similarity; the other is the structure-based approach which assumes additional structural similarity by using idealized mechanical models to account for the specific anatomy and expected loading conditions. Given the use of exterior body dimensions and a uniform Young’s modulus, the two approaches showed close values of the scaling factors for most body regions, with 1.5 % difference on force scaling factors and 13.5 % difference on moment scaling factors, on average. One exception was on the thoracic modeling, with 19.3 % difference on the scaling factor of the deflection. Two 6-year-old child models were generated from a baseline adult model as application example and were evaluated using recent biomechanical data from cadaveric pediatric experiments. The scaled models predicted similar impact responses of the thorax and lower extremity, which were within the experimental corridors; and suggested further consideration of age-specific structural change of the pelvis. Towards improved scaling methods to develop biofidelic human models, this comparative analysis suggests further investigation on interior anatomical geometry and detailed biological material properties associated with the demographic range of the population.  相似文献   
916.
The introduction of moving loads in the Floating Frame of Reference Formulation is presented. We derive the kinematics and governing equations of motion of a general flexible multibody system and their extension to moving loads. The equivalence of convective effects with Coriolis and centripetal forces is shown. These effects are measured numerically and their significance in moving loads traveling at high speed is confirmed. A method is presented to handle discontinuities when moving loads separate from the flexible structure. The method is extended from beam models to general flexible structures obtained by means of the Finite Element Method. An interpolation method for the deformation field of the modal representation of these bodies is introduced.The work is concluded by application of the method to modern mechanical problems in numerical simulations.  相似文献   
917.
Dynamic time warping (DTW) has proven itself to be an exceptionally strong distance measure for time series. DTW in combination with one-nearest neighbor, one of the simplest machine learning methods, has been difficult to convincingly outperform on the time series classification task. In this paper, we present a simple technique for time series classification that exploits DTW’s strength on this task. But instead of directly using DTW as a distance measure to find nearest neighbors, the technique uses DTW to create new features which are then given to a standard machine learning method. We experimentally show that our technique improves over one-nearest neighbor DTW on 31 out of 47 UCR time series benchmark datasets. In addition, this method can be easily extended to be used in combination with other methods. In particular, we show that when combined with the symbolic aggregate approximation (SAX) method, it improves over it on 37 out of 47 UCR datasets. Thus the proposed method also provides a mechanism to combine distance-based methods like DTW with feature-based methods like SAX. We also show that combining the proposed classifiers through ensembles further improves the performance on time series classification.  相似文献   
918.
Some supervised tasks are presented with a numerical output but decisions have to be made in a discrete, binarised, way, according to a particular cutoff. This binarised regression task is a very common situation that requires its own analysis, different from regression and classification—and ordinal regression. We first investigate the application cases in terms of the information about the distribution and range of the cutoffs and distinguish six possible scenarios, some of which are more common than others. Next, we study two basic approaches: the retraining approach, which discretises the training set whenever the cutoff is available and learns a new classifier from it, and the reframing approach, which learns a regression model and sets the cutoff when this is available during deployment. In order to assess the binarised regression task, we introduce context plots featuring error against cutoff. Two special cases are of interest, the \( UCE \) and \( OCE \) curves, showing that the area under the former is the mean absolute error and the latter is a new metric that is in between a ranking measure and a residual-based measure. A comprehensive evaluation of the retraining and reframing approaches is performed using a repository of binarised regression problems created on purpose, concluding that no method is clearly better than the other, except when the size of the training data is small.  相似文献   
919.
This paper describes the design and implementation of a computational model for Arabic natural language semantics, a semantic parser for capturing the deep semantic representation of Arabic text. The parser represents a major part of an Interlingua-based machine translation system for translating Arabic text into Sign Language. The parser follows a frame-based analysis to capture the overall meaning of Arabic text into a formal representation suitable for NLP applications that need for deep semantics representation, such as language generation and machine translation. We will show the representational power of this theory for the semantic analysis of texts in Arabic, a language which differs substantially from English in several ways. We will also show that the integration of WordNet and FrameNet in a single unified knowledge resource can improve disambiguation accuracy. Furthermore, we will propose a rule based algorithm to generate an equivalent Arabic FrameNet, using a lexical resource alignment of FrameNet1.3 LUs and WordNet3.0 synsets for English Language. A pilot study of motion and location verbs was carried out in order to test our system. Our corpus is made up of more than 2000 Arabic sentences in the domain of motion events collected from Algerian first level educational Arabic books and other relevant Arabic corpora.  相似文献   
920.
As part of information retrieval systems (IRS) and in the context of the use of ontologies for documents and queries indexing, we propose and evaluate in this paper the contribution of this approach applied to Arabic texts. To do this we indexed a corpus of Arabic text using Arabic WordNet. The disambiguation of words was performed by applying the Lesk algorithm. The results obtained by our experiment allowed us to deduct the contribution of this approach in IRS for Arabic texts.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号