首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   38902篇
  免费   3598篇
  国内免费   2443篇
电工技术   3427篇
技术理论   1篇
综合类   3545篇
化学工业   1908篇
金属工艺   3659篇
机械仪表   7108篇
建筑科学   1192篇
矿业工程   2004篇
能源动力   586篇
轻工业   3021篇
水利工程   389篇
石油天然气   641篇
武器工业   296篇
无线电   2587篇
一般工业技术   2547篇
冶金工业   1585篇
原子能技术   89篇
自动化技术   10358篇
  2024年   229篇
  2023年   757篇
  2022年   1324篇
  2021年   1429篇
  2020年   1421篇
  2019年   1050篇
  2018年   923篇
  2017年   1151篇
  2016年   1336篇
  2015年   1478篇
  2014年   2477篇
  2013年   1993篇
  2012年   2911篇
  2011年   3022篇
  2010年   2191篇
  2009年   2203篇
  2008年   2035篇
  2007年   2630篇
  2006年   2508篇
  2005年   2137篇
  2004年   1689篇
  2003年   1476篇
  2002年   1242篇
  2001年   1072篇
  2000年   872篇
  1999年   672篇
  1998年   519篇
  1997年   429篇
  1996年   333篇
  1995年   312篇
  1994年   260篇
  1993年   183篇
  1992年   123篇
  1991年   98篇
  1990年   87篇
  1989年   94篇
  1988年   80篇
  1987年   33篇
  1986年   28篇
  1985年   18篇
  1984年   14篇
  1983年   23篇
  1982年   17篇
  1981年   6篇
  1980年   6篇
  1979年   8篇
  1978年   6篇
  1963年   4篇
  1959年   4篇
  1957年   3篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
921.
We present a single-machine problem with the unequal release times under learning effect and deteriorating jobs when the objective is minimizing the makespan. In this study, we introduced a scheduling model with unequal release times in which both job deterioration and learning exist simultaneously. By the effects of learning and deterioration, we mean that the processing time of a job is defined by increasing function of its execution start time and position in the sequence. A branch-and-bound algorithm incorporating with several dominance properties and lower bounds is developed to derive the optimal solution. A heuristic algorithm is proposed to obtain a near-optimal solution. The computational experiments show that the branch-and-bound algorithm can solve instances up to 30 jobs, and the average error percentage of the proposed heuristic is less than 0.16%.  相似文献   
922.
This paper studies a generalization of the order acceptance and scheduling problem in a single-machine environment where a pool consisting of firm planned orders as well as potential orders is available from which an over-demanded company can select. The capacity available for processing the accepted orders is limited and each order is characterized by a known processing time, delivery date, revenue and a weight representing a penalty per unit-time delay beyond the delivery date. We prove that the existence of a constant-factor approximation algorithm for this problem is unlikely. We propose two linear formulations that are solved using an IP solver and we devise two exact branch-and-bound procedures able to solve instances with up to 50 jobs within reasonable CPU times. We compare the efficiency and quality of the results obtained using the different solution approaches.  相似文献   
923.
Feature selection for text categorization is a well-studied problem and its goal is to improve the effectiveness of categorization, or the efficiency of computation, or both. The system of text categorization based on traditional term-matching is used to represent the vector space model as a document; however, it needs a high dimensional space to represent the document, and does not take into account the semantic relationship between terms, which leads to a poor categorization accuracy. The latent semantic indexing method can overcome this problem by using statistically derived conceptual indices to replace the individual terms. With the purpose of improving the accuracy and efficiency of categorization, in this paper we propose a two-stage feature selection method. Firstly, we apply a novel feature selection method to reduce the dimension of terms; and then we construct a new semantic space, between terms, based on the latent semantic indexing method. Through some applications involving the spam database categorization, we find that our two-stage feature selection method performs better.  相似文献   
924.
In this paper, we propose a methodology for training a new model of artificial neural network called the generalized radial basis function (GRBF) neural network. This model is based on generalized Gaussian distribution, which parametrizes the Gaussian distribution by adding a new parameter τ. The generalized radial basis function allows different radial basis functions to be represented by updating the new parameter τ. For example, when GRBF takes a value of τ=2, it represents the standard Gaussian radial basis function. The model parameters are optimized through a modified version of the extreme learning machine (ELM) algorithm. In the methodology proposed (MELM-GRBF), the centers of each GRBF were taken randomly from the patterns of the training set and the radius and τ values were determined analytically, taking into account that the model must fulfil two constraints: locality and coverage. An thorough experimental study is presented to test its overall performance. Fifteen datasets were considered, including binary and multi-class problems, all of them taken from the UCI repository. The MELM-GRBF was compared to ELM with sigmoidal, hard-limit, triangular basis and radial basis functions in the hidden layer and to the ELM-RBF methodology proposed by Huang et al. (2004) [1]. The MELM-GRBF obtained better results in accuracy than the corresponding sigmoidal, hard-limit, triangular basis and radial basis functions for almost all datasets, producing the highest mean accuracy rank when compared with these other basis functions for all datasets.  相似文献   
925.
Rolling element bearing fault diagnosis using wavelet transform   总被引:2,自引:0,他引:2  
This paper is focused on fault diagnosis of ball bearings having localized defects (spalls) on the various bearing components using wavelet-based feature extraction. The statistical features required for the training and testing of artificial intelligence techniques are calculated by the implementation of a wavelet based methodology developed using Minimum Shannon Entropy Criterion. Seven different base wavelets are considered for the study and Complex Morlet wavelet is selected based on minimum Shannon Entropy Criterion to extract statistical features from wavelet coefficients of raw vibration signals. In the methodology, firstly a wavelet theory based feature extraction methodology is developed that demonstrates the information of fault from the raw signals and then the potential of various artificial intelligence techniques to predict the type of defect in bearings is investigated. Three artificial intelligence techniques are used for faults classifications, out of which two are supervised machine learning techniques i.e. support vector machine, learning vector quantization and other one is an unsupervised machine learning technique i.e. self-organizing maps. The fault classification results show that the support vector machine identified the fault categories of rolling element bearing more accurately and has a better diagnosis performance as compared to the learning vector quantization and self-organizing maps.  相似文献   
926.
Support vector machines (SVMs) are theoretically well-justified machine learning techniques, which have also been successfully applied to many real-world domains. The use of optimization methodologies plays a central role in finding solutions of SVMs. This paper reviews representative and state-of-the-art techniques for optimizing the training of SVMs, especially SVMs for classification. The objective of this paper is to provide readers an overview of the basic elements and recent advances for training SVMs and enable them to develop and implement new optimization strategies for SVM-related research at their disposal.  相似文献   
927.
Face recognition based on extreme learning machine   总被引:2,自引:0,他引:2  
Extreme learning machine (ELM) is an efficient learning algorithm for generalized single hidden layer feedforward networks (SLFNs), which performs well in both regression and classification applications. It has recently been shown that from the optimization point of view ELM and support vector machine (SVM) are equivalent but ELM has less stringent optimization constraints. Due to the mild optimization constraints ELM can be easy of implementation and usually obtains better generalization performance. In this paper we study the performance of the one-against-all (OAA) and one-against-one (OAO) ELM for classification in multi-label face recognition applications. The performance is verified through four benchmarking face image data sets.  相似文献   
928.
Indoor location estimation based on Wi-Fi has attracted more and more attention from both research and industry fields. It brings two significant challenges. One is requiring a vast amount of labeled calibration data. The other is real-time training and testing for location estimation task. Traditional machine learning methods cannot get high performance in both aspects. This paper proposed a novel semi-supervised learning method SELM (semi-supervised extreme learning machine) and applied it to sparse calibrated location estimation. There are two advantages of the proposed SELM. First, it employs graph Laplacian regularization to import large number of unlabeled samples which can dramatically reduce labeled calibration samples. Second, it inherits the good property of ELM on extreme training and testing speed. Comparative experiments show that with same number of labeled samples, our method outperforms original ELM and back propagation (BP) network, especially in the case that the calibration data is very sparse.  相似文献   
929.
930.

Context

Software developers spend considerable effort implementing auxiliary functionality used by the main features of a system (e.g., compressing/decompressing files, encryption/decription of data, scaling/rotating images). With the increasing amount of open source code available on the Internet, time and effort can be saved by reusing these utilities through informal practices of code search and reuse. However, when this type of reuse is performed in an ad hoc manner, it can be tedious and error-prone: code results have to be manually inspected and integrated into the workspace.

Objective

In this paper we introduce and evaluate the use of test cases as an interface for automating code search and reuse. We call our approach Test-Driven Code Search (TDCS). Test cases serve two purposes: (1) they define the behavior of the desired functionality to be searched; and (2) they test the matching results for suitability in the local context. We also describe CodeGenie, an Eclipse plugin we have developed that performs TDCS using a code search engine called Sourcerer.

Method

Our evaluation consists of two studies: an applicability study with 34 different features that were searched using CodeGenie; and a performance study comparing CodeGenie, Google Code Search, and a manual approach.

Results

Both studies present evidence of the applicability and good performance of TDCS in the reuse of auxiliary functionality.

Conclusion

This paper presents an approach to source code search and its application to the reuse of auxiliary functionality. Our exploratory evaluation shows promising results, which motivates the use and further investigation of TDCS.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号