首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   93881篇
  免费   17650篇
  国内免费   13490篇
电工技术   9558篇
技术理论   4篇
综合类   11144篇
化学工业   2275篇
金属工艺   2127篇
机械仪表   7228篇
建筑科学   1832篇
矿业工程   1109篇
能源动力   1681篇
轻工业   1261篇
水利工程   1207篇
石油天然气   930篇
武器工业   1280篇
无线电   15140篇
一般工业技术   5702篇
冶金工业   2497篇
原子能技术   294篇
自动化技术   59752篇
  2024年   1251篇
  2023年   3340篇
  2022年   5550篇
  2021年   5630篇
  2020年   5316篇
  2019年   4041篇
  2018年   3564篇
  2017年   3803篇
  2016年   4245篇
  2015年   4855篇
  2014年   6376篇
  2013年   6012篇
  2012年   7260篇
  2011年   7837篇
  2010年   6327篇
  2009年   6409篇
  2008年   6674篇
  2007年   7077篇
  2006年   5669篇
  2005年   4728篇
  2004年   3787篇
  2003年   3050篇
  2002年   2380篇
  2001年   1814篇
  2000年   1548篇
  1999年   1224篇
  1998年   1023篇
  1997年   834篇
  1996年   657篇
  1995年   480篇
  1994年   398篇
  1993年   324篇
  1992年   294篇
  1991年   150篇
  1990年   179篇
  1989年   140篇
  1988年   96篇
  1987年   49篇
  1986年   75篇
  1985年   40篇
  1984年   35篇
  1983年   33篇
  1982年   34篇
  1981年   24篇
  1980年   27篇
  1965年   24篇
  1964年   25篇
  1963年   23篇
  1959年   23篇
  1955年   23篇
排序方式: 共有10000条查询结果,搜索用时 0 毫秒
1.
Competition forces manufacturing systems to be flexible and to increase product variety and process complexity. These tasks depend on the flexible design of a bill of materials (BOM), one of the most important inputs in manufacturing planning and control systems. Product variety forces systems to generate BOMs with regard to product properties through a BOM pattern. A variant bill of materials provides a structure to manage product variability. In this study, an algorithm is designed to build a BOM pattern using computer-aided design and computer-aided manufacturing (CAD/CAM) data, and another algorithm is designed to generate variants with regard to product specifications. Genetic algorithm is used to generate new products to provide high product variability for testing algorithms. After the test, both algorithms are applied to a real industry problem. The BOM pattern is built automatically using CAD/CAM data, and variants are generated with regard to the pattern, and the results are discussed.  相似文献   
2.
Redundancy allocation problem (RAP) is one of the best-developed problems in reliability engineering studies. This problem follows to optimize the reliability of a system containing s sub-systems under different constraints, including cost, weight, and volume restrictions using redundant components for each sub-system. Various solving methodologies have been used to optimize this problem, including exact, heuristic, and meta-heuristic algorithms. In this paper, an efficient multi-objective meta-heuristic algorithm based on simulated annealing (SA) is developed to solve multi-objective RAP (MORAP). This algorithm is knowledge-based archive multi-objective simulated annealing (KBAMOSA). KBAMOSA applies a memory matrix to reinforce the neighborhood structure to achieve better quality solutions. The results analysis and comparisons demonstrate the performance of the proposed algorithm for solving MORAP.  相似文献   
3.
This paper considers the scheduling problem of minimizing earliness–tardiness (E/T) on a single batch processing machine with a common due date. The problem is extended to the environment of non-identical job sizes. First, a mathematical model is formulated, which is tested effectively under IBM ILOG CPLEX using the constraint programming solver. Then several optimal properties are given to schedule batches effectively, and by introducing the concept of ARB (Attribute Ratio of Batch), it is proven that the ARB of each batch should be made as small as possible in order to minimize the objective, designed as the heuristic information for assigning jobs into batches. Based on these properties, a heuristic algorithm MARB (Minimum Attribute Ratio of Batch) for batch forming is proposed, and a hybrid genetic algorithm is developed for the problem under study by combining GA (genetic algorithm) with MARB. Experimental results demonstrate that the proposed algorithm outperforms other algorithms in the literature, both for small and large problem instances.  相似文献   
4.
Although greedy algorithms possess high efficiency, they often receive suboptimal solutions of the ensemble pruning problem, since their exploration areas are limited in large extent. And another marked defect of almost all the currently existing ensemble pruning algorithms, including greedy ones, consists in: they simply abandon all of the classifiers which fail in the competition of ensemble selection, causing a considerable waste of useful resources and information. Inspired by these observations, an interesting greedy Reverse Reduce-Error (RRE) pruning algorithm incorporated with the operation of subtraction is proposed in this work. The RRE algorithm makes the best of the defeated candidate networks in a way that, the Worst Single Model (WSM) is chosen, and then, its votes are subtracted from the votes made by those selected components within the pruned ensemble. The reason is because, for most cases, the WSM might make mistakes in its estimation for the test samples. And, different from the classical RE, the near-optimal solution is produced based on the pruned error of all the available sequential subensembles. Besides, the backfitting step of RE algorithm is replaced with the selection step of a WSM in RRE. Moreover, the problem of ties might be solved more naturally with RRE. Finally, soft voting approach is employed in the testing to RRE algorithm. The performances of RE and RRE algorithms, and two baseline methods, i.e., the method which selects the Best Single Model (BSM) in the initial ensemble, and the method which retains all member networks of the initial ensemble (ALL), are evaluated on seven benchmark classification tasks under different initial ensemble setups. The results of the empirical investigation show the superiority of RRE over the other three ensemble pruning algorithms.  相似文献   
5.
In this research, we propose a novel framework referred to as collective game behavior decomposition where complex collective behavior is assumed to be generated by aggregation of several groups of agents following different strategies and complexity emerges from collaboration and competition of individuals. The strategy of an agent is modeled by certain simple game theory models with limited information. Genetic algorithms are used to obtain the optimal collective behavior decomposition based on history data. The trained model can be used for collective behavior prediction. For modeling individual behavior, two simple games, the minority game and mixed game are investigated in experiments on the real-world stock prices and foreign-exchange rate. Experimental results are presented to show the effectiveness of the new proposed model.  相似文献   
6.
The load applied to a machine tool feed drive changes during the machining process as material is removed. This load change alters the Coulomb friction of the feed drive. Because Coulomb friction accounts for a large part of the total friction the friction compensation control accuracy of the feed drives is limited if this nonlinear change in the applied load is not considered. This paper presents a new friction compensation method that estimates the machine tool load in real time and considers its effect on friction characteristics. A friction observer based on a Kalman filter with load estimation is proposed for friction compensation control considering the applied load change. A specially designed feed drive testbed that enables the applied load to be modified easily was constructed for experimental verification. Control performance and friction estimation accuracy are demonstrated experimentally using the testbed.  相似文献   
7.
Chaos optimization algorithm (COA) utilizes the chaotic maps to generate the pseudo-random sequences mapped as the decision variables for global optimization applications. A kind of parallel chaos optimization algorithm (PCOA) has been proposed in our former studies to improve COA. The salient feature of PCOA lies in its pseudo-parallel mechanism. However, all individuals in the PCOA search independently without utilizing the fitness and diversity information of the population. In view of the limitation of PCOA, a novel PCOA with migration and merging operation (denoted as MMO-PCOA) is proposed in this paper. Specifically, parallel individuals are randomly selected to be conducted migration and merging operation with the so far parallel solutions. Both migration and merging operation exchange information within population and produce new candidate individuals, which are different from those generated by stochastic chaotic sequences. Consequently, a good balance between exploration and exploitation can be achieved in the MMO-PCOA. The impacts of different one-dimensional maps and parallel numbers on the MMO-PCOA are also discussed. Benchmark functions and parameter identification problems are used to test the performance of the MMO-PCOA. Simulation results, compared with other optimization algorithms, show the superiority of the proposed MMO-PCOA algorithm.  相似文献   
8.
This paper is the second one of the two papers entitled “Weighted Superposition Attraction (WSA) Algorithm”, which is about the performance evaluation of the WSA algorithm in solving the constrained global optimization problems. For this purpose, the well-known mechanical design optimization problems, design of a tension/compression coil spring, design of a pressure vessel, design of a welded beam and design of a speed reducer, are selected as test problems. Since all these problems were formulated as constrained global optimization problems, WSA algorithm requires a constraint handling method for tackling them. For this purpose we have selected 6 formerly developed constraint handling methods for adapting into WSA algorithm and analyze the effect of the used constraint handling method on the performance of the WSA algorithm. In other words, we have the aim of producing concluding remarks over the performance and robustness of the WSA algorithm through a set of computational study in solving the constrained global optimization problems. Computational study indicates the robustness and the effectiveness of the WSA in terms of obtained results, reached level of convergence and the capability of coping with the problems of premature convergence, trapping in a local optima and stagnation.  相似文献   
9.
We present an optimization-based unsupervised approach to automatic document summarization. In the proposed approach, text summarization is modeled as a Boolean programming problem. This model generally attempts to optimize three properties, namely, (1) relevance: summary should contain informative textual units that are relevant to the user; (2) redundancy: summaries should not contain multiple textual units that convey the same information; and (3) length: summary is bounded in length. The approach proposed in this paper is applicable to both tasks: single- and multi-document summarization. In both tasks, documents are split into sentences in preprocessing. We select some salient sentences from document(s) to generate a summary. Finally, the summary is generated by threading all the selected sentences in the order that they appear in the original document(s). We implemented our model on multi-document summarization task. When comparing our methods to several existing summarization methods on an open DUC2005 and DUC2007 data sets, we found that our method improves the summarization results significantly. This is because, first, when extracting summary sentences, this method not only focuses on the relevance scores of sentences to the whole sentence collection, but also the topic representative of sentences. Second, when generating a summary, this method also deals with the problem of repetition of information. The methods were evaluated using ROUGE-1, ROUGE-2 and ROUGE-SU4 metrics. In this paper, we also demonstrate that the summarization result depends on the similarity measure. Results of the experiment showed that combination of symmetric and asymmetric similarity measures yields better result than their use separately.  相似文献   
10.
Steganography is the science of hiding secret message in an appropriate digital multimedia in such a way that the existence of the embedded message should be invisible to anyone apart from the sender or the intended recipient. This paper presents an irreversible scheme for hiding a secret image in the cover image that is able to improve both the visual quality and the security of the stego-image while still providing a large embedding capacity. This is achieved by a hybrid steganography scheme incorporates Noise Visibility Function (NVF) and an optimal chaotic based encryption scheme. In the embedding process, first to reduce the image distortion and to increase the embedding capacity, the payload of each region of the cover image is determined dynamically according to NVF. NVF analyzes the local image properties to identify the complex areas where more secret bits should be embedded. This ensures to maintain a high visual quality of the stego-image as well as a large embedding capacity. Second, the security of the secret image is brought about by an optimal chaotic based encryption scheme to transform the secret image into an encrypted image. Third, the optimal chaotic based encryption scheme is achieved by using a hybrid optimization of Particle Swarm Optimization (PSO) and Genetic Algorithm (GA) which is allowing us to find an optimal secret key. The optimal secret key is able to encrypt the secret image so as the rate of changes after embedding process be decreased which results in increasing the quality of the stego-image. In the extracting process, the secret image can be extracted from the stego-image losslessly without referring to the original cover image. The experimental results confirm that the proposed scheme not only has the ability to achieve a good trade-off between the payload and the stego-image quality, but also can resist against the statistics and image processing attacks.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号