首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   84787篇
  免费   1109篇
  国内免费   406篇
电工技术   776篇
综合类   2317篇
化学工业   11717篇
金属工艺   4787篇
机械仪表   3056篇
建筑科学   2204篇
矿业工程   567篇
能源动力   1163篇
轻工业   3820篇
水利工程   1279篇
石油天然气   349篇
无线电   9302篇
一般工业技术   16466篇
冶金工业   2708篇
原子能技术   257篇
自动化技术   25534篇
  2023年   10篇
  2022年   28篇
  2021年   72篇
  2020年   50篇
  2019年   48篇
  2018年   14490篇
  2017年   13423篇
  2016年   10003篇
  2015年   646篇
  2014年   277篇
  2013年   300篇
  2012年   3232篇
  2011年   9498篇
  2010年   8355篇
  2009年   5608篇
  2008年   6848篇
  2007年   7834篇
  2006年   152篇
  2005年   1249篇
  2004年   1153篇
  2003年   1192篇
  2002年   560篇
  2001年   106篇
  2000年   187篇
  1999年   67篇
  1998年   82篇
  1997年   43篇
  1996年   52篇
  1995年   18篇
  1994年   22篇
  1993年   16篇
  1992年   17篇
  1991年   28篇
  1988年   14篇
  1969年   24篇
  1968年   43篇
  1967年   34篇
  1966年   42篇
  1965年   44篇
  1964年   13篇
  1963年   28篇
  1962年   23篇
  1961年   19篇
  1960年   31篇
  1959年   37篇
  1958年   38篇
  1957年   36篇
  1956年   34篇
  1955年   63篇
  1954年   68篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
991.
This work introduces and establishes a new model for cache management, where clients suggest preferences regarding their expectations for the time they are willing to wait, and the level of obsolescence they are willing to tolerate. The cache uses these preferences to decide upon entrance and exit of objects to and from its storage, and select the best copy of requested object among all available copies (fresh, cached, remote). We introduce three replacement policies, each evicts objects based on ongoing scores, considering users’ preferences combined with other objects’ properties such as size, obsolescence rate and popularity. Each replacement algorithm follows a different strategy: (a) an optimal solution that use dynamic programming approach to find the best objects to be kept (b) another optimal solution that use branch and bound approach to find the worst objects to be thrown out (c) an algorithm that use heuristic approach to efficiently select the objects to be evicted. Using these replacement algorithms the cache is able to keep the objects that are best suited for users preferences and dump the other objects. We compare our proposed algorithms to the Least-Recently-Used algorithm, and provide evidence to the advantages of our algorithms providing better service to cache’s users with less burden on network resources and reduced workloads on origin servers.  相似文献   
992.
Knowledge collaboration (KC) is an important strategy measure to improve knowledge management, focusing on not only efficiency of knowledge cooperation, but also adding value of intellectual capital and social capital. In virtual teams, many factors, such as team’s network characteristics, collaborative culture, and individual collaborative intention, affect the performance of KC. By discussing the nature of KC, this paper presents that the performance of can be measured from two aspects: effectiveness of collaboration and efficiency of cooperation. Among them, effectiveness of collaboration is measured through value added and efficiency of cooperation is measured through accuracy and timeliness. Then the paper discusses the factors affecting the performance of KC from network characteristics, individual attributes and team attributes. The results show that network characteristics, individual attributes and team attributes in virtual team have significant impacts on the performance of KC.  相似文献   
993.
In many practical situations, some of the attribute values for an object may be interval and set-valued. The interval and set-valued information systems have been introduced. According to the semantic relation of attribute values, interval and set-valued information systems can be classified into two categories, disjunctive (type 1) and conjunctive (type 2) systems. This paper mainly focuses on semantic interpretation of type 1. Then, a new fuzzy preference relation for interval and set-valued information systems is defined. Moreover, based on the new fuzzy preference relation, the concepts of fuzzy information entropy, fuzzy rough entropy, fuzzy knowledge granulation and fuzzy granularity measure are studied and relationships between entropy measures and granularity measures are investigated. Finally, an illustrative example to substantiate the theoretical arguments is given. These results may supply a further understanding of the essence of uncertainty in interval and set-valued information systems.  相似文献   
994.
Binary relations play an important role in rough set theory. This paper investigates the similarity of binary relations based on L-fuzzy topologies, where L is a boolean algebra. First, rough approximations based on a boolean algebra are proposed through successor neighborhoods on binary relations. Next, L-fuzzy topologies induced by binary relations are investigated. Finally, similarity of binary relations is introduced by using the L-fuzzy topologies and the fact that every binary relation is solely similar to some preorder relation is proved. It is worth mentioning that similarity of binary relations are both originated in the L-fuzzy topology and independent of the L-fuzzy topology.  相似文献   
995.
The classical operational law of uncertain variables proposed by Liu makes an important contribution to the development of the uncertainty theory in both theories and applications. It provides a powerful and practical approach for calculating the uncertainty distribution of strictly monotone function of uncertain variables. However, the restriction on strictly monotone functions of the operational law limits its applications since many practical problems cannot be modeled by strictly monotone functions but general monotone functions. Therefore, an extension of the original operational law is needed. For this purpose, some properties concerning the uncertainty distributions of monotone functions of uncertain variables as well as the generalized inverse uncertainty distributions are presented first in this paper. On the basis of these discussions, a generalized operational law is proposed as a natural extension of the original operational law. Then the uncertainty distribution of a general monotone function of independent regular uncertain variables can be derived, which is analogous to the way that suggested by the original operational law for dealing with strictly monotone functions. Furthermore, as an application of the generalized operational law, a theorem for calculating the expected values of general monotone functions of uncertain variables is presented as well.  相似文献   
996.
Differential evolution (DE) is a well-known optimization approach to deal with nonlinear and complex optimization problems. However, many real-world optimization problems are constrained problems that involve equality and inequality constraints. DE with constraint handling techniques, named constrained differential evolution (CDE), can be used to solve constrained optimization problems. In this paper, we propose a new CDE framework that uses generalized opposition-based learning (GOBL), named GOBL-CDE. In GOBL-CDE, firstly, the transformed population is generated using general opposition-based learning in the population initialization. Secondly, the transformed population and the initial population are merged and only half of the best individuals are selected to compose the new initial population to proceed mutation, crossover, and selection. Lastly, based on a jumping probability, the transformed population is calculated again after generating new populations, and the fittest individuals are selected to compose new population from the union of the current population and the transformed population. The GOBL-CDE framework can be applied to most CDE variants. As examples, in this study, the framework is applied to two popular representative CDE variants, i.e., rank-iMDDE and \(\varepsilon \)DEag. Experiment results on 24 benchmark functions from CEC’2006 and 18 benchmark functions from CEC’2010 show that the proposed framework is an effective approach to enhance the performance of CDE algorithms.  相似文献   
997.
MOEA/D is one of the promising evolutionary algorithms for multi- and many-objective optimization. To improve the search performance of MOEA/D, this work focuses on the solution update method in the conventional MOEA/D and proposes its alternative, the chain-reaction solution update. The proposed method is designed to maintain and improve the variable (genetic) diversity in the population by avoiding duplication of solutions in the population. In addition, the proposed method determines the order of existing solutions to be updated depending on the location of each offspring in the objective space. Furthermore, when an existing solution in the population is replaced by a new offspring, the proposed method tries to reutilize the existing solution for other search directions by recursively performing the proposed chain-reaction update procedure. This work uses discrete knapsack and continuous WFG4 problems with 2–8 objectives. Experimental results using knapsack problems show the proposed chain-reaction update contributes to improving the search performance of MOEA/D by enhancing the diversity of solutions in the objective space. In addition, experimental results using WFG4 problems show that the search performance of MOEA/D can be further improved using the proposed method.  相似文献   
998.
Web services, which can be described as functionality modules invoked over a network as part of a larger application are often used in software development. Instead of occasionally incorporating some of these services in an application, they can be thought of as fundamental building blocks that are combined in a process known as Web service composition. Manually creating compositions from a large number of candidate services is very time consuming, and developing techniques for achieving this objective in an automated manner becomes an active research field. One promising group of techniques encompasses evolutionary computing, which can effectively tackle the large search spaces characteristic of the composition problem. Therefore, this paper proposes the use of genetic programming for Web service composition, investigating three variations to ensure the creation of functionally correct solutions that are also optimised according to their quality of service. A variety of comparisons are carried out between these variations and two particle swarm optimisation approaches, with results showing that there is likely a trade-off between execution time and the quality of solutions when employing genetic programming and particle swarm optimisation. Even though genetic programming has a higher execution time for most datasets, the results indicate that it scales better than particle swarm optimisation.  相似文献   
999.
Scalability is a main and urgent problem in evolvable hardware (EHW) field. For the design of large circuits, an EHW method with a decomposition strategy is able to successfully find a solution, but requires a large complexity and evolution time. This study aims to optimize the decomposition on large-scale circuits so that it provides a solution for the EHW method to scalability and improves the efficiency. This paper proposes a projection-based decomposition (PD), together with Cartesian genetic programming (CGP) as an EHW system namely PD-CGP, to design relatively large circuits. PD gradually decomposes a Boolean function by adaptively projecting it onto the property of variables, which makes the complexity and number of sub-logic blocks minimized. CGP employs an evolutionary strategy to search for the simple and compact solutions of these sub-blocks. The benchmark circuits from the MCNC library, \(n\)-parity circuits, and arithmetic circuits are used in the experiment to prove the ability of PD-CGP in solving scalability and efficiency. The results illustrate that PD-CGP is superior to 3SD-ES in evolving large circuits in terms of complexity reduction. PD-CGP also outperforms GDD+GA in evolving relatively large arithmetic circuits. Additionally, PD-CGP successfully evolves larger \(n\)-even-parity and arithmetic circuits, which have not done by other approaches.  相似文献   
1000.
Minimal attribute reduction plays an important role in rough set. Heuristic algorithms are proposed in literature reviews to get a minimal reduction and yet an unresolved issue is that many redundancy non-empty elements involving duplicates and supersets exist in discernibility matrix. To be able to eliminate the related redundancy and pointless elements, in this paper, we propose a compactness discernibility information tree (CDI-tree). The CDI-tree has the ability to map non-empty elements into one path and allow numerous non-empty elements share the same prefix, which is recognized as a compact structure to store non-empty elements in discernibility matrix. A complete algorithm is presented to address Pawlak reduction based on CDI-tree. The experiment results reveal that the proposed algorithm is more efficient than the benchmark algorithms to find out a minimal attribute reduction.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号