首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   10878篇
  免费   726篇
  国内免费   566篇
电工技术   482篇
综合类   811篇
化学工业   171篇
金属工艺   117篇
机械仪表   553篇
建筑科学   216篇
矿业工程   67篇
能源动力   183篇
轻工业   55篇
水利工程   70篇
石油天然气   47篇
武器工业   52篇
无线电   923篇
一般工业技术   809篇
冶金工业   119篇
原子能技术   29篇
自动化技术   7466篇
  2024年   43篇
  2023年   74篇
  2022年   110篇
  2021年   111篇
  2020年   174篇
  2019年   169篇
  2018年   135篇
  2017年   269篇
  2016年   278篇
  2015年   372篇
  2014年   504篇
  2013年   732篇
  2012年   607篇
  2011年   751篇
  2010年   490篇
  2009年   749篇
  2008年   812篇
  2007年   831篇
  2006年   747篇
  2005年   600篇
  2004年   547篇
  2003年   492篇
  2002年   420篇
  2001年   333篇
  2000年   293篇
  1999年   221篇
  1998年   217篇
  1997年   198篇
  1996年   130篇
  1995年   101篇
  1994年   95篇
  1993年   71篇
  1992年   72篇
  1991年   42篇
  1990年   34篇
  1989年   40篇
  1988年   44篇
  1987年   25篇
  1986年   22篇
  1985年   32篇
  1984年   48篇
  1983年   27篇
  1982年   21篇
  1981年   21篇
  1980年   18篇
  1979年   12篇
  1978年   8篇
  1977年   11篇
  1976年   9篇
  1975年   4篇
排序方式: 共有10000条查询结果,搜索用时 12 毫秒
51.
In this paper we present an efficient technique for piecewise cubic Bézier approximation of digitized curve. An adaptive breakpoint detection method divides a digital curve into a number of segments and each segment is approximated by a cubic Bézier curve so that the approximation error is minimized. Initial approximated Bézier control points for each of the segments are obtained by interpolation technique i.e. by the reverse recursion of De Castaljau's algorithm. Two methods, two-dimensional logarithmic search algorithm (TDLSA) and an evolutionary search algorithm (ESA), are introduced to find the best-fit Bézier control points from the approximate interpolated control points. ESA based refinement is proved to be better experimentally. Experimental results show that Bézier approximation of a digitized curve is much more accurate and uses less number of points compared to other approximation techniques.  相似文献   
52.
The concept of combinatorial objects is formalized. It allows strict definition of a combinatorial optimization problem (COP). An efficient metaheuristic method to solve COPs (H-method) is considered. It includes stochastic local search algorithms as a built-in procedure. A parallel implementation of the H-method is set forth and analyzed. The results from a numerical experiment and solution of some well-known COPs on personal computers and on the SKIT cluster supercomputer are presented. The study was supported by INTAS (Project 06-1000017-8909). __________ Translated from Kibernetika i Sistemnyi Analiz, No. 6, pp. 70–79, November–December 2007.  相似文献   
53.
Optimal implementations of UPGMA and other common clustering algorithms   总被引:2,自引:0,他引:2  
In this work we consider hierarchical clustering algorithms, such as UPGMA, which follow the closest-pair joining scheme. We survey optimal O(n2)-time implementations of such algorithms which use a ‘locally closest’ joining scheme, and specify conditions under which this relaxed joining scheme is equivalent to the original one (i.e. ‘globally closest’).  相似文献   
54.
In this paper, we present the results of an experiment in which a collection of simulated robots that have been evolved for the ability to solve a collective navigation problem develop a communication system that allows them to co-operate better. The analysis of the results obtained indicates how evolving robots develop a non-trivial communication system and exploit different communication modalities. The results also indicate how the possibility of co-adapting the robots’ individual and social/communicative behaviour plays a key role in the development of progressively more complex and effective individuals.  相似文献   
55.
Intensity modulated radiation therapy (IMRT) is one of the most effective modalities for modern cancer treatment. The key to successful IMRT treatment hinges on the delivery of a two-dimensional discrete radiation intensity matrix using a device called a multileaf collimator (MLC). Mathematically, the delivery of an intensity matrix using an MLC can be viewed as the problem of representing a non-negative integral matrix (i.e., the intensity matrix) by a linear combination of certain special non-negative integral matrices called segments, where each such segment corresponds to one of the allowed states of the MLC. The problem of representing the intensity matrix with the minimum number of segments is known to be NP-complete. In this paper, we present two approximation algorithms for this matrix representation problem. To the best of our knowledge, these are the first algorithms to achieve non-trivial performance guarantees for multi-row intensity matrices.  相似文献   
56.
Real-time homogenous translucent material editing   总被引:4,自引:0,他引:4  
This paper presents a novel method for real-time homogenous translucent material editing under fixed illumination. We consider the complete analytic BSSRDF model proposed by Jensen et al. [ [JMLH01] ], including both multiple scattering and single scattering. Our method allows the user to adjust the analytic parameters of BSSRDF and provides high-quality, real-time rendering feedback. Inspired by recently developed Precomputed Radiance Transfer (PRT) techniques, we approximate both the multiple scattering diffuse reflectance function and the single scattering exponential attenuation function in the analytic model using basis functions, so that re-computing the outgoing radiance at each vertex as parameters change reduces to simple dot products. In addition, using a non-uniform piecewise polynomial basis, we are able to achieve smaller approximation error than using bases adopted in previous PRT-based works, such as spherical harmonics and wavelets. Using hardware acceleration, we demonstrate that our system generates images comparable to [ [JMLH01] ]at real-time frame-rates.  相似文献   
57.
An actor-critic algorithm for constrained Markov decision processes   总被引:2,自引:0,他引:2  
An actor-critic type reinforcement learning algorithm is proposed and analyzed for constrained controlled Markov decision processes. The analysis uses multiscale stochastic approximation theory and the envelope theorem' of mathematical economics.  相似文献   
58.
Lee and Batcher have designed networks that efficiently merge k separately provided sorted sequences of known lengths totalling n. We show that the design is still possible, and in fact easier to describe, if we do not make use of the lengths, or even the directions of monotonicity, of the individual sequences—the sequences can be provided in a single undelimited concatenation of length n. The depth of the simplest resulting network to sort sequences that are “k-tonic” and of length n is , generalizing Batcher's 1968 results for the extreme values of k (k=2 corresponding to merging, and k=n/2 corresponding to general sorting).The exposition is self-contained and can serve even as an introduction to sorting networks and Batcher's results.  相似文献   
59.
W.A.  H.J. 《Pattern recognition》1995,28(12):1985-1992
A fast digital Radon transform based on recursively defined digital straight lines is described, which has the sequential complexity of N2 log N additions for an N × N image. This transform can be used to evaluate the Hough transform to detect straight lines in a digital image. Whilst a parallel implementation of the Hough transform algorithm is difficult because of global memory access requirements, the fast digital Radon transform is vectorizable and therefore well suited for parallel computation. The structure of the fast algorithm is shown to be quite similar to the FFT algorithm for decimation in frequency. It is demonstrated that even for sequential computation the fast Radon transform is an attractive alternative to the classical Hough transform algorithm.  相似文献   
60.
Young 《Algorithmica》2002,33(3):371-383
Abstract. Consider the following file caching problem: in response to a sequence of requests for files, where each file has a specified size and retrieval cost , maintain a cache of files of total size at most some specified k so as to minimize the total retrieval cost. Specifically, when a requested file is not in the cache, bring it into the cache and pay the retrieval cost, and remove other files from the cache so that the total size of files remaining in the cache is at most k . This problem generalizes previous paging and caching problems by allowing objects of arbitrary size and cost, both important attributes when caching files for world-wide-web browsers, servers, and proxies. We give a simple deterministic on-line algorithm that generalizes many well-known paging and weighted-caching strategies, including least-recently-used, first-in-first-out, flush-when-full, and the balance algorithm. On any request sequence, the total cost incurred by the algorithm is at most k/(k-h+1) times the minimum possible using a cache of size h ≤ k . For any algorithm satisfying the latter bound, we show it is also the case that for most choices of k , the retrieval cost is either insignificant or at most a constant (independent of k ) times the optimum. This helps explain why competitive ratios of many on-line paging algorithms have been typically observed to be constant in practice.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号