首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   3861篇
  免费   491篇
  国内免费   309篇
电工技术   236篇
综合类   408篇
化学工业   64篇
金属工艺   56篇
机械仪表   180篇
建筑科学   110篇
矿业工程   56篇
能源动力   89篇
轻工业   23篇
水利工程   69篇
石油天然气   31篇
武器工业   35篇
无线电   241篇
一般工业技术   220篇
冶金工业   167篇
原子能技术   8篇
自动化技术   2668篇
  2023年   11篇
  2022年   37篇
  2021年   38篇
  2020年   73篇
  2019年   104篇
  2018年   92篇
  2017年   76篇
  2016年   130篇
  2015年   127篇
  2014年   173篇
  2013年   292篇
  2012年   234篇
  2011年   322篇
  2010年   296篇
  2009年   306篇
  2008年   334篇
  2007年   345篇
  2006年   301篇
  2005年   246篇
  2004年   197篇
  2003年   188篇
  2002年   131篇
  2001年   91篇
  2000年   69篇
  1999年   68篇
  1998年   67篇
  1997年   37篇
  1996年   33篇
  1995年   37篇
  1994年   28篇
  1993年   26篇
  1992年   18篇
  1991年   12篇
  1990年   11篇
  1989年   16篇
  1988年   5篇
  1987年   8篇
  1986年   5篇
  1985年   8篇
  1984年   7篇
  1983年   10篇
  1982年   7篇
  1981年   6篇
  1980年   4篇
  1979年   7篇
  1978年   4篇
  1977年   3篇
  1959年   2篇
  1958年   2篇
  1957年   2篇
排序方式: 共有4661条查询结果,搜索用时 15 毫秒
71.
在处理海量数据集时,由于单台计算机的处理能力有限,利用传统的聚类算法难以在有效的时间内获得聚类结果。在基于密度和自适应密度可达聚类算法的基础上,提出一种并行聚类算法。理论和实验结果证明该算法具有接近线性的加速比,能够有效地处理大规模的数据集。  相似文献   
72.
手写签名认证是人工智能领域的一个重要课题,对它的深入研究具有重要的现实意义。文章采用基于概率统计的方法表示手写签名的速度特征,用vague集相似度表示书写压力特征,并通过细分的拐点表示笔画特征进行签名认证,最后,构造了一个在线手写签名认证系统。实验证明了该方法的有效性。  相似文献   
73.
Generalized rough sets over fuzzy lattices   总被引:2,自引:0,他引:2  
This paper studies generalized rough sets over fuzzy lattices through both the constructive and axiomatic approaches. From the viewpoint of the constructive approach, the basic properties of generalized rough sets over fuzzy lattices are obtained. The matrix representation of the lower and upper approximations is given. According to this matrix view, a simple algorithm is obtained for computing the lower and upper approximations. As for the axiomatic approach, a set of axioms is constructed to characterize the upper approximation of generalized rough sets over fuzzy lattices.  相似文献   
74.
This paper presents a new extension of Gaussian mixture models (GMMs) based on type-2 fuzzy sets (T2 FSs) referred to as T2 FGMMs. The estimated parameters of the GMM may not accurately reflect the underlying distributions of the observations because of insufficient and noisy data in real-world problems. By three-dimensional membership functions of T2 FSs, T2 FGMMs use footprint of uncertainty (FOU) as well as interval secondary membership functions to handle GMMs uncertain mean vector or uncertain covariance matrix, and thus GMMs parameters vary anywhere in an interval with uniform possibilities. As a result, the likelihood of the T2 FGMM becomes an interval rather than a precise real number to account for GMMs uncertainty. These interval likelihoods are then processed by the generalized linear model (GLM) for classification decision-making. In this paper we focus on the role of the FOU in pattern classification. Multi-category classification on different data sets from UCI repository shows that T2 FGMMs are consistently as good as or better than GMMs in case of insufficient training data, and are also insensitive to different areas of the FOU. Based on T2 FGMMs, we extend hidden Markov models (HMMs) to type-2 fuzzy HMMs (T2 FHMMs). Phoneme classification in the babble noise shows that T2 FHMMs outperform classical HMMs in terms of the robustness and classification rate. We also find that the larger area of the FOU in T2 FHMMs with uncertain mean vectors performs better in classification when the signal-to-noise ratio is lower.  相似文献   
75.
We present the software library STXXL that is an implementation of the C++ standard template library (STL) for processing huge data sets that can fit only on hard disks. It supports parallel disks, overlapping between disk I/O and computation and it is the first I/O‐efficient algorithm library that supports the pipelining technique that can save more than half of the I/Os. STXXL has been applied both in academic and industrial environments for a range of problems including text processing, graph algorithms, computational geometry, Gaussian elimination, visualization, and analysis of microscopic images, differential cryptographic analysis, etc. The performance of STXXL and its applications are evaluated on synthetic and real‐world inputs. We present the design of the library, how its performance features are supported, and demonstrate how the library integrates with STL. Copyright © 2007 John Wiley & Sons, Ltd.  相似文献   
76.
A new likelihood based AR approximation is given for ARMA models. The usual algorithms for the computation of the likelihood of an ARMA model require O(n) flops per function evaluation. Using our new approximation, an algorithm is developed which requires only O(1) flops in repeated likelihood evaluations. In most cases, the new algorithm gives results identical to or very close to the exact maximum likelihood estimate (MLE). This algorithm is easily implemented in high level quantitative programming environments (QPEs) such as Mathematica, MatLab and R. In order to obtain reasonable speed, previous ARMA maximum likelihood algorithms are usually implemented in C or some other machine efficient language. With our algorithm it is easy to do maximum likelihood estimation for long time series directly in the QPE of your choice. The new algorithm is extended to obtain the MLE for the mean parameter. Simulation experiments which illustrate the effectiveness of the new algorithm are discussed. Mathematica and R packages which implement the algorithm discussed in this paper are available [McLeod, A.I., Zhang, Y., 2007. Online supplements to “Faster ARMA Maximum Likelihood Estimation”, 〈http://www.stats.uwo.ca/faculty/aim/2007/faster/〉]. Based on these package implementations, it is expected that the interested researcher would be able to implement this algorithm in other QPEs.  相似文献   
77.
This paper discusses the problem of modeling on triangulated surfaces with geodesic curves. In the first part of the paper we define a new class of curves, called geodesic Bézier curves, that are suitable for modeling on manifold triangulations. As a natural generalization of Bézier curves, the new curves are as smooth as possible. In the second part we discuss the construction of C 0 and C 1 piecewise Bézier splines. We also describe how to perform editing operations, such as trimming, using these curves. Special care is taken to achieve interactive rates for modeling tasks. The third part is devoted to the definition and study of convex sets on triangulated surfaces. We derive the convex hull property of geodesic Bézier curves.
Luiz VelhoEmail:
  相似文献   
78.
The aim of this paper is to study the invariant and attracting sets of impulsive delay difference equations with continuous variables. Some criteria for the invariant and attracting sets are obtained by using the decomposition approach and delay difference inequalities with impulsive initial conditions.  相似文献   
79.
The covering generalized rough sets are an improvement of traditional rough set model to deal with more complex practical problems which the traditional one cannot handle. It is well known that any generalization of traditional rough set theory should first have practical applied background and two important theoretical issues must be addressed. The first one is to present reasonable definitions of set approximations, and the second one is to develop reasonable algorithms for attributes reduct. The existing covering generalized rough sets, however, mainly pay attention to constructing approximation operators. The ideas of constructing lower approximations are similar but the ideas of constructing upper approximations are different and they all seem to be unreasonable. Furthermore, less effort has been put on the discussion of the applied background and the attributes reduct of covering generalized rough sets. In this paper we concentrate our discussion on the above two issues. We first discuss the applied background of covering generalized rough sets by proposing three kinds of datasets which the traditional rough sets cannot handle and improve the definition of upper approximation for covering generalized rough sets to make it more reasonable than the existing ones. Then we study the attributes reduct with covering generalized rough sets and present an algorithm by using discernibility matrix to compute all the attributes reducts with covering generalized rough sets. With these discussions we can set up a basic foundation of the covering generalized rough set theory and broaden its applications.  相似文献   
80.
关联规则的快速提取算法   总被引:1,自引:0,他引:1  
针对基于频繁项集的关联规则挖掘算法效率低,需要多次扫描数据库且生成冗余候选项集问题,该文利用频繁项集的Aprior性质和概念格的基本思想提出一种关联规则提取算法,利用极大频繁项集来进行规则提取,去除了多数冗余的候选项集,提高了提取效率。  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号