首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   59402篇
  免费   7470篇
  国内免费   5987篇
电工技术   5212篇
技术理论   9篇
综合类   6531篇
化学工业   1388篇
金属工艺   693篇
机械仪表   3596篇
建筑科学   2455篇
矿业工程   1589篇
能源动力   794篇
轻工业   765篇
水利工程   1421篇
石油天然气   2918篇
武器工业   805篇
无线电   9047篇
一般工业技术   2864篇
冶金工业   1103篇
原子能技术   738篇
自动化技术   30931篇
  2024年   319篇
  2023年   859篇
  2022年   1754篇
  2021年   2174篇
  2020年   2320篇
  2019年   1794篇
  2018年   1602篇
  2017年   1991篇
  2016年   2235篇
  2015年   2550篇
  2014年   4078篇
  2013年   3660篇
  2012年   4351篇
  2011年   4705篇
  2010年   3630篇
  2009年   3567篇
  2008年   4050篇
  2007年   4587篇
  2006年   3875篇
  2005年   3500篇
  2004年   3010篇
  2003年   2539篇
  2002年   1932篇
  2001年   1483篇
  2000年   1265篇
  1999年   915篇
  1998年   709篇
  1997年   593篇
  1996年   488篇
  1995年   438篇
  1994年   349篇
  1993年   249篇
  1992年   175篇
  1991年   178篇
  1990年   134篇
  1989年   108篇
  1988年   85篇
  1987年   68篇
  1986年   59篇
  1985年   86篇
  1984年   58篇
  1983年   71篇
  1982年   60篇
  1981年   37篇
  1980年   21篇
  1979年   34篇
  1978年   12篇
  1977年   20篇
  1976年   13篇
  1959年   7篇
排序方式: 共有10000条查询结果,搜索用时 15 毫秒
991.
This paper analyzes the execution behavior of “No Random Accesses” (NRA) and determines the depths to which each sorted file is scanned in growing phase and shrinking phase of NRA respectively. The analysis shows that NRA needs to maintain a large quantity of candidate tuples in growing phase on massive data. Based on the analysis, this paper proposes a novel top-k algorithm Top-K with Early Pruning (TKEP) which performs early pruning in growing phase. General rule and mathematical analysis for early pruning are presented in this paper. The theoretical analysis shows that early pruning can prune most of the candidate tuples. Although TKEP is an approximate method to obtain the top-k result, the probability for correctness is extremely high. Extensive experiments show that TKEP has a significant advantage over NRA.  相似文献   
992.
993.
Increasingly large amount of multidimensional data are being generated on a daily basis in many applications. This leads to a strong demand for learning algorithms to extract useful information from these massive data. This paper surveys the field of multilinear subspace learning (MSL) for dimensionality reduction of multidimensional data directly from their tensorial representations. It discusses the central issues of MSL, including establishing the foundations of the field via multilinear projections, formulating a unifying MSL framework for systematic treatment of the problem, examining the algorithmic aspects of typical MSL solutions, and categorizing both unsupervised and supervised MSL algorithms into taxonomies. Lastly, the paper summarizes a wide range of MSL applications and concludes with perspectives on future research directions.  相似文献   
994.
Heterogeneous datasets arise naturally in most applications due to the use of a variety of sensors and measuring platforms. Such datasets can be heterogeneous in terms of the error characteristics and sensor models. Treating such data is most naturally accomplished using a Bayesian or model-based geostatistical approach; however, such methods generally scale rather badly with the size of dataset, and require computationally expensive Monte Carlo based inference. Recently within the machine learning and spatial statistics communities many papers have explored the potential of reduced rank representations of the covariance matrix, often referred to as projected or fixed rank approaches. In such methods the covariance function of the posterior process is represented by a reduced rank approximation which is chosen such that there is minimal information loss. In this paper a sequential Bayesian framework for inference in such projected processes is presented. The observations are considered one at a time which avoids the need for high dimensional integrals typically required in a Bayesian approach. A C++ library, gptk, which is part of the INTAMAP web service, is introduced which implements projected, sequential estimation and adds several novel features. In particular the library includes the ability to use a generic observation operator, or sensor model, to permit data fusion. It is also possible to cope with a range of observation error characteristics, including non-Gaussian observation errors. Inference for the covariance parameters is explored, including the impact of the projected process approximation on likelihood profiles. We illustrate the projected sequential method in application to synthetic and real datasets. Limitations and extensions are discussed.  相似文献   
995.
Electrical borehole wall images represent micro-resistivity measurements at the borehole wall. The lithology reconstruction is often based on visual interpretation done by geologists. This analysis is very time-consuming and subjective. Different geologists may interpret the data differently. In this work, linear discriminant analysis (LDA) in combination with texture features is used for an automated lithology reconstruction of ODP (Ocean Drilling Program) borehole 1203A drilled during Leg 197. Six rock groups are identified by their textural properties in resistivity data obtained by a Formation MircoScanner (FMS). Although discriminant analysis can be used for multi-class classification, non-optimal decision criteria for certain groups could emerge. For this reason, we use a combination of 2-class (binary) classifiers to increase the overall classification accuracy. The generalization ability of the combined classifiers is evaluated and optimized on a testing dataset where a classification rate of more than 80% for each of the six rock groups is achieved. The combined, trained classifiers are then applied on the whole dataset obtaining a statistical reconstruction of the logged formation. Compared to a single multi-class classifier the combined binary classifiers show better classification results for certain rock groups and more stable results in larger intervals of equal rock type.  相似文献   
996.
An open-source software including an easy-to-use graphical user interface (GUI) has been developed for processing, modeling and mapping of gravity and magnetic data. The program, called Potensoft, is a set of functions written in MATLAB. The most common application of Potensoft is spatial and frequency domain filtering of gravity and magnetic data. The GUI helps the user easily change all the required parameters. One of the major advantages of the program is to display the input and processed maps in a preview window, thereby allowing the user to track the results during the ongoing process. Source codes can be modified depending on the users' goals. This paper discusses the main features of the program and its capabilities are demonstrated by means of illustrative examples. The main objective is to introduce and ensure usage of the developed package for academic, teaching and professional purposes.  相似文献   
997.
Both the overhearing and overhearing avoidance in a densely distributed sensor network may inevitably incur considerable power consumption. In this paper we propose a so-called CCS-MAC (collaborative compression strategy-based MAC) MAC protocol which facilitates to exploit those overheard data that is treated useless in traditional MAC protocols for the purpose of cost and energy savings. Particularly the CCS-MAC enables different sensor nodes to perform data compression cooperatively with regard to those overheard data, so that the redundancy of data prepared for the link layer transmission can be totally eliminated at the earliest. The problem of collaborative compression is analyzed and discussed along with a corresponding linear programming model formulated. Based on it a heuristic node-selection algorithm with a time complexity of (O(N2)) is proposed to the solve the linear programming problem. The node-selection algorithm is implemented in CCS-MAC at each sensor node in a distributed manner. The experiment results verify that the proposed CCS-MAC scheme can achieve a significant energy savings so as to prolong the lifetime of the sensor networks so far.  相似文献   
998.
Schema integration aims to create a mediated schema as a unified representation of existing heterogeneous sources sharing a common application domain. These sources have been increasingly written in XML due to its versatility and expressive power. Unfortunately, these sources often use different elements and structures to express the same concepts and relations, thus causing substantial semantic and structural conflicts. Such a challenge impedes the creation of high-quality mediated schemas and has not been adequately addressed by existing integration methods. In this paper, we propose a novel method, named XINTOR, for automating the integration of heterogeneous schemas. Given a set of XML sources and a set of correspondences between the source schemas, our method aims to create a complete and minimal mediated schema: it completely captures all of the concepts and relations in the sources without duplication, provided that the concepts do not overlap. Our contributions are fourfold. First, we resolve structural conflicts inherent in the source schemas. Second, we introduce a new statistics-based measure, called path cohesion, for selecting concepts and relations to be a part of the mediated schema. The path cohesion is statistically computed based on multiple path quality dimensions such as average path length and path frequency. Third, we resolve semantic conflicts by augmenting the semantics of similar concepts with context-dependent information. Finally, we propose a novel double-layered mediated schema to retain a wider range of concepts and relations than existing mediated schemas, which are at best either complete or minimal, but not both. Performed on both real and synthetic datasets, our experimental results show that XINTOR outperforms existing methods with respect to (i) the mediated-schema quality using precision, recall, F-measure, and schema minimality; and (ii) the execution performance based on execution time and scale-up performance.  相似文献   
999.
Clustering of data in an uncertain environment can result into different partitions of the data at different points in time. Therefore, the initial formed clusters of non-stationary data can adapt over time which means that feature vectors associated with different clusters can follow different migration types to and from other clusters. This paper investigates different data migration types and proposes a technique to generate artificial non-stationary data which follows different migration types. Furthermore, the paper proposes clustering performance measures which are more applicable to measure the clustering quality in a non-stationary environment compared to the clustering performance measures for stationary environments. The proposed clustering performance measures in this paper are then used to compare the clustering results of three network based artificial immune models, since the adaptability and self-organising behaviour of the natural immune system inspired the modelling of network based artificial immune models for clustering of non-stationary data.  相似文献   
1000.
Almost all subspace clustering algorithms proposed so far are designed for numeric datasets. In this paper, we present a k-means type clustering algorithm that finds clusters in data subspaces in mixed numeric and categorical datasets. In this method, we compute attributes contribution to different clusters. We propose a new cost function for a k-means type algorithm. One of the advantages of this algorithm is its complexity which is linear with respect to the number of the data points. This algorithm is also useful in describing the cluster formation in terms of attributes contribution to different clusters. The algorithm is tested on various synthetic and real datasets to show its effectiveness. The clustering results are explained by using attributes weights in the clusters. The clustering results are also compared with published results.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号