首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到19条相似文献,搜索用时 203 毫秒
1.
用倍增技术在带有Wormhole路由技术的n×n二维网孔机器上提出了时间复杂度为O(log2n)的连通分量和传递闭包并行算法,并在此基础上提出了一个时间复杂度为O(log3n)的最小生成树并行算法.这些都改进了Store-and-Forward路由技术下的时间复杂度下界O(n).同其他运行在非总线连接分布式存储并行计算机上的算法相比,此连通分量和传递闭包算法的时间复杂度是最优的.  相似文献   

2.
本文讨论了动态矩形交查询算法.文中介绍了两个半动态矩形查询的新算法,它们分别基于一维数据结构和二维数据结构.一维查询算法的查询时间复杂度是O(logMk′),更新时间复杂度是O(logMlogn),空间复杂度是OnlogM/).二维查询算法的查询时间复杂度是O(log2Mk),更新时间复杂度是O(log2Mlogn),空间复杂度是Onlog2M).本文分别实现了这两个算法,通过对它们的性能进行比较,发现一维查询算法是一种高效、实用的算法.  相似文献   

3.
一种时间复杂度最优的精确串匹配算法   总被引:14,自引:2,他引:12  
现有的串匹配算法通常以模式长度作为滑动窗口大小.在窗口移动后,往往会丢弃掉一些已扫描正文的信息.提出了LDM(linear DAWG matching)串匹配算法,该算法将正文分为[n/m]个相互重叠、大小为2m-1的扫描窗口.在每个扫描窗口内,算法批量地尝试m个可能位置,首先使用反向后缀自动机从窗口中间位置向前扫描模式前缀;若成功,则再使用正向有限状态自动机从中间位置向后扫描剩余的模式后缀.分析证明,LDM算法的最差、最好、平均时间复杂度分别达到了理论最好结果:O(n),O(n/m),O(n(1ogσm)/m).实际性能测试也验证了平均时间复杂度最优这一理论结果.而且,对于在较大字母表下查找短模式的情况,LDM算法速度在被测试算法中最快.总之,LDM算法不但适合进行离线模式匹配,而且还特别适合需要进行在线高速匹配的应用.  相似文献   

4.
数据仓库系统中层次式Cube存储结构   总被引:11,自引:0,他引:11       下载免费PDF全文
高宏  李建中  李金宝 《软件学报》2003,14(7):1258-1266
区域查询是数据仓库上支持联机分析处理(on-line analytical processing,简称OLAP)的重要操作.近几年,人们提出了一些支持区域查询和数据更新的Cube存储结构.然而这些存储结构的空间复杂性和时间复杂性都很高,难以在实际中使用.为此,提出了一种层次式Cube存储结构HDC(hierarchical data cube)及其上的相关算法.HDC上区域查询的代价和数据更新代价均为O(logdn),综合性能为O((logn)2d)(使用CqCu模型)或O(K(logn)d)(使用Cqnq+Cunu模型).理论分析与实验表明,HDC的区域查询代价、数据更新代价、空间代价以及综合性能都优于目前所有的Cube存储结构.  相似文献   

5.
沈一飞  陈国良  张强锋 《软件学报》2007,18(11):2683-2690
分别在两种重要并行计算模型中给出计算有向基因组排列的反转距离新的并行算法.基于Hannenhalli和Pevzner理论,分3个主要部分设计并行算法:构建断点图、计算断点图中圈数、计算断点图中障碍的数目.在CREW-PRAM模型上,算法使用O(n2)处理器,时间复杂度为O(log2n);在基于流水光总线的可重构线性阵列系统(linear array with a reconfigurable pipelined bus system, LARPBS)模型上,算法使用O(n3)处理器,计算时间复杂度为O(logn).  相似文献   

6.
谢民主  陈建二  王建新 《软件学报》2007,18(9):2070-2082
个体单体型MSR(minimum SNP removal)问题是指如何利用个体的基因测序片断数据去掉最少的SNP(single-nucleotide polymorphisms)位点,以确定该个体单体型的计算问题.对此问题,Bafna等人提出了时间复杂度为O(2kn2m)的算法,其中,m为DNA片断总数,n为SNP位点总数,k为片断中洞(片断中的空值位点)的个数.由于一个Mate-Pair片段中洞的个数可以达到100,因此,在片段数据中有Mate-Pair的情况下,Bafna的算法通常是不可行的.根据片段数据的特点提出了一个时间复杂度为O((n-1)(k1-1)k222h+(k1+1)2h+nk2+mk1)的新算法,其中,k1为一个片断覆盖的最大SNP位点数(不大于n),k2为覆盖同一SNP位点的片段的最大数(通常不大于19),h为覆盖同一SNP位点且在该位点取空值的片断的最大数(不大于k2).该算法的时间复杂度与片断中洞的个数的最大值k没有直接的关系,在有Mate-Pair片断数据的情况下仍然能够有效地进行计算,具有良好的可扩展性和较高的实用价值.  相似文献   

7.
杨智应  雷向欣  朱洪 《软件学报》2006,17(10):2057-2062
算法的平滑复杂度能够更合理地反映算法的实际性能.在运行高斯算法求解线性系统过程中,矩阵条件数是导致求解误差偏大的一个因素.Sankar等人用0-保留高斯扰动进行对称矩阵条件数平滑分析.然而,Sankar等人给出的平滑复杂度过高而且复杂.为了解决这个问题,首先提出了两个关键的不等式;然后将这两个不等式用于对称矩阵条件教的平滑分析,得到更简单、更低的平滑复杂度;并利用该结果对高斯算法求解精度进行平滑分析,从而得到更低的平滑复杂度.  相似文献   

8.
确定任意多边形凸凹顶点的算法   总被引:21,自引:0,他引:21  
周培德 《软件学报》1995,6(5):276-279
本文提出一种确定任意多边形凸凹顶点的算法.该算法的时间复杂性为O(n2logn)次乘法和O(n2)次比较.  相似文献   

9.
网格多处理机的一种改进的子网分配算法   总被引:7,自引:0,他引:7  
张艳  孙世新  彭文钦 《软件学报》2001,12(8):1250-1257
子网分配问题是指识别并分配一个空闲的、满足指定大小要求的节点机.首先,提出了网格结构中一种新的具有O(N2a·1og2Na)时间复杂度的空闲子网搜索算法,它优于现有的O(N3a)时间复杂度的搜索算法.然后,用该算法对基于保留因子的最佳匹配类子网分配算法——RF(reservation factor)算法进行了改进,得到了  相似文献   

10.
在EREW PRAM(exclusive-read and exclusive-write parallel random access machine)并行计算模型上,对范围很广的一类无向图的边极大匹配问题,给出时间复杂性为O(logn),使用O((n+m)/logn)处理器的最佳、高速并行算法.  相似文献   

11.
With the first human DNA being decoded into a sequence of about 2.8 billion characters, much biological research has been centered on analyzing this sequence. Theoretically speaking, it is now feasible to accommodate an index for human DNA in the main memory so that any pattern can be located efficiently. This is due to the recent breakthrough on compressed suffix arrays, which reduces the space requirement from O(n log n) bits to O(n) bits. However, constructing compressed suffix arrays is still not an easy task because we still have to compute suffix arrays first and need a working memory of O(n log n) bits (i.e., more than 13 gigabytes for human DNA). This paper initiates the study of constructing compressed suffix arrays directly from the text. The main contribution is a construction algorithm that uses only O(n) bits of working memory, and the time complexity is O(n log n). Our construction algorithm is also time and space efficient for texts with large alphabets such as Chinese or Japanese. Precisely, when the alphabet size is |Σ|, the working space is O(n log |Σ|) bits, and the time complexity remains O(n log n), which is independent of |Σ|.  相似文献   

12.
Performing computations with a low-bit number representation results in a faster implementation that uses less silicon, and hence allows an algorithm to be implemented in smaller and cheaper processors without loss of performance. We propose a novel formulation to efficiently exploit the low (or non-standard) precision number representation of some computer architectures when computing the solution to constrained LQR problems, such as those that arise in predictive control. The main idea is to include suitably-defined decision variables in the quadratic program, in addition to the states and the inputs, to allow for smaller roundoff errors in the solver. This enables one to trade off the number of bits used for data representation against speed and/or hardware resources, so that smaller numerical errors can be achieved for the same number of bits (same silicon area). Because of data dependencies, the algorithm complexity, in terms of computation time and hardware resources, does not necessarily increase despite the larger number of decision variables. Examples show that a 10-fold reduction in hardware resources is possible compared to using double precision floating point, without loss of closed-loop performance.  相似文献   

13.
A matrix A of size m×n containing items from a totally ordered universe is termed monotone if, for every i, j, 1⩽i2. In case m=nϵ for some constant ϵ, (0<ϵ⩽1), our algorithm runs in O(log log n) time  相似文献   

14.
节点导纳矩阵是一个稀疏矩阵,短路电流计算需要对导纳矩阵数据进行查询。为了既能保持快速按行列查询元素数值,又进一步提高按数值查询其所在行列的效率,以便于存储调用及后续矩阵的处理,提出构建高度平衡二叉树的改进十字链表方法,即在十字链表存储的基础上,拓展存储数据结点指针域,形成平衡二叉树,将高度维持在(O(log2n)),平均查找长度也可维持在(O(log2n)),大大降低操作时间复杂度,提高数值查询效率。同时,为保证测试结果的公平性,把构建高度平衡二叉树的时间计入总时间,以进行对比。通过相应算例,验证了该改进方法的高效性。  相似文献   

15.
The minimization of a linear cost function subject to the condition that some matrix polynomials depending linearly on the decision variables are sums of squares of matrix polynomials (SOS) is known as SOS programming. This paper proposes an analysis of the complexity of SOS programming, in particular of the number of linear matrix inequality (LMI) scalar variables required for establishing whether a matrix polynomial is SOS. This number is analyzed for real and complex matrix polynomials, in the general case and in the case of some exact reductions achievable for some classes of matrix polynomials. An analytical formula is proposed in each case in order to provide this number as a function of the number of variables, degree and size of the matrix polynomials. Some tables reporting this number are also provided as reference for the reader. Two applications in control systems are presented in order to show the usefulness of the proposed results.  相似文献   

16.
针对压缩传感中高维投影计算采用稀疏性较差的普通随机测量矩阵,从而导致计算复杂度高,重构性能不佳这一难题,提出一种基于二分图邻接矩阵的压缩传感图像快速重建算法。该算法在满足测量矩阵的RIP条件下,充分利用二分图邻接矩阵的稀疏性与二值性,将时间复杂度由传统算法的O(N·logN)降低至O(N)。实验结果表明,算法在保证图像重构质量情况下大大提高了运算性能,尤其对于色彩(灰度)变化平缓图像,该算法性能更加优越。  相似文献   

17.
基于广义图像灰度共生矩阵的图像检索方法   总被引:9,自引:0,他引:9  
当图像发生旋转或大小改变显著时,用已有的基于灰度共生矩阵的图像检索方法,不能很好地给出检索结果,在此基础上,该文提出一种基于广义图像灰度共生矩阵的图像检索方法。该方法将原图像作平滑处理得到平滑图像,然后将原图像和平滑图像组合起来得到广义图像灰度共生矩阵,提取该矩阵的统计特征量后,将其组成向量并归一化后用于检索。该方法引入了图像的空间信息,对于图像旋转和尺寸变化均不敏感。实验结果与性能比较表明,新方法的效果优于单纯的灰度共生矩阵法。  相似文献   

18.
现有的均匀交换结构只能单独支持速率保证业务或单独支持尽力而为业务.本文研究如何利用均匀交换结构实现对多类型业务同时支持问题.提出采用速率分配和改变缓存队列组织结构的方法实现多类型业务的支持;针对该方法所产生的交叉节点缓存最大占用增大问题,提出两级均匀调度策略,并进一步对均匀速率调度算法做出改进,降低存储复杂度.通过理论分析和仿真实验证明了在支持多类型业务的情况下均匀交换结构的复杂度没有提升.  相似文献   

19.
QuickHeapsort is a combination of Quicksort and Heapsort. We show that the expected number of comparisons for QuickHeapsort is always better than for Quicksort if a usual median-of-constant strategy is used for choosing pivot elements. In order to obtain the result we present a new analysis for QuickHeapsort splitting it into the analysis of the partition-phases and the analysis of the heap-phases. This enables us to consider samples of non-constant size for the pivot selection and leads to better theoretical bounds for the algorithm. Furthermore, we introduce some modifications of QuickHeapsort. We show that for every input the expected number of comparisons is at most \(n\log _{2}n - 0.03n + o(n)\) for the in-place variant. If we allow n extra bits, then we can lower the bound to \( n\log _{2} n -0.997 n+ o (n)\). Thus, spending n extra bits we can save more that 0.96n comparisons if n is large enough. Both estimates improve the previously known results. Moreover, our non-in-place variant does essentially use the same number of comparisons as index based Heapsort variants and Relaxed-Weak-Heapsort which use \( n\log _{2}n -0.9 n+ o (n)\) comparisons in the worst case. However, index based Heapsort variants and Relaxed-Weak-Heapsort require \({\Theta }(n\log n)\) extra bits whereas we need n bits only. Our theoretical results are upper bounds and valid for every input. Our computer experiments show that the gap between our bounds and the actual values on random inputs is small. Moreover, the computer experiments establish QuickHeapsort as competitive with Quicksort in terms of running time.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号