首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
矢量量化的初始码书算法   总被引:2,自引:0,他引:2       下载免费PDF全文
矢量量化的初始码书设计是很重要的,影响或决定着其后码书形成算法的迭代次数和最终的码书质量。针对原有的初始码书算法在性能上随机性强与信源匹配程度不高的问题,提出一种对于训练矢量实施基于分量的和值排序,然后做分离平均的初始码书形成算法。算法使用了矢量的特征量,脱离了对于图像结构因数的依赖,能产生鲁棒性较好的初始码书。实验证明了该方法的有效性,与LBG算法结合可进一步提高码书质量。  相似文献   

2.
Recently, medical image compression becomes essential to effectively handle large amounts of medical data for storage and communication purposes. Vector quantization (VQ) is a popular image compression technique, and the commonly used VQ model is Linde–Buzo–Gray (LBG) that constructs a local optimal codebook to compress images. The codebook construction was considered as an optimization problem, and a bioinspired algorithm was employed to solve it. This article proposed a VQ codebook construction approach called the L2‐LBG method utilizing the Lion optimization algorithm (LOA) and Lempel Ziv Markov chain Algorithm (LZMA). Once LOA constructed the codebook, LZMA was applied to compress the index table and further increase the compression performance of the LOA. A set of experimentation has been carried out using the benchmark medical images, and a comparative analysis was conducted with Cuckoo Search‐based LBG (CS‐LBG), Firefly‐based LBG (FF‐LBG) and JPEG2000. The compression efficiency of the presented model was validated in terms of compression ratio (CR), compression factor (CF), bit rate, and peak signal to noise ratio (PSNR). The proposed L2‐LBG method obtained a higher CR of 0.3425375 and PSNR value of 52.62459 compared to CS‐LBG, FA‐LBG, and JPEG2000 methods. The experimental values revealed that the L2‐LBG process yielded effective compression performance with a better‐quality reconstructed image.  相似文献   

3.
针对LBG算法初始码本随机选取后易出现空胞腔、易陷入局部极小、迭代次数大等缺陷,本文依据模糊聚类理论引入了矢量量化码本设计训练的模糊聚类与LBG级联算法:先用模糊聚类算法训练码本,将训练得到的码本作为传统LBG算法的初始码本,再用传统LBG算法训练.论述了模糊聚类和LBG联合算法的原理与方法;用该算法分剐训练了语音线性...  相似文献   

4.
Traditional LBG algorithm is a pure iterative optimization procedure to achieve the vector quantization (VQ) codebook, where an initial codebook is continually refined at every iteration to reduce the distortion between code-vectors and a given training data set. However, such interactive type learning algorithms will easily direct final results converging toward the local optimization while the high quality of the initial codebook is not available. In this article, an efficient heuristic-based learning method, called novel particle swarm optimization (NPSO), is proposed to design the proper codebook of VQ scheme that can develop the image compression system. To improve the performance of the basic PSO, the centroid updating machine applies the one step-size gradient descent learning step in the heuristic learning procedure. Additionally, the presented NPSO with advantages of the centroid updating machine is proposed to quickly achieve the near-optimal reconstructive image. For demonstrating the proposed NPSO learning scheme, the image with several horizontal grey bars is first applied to present the efficiency of the NPSO learning mechanism. LBG and NPSO learning methods are also applied to test the reconstructing performance in several type images “Lena,” “Airplane,” “Cameraman”, and “peppers.” In our experiments, the NPSO learning algorithm provides the higher performance than conventional LBG methods in the application of building image compression system.  相似文献   

5.
The vector quantization (VQ) was a powerful technique in the applications of digital image compression. The traditionally widely used method such as the Linde–Buzo–Gray (LBG) algorithm always generated local optimal codebook. Recently, particle swarm optimization (PSO) is adapted to obtain the near-global optimal codebook of vector quantization. An alternative method, called the quantum particle swarm optimization (QPSO) had been developed to improve the results of original PSO algorithm. In this paper, we applied a new swarm algorithm, honey bee mating optimization, to construct the codebook of vector quantization. The results were compared with the other three methods that are LBG, PSO–LBG and QPSO–LBG algorithms. Experimental results showed that the proposed HBMO–LBG algorithm is more reliable and the reconstructed images get higher quality than those generated from the other three methods.  相似文献   

6.
许允喜  俞一彪 《计算机应用》2008,28(2):339-341,
矢量量化(VQ)方法是文本无关说话人识别中广泛应用的建模方法之一,它的主要问题是码本设计问题。语音特征参数是高维数据,样本分布复杂,因此码本设计的难度也很大,传统的LBG算法只能获得局部最优的码本。提出一种VQ码本设计的新方法,将小生境技术与K-均值算法融入到免疫算法训练过程中,形成混合免疫算法,采用针对高维数据聚类的改进变异算子,降低了随机变异的盲目性,增强群体的全局及局部搜索能力,同时通过接种疫苗提高算法的收敛速度。说话人识别实验表明,与传统LBG和基于混合遗传算法的VQ码本设计方法相比,该方法可以得到更优的模型参数,使得系统的识别率进一步提高。  相似文献   

7.
An important task of speaker verification is to generate speaker specific models and match an input speaker’s utterance with these models. This paper focuses on comparing the performance of text dependent speaker verification system using Mel Frequency Cepstral Coefficients feature and different Vector Quantization (VQ) based speaker modelling techniques to generate the speaker specific models. Speaker-specific information is mainly represented by spectral features and using these features we have developed the model which serves as an important entity for determining the claimed identity of the speaker. In the modelling part, we used Linde, Buzo, Gray (LBG) VQ, proposed adaptive LBG VQ and Fuzzy C Means (FCM) VQ for generating speaker specific model. The experimental results that are performed on microphonic database shows that accuracy significantly depends on the size of the codebook in all VQ techniques, and on FCM VQ accuracy also depend on the value of learning parameter of the objective function. Experiment results shows that how the accuracy of speaker verification system is depend on different representations of the codebook, different size of codebook in VQ modelling techniques and learning parameter in FCM VQ.  相似文献   

8.
采用遗传算法的文本无关说话人识别   总被引:1,自引:0,他引:1  
为解决在说话人识别方法的矢量量化(Vector Quantization,VQ)系统中,K-均值法的码本设计很容易陷入局部最优,而且初始码本的选取对最佳码本设计影响很大的问题,将遗传算法(Genetic Algorithm,GA)与基于非参数模型的VQ相结合,得到1种VQ码本设计的GA-K算法.该算法利用GA的全局优化能力得到最优的VQ码本,避免LBG算法极易收敛于局部最优点的问题;通过GA自身参数,结合K-均值法收敛速度快的优点,搜索出训练矢量空间中全局最优的码本.实验结果表明,GA-K算法优于LBG算法,可以很好地协调收敛性和识别率之间的关系.  相似文献   

9.
This article develops an evolutional fuzzy particle swarm optimization (FPSO) learning algorithm to self extract the near optimum codebook of vector quantization (VQ) for carrying on image compression. The fuzzy particle swarm optimization vector quantization (FPSOVQ) learning schemes, combined advantages of the adaptive fuzzy inference method (FIM), the simple VQ concept and the efficient particle swarm optimization (PSO), are considered at the same time to automatically create near optimum codebook to achieve the application of image compression. The FIM is known as a soft decision to measure the relational grade for a given sequence. In our research, the FIM is applied to determine the similar grade between the codebook and the original image patterns. In spite of popular usage of Linde–Buzo–Grey (LBG) algorithm, the powerful evolutional PSO learning algorithm is taken to optimize the fuzzy inference system, which is used to extract appropriate codebooks for compressing several input testing grey-level images. The proposed FPSOVQ learning scheme compared with LBG based VQ learning method is presented to demonstrate its great result in several real image compression examples.  相似文献   

10.
《Parallel Computing》2002,28(7-8):1079-1093
Vector quantization (VQ) is a widely used algorithm in speech and image data compression. One of the problems of the VQ methodology is that it requires large computation time especially for large codebook size. This paper addresses two issues. The first deals with the parallel construction of the VQ codebook which can drastically reduce the training time. A master/worker parallel implementation of a VQ algorithm is proposed. The algorithm is executed on the DM-MIMD Alex AVX-2 machine using a pipeline architecture. The second issue deals with the ability of accurately predicting the machine performance. Using communication and computation models, a comparison between expected and real performance is carried out. Results show that the two models can accurately predict the performance of the machine for image data compression. Analysis of metrics normally used in parallel realization is conducted.  相似文献   

11.
This paper evaluates the impact of three special forms of the Minkowski metric (Euclidean, City Block, and Chebychev distances) on the performance of the conventional vector quantization (VQ) and Gaussian mixture model (GMM) based closed-set text-independent speaker recognition systems, in terms of recognition rate and confidence on decisions. For the VQ based system, evaluations are carried out using the two most common clustering algorithms, LBG and K-means, and it is revealed which clustering algorithm and distance pair should be used to exploit the best attribute of both to achieve the best recognition rate for a given codebook size. In the case of GMM based system, we introduce the metrics into the GMM using a concatenation of the LBG and K-means algorithms in estimating the initial mean vectors, to which the system performance is sensitive, and explore their impact on system performance. We also make comparison of results obtained from evaluations on clean speech (TIMIT) and telephone speech databases (NTIMIT and NIST2001) with the modern classifiers VQ-UBM and GMM-UBM. It is found that there are cases where conventional VQ based system outperforms the modern systems. Moreover, the impact of distance metrics on the performance of the conventional and modern systems depends on the recognition task imposed (verification/identification).  相似文献   

12.
为了克服低速率声码器因清浊音硬判决、粗判决而导致解码语音有帧过渡等不自然感的缺陷,在分析比较目前主流声码器编码算法中激励参数提取和量化算法的基础上,将模糊数学中的隶属度概念引入语音子带清浊音描述中,提出了5维的浊音隶属度矢量概念,用于精细描述语音丰富的激励信息;介绍了浊音隶属度矢量的提取算法;提出了矢量量化码本的模糊聚类与LBG级联训练算法(F-LBG);用提取算法提取、建立了浊音隶属度码本的训练样本集,用F-LBG训练了浊音隶属度码本;将提取算法和F-LBG法训练得到的浊音隶属度码本分别应用于正弦激励声码器、混合激励声码器和同态声码器进行语音编、解码仿真;结果表明,用浊音隶属度矢量描述和合成语音激励信号的算法,具有较高的准确性和较强的噪声鲁棒性。  相似文献   

13.
一种高效体数据压缩算法及其在地震数据处理中的应用   总被引:2,自引:0,他引:2  
采用可编程图形硬件对大规模体数据进行直接体绘制时常常受到图形卡容量的限制,导致数据在内存与显存之间频繁交换,从而成为绘制的瓶颈.为此,提出一种大规模体数据矢量量化压缩算法.首先对体数据分块,并依据块内数据平均梯度值是否为0对该块进行分类;然后用3层结构表示梯度值非0的块,对其中次高层和最高层采用基于主分量分析分裂法产生初始码书,用LBG算法进行码书优化和量化,而对最低层以及梯度值为0的块采用定比特量化.实验结果表明,在保证较好图像重构质量的前提下,该算法可获得50倍以上的压缩比和更快的解压速度.  相似文献   

14.
基于方差归一化失真测度的改进的LBG算法   总被引:3,自引:1,他引:2  
矢量量化(VQ)技术在话者识别系统中得到了广泛的应用。 VQ码本的产生通常采用 LBG算法,失真测度则为对矢量的各分量等权重的欧氏距离。在话者识别系统中特征矢量的各个分量的分布是有差别的,且对于不同的话者,这种差别的程度又是不一样的。由于不同分布的各维参数对话者识别的有效性各不相同,因此,文章提出了一种能反映这种有效性差别的失真测度,即:方差归一化失真测度。以该失真测度为基础,并结合时序相关的初始码本设计方法及有效的零胞腔处理技术,文章提出了改进的LBG算法,同时利用该算法训练出改进的VQ话者模型,并进行了话者识别实验。  相似文献   

15.
针对基于主分量分析和遗传算法的码书设计算法中当码书大小超过64时码书性能下降的问题,提出了一种改进的码书设计算法.首先采用主分量分析对训练矢量降维以减少计算复杂度,然后利用遗传算法的全局优化能力计算得到接近全局最优的码书.实验结果表明,与原算法和经典的LBG算法相比,文中算法所生成的码书性能有了明显提高,而且计算时间也少于LBG算法.  相似文献   

16.
有序抖动半调图像压缩算法   总被引:1,自引:0,他引:1  
刘欣  耿烨  李智杰 《计算机应用》2011,31(1):154-155
结合有序抖动半调图像特性,针对原有矢量量化初始码书算法存在非典型码字、运算量大和空间分布不均匀等欠缺,提出一种能在训练集中均匀分布的初始码书生成方法,应用到LBG算法中,并结合无损压缩方法,进行二次压缩。实验表明,本方法压缩比高、实现简单,且解码后图像视觉效果得到明显改善。  相似文献   

17.
We address the problem of speech compression at very low rates, with the short-term spectrum compressed to less than 20 bits per frame. Current techniques apply structured vector quantization (VQ) to the short-term synthesis filter coefficients to achieve rates of the order of 24 to 26 bits per frame. In this paper we show that temporal correlations in the VQ index stream can be introduced by dynamic codebook ordering, and that these correlations can be exploited by lossless coding approaches to reduce the number of bits per frame of the VQ scheme. The use of lossless coding ensures that no additional distortion is introduced, unlike other interframe techniques. We then detail two constructive algorithms which are able to exploit this redundancy. The first method is a delayed-decision approach, which dynamically adapts the VQ codebook to allow for efficient entropy coding of the index stream. The second is based on a vector subcodebook approach and does not incur any additional delay. Experimental results are presented for both methods to validate the approach.  相似文献   

18.
一种基于小波变换的图像压缩方法   总被引:8,自引:0,他引:8  
提出一种基于小波变换的灰度图像数据压缩编码方法,基本思路是利用小波变换实现图像的多分辨分解,用矢量量化(VQ)对分解后的图像进行编码,在矢量量化LBG算法的初始码书的选取中根据矢量中各分量的特性提出一种改进的随机选取法,避免了可能的胞腔不均现象,提高了码书的质量,而且重构的图像质量也有所提高。  相似文献   

19.
A self-organizing map (SOM) approach for vector quantization (VQ) over wireless channels is presented. We introduce a soft decoding SOM-based robust VQ (RVQ) approach with performance comparable to that of the conventional channel optimized VQ (COVQ) approach. In particular, our SOM approach avoids the time-consuming index assignment process in traditional RVQs and does not require a reliable feedback channel for COVQ-like training. Simulation results show that our approach can offer potential performance gain over the conventional COVQ approach. For data sources with Gaussian distribution, the gain of our approach is demonstrated to be in the range of 1–4 dB. For image data, our approach gives a performance comparable to a sufficiently trained COVQ, and is superior with a similar number of training epoches. To further improve the performance, a SOM–based COVQ approach is also discussed.  相似文献   

20.
The conventional channel-optimized vector quantization (COVQ) is very powerful in the protection of vector quantization (VQ) data over noisy channels. However, it suffers from the time consuming training process. A soft decoding self-organizing map (SOM) approach for VQ over noisy channels is presented. Compared with the COVQ approach, it does not require a long training time. For AWGN and fading channels, the distortion of the proposed approach is comparable to that of COVQ. Simulation confirmed that our proposed approach is a fast and practical method for VQ over noisy channels.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号