首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 187 毫秒
1.
A progressive image transmission scheme based on iterative transform coding structure is proposed for application in interactive image communication over low-bandwidth channels. The scheme not only provides progressive transmission, but also guarantees lossless reproduction combined with a degree of compression. The image to be transmitted undergoes an orthogonal transform, and the transform coefficients are quantized (scalar or vector) before transmission. The novelty is that the residual error array due to quantization is iteratively fedback and requantized (scalar or vector); the coded residual error information is progressively transmitted and utilized in reconstructing the successive approximations. It is shown that the average reconstruction error variance converges to zero as the number of iterative stages approaches infinity. In practice, lossless reproduction can be achieved with a small number of iterations by using an entropy coder on the final residual-error image. Computer simulation results demonstrate the effectiveness of the technique  相似文献   

2.
Advances in residual vector quantization (RVQ) are surveyed. Definitions of joint encoder optimality and joint decoder optimality are discussed. Design techniques for RVQs with large numbers of stages and generally different encoder and decoder codebooks are elaborated and extended. Fixed-rate RVQs, and variable-rate RVQs that employ entropy coding are examined. Predictive and finite state RVQs designed and integrated into neural-network based source coding structures are revisited. Successive approximation RVQs that achieve embedded and refinable coding are reviewed. A new type of successive approximation RVQ that varies the instantaneous block rate by using different numbers of stages on different blocks is introduced and applied to image waveforms, and a scalar version of the new residual quantizer is applied to image subbands in an embedded wavelet transform coding system.  相似文献   

3.
4.
Recently deep learning has been introduced to the field of image compression. In this paper, we present a hybrid coding framework that combines entropy coding, deep learning, and traditional coding framework. In the base layer of the encoding, we use convolutional neural networks to learn the latent representation and importance map of the original image respectively. The importance map is then used to guide the bit allocation of the latent representation. A context model is also developed to help the entropy coding after the masked quantization. Another network is used to get a coarse reconstruction of the image in the base layer. The residual between the input and the coarse reconstruction is then obtained and encoded by the traditional BPG codec as the enhancement layer of the bit stream. We only need to train a basic model and the proposed scheme can realize image compression at different bit rates, thanks to the use of the traditional codec. Experimental results using the Kodak, Urban100 and BSD100 datasets show that the proposed scheme outperforms many deep learning-based methods and traditional codecs including BPG in MS-SSIM metric across a wide range of bit rates. It also exceeds some latest hybrid schemes in RGB444 domain on Kodak dataset in both PSNR and MS-SSIM metrics.  相似文献   

5.
6.
7.
We present a wavelet-based, high performance, hierarchical scheme for image matching which includes (1) dynamic detection of interesting points as feature points at different levels of subband images via the wavelet transform, (2) adaptive thresholding selection based on compactness measures of fuzzy sets in image feature space, and (3) a guided searching strategy for the best matching from coarse level to fine level. In contrast to the traditional parallel approaches which rely on specialized parallel machines, we explored the potential of distributed systems for parallelism. The proposed image matching algorithms were implemented on a network of workstation clusters using parallel virtual machine (PVM). The results show that our wavelet-based hierarchical image matching scheme is efficient and effective for object recognition.  相似文献   

8.
In this paper, we propose a coding algorithm for still images using vector quantization (VQ) and fractal approximation, in which low-frequency components of an input image are approximated by VQ, and its residual is coded by fractal mapping. The conventional fractal coding algorithms indirectly used the gray patterns of an original image with contraction mapping, whereas the proposed fractal coding method employs an approximated and then decimated image as a domain pool and uses its gray patterns. Thus, the proposed algorithm utilizes fractal approximation without the constraint of contraction mapping. For approximation of an original image, we employ the discrete cosine transform (DCT) rather than conventional polynomial-based transforms. In addition, for variable blocksize segmentation, we use the fractal dimension of a block that represents the roughness of the gray surface of a region. Computer simulations with several test images show that the proposed method shows better performance than the conventional fractal coding methods for encoding still pictures.  相似文献   

9.
研究了二代小波用于任意形状图像目标变换编码的可行性。文中采用一种需要对目标作适当的延拓后再作变换的方法,另一种则在修改二代小波提升算法的基础上获得一种不需延拓的形状自适应变换方法。实验结果表明,这二种方法与Katata使用的一代小波的方法相比,有更好的压缩性能。  相似文献   

10.
Fractal coding of subbands with an oriented partition   总被引:1,自引:0,他引:1  
We propose a new image compression scheme based on fractal coding of the coefficients of a wavelet transform, in order to take into account the self-similarity observed in each subband. The original image is first decomposed into subbands containing information in different spatial directions and at different scales, using an orthogonal wavelet-generated filter bank. Subbands are encoded using local iterated function systems (LIFS), with range and domain blocks presenting horizontal or vertical directionalities. Their sizes are defined according to the correlation lengths and resolution of each subband. The edge degradation and the blocking effects encountered at low bit-rates using conventional LIFS algorithm are reduced with this approach. The computation complexity is also greatly decreased by a 12:1 factor in comparison to fractal coding of the full resolution image. The proposed method is applied to standard test images. The comparison with other fractal coding approaches and with JPEG shows an important increase in terms of PPSNR/bit-rate. Especially for images presenting a privileged directionality, the use of adaptive partitions results in about 3 dB improvement in PPSNR. We also discuss the distorsion versus rate improvement obtained on high-frequency subbands when fractal coding instead of pyramidal vector quantization is used. Our approach achieves a real gain in PPSNR for low bit-rates between 0.3 and 1.2 bpp.  相似文献   

11.
刘美琴  赵耀 《电子学报》2010,38(3):658-663
本文在基于分形图像编码的多描述编码方案(MDFIC)的基础上,做了两处改进:一是引入提出的基于FGSE的快速分形图像编码算法,以提高编码速度;二是利用相邻值域块均值间的相关性,以减少比特率。实验结果表明,本方案与MDFIC相比,可以在保持解码图像质量几乎不变的情况下,提高编码速度和减少比特率,并具有较高的鲁棒性。  相似文献   

12.
基于帧间去相关的超光谱图像压缩方法   总被引:7,自引:1,他引:6  
针对超光谱图像的特点和硬件实现的实际需要,提出了一种基于小波变换的前向预测帧间去相关超光谱图像压缩算法。通过图像匹配和帧间去相关,消除超光谱图像帧间的冗余,对残差图像的压缩采用基于小波变换的快速位平面结合自适应算术编码的压缩算法,按照率失真准则控制输出码流,实现了对超光谱图像的高保真压缩。通过实验证明了该方案的有效性,基于小波变换的快速位平面结合自适应算术编码的压缩算法速度优于SPIHT,而且易于硬件实现。  相似文献   

13.
The paper presents an adaptive scheme for image-data compression. It is a region-based approach that suitably integrates two different approaches to image coding, vector quantization (VQ) and polynomial approximation (PA). The scheme is adaptive from the point of view of the human observer: the perceptually most significant areas are those near edges or details. In smoothed areas, PA can be used with notable results, but there VQ must be employed to ensure high fidelity. The two techniques exhibit a complementarity in both advantages and drawbacks. PA is not efficient in compressing high-frequency areas, but yields the best results when applied to highly correlated data. VQ is unable to reach high-compression ratios because of its low adaptability, but is quite suitable for compressing uncorrelated data. The means to achieve the integration of the two techniques is a control image containing information about edge and texture locations. In the paper, edge encoding and restoration are also addressed, which are closely related to the proposed hybrid scheme; block prediction is also utilized to further reduce the residual redundancy between VQ blocks. The exploitation of the best features of both approaches results in high compression factors, and in perceivable good quality. In particular, bit rates range from 0.15 to 0.07 bpp. Main applications of this compression scheme are in the areas of very-low bit rate image transmission and image archiving  相似文献   

14.
Image coding using dual-tree discrete wavelet transform   总被引:2,自引:0,他引:2  
In this paper, we explore the application of 2-D dual-tree discrete wavelet transform (DDWT), which is a directional and redundant transform, for image coding. Three methods for sparsifying DDWT coefficients, i.e., matching pursuit, basis pursuit, and noise shaping, are compared. We found that noise shaping achieves the best nonlinear approximation efficiency with the lowest computational complexity. The interscale, intersubband, and intrasubband dependency among the DDWT coefficients are analyzed. Three subband coding methods, i.e., SPIHT, EBCOT, and TCE, are evaluated for coding DDWT coefficients. Experimental results show that TCE has the best performance. In spite of the redundancy of the transform, our DDWT _ TCE scheme outperforms JPEG2000 up to 0.70 dB at low bit rates and is comparable to JPEG2000 at high bit rates. The DDWT _TCE scheme also outperforms two other image coders that are based on directional filter banks. To further improve coding efficiency, we extend the DDWT to an anisotropic dual-tree discrete wavelet packets (ADDWP), which incorporates adaptive and anisotropic decomposition into DDWT. The ADDWP subbands are coded with TCE coder. Experimental results show that ADDWP _ TCE provides up to 1.47 dB improvement over the DDWT _TCE scheme, outperforming JPEG2000 up to 2.00 dB. Reconstructed images of our coding schemes are visually more appealing compared with DWT-based coding schemes thanks to the directionality of wavelets.  相似文献   

15.
This article addresses the use of a joint source-channel coding strategy for enhancing the error resilience of images transmitted over a binary channel with additive Markov noise. In this scheme, inherent or residual (after source coding) image redundancy is exploited at the receiver via a maximum a posteriori (MAP) channel detector. This detector, which is optimal in terms of minimizing the probability of error, also exploits the larger capacity of the channel with memory as opposed to the interleaved (memoryless) channel. We first consider MAP channel decoding of uncompressed two-tone and bit-plane encoded grey-level images. Next, we propose a scheme relying on unequal error protection and MAP detection for transmitting grey-level images compressed using the discrete cosine transform (DCT), zonal coding, and quantization. Experimental results demonstrate that for various overall (source and channel) operational rates, significant performance improvements can be achieved over interleaved systems that do not incorporate image redundancy.  相似文献   

16.
We seek to evaluate the efficiency of hybrid transform/ DPCM interframe image coding relative to an optimal scheme that minimizes the mean-squared error in encoding a stationary Gaussian image sequence. The stationary assumption leads us to use the asymptotically optimal discrete Fourier transform (DFT) on the full frame of an image. We encode an actual image sequence with full-frame DFT/DPCM at several rates and compare it to previous interframe coding results with the same sequence. We also encode a single frame at these same rates using a full-frame DFT to demonstrate the inherent coding gains of interframe transform DPCM over intraframe coding. We then generate a pseudorandom image sequence with precise Gauss-Markov statistics and encode it by hybrid full-frame DFT/DPCM at various rates. We compare the signal-to-noise ratios (SNR's) of these reconstructions to the optimal ones calculated from the rate-distortion function. We conclude that in a medium rate range below 1 bit/pel/frame where reconstructions for hybrid transform/ DPCM may be unsatisfactory, there is enough margin for improvement to consider more sophisticated coding schemes.  相似文献   

17.
In this paper, we present a two-stage near-lossless compression scheme. It belongs to the class of "lossy plus residual coding" and consists of a wavelet-based lossy layer followed by arithmetic coding of the quantized residual to guarantee a given L(infinity) error bound in the pixel domain. We focus on the selection of the optimum bit rate for the lossy layer to achieve the minimum total bit rate. Unlike other similar lossy plus lossless approaches using a wavelet-based lossy layer, the proposed method does not require iteration of decoding and inverse discrete wavelet transform in succession to locate the optimum bit rate. We propose a simple method to estimate the optimal bit rate, with a theoretical justification based on the critical rate argument from the rate-distortion theory and the independence of the residual error.  相似文献   

18.
19.
一种高效的多描述容错编码方案   总被引:2,自引:0,他引:2  
针对JPEG2000标准所提供的容错编码工具的局限性,提出了把多描述编码运用于小波变换系数当中,产生图像的多个压缩编码描述的方案,该方案进一步提高压缩编码图像的容错性能。实验表明,此方案提供了更好的容错性能,非常适于网络中的图像传输。  相似文献   

20.
Recently deep learning-based methods have been applied in image compression and achieved many promising results. In this paper, we propose an improved hybrid layered image compression framework by combining deep learning and the traditional image codecs. At the encoder, we first use a convolutional neural network (CNN) to obtain a compact representation of the input image, which is losslessly encoded by the FLIF codec as the base layer of the bit stream. A coarse reconstruction of the input is obtained by another CNN from the reconstructed compact representation. The residual between the input and the coarse reconstruction is then obtained and encoded by the H.265/HEVC-based BPG codec as the enhancement layer of the bit stream. Experimental results using the Kodak and Tecnick datasets show that the proposed scheme outperforms the state-of-the-art deep learning-based layered coding scheme and traditional codecs including BPG in both PSNR and MS-SSIM metrics across a wide range of bit rates, when the images are coded in the RGB444 domain.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号