共查询到20条相似文献,搜索用时 0 毫秒
1.
A method is presented for the joint source-channel coding optimization of a scheme based on the two-dimensional block cosine transform when the output of the encoder is to be transmitted via a memoryless binary symmetric channel. The authors' approach involves an iterative algorithm for the design of the quantizers (in the presence of channel errors) used for encoding the transform coefficients. This algorithm produces a set of locally optimum (in the mean-squared error sense) quantizers and the corresponding binary codeword assignment for the assumed transform coefficient statistics. To determine the optimum bit assignment among the transform coefficients, the authors have used an algorithm based on the steepest descent method, which, under certain convexity conditions on the performance of the channel-optimized quantizers, yields the optimal bit allocation. Simulation results for the performance of this locally optimum system over noisy channels have been obtained, and appropriate comparisons with a reference system designed for no channel errors have been made. It is shown that substantial performance improvements can be obtained by using this scheme. Furthermore, theoretically predicted results and rate distortion-theoretic bounds for an assumed two-dimensional image model are provided 相似文献
2.
3.
Wiener filters in canonical coordinates for transform coding,filtering, and quantizing 总被引:2,自引:0,他引:2
Canonical correlations are used to decompose the Wiener filter into a whitening transform coder, a canonical filter, and a coloring transform decoder. The outputs of the whitening transform coder are called canonical coordinates; these are the coordinates that are reduced in rank and quantized in our finite-precision version of the Gauss-Markov theorem. Canonical correlations are, in fact, cosines of the canonical, angles between a source vector and a measurement vector. They produce new formulas for error covariance, spectral flatness, and entropy 相似文献
4.
Effros M. Hanying Feng Zeger K. 《IEEE transactions on information theory / Professional Technical Group on Information Theory》2004,50(8):1605-1619
We examine the performance of the Karhunen-Loeve transform (KLT) for transform coding applications. The KLT has long been viewed as the best available block transform for a system that orthogonally transforms a vector source, scalar quantizes the components of the transformed vector using optimal bit allocation, and then inverse transforms the vector. This paper treats fixed-rate and variable-rate transform codes of non-Gaussian sources. The fixed-rate approach uses an optimal fixed-rate scalar quantizer to describe the transform coefficients; the variable-rate approach uses a uniform scalar quantizer followed by an optimal entropy code, and each quantized component is encoded separately. Earlier work shows that for the variable-rate case there exist sources on which the KLT is not unique and the optimal quantization and coding stage matched to a "worst" KLT yields performance as much as 1.5 dB worse than the optimal quantization and coding stage matched to a "best" KLT. In this paper, we strengthen that result to show that in both the fixed-rate and the variable-rate coding frameworks there exist sources for which the performance penalty for using a "worst" KLT can be made arbitrarily large. Further, we demonstrate in both frameworks that there exist sources for which even a best KLT gives suboptimal performance. Finally, we show that even for vector sources where the KLT yields independent coefficients, the KLT can be suboptimal for fixed-rate coding. 相似文献
5.
A robust quantizer is developed for encoding memoryless sources and transmission over the binary symmetric channel (BSC). The system combines channel optimized scalar quantization (COSQ) with all-pass filtering, the latter performed using a binary phase-scrambling/descrambling method. Applied to a broad class of sources, the robust quantizer achieves the same performance as the Gaussian COSQ for the memoryless Gaussian source. This quantizer is used in image coding for transmission over a BSC. The peak signal-to-noise ratio (PSNR) performance degrades gracefully as the channel bit error rate increases. 相似文献
6.
Jianping Pan 《IEEE transactions on image processing》1999,8(9):1175-1182
This paper introduces vector-scalar classification (VSC) for discrete cosine transform (DCT) coding of images. Two main characteristics of VSC differentiate it from previously proposed classification methods. First, pattern classification is effectively performed in the energy domain of the DCT subvectors using vector quantization. Second, the subvectors, instead of the DCT vectors, are mapped into a prescribed number of classes according to a pattern-to-class link established by scalar quantization. Simulation results demonstrate that the DCT coding systems based on VSC are superior to the other proposed DCT coding systems and are competitive compared to the best subband and wavelet coding systems reported in the literature. 相似文献
7.
We consider the joint source–channel coding problem of stereo video transmitted over AWGN and flat Rayleigh fading channels. Multiview coding (MVC) is used to encode the source, as well as a type of spatial scalable MVC. Our goal is to minimize the total number of bits, which is the sum of the number of source bits and the number of forward error correction bits, under the constraints that the quality of the left and right views must each be greater than predetermined PSNR thresholds at the receiver. We first consider symmetric coding, for which the quality thresholds are equal. Following binocular suppression theory, we also consider asymmetric coding, for which the quality thresholds are unequal. The optimization problem is solved using both equal error protection (EEP) and a proposed unequal error protection (UEP) scheme. An estimate of the expected end-to-end distortion of the two views is formulated for a packetized MVC bitstream over a noisy channel. The UEP algorithm uses these estimates for packet rate allocation. Results for various scenarios, including non-scalable/scalable MVC, symmetric/asymmetric coding, and UEP/EEP, are provided for both AWGN and flat Rayleigh fading channels. The UEP bit savings compared to EEP are given, and the performances of different scenarios are compared for a set of stereo video sequences. 相似文献
8.
Ayanoglu E. 《IEEE transactions on information theory / Professional Technical Group on Information Theory》1990,36(6):1450-1452
Problems in optimal multidimensional quantization of sources corrupted by noise are addressed. Expressions for the optimum quantizer values and the optimum quantization rule for the weighted squared error distortion measure are found and calculated for the Gaussian signal in additive independent Gaussian noise problem. Some properties of the optimum quantizer, and its relations with the optimal estimator for the general problem, are derived 相似文献
9.
This work addresses the problem of designing turbo codes for nonuniform binary memoryless or independent and identically distributed (i.i.d.) sources over noisy channels. The extrinsic information in the decoder is modified to exploit the source redundancy in the form of nonuniformity; furthermore, the constituent encoder structure is optimized for the considered nonuniform i.i.d. source to further enhance the system performance. Some constituent encoders are found to substantially outperform Berrou's (1996) (37, 21) encoder. Indeed, it is shown that the bit error rate (BER) performance of the newly designed turbo codes is greatly improved as significant coding gains are obtained 相似文献
10.
Demin Wang Liang Zhang André Vincent Filippo Speranza 《IEEE transactions on image processing》2006,15(8):2413-2421
The conventional two-dimensional wavelet transform used in existing image coders is usually performed through one-dimensional (1-D) filtering in the vertical and horizontal directions, which cannot efficiently represent edges and lines in images. The curved wavelet transform presented in this paper is carried out by applying 1-D filters along curves, rather than being restricted to vertical and horizontal straight lines. The curves are determined based on image content and are usually parallel to edges and lines in the image to be coded. The pixels along these curves can be well represented by a small number of wavelet coefficients. The curved wavelet transform is used to construct a new image coder. The code-stream syntax of the new coder is the same as that of JPEG2000, except that a new marker segment is added to the tile headers. Results of image coding and subjective quality assessment show that the new image coder performs better than, or as well as, JPEG2000. It is particularly efficient for images that contain sharp edges and can provide a PSNR gain of up to 1.67 dB for natural images compared with JPEG2000. 相似文献
11.
Hadamard transform image coding 总被引:1,自引:0,他引:1
The introduction of the fast Fourier transform algorithm has led to the development of the Fourier transform image coding technique whereby the two-dimensional Fourier transform of an image is transmitted over a channel rather than the image itself. This devlopement has further led to a related image coding technique in which an image is transformed by a Hadamard matrix operator. The Hadamard matrix is a square array of plus and minus ones whose rows and columns are orthogonal to one another. A high-speed computational algorithm, similar to the fast Fourier transform algorithm, which performs the Hadamard transformation has been developed. Since only real number additions and subtractions are required with the Hadamard transform, an order of magnitude speed advantage is possible compared to the complex number Fourier transform. Transmitting the Hadamard transform of an image rather than the spatial representation of the image provides a potential toleration to channel errors and the possibility of reduced bandwidth transmission. 相似文献
12.
The optimal linear block transform for coding images is well known to be the Karhunen-Loeve transformation (KLT). However, the assumption of stationarity in the optimality condition is far from valid for images. Images are composed of regions whose local statistics may vary widely across an image. While the use of adaptation can result in improved performance, there has been little investigation into the optimality of the criterion upon which the adaptation is based. In this paper we propose a new transform coding method in which the adaptation is optimal. The system is modular, consisting of a number of modules corresponding to different classes of the input data. Each module consists of a linear transformation, whose bases are calculated during an initial training period. The appropriate class for a given input vector is determined by the subspace classifier. The performance of the resulting adaptive system is shown to be superior to that of the optimal nonadaptive linear transformation. This method can also be used as a segmentor. The segmentation it performs is independent of variations in illumination. In addition, the resulting class representations are analogous to the arrangement of the directionally sensitive columns in the visual cortex. 相似文献
13.
14.
We introduce the eidochromatic transform as a tool for improved lossy coding of color images. Many current image-coding formats (such as JPEG 2000) utilize both a color-component transform (relating values of different image components at a single location) and a wavelet or other spatial transform (relating values of a single-image component at proximate, but different image locations). The eidochromatic transform further reduces redundancy by relating image values simultaneously across color components and in the two spatial dimensions. Our approach is to introduce an additional transform step following the color-component and spatial transforms. In tests, this step reduced the overall static entropy of the chrominance components of quantized transformed images by up to 40% or more. Combined with JPEG 2000's modeling and coding method, the eidochromatic transform was found to reduce the size of lossily coded color images by up to 27% overall. 相似文献
15.
Eizaburo Doi Doru C Balcan Michael S Lewicki 《IEEE transactions on image processing》2007,16(2):442-452
We address the problem of robust coding in which the signal information should be preserved in spite of intrinsic noise in the representation. We present a theoretical analysis for 1- and 2-D cases and characterize the optimal linear encoder and decoder in the mean-squared error sense. Our analysis allows for an arbitrary number of coding units, thus including both under- and over-complete representations, and provides insights into optimal coding strategies. In particular, we show how the form of the code adapts to the number of coding units and to different data and noise conditions in order to achieve robustness. We also present numerical solutions of robust coding for high-dimensional image data, demonstrating that these codes are substantially more robust than other linear image coding methods such as PCA, ICA, and wavelets. 相似文献
16.
Chi-Hsi Su Hsueh-Ming Hang Cho-Ho Wei 《Electronics letters》2000,36(4):311-312
A robust quantiser design for image coding is presented. The proposed quantiser can be viewed as the combination of compound of a quantiser, a variable length code (VLC) coder, and a channel coder. Simulation results show that our proposed scheme has a graceful distortion behaviour within the designed noise range 相似文献
17.
Because the Radon transform is a smoothing transform, any noise in the Radon data becomes magnified when the inverse Radon transform is applied. Among the methods used to deal with this problem is the wavelet-vaguelette decomposition (WVD) coupled with wavelet shrinkage, as introduced by Donoho (1995). We extend several results of Donoho and others here. First, we introduce a new sufficient condition on wavelets to generate a WVD. For a general homogeneous operator, whose class includes the Radon transform, we show that a variant of Donoho's method for solving inverse problems can be derived as the exact minimizer of a variational problem that uses a Besov norm as the smoothing functional. We give a new proof of the rate of convergence of wavelet shrinkage that allows us to estimate rather sharply the best shrinkage parameter needed to recover an image from noise-corrupted data. We conduct tomographic reconstruction computations that support the hypothesis that near-optimal shrinkage parameters can be derived if one can estimate only two Besov-space parameters about an image f. Both theoretical and experimental results indicate that our choice of shrinkage parameters yields uniformly better results than Kolaczyk's (1996) variant of Donoho's method and the classical filtered backprojection method. 相似文献
18.
A new design procedure for shape-gain vector quantizers (SGVQs) which leads to substantially improved robustness against channel errors without increasing the computational complexity is proposed. This aim is achieved by including the channel transition probabilities in the design procedure, leading to an improved assignment of binary codewords to the coding regions as well as a change of partition and centroids. In contrast to conventional design, negative gain values are also permitted. The new design procedure is applied to adaptive transform image coding. Simulation results are compared with those obtained by the conventional design procedure. The new algorithm is particularly useful for heavily distorted or fading channels 相似文献
19.
《Signal Processing Magazine, IEEE》2001,18(5):9-21
Discusses various aspects of transform coding, including: source coding, constrained source coding, the standard theoretical model for transform coding, entropy codes, Huffman codes, quantizers, uniform quantization, bit allocation, optimal transforms, transforms visualization, partition cell shapes, autoregressive sources, transform optimization, synthesis transform optimization, orthogonality and independence, and departures form the standard model 相似文献
20.
Aas K.C. Mullis C.T. 《IEEE transactions on information theory / Professional Technical Group on Information Theory》1996,42(4):1179-1192
Knowledge of the power spectrum of a stationary random sequence can be used for quantizing the signal efficiently and with minimum mean-squared error. A multichannel filter is used to transform the random sequence into an intermediate set of variables that are quantized using independent scalar quantizers, and then inverse-filtered, producing a quantized version of the original sequence. Equal word-length and optimal word-length quantization at high bit rates is considered. An analytical solution for the filter that minimizes the mean-squared quantization error is obtained in terms of its singular value decomposition. The performance is characterized by a set of invariants termed second-order modes, which are derived from the eigenvalue decomposition of the matrix-valued power spectrum. A more general rank-reduced model is used for decreasing distortion by introducing bias. The results are specialized to the case when the vector-valued time series is obtained from a scalar random sequence, which gives rise to a filter bank model for quantization. The asymptotic performance of such a subband coder is derived and shown to coincide with the asymptotic bound for transform coding. Quantization employing a single scalar pre- and postfilter, traditional transform coding using a square linear transformation, and subband coding in filter banks, arise as special cases of the structure analyzed here 相似文献