共查询到20条相似文献,搜索用时 578 毫秒
1.
应用于多视点视频压缩的多参考B帧快速编码算法 总被引:2,自引:0,他引:2
基于H.264平台就多视点视频压缩中多参考帧技术存在计算复杂度高的现状,根据多视点视频序列特性,提出了一种针对多参考B帧的快速多视点视频编码压缩算法。新算法改进了多参考帧列表策略,并结合多视点视频序列的时空域相关性特性,提出了适用于多视点视频编码的参考帧列表调整策略。此外,还针对多种块模式编码引入了块模式预判策略。新算法在保证率失真性能的基础上,有效地降低了计算复杂度,提高了多视点视频编码速度。虽然本文是基于M-Picture编码结构实现了新算法,但它同样适用于其它应用多参考B帧预测技术的多视点视频编码结构中。 相似文献
2.
相对于以前其他的标准,H.264标准在分层编码、帧内/帧间预测编码、多帧参考、预测精度等技术方面做了巨大的改进。因此,在TMS320DM642平台上实现H.264基档次编码器的移植与优化显得格外实用和必要。基于对DSP平台的结构特性和H.264的计算复杂度分析,主要从以下3个方面对H.264编码器进行了优化:核心算法、数据传输和存储器/Cache使用。实验结果表明,对于CIF格式的视频序列,最优化后的H.264编码器能够达到每秒高于24帧的编码速度,满足了视频处理对于实时性的要求。 相似文献
3.
4.
Side-by-Side(SbS)是立体视频常用的格式。传统的视频编码方法利用帧内预测或帧间预测对视频进行压缩编码。基于H.264视频压缩编码标准,针对SbS格式的立体视频提出了一种基于多参考帧方式的视间预测编码方法,并在JM测试代码基础上将其实现。实验结果表明,加入视间预测编码后,以很小的PSNR损失为代价,使得SbS视频的压缩效率提高了大约10%~20%。 相似文献
5.
Side-by-Side(SbS)是立体视频常用的格式。传统的视频编码方法利用帧内预测或帧间预测对视频进行压缩编码。本文基于H.264视频压缩编码标准,针对SbS格式的立体视频提出了一种基于多参考帧方式的视间预测编码方法,并在JM测试代码基础上将其实现。实验结果表明,加入视间预测编码后,以很小的PSNR损失为代价,使得SbS视频的压缩效率提高了大约10%至20%。 相似文献
6.
7.
预测结构是多视点视频编码(Multi-View Video Coding,MVC)研究的主要内容之一。MVC目前采用HHI(Heinrich-Hertz-Institute)提出的分层次B帧预测结构(HBP),比联播预测结构获得了更好的压缩效率。分析了多种预测结构,并针对平行摄像机采集的多视点视频序列,提出了一种新的预测结构AS_EIPP,该结构充分利用相邻视点间的相关性和多参考帧模式,进一步提高了压缩效率。在多视点视频软件测试平台JMVC8.3上进行验证,实验结果表明:新的预测结构在保证重建视频质量基本不变的前提下,压缩效率比HBP预测结构提高了1%~4%。 相似文献
8.
介绍了中国音视频编码技术标准AVS中的帧内预测和基于宏块的帧场自适应编码技术.通过分析其中的5种帧内预测模式的空间分布特点,提出了一种新的隔行视频帧内预测模式.实验表明该模式可以在标清、高清隔行视频的帧内预测上获得平均0.24 dB以上的信噪比增益,平均减少3.1%的码流,是一种有效提高编码效率的预测模式. 相似文献
9.
在H.264/AVC视频压缩标准采用的几项关键技术中,使用多参考帧预测可以增加最佳匹配块的检索概率,进而大大提高了编码效率.受B帧直接( direct)预测模式的启发,提出了一种新的基于扩展帧的多参考帧预测方法,由多参考帧中的共同位置块及其参考块扩展得到一个新的抽取帧,增加了原始序列的时域分辨率,使得扩展帧更加接近当前帧,提高了运动估计中最佳匹配的检索概率,进而提升了编码效率.仿真结果证实该方法的编码性能好于H.264/AVC参考软件. 相似文献
10.
林琪赵海武王国中滕国伟李国平 《电视技术》2014,(17):76-79
新一代具有自主知识产权的视频编码标准AVS2正在制定当中,其相对于上一代标准而言具有更高的编码效率,但由于使用了多划分模式、多参考帧等技术,其编码复杂度也大幅增加。为了提高AVS2帧间编码速度,提出了一种快速帧间预测算法,在使用非对称运动划分(Asymmetric Motion Partitions,AMP)快速决策算法的基础上,深度挖掘同一编码单元下相邻预测单元之间的相关性,利用邻近预测单元来预测当前块的参考帧,通过判断预测单元的运动矢量残差的曼哈顿距离,以确定最佳参考帧,从而减小参与计算的参考帧数目。实验结果表明,该算法在图像质量和比特率基本保持不变的情况下,有效地提高了AVS2的编码速度。 相似文献
11.
Yu-Chen SunChun-Jen Tsai 《Journal of Visual Communication and Image Representation》2012,23(3):535-548
In this paper, we propose a perceptual-based distributed video coding (DVC) technique. Unlike traditional video codecs, DVC applies video prediction process at the decoder side using previously received frames. The predicted video frames (i.e., side information) contain prediction errors. The encoder then transmits error-correcting parity bits to the decoder to reconstruct the video frames from side information. However, channel codes based on i.i.d. noise models are not always efficient in correcting video prediction errors. In addition, some of the prediction errors do not cause perceptible visual distortions. From perceptual coding point of view, there is no need to correct such errors. This paper proposes a scheme for the decoder to perform perceptual quality analysis on the predicted side information. The decoder only requests parity bits to correct visually sensitive errors. More importantly, with the proposed technique, key frames can be encoded at higher rates while still maintaining consistent visual quality across the video sequence. As a result, even the objective PSNR measure of the decoded video sequence will increase too. Experimental results show that the proposed technique improves the R-D performance of a transform domain DVC codec both subjectively and objectively. Comparisons with a well-known DVC codec show that the proposed perceptual-based DVC coding scheme is very promising for distributed video coding framework. 相似文献
12.
A dynamic bandwidth allocation strategy to support variable bit rate (VBR) video traffic is proposed. This strategy predicts the bandwidth requirements for future frames using adaptive linear prediction that minimizes the mean square error. The adaptive technique does not require any prior knowledge of the traffic statistics nor assume stationarity. Analyses using six one-half-hour video tracts indicate that prediction errors for the bandwidth required for the next frames and group of pictures (GOP) are almost white noise or short memory. The performance of the strategy is studied using renegotiated constant bit rate (RCBR) network service model and methods that control the tradeoff between the number of renegotiations and network utilization are proposed. Simulation results using MPEG-I video traces for predicting GOP rates show that the queue size is reduced by a factor of 15-160 and the network utilization is increased between 190%-300% as compared to a fixed service rate. Results also show that even when renegotiations occur on the average in tens of seconds, the queue size is reduced by a factor between 16-30 相似文献
13.
14.
Compression of captured video frames is crucial for saving the power in wireless capsule endoscopy (WCE). A low complexity encoder is desired to limit the power consumption required for compressing the WCE video. Distributed video coding (DVC) technique is best suitable for designing a low complexity encoder. In this technique, frames captured in RGB colour space are converted into YCbCr colour space. Both Y and CbCr representing luma and chroma components of the Wyner–Ziv (WZ) frames are processed and encoded in existing DVC techniques proposed for WCE video compression. In the WCE video, consecutive frames exhibit more similarity in texture and colour properties. The proposed work uses these properties to present a method for processing and encoding only the luma component of a WZ frame. The chroma components of the WZ frame are predicted by an encoder–decoder based deep chroma prediction model at the decoder by matching luma and texture information of the keyframe and WZ frame. The proposed method reduces the computations required for encoding and transmitting of WZ chroma component. The results show that the proposed DVC with a deep chroma prediction model performs better when compared to motion JPEG and existing DVC systems for WCE at the reduced encoder complexity. 相似文献
15.
《Journal of Visual Communication and Image Representation》2008,19(7):437-449
Since the quality of compressed video is vulnerable to errors, video transmission over unreliable Internet is very challenging today. Two-Hypothesis Motion-Compensated Prediction (THMCP) has been shown to have Error Resilience (ER) capability for video transmission, where each macroblock is predicted from its previous two frames. In this paper, we propose a novel ER approach named Alternate Motion-Compensated Prediction (AMCP). In addition to two-hypothesis prediction, one-hypothesis prediction is alternately used. We use some schemes to determine which kind of prediction should be used, so that in some cases of loss, the propagated error can be first decreased to some extent before it spreads to the subsequent frames. As a result, the expected converged error is less than that obtained from THMCP with fixed weights (THMCPF). Both analysis and simulation results are given to show that AMCP performs better than THMCPF, in terms of both compression efficiency and ER capability. 相似文献
16.
Sik-Ho TsangYui-Lam Chan Wan-Chi Siu 《Journal of Visual Communication and Image Representation》2012,23(2):264-270
A novel algorithm for coding flash scenes is proposed. In principle, flash scenes can be detected by analyzing the histogram differences between frames. The proposed algorithm then suggests an adaptive coding order technique for increasing the efficiency of video coding by taking account of characteristics of flash scenes in video contents. The use of adaptive coding technique also benefits to enhance the accuracy of derived motion vectors for determination of weighting parameter sets. Experimental results show that a significant improvement of coding performance in terms of bitrate and PSNR can be achieved in comparison with the conventional weighted prediction algorithms. 相似文献
17.
18.
为了提高关键帧提取的准确率,改善视频摘要的质量,提出了一种HEVC压缩域的视频摘要关键帧提取方法。首先,对视频序列进行编解码,在解码中统计HEVC帧内编码PU块的亮度预测模式数目。然后,特征提取是利用统计得到的模式数目构建成模式特征向量,并将其作为视频帧的纹理特征用于关键帧的提取。最后,利用融合迭代自组织数据分析算法(ISODATA)的自适应聚类算法对模式特征向量进行聚类,在聚类结果中选取每个类内中间向量对应的帧作为候选关键帧,并通过相似度对候选关键帧进行再次筛选,剔除冗余帧,得到最终的关键帧。实验结果表明,在Open Video Project数据集上进行的大量实验验证,该方法提取关键帧的精度为79.9%、召回率达到93.6%、F-score为86.2%,有效地改善了视频摘要的质量。 相似文献
19.
Zhang J. Ahmad M.O. Swamy M.N.S. 《Vision, Image and Signal Processing, IEE Proceedings -》1998,145(6):399-407
The conventional overlapped block motion compensation (OBMC) technique, although effective in reducing the blocking effect for fixed-size partitioned frames, cannot be readily used for a frame partitioned region-wise or using variable-size blocks. The generalised OBMC (GOBMC) method, even though generally applicable, is not very effective in reducing the blocking effect and prediction error. Two windowing techniques are presented to reduce the blocking effect for a frame partitioned region-wise or using variable-size blocks. In the first technique, a virtual re-partitioning operation is employed, which maps a partitioned frame into its corresponding fully partitioned frame at the bottom level of the quadtree so that each resulting block has eight neighbouring blocks. In the second technique, the virtual re-partitioning operation for a given block (region) is carried out adaptively and performed locally; however, blocks (regions) need not always be virtually partitioned to the bottom level. Compared to the GOBMC method, the proposed techniques make use of more pixels in the close vicinity of boundaries of the regions in the windowing operation. Simulation results are included of applying the proposed techniques on a number of MPEG video sequences. These results indicate that the proposed techniques are superior to the GOBMC method in terms of reducing the prediction error as well as the blocking effect 相似文献
20.
Naveed Ejaz Tayyab Bin Tariq Sung Wook Baik 《Journal of Visual Communication and Image Representation》2012,23(7):1031-1040
Video summarization is a method to reduce redundancy and generate succinct representation of the video data. One of the mechanisms to generate video summaries is to extract key frames which represent the most important content of the video. In this paper, a new technique for key frame extraction is presented. The scheme uses an aggregation mechanism to combine the visual features extracted from the correlation of RGB color channels, color histogram, and moments of inertia to extract key frames from the video. An adaptive formula is then used to combine the results of the current iteration with those from the previous. The use of the adaptive formula generates a smooth output function and also reduces redundancy. The results are compared to some of the other techniques based on objective criteria. The experimental results show that the proposed technique generates summaries that are closer to the summaries created by humans. 相似文献