首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 578 毫秒
1.
应用于多视点视频压缩的多参考B帧快速编码算法   总被引:2,自引:0,他引:2  
基于H.264平台就多视点视频压缩中多参考帧技术存在计算复杂度高的现状,根据多视点视频序列特性,提出了一种针对多参考B帧的快速多视点视频编码压缩算法。新算法改进了多参考帧列表策略,并结合多视点视频序列的时空域相关性特性,提出了适用于多视点视频编码的参考帧列表调整策略。此外,还针对多种块模式编码引入了块模式预判策略。新算法在保证率失真性能的基础上,有效地降低了计算复杂度,提高了多视点视频编码速度。虽然本文是基于M-Picture编码结构实现了新算法,但它同样适用于其它应用多参考B帧预测技术的多视点视频编码结构中。  相似文献   

2.
相对于以前其他的标准,H.264标准在分层编码、帧内/帧间预测编码、多帧参考、预测精度等技术方面做了巨大的改进。因此,在TMS320DM642平台上实现H.264基档次编码器的移植与优化显得格外实用和必要。基于对DSP平台的结构特性和H.264的计算复杂度分析,主要从以下3个方面对H.264编码器进行了优化:核心算法、数据传输和存储器/Cache使用。实验结果表明,对于CIF格式的视频序列,最优化后的H.264编码器能够达到每秒高于24帧的编码速度,满足了视频处理对于实时性的要求。  相似文献   

3.
国际视频编码最新标准H.266/VVC在H.265/HEVC基础上,采用新的编码工具来提升编码效率。帧内预测能有效的去除像素间的空间冗余,是视频编码的关键环节。H.266/VVC在传统帧内编码技术基础上针对新的应用需求进一步扩展,并提出了新的编码技术。本文研究H.266/VVC帧内预测关键技术的最新进展,对比了与H.265/HEVC帧内预测技术的主要差异,分析了编码技术的性能和复杂度,为我国新一代视频编码标准帧内预测技术的选择提供参考。  相似文献   

4.
王凌杰  周军 《电视技术》2012,36(5):12-14
Side-by-Side(SbS)是立体视频常用的格式。传统的视频编码方法利用帧内预测或帧间预测对视频进行压缩编码。基于H.264视频压缩编码标准,针对SbS格式的立体视频提出了一种基于多参考帧方式的视间预测编码方法,并在JM测试代码基础上将其实现。实验结果表明,加入视间预测编码后,以很小的PSNR损失为代价,使得SbS视频的压缩效率提高了大约10%~20%。  相似文献   

5.
Side-by-Side(SbS)是立体视频常用的格式。传统的视频编码方法利用帧内预测或帧间预测对视频进行压缩编码。本文基于H.264视频压缩编码标准,针对SbS格式的立体视频提出了一种基于多参考帧方式的视间预测编码方法,并在JM测试代码基础上将其实现。实验结果表明,加入视间预测编码后,以很小的PSNR损失为代价,使得SbS视频的压缩效率提高了大约10%至20%。  相似文献   

6.
在多视点视频编码中,引入了视点间运动补偿预测技术,而在视点间预测过程中,由于视差矢量的使用,使得GoP长度对编码效率的影响与一般H.264/AVC编码有所不同。在分析了可分级B帧的预测结构对多视点视频编码性能的影响基础上,通过对不同视频序列设置不同GoP的大小,经过多次反复实验,而后对实验结果进行分析,找到了在一定数值范围内的GoP。通过该GoP得到了较好的PSNR和模值较小的MV或者DV,节约了一定的编码时间。  相似文献   

7.
雷海军  杨辉  何业军 《电视技术》2012,36(18):32-35
预测结构是多视点视频编码(Multi-View Video Coding,MVC)研究的主要内容之一。MVC目前采用HHI(Heinrich-Hertz-Institute)提出的分层次B帧预测结构(HBP),比联播预测结构获得了更好的压缩效率。分析了多种预测结构,并针对平行摄像机采集的多视点视频序列,提出了一种新的预测结构AS_EIPP,该结构充分利用相邻视点间的相关性和多参考帧模式,进一步提高了压缩效率。在多视点视频软件测试平台JMVC8.3上进行验证,实验结果表明:新的预测结构在保证重建视频质量基本不变的前提下,压缩效率比HBP预测结构提高了1%~4%。  相似文献   

8.
介绍了中国音视频编码技术标准AVS中的帧内预测和基于宏块的帧场自适应编码技术.通过分析其中的5种帧内预测模式的空间分布特点,提出了一种新的隔行视频帧内预测模式.实验表明该模式可以在标清、高清隔行视频的帧内预测上获得平均0.24 dB以上的信噪比增益,平均减少3.1%的码流,是一种有效提高编码效率的预测模式.  相似文献   

9.
在H.264/AVC视频压缩标准采用的几项关键技术中,使用多参考帧预测可以增加最佳匹配块的检索概率,进而大大提高了编码效率.受B帧直接( direct)预测模式的启发,提出了一种新的基于扩展帧的多参考帧预测方法,由多参考帧中的共同位置块及其参考块扩展得到一个新的抽取帧,增加了原始序列的时域分辨率,使得扩展帧更加接近当前帧,提高了运动估计中最佳匹配的检索概率,进而提升了编码效率.仿真结果证实该方法的编码性能好于H.264/AVC参考软件.  相似文献   

10.
新一代具有自主知识产权的视频编码标准AVS2正在制定当中,其相对于上一代标准而言具有更高的编码效率,但由于使用了多划分模式、多参考帧等技术,其编码复杂度也大幅增加。为了提高AVS2帧间编码速度,提出了一种快速帧间预测算法,在使用非对称运动划分(Asymmetric Motion Partitions,AMP)快速决策算法的基础上,深度挖掘同一编码单元下相邻预测单元之间的相关性,利用邻近预测单元来预测当前块的参考帧,通过判断预测单元的运动矢量残差的曼哈顿距离,以确定最佳参考帧,从而减小参与计算的参考帧数目。实验结果表明,该算法在图像质量和比特率基本保持不变的情况下,有效地提高了AVS2的编码速度。  相似文献   

11.
In this paper, we propose a perceptual-based distributed video coding (DVC) technique. Unlike traditional video codecs, DVC applies video prediction process at the decoder side using previously received frames. The predicted video frames (i.e., side information) contain prediction errors. The encoder then transmits error-correcting parity bits to the decoder to reconstruct the video frames from side information. However, channel codes based on i.i.d. noise models are not always efficient in correcting video prediction errors. In addition, some of the prediction errors do not cause perceptible visual distortions. From perceptual coding point of view, there is no need to correct such errors. This paper proposes a scheme for the decoder to perform perceptual quality analysis on the predicted side information. The decoder only requests parity bits to correct visually sensitive errors. More importantly, with the proposed technique, key frames can be encoded at higher rates while still maintaining consistent visual quality across the video sequence. As a result, even the objective PSNR measure of the decoded video sequence will increase too. Experimental results show that the proposed technique improves the R-D performance of a transform domain DVC codec both subjectively and objectively. Comparisons with a well-known DVC codec show that the proposed perceptual-based DVC coding scheme is very promising for distributed video coding framework.  相似文献   

12.
A dynamic bandwidth allocation strategy to support variable bit rate (VBR) video traffic is proposed. This strategy predicts the bandwidth requirements for future frames using adaptive linear prediction that minimizes the mean square error. The adaptive technique does not require any prior knowledge of the traffic statistics nor assume stationarity. Analyses using six one-half-hour video tracts indicate that prediction errors for the bandwidth required for the next frames and group of pictures (GOP) are almost white noise or short memory. The performance of the strategy is studied using renegotiated constant bit rate (RCBR) network service model and methods that control the tradeoff between the number of renegotiations and network utilization are proposed. Simulation results using MPEG-I video traces for predicting GOP rates show that the queue size is reduced by a factor of 15-160 and the network utilization is increased between 190%-300% as compared to a fixed service rate. Results also show that even when renegotiations occur on the average in tens of seconds, the queue size is reduced by a factor between 16-30  相似文献   

13.
李林格  张恋  王洁  周巧  张昊 《电视技术》2016,40(11):18-24
帧内预测在视频编码中是非常重要的模块.在视频实时编码与传输过程中,场景切换会经常出现.此时,一般会采用全Ⅰ帧编码.研究发现,即使是全Ⅰ帧编码,也往往会非常耗时.基于编码单元深度范围和帧内预测中候选预测方向个数研究了HEVC编码器的复杂度控制问题.针对不同的目标编码复杂度,算法自适应地选择不同的方法来优化编码过程.实验结果表明,该算法在保证视频质量的前提下实现了对不同复杂度目标的控制.  相似文献   

14.
Compression of captured video frames is crucial for saving the power in wireless capsule endoscopy (WCE). A low complexity encoder is desired to limit the power consumption required for compressing the WCE video. Distributed video coding (DVC) technique is best suitable for designing a low complexity encoder. In this technique, frames captured in RGB colour space are converted into YCbCr colour space. Both Y and CbCr representing luma and chroma components of the Wyner–Ziv (WZ) frames are processed and encoded in existing DVC techniques proposed for WCE video compression. In the WCE video, consecutive frames exhibit more similarity in texture and colour properties. The proposed work uses these properties to present a method for processing and encoding only the luma component of a WZ frame. The chroma components of the WZ frame are predicted by an encoder–decoder based deep chroma prediction model at the decoder by matching luma and texture information of the keyframe and WZ frame. The proposed method reduces the computations required for encoding and transmitting of WZ chroma component. The results show that the proposed DVC with a deep chroma prediction model performs better when compared to motion JPEG and existing DVC systems for WCE at the reduced encoder complexity.  相似文献   

15.
Since the quality of compressed video is vulnerable to errors, video transmission over unreliable Internet is very challenging today. Two-Hypothesis Motion-Compensated Prediction (THMCP) has been shown to have Error Resilience (ER) capability for video transmission, where each macroblock is predicted from its previous two frames. In this paper, we propose a novel ER approach named Alternate Motion-Compensated Prediction (AMCP). In addition to two-hypothesis prediction, one-hypothesis prediction is alternately used. We use some schemes to determine which kind of prediction should be used, so that in some cases of loss, the propagated error can be first decreased to some extent before it spreads to the subsequent frames. As a result, the expected converged error is less than that obtained from THMCP with fixed weights (THMCPF). Both analysis and simulation results are given to show that AMCP performs better than THMCPF, in terms of both compression efficiency and ER capability.  相似文献   

16.
A novel algorithm for coding flash scenes is proposed. In principle, flash scenes can be detected by analyzing the histogram differences between frames. The proposed algorithm then suggests an adaptive coding order technique for increasing the efficiency of video coding by taking account of characteristics of flash scenes in video contents. The use of adaptive coding technique also benefits to enhance the accuracy of derived motion vectors for determination of weighting parameter sets. Experimental results show that a significant improvement of coding performance in terms of bitrate and PSNR can be achieved in comparison with the conventional weighted prediction algorithms.  相似文献   

17.
18.
为了提高关键帧提取的准确率,改善视频摘要的质量,提出了一种HEVC压缩域的视频摘要关键帧提取方法。首先,对视频序列进行编解码,在解码中统计HEVC帧内编码PU块的亮度预测模式数目。然后,特征提取是利用统计得到的模式数目构建成模式特征向量,并将其作为视频帧的纹理特征用于关键帧的提取。最后,利用融合迭代自组织数据分析算法(ISODATA)的自适应聚类算法对模式特征向量进行聚类,在聚类结果中选取每个类内中间向量对应的帧作为候选关键帧,并通过相似度对候选关键帧进行再次筛选,剔除冗余帧,得到最终的关键帧。实验结果表明,在Open Video Project数据集上进行的大量实验验证,该方法提取关键帧的精度为79.9%、召回率达到93.6%、F-score为86.2%,有效地改善了视频摘要的质量。   相似文献   

19.
The conventional overlapped block motion compensation (OBMC) technique, although effective in reducing the blocking effect for fixed-size partitioned frames, cannot be readily used for a frame partitioned region-wise or using variable-size blocks. The generalised OBMC (GOBMC) method, even though generally applicable, is not very effective in reducing the blocking effect and prediction error. Two windowing techniques are presented to reduce the blocking effect for a frame partitioned region-wise or using variable-size blocks. In the first technique, a virtual re-partitioning operation is employed, which maps a partitioned frame into its corresponding fully partitioned frame at the bottom level of the quadtree so that each resulting block has eight neighbouring blocks. In the second technique, the virtual re-partitioning operation for a given block (region) is carried out adaptively and performed locally; however, blocks (regions) need not always be virtually partitioned to the bottom level. Compared to the GOBMC method, the proposed techniques make use of more pixels in the close vicinity of boundaries of the regions in the windowing operation. Simulation results are included of applying the proposed techniques on a number of MPEG video sequences. These results indicate that the proposed techniques are superior to the GOBMC method in terms of reducing the prediction error as well as the blocking effect  相似文献   

20.
Video summarization is a method to reduce redundancy and generate succinct representation of the video data. One of the mechanisms to generate video summaries is to extract key frames which represent the most important content of the video. In this paper, a new technique for key frame extraction is presented. The scheme uses an aggregation mechanism to combine the visual features extracted from the correlation of RGB color channels, color histogram, and moments of inertia to extract key frames from the video. An adaptive formula is then used to combine the results of the current iteration with those from the previous. The use of the adaptive formula generates a smooth output function and also reduces redundancy. The results are compared to some of the other techniques based on objective criteria. The experimental results show that the proposed technique generates summaries that are closer to the summaries created by humans.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号