首页 | 本学科首页   官方微博 | 高级检索  
     


Video saliency detection incorporating temporal information in compressed domain
Affiliation:1. School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China;2. School of Mathematics and Computer Science, Fujian Normal University, Fuzhou 350001, China;1. 28 Xianning West Road, Xi?an Jiaotong University, China;2. 475 Northwestern Ave, Purdue University, USA;3. 2201 West End Ave, Vanderbilt University, USA;1. Moscow Institute of Physics and Technology (State University), Moscow Region, Dolgoprudny, 141701, Russia;2. V.L.Talrose Institute for Energy Problems of Chemical Physics, Russian Academy of Sciences, Moscow, 119334, Russia;3. Institute of Biomedical Chemistry, Moscow, 119121, Russia;1. Vellalar College for Women(Autonomous), Erode - 638012, Tamil Nadu, India;2. Govt. Arts and Science College, Komarapalayam, Namakkal District, Tamil Nadu, India;1. School of Computing, Edinburgh Napier University, Edinburgh, EH10 5DT, UK;2. School of Computing, Dublin City University, Dublin, Ireland
Abstract:Saliency detection is widely used to pick out relevant parts of a scene as visual attention regions for various image/video applications. Since video is increasingly being captured, moved and stored in compressed form, there is a need for detecting video saliency directly in compressed domain. In this study, a compressed video saliency detection algorithm is proposed based on discrete cosine transformation (DCT) coefficients and motion information within a visual window. Firstly, DCT coefficients and motion information are extracted from H.264 video bitstream without full decoding. Due to a high quantization parameter setting in encoder, skip/intra is easily chosen as the best prediction mode, resulting in a large number of blocks with zero motion vector and no residual existing in video bitstream. To address these problems, the motion vectors of skip/intra coded blocks are calculated by interpolating its surroundings. In addition, a visual window is constructed to enhance the contrast of features and to avoid being affected by encoder. Secondly, after spatial and temporal saliency maps being generated by the normalized entropy, a motion importance factor is imposed to refine the temporal saliency map. Finally, a variance-like fusion method is proposed to dynamically combine these maps to yield the final video saliency map. Experimental results show that the proposed approach significantly outperforms other state-of-the-art video saliency detection models.
Keywords:Compressed domain  Video saliency detection  Visual window  Motion importance factor
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号