首页 | 本学科首页   官方微博 | 高级检索  
     


Adaptive weighted fusion with new spatial and temporal fingerprints for improved video copy detection
Affiliation:1. Image and Video Systems Lab, Department of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Yuseong-Gu, Daejeon 305-701, Republic of Korea;2. Cyber Security-Convergence Research Laboratory, Electronics and Telecommunications Research Institute (ETRI), 218 Gajeongno, Yuseong-gu, Daejeon 305-700, Republic of Korea;1. Control and Computer Engineering Department, Politecnico di Torino, corso Duca degli Abruzzi 24, 10129 Torino, Italy;2. College of Electronics and Information Engineering, Sichuan University, No. 24 South Section 1, Yihuan Road, 610065 Chengdu, China;1. Institute for Electronics and Telecommunication of Rennes IETR/INSA, 20 avenue des Buttes de Coesmes, Rennes, France;2. IRISA Lagadic Team, Campus universitaire de Beaulieu, 263 Avenue du Général Leclerc, France;3. Technicolor Research and Innovation, avenue des Champs Blancs, Cesson Sevigné, France
Abstract:In this paper, we propose a new and novel modality fusion method designed for combining spatial and temporal fingerprint information to improve video copy detection performance. Most of the previously developed methods have been limited to use only pre-specified weights to combine spatial and temporal modality information. Hence, previous approaches may not adaptively adjust the significance of the temporal fingerprints that depends on the difference between the temporal variances of compared videos, leading to performance degradation in video copy detection. To overcome the aforementioned limitation, the proposed method has been devised to extract two types of fingerprint information: (1) spatial fingerprint that consists of the signs of DCT coefficients in local areas in a keyframe and (2) temporal fingerprint that computes the temporal variances in local areas in consecutive keyframes. In addition, the so-called temporal strength measurement technique is developed to quantitatively represent the amount of the temporal variances; it can be adaptively used to consider the significance of compared temporal fingerprints. The experimental results show that the proposed modality fusion method outperforms other state-of-the-arts fusion methods and popular spatio-temporal fingerprints in terms of video copy detection. Furthermore, the proposed method can save 39.0%, 25.1%, and 46.1% time complexities needed to perform video fingerprint matching without a significant loss of detection accuracy for our synthetic dataset, TRECVID 2009 CCD Task, and MUSCLE-VCD 2007, respectively. This result indicates that our proposed method can be readily incorporated into the real-life video copy detection systems.
Keywords:Video copy detection  Video sequence matching  Modality fusion  Video fingerprint  Weighted adaptive fusion  Spatial and temporal information
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号