首页 | 本学科首页   官方微博 | 高级检索  
     


An iterative adaptive multi-modal stereo-vision method using mutual information
Affiliation:1. Beijing Key Laboratory of Digital Media, School of Computer Science and Engineering, Beihang University, Beijing 100191, China;2. State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing 100191, China;1. Department of Electrical Engineering, School of Electronic Information, Wuhan University, Wuhan 430072, Hubei, PR China;2. School of International Software, Wuhan University, Wuhan 430072, Hubei, PR China;1. School of Electronic and Information Engineering, Xi’an Jiaotong University, Xi’an 710049, PR China;2. State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210093, PR China;1. Institute for Infocomm Research, Agency for Science, Technology and Research, Singapore;2. School of Remote Sensing and Information Engineering, Wuhan University, China;3. Nanyang Technological University, Singapore;1. School of Computer Engineering, Nanyang Technological University, 639798, Singapore;2. School of Information Technology, Jiangxi University of Finance and Economics, Nanchang 330032, Jiangxi, China
Abstract:We propose a method for computing disparity maps from a multi-modal stereo-vision system composed of an infrared–visible camera pair. The method uses mutual information (MI) as the basic similarity measure where a segment-based adaptive windowing mechanism is proposed along with a novel MI computation surface with joint prior probabilities incorporated. The computed cost confidences are aggregated using a novel adaptive cost aggregation method, and the resultant minimum cost disparities in segments are plane-fitted in their respective segments which are iteratively refined by merging and splitting segments reducing dependency to initial segmentation. Finally, the estimated disparities are iteratively refined by repeating all the steps. On an artificially-modified version of the Middlebury dataset and a Kinect dataset that we created in this study, we show that (i) our proposal improves the quality of existing MI formulation, and (ii) our method can provide depth comparable to the quality of Kinect depth data.
Keywords:Multi-modal stereo-vision  Mutual information  Adaptive windowing  Adaptive cost aggregation  Iterative stereo  Plane fitting  RGB-D  Middleburry dataset
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号