首页 | 本学科首页   官方微博 | 高级检索  
     


Multi-scale convolutional neural network for multi-focus image fusion
Affiliation:1. Department of Biomedical Engineering, Hefei University of Technology, Hefei 230009, China;2. Department of Automation, University of Science and Technology of China, Hefei 230026, China;1. Image Processing Center, Beijing University of Aeronautics and Astronautics, Beijing 100191, China;2. State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing 100191, China;1. Key Laboratory of Electronic Equipment Structure Design, Ministry of Education, Xidian University, Xi''an, Shaanxi 710071,China;2. Center for Complex Systems, School of Mechano-electronic Engineering, Xidian University, Xi''an Shaanxi 710071, China;3. WMG Data Science, University of Warwick, Coventry CV4 AL7, U.K;1. Key Laboratory of Electronic Equipment Structure Design, Ministry of Education, Xidian University, Xi''an, Shaanxi710071, China;2. Center for Complex Systems, School of Mechano-electronic Engineering, Xidian University, Xi''an, Shaanxi710071, China;3. School of Electronic Engineering, Xidian University, Xi''an, Shaanxi710071, China;4. Computer Science Department, Aberystwyth University, Aberystwyth SY23 3FL, United Kingdom
Abstract:In this study, we present new deep learning (DL) method for fusing multi-focus images. Current multi-focus image fusion (MFIF) approaches based on DL methods mainly treat MFIF as a classification task. These methods use a convolutional neural network (CNN) as a classifier to identify pixels as focused or defocused pixels. However, due to unavailability of labeled data to train networks, existing DL-based supervised models for MFIF add Gaussian blur in focused images to produce training data. DL-based unsupervised models are also too simple and only applicable to perform fusion tasks other than MFIF. To address the above issues, we proposed a new MFIF method, which aims to learn feature extraction, fusion and reconstruction components together to produce a complete unsupervised end-to-end trainable deep CNN. To enhance the feature extraction capability of CNN, we introduce a Siamese multi-scale feature extraction module to achieve a promising performance. In our proposed network we applied multiscale convolutions along with skip connections to extract more useful common features from a multi-focus image pair. Instead of using basic loss functions to train the CNN, our model utilizes structure similarity (SSIM) measure as a training loss function. Moreover, the fused images are reconstructed in a multiscale manner to guarantee more accurate restoration of images. Our proposed model can process images with variable size during testing and validation. Experimental results on various test images validate that our proposed method yields better quality fused images that are superior to the fused images generated by compared state-of-the-art image fusion methods.
Keywords:
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号