首页 | 本学科首页   官方微博 | 高级检索  
     

融合残差上下文编码和路径增强的视杯视盘分割
引用本文:梅华威,尚虹霖,苏攀,刘艳平. 融合残差上下文编码和路径增强的视杯视盘分割[J]. 中国图象图形学报, 2024, 29(3): 637-654
作者姓名:梅华威  尚虹霖  苏攀  刘艳平
作者单位:华北电力大学控制与计算机工程学院, 保定 071003;复杂能源系统智能计算教育部工程研究中心, 保定 071003;华北电力大学控制与计算机工程学院, 保定 071003;河北省能源电力知识计算重点实验室, 保定 071000
基金项目:国家自然科学基金项目(61906181);中央高校基本科研业务费专项资金资助(2021MS094);河北省高等学校科学技术研究项目(QN2023181)
摘    要:目的 从眼底图像中分割视盘和视杯对于眼部疾病智能诊断来说是一项重要工作,U-Net及变体模型已经广泛应用在视杯盘分割任务中。由于连续的卷积与池化操作容易引起空间信息损失,导致视盘和视杯分割精度差且效率低。提出了融合残差上下文编码和路径增强的深度学习网络RCPA-Net,提升了分割结果的准确性与连续性。方法 采用限制对比度自适应直方图均衡方法处理输入图像,增强对比度并丰富图像信息。特征编码模块以ResNet34(residual neural network)为骨干网络,通过引入残差递归与注意力机制使模型更关注感兴趣区域,采用残差空洞卷积模块捕获更深层次的语义特征信息,使用路径增强模块在浅层特征中获得精确的定位信息来增强整个特征层次。本文还提出了一种新的多标签损失函数用于提高视盘视杯与背景区域的像素比例并生成最终的分割图。结果 在4个数据集上与多种分割方法进行比较,在ORIGA(online retinal fundus image database for glaucoma analysis)数据集中,本文方法对视盘分割的JC(Jaccard)指数为0.939 1,F-measure为...

关 键 词:视杯视盘分割  深度学习  注意力机制  残差空洞卷积  路径增强
收稿时间:2023-03-24
修稿时间:2023-06-19

Optic disc and cup segmentation with combined residual context encoding and path augmentation
Mei Huawei,Shang Honglin,Su Pan,Liu Yanping. Optic disc and cup segmentation with combined residual context encoding and path augmentation[J]. Journal of Image and Graphics, 2024, 29(3): 637-654
Authors:Mei Huawei  Shang Honglin  Su Pan  Liu Yanping
Affiliation:School of Control and Computer Engineering, North China Electric Power University, Baoding 071003, China;Engineering Research Center of Intelligent Computing for Complex Energy Systems, Ministry of Education, Baoding 071003, China;School of Control and Computer Engineering, North China Electric Power University, Baoding 071003, China;Hebei Key Laboratory of Knowledge Computing for Energy and Power, Baoding 071000, China
Abstract:Objective Ophthalmic image segmentation is an important part of medical image analysis. Among these, optic disc(OD) and optic cup(OC) segmentation are crucial technologies for the intelligent diagnosis of glaucoma, which can cause irreversible damage to the eyes and is the second leading cause of blindness worldwide. The primary glaucoma screening method is the evaluation of OD and OC based on fundus images. The cup disc ratio(CDR) is one of the most representative glaucoma detection features. In general, eyes with CDR greater than 0. 65 are considered to have glaucoma. With the continuous development of deep learning, U-Net and its variant models, including superpixel classification and edge segmentation, have been widely used in OD and OC segmentation tasks. However, the segmentation accuracy of OD and OC is limited, and their efficiency is low during training due to the loss of spatial information caused by continuous convolution and pooling operations. To improve the accuracy and training efficiency of OD and OC segmentation, we proposed the residual context path augmentation U-Net(RCPA-Net), which can capture deeper semantic feature information and solve the problem of unclear edge localization. Method RCPA-Net includes three modules:feature coding module(FCM), residual atrous convolution(RAC) module, and path augmentation module(PAM). First, the FCM block adopts the ResNet34 network as the backbone network. By introducing the residual module and attention mechanism, the model is enabled to focus on the region of interest, and the efficient channel attention(ECA) is adopted to the squeeze and excitation(SE) module. The ECA module is an efficient channel attention module that avoids dimensionality reduction and captures cross-channel features effectively. Second, the RAC block is used to obtain the context feature information of a wider layer. Inspired by Inception-V4 and context encoder network (CE-Net), we fuse cavity convolution into the inception series network and stack convolution blocks. Traditional convolution is replaced with cavity convolution, such that the receptive field increases while the number of parameters remains the same. Finally, to shorten the information path between the lowlevel and top-level features, the PAM block uses an accurate low-level positioning signal and lateral connection to enhance the entire feature hierarchy. To solve the problem of extremely unbalanced pixels and generate the final segmentation map, we propose a new multi-label loss function based on the dice coefficient and focal loss. This function improves the pixel ratio between the OD/OC and background regions. In addition, we enhance the training data by flipping the image and adjusting the ratio of length and width. Then, the input images are processed using the contrast-limited adaptive histogram equalization method, and each resultant image is fused with its original one and then averaged to form a new three-channel image. This step aims to enhance image contrast and enrich image information. In the experimental stage, we use Adam optimization instead of the stochastic gradient descent method to optimize the model. The number of samples selected for each training stage is eight, and the weight decay is 0. 000 1. During training, the learning rate is adjusted adaptively in accordance with the number of samples selected each time. In outputting the prediction results, the maximum connected region in OD and OC is selected to obtain the final segmentation result. Result Four datasets(ORIGA, Drishti-GS1, Refuge, and RIM-ONE-R1) are employed to validate the performance of the proposed method. Then, the results are compared with various state-of-the-art methods, including U-Net, M-Net, and CE-Net. The ORIGA dataset contains 650 color fundus images of 3 072 × 2 048 pixels, and the ratio of the training set to the test set is 1:1 during the experiment. The Drishti-GS1 dataset contains 101 images, including 31 normal images and 70 diseased images. The fundus images are divided into two datasets, Groups A and B, which include 50 training samples and 51 testing samples, respectively. The 400 fundus images in the Refuge dataset are also divided into two datasets. Group A includes 320 training samples, while Group B includes 80 testing samples. The Jaccard index and F-measure score are used in the experimentation to evaluate the results of OD and OC segmentation. The results indicate that in the ORIGA dataset, the Jaccard index and F-measure of the proposed method in OD/OC segmentation are 0. 939 1/0. 794 8 and 0. 968 6/0. 885 5, respectively. In the DrishtiGS1 dataset, the results in OD/OC segmentation are 0. 951 3/0. 863 3 and 0. 975 0/0. 926 6, respectively. In the Refuge dataset, the results are 0. 929 8/0. 828 8 and 0. 963 6/0. 906 3, respectively. In the RIM-ONE-R1 dataset, the results of OD segmentation are 0. 929 0 and 0. 962 8. The results of the proposed method on the four datasets are all better than those of its counterparts, and the performance of the network is significantly improved. In addition, we conduct ablation experiments for the primary modules proposed in the network, where we perform comparative experiments with respect to the location of the modules, the parameters in the model, and other factors. The results of the ablation experiments demonstrate the effectiveness of each proposed module in RCPA-Net. Conclusion In this study, we propose RCPA-Net, which combines the advantages of deep segmentation models. The images predicted using RCPA-Net are closer to the real results, providing more accurate segmentation of OD and OC than several state-of-the-art methods. The experimentation demonstrates the high effectiveness and generalization ability of RCPA-Net.
Keywords:optic disc and optic cup segmentation  deep learning  attention mechanism  residual atrous convolution  path augmentation
点击此处可从《中国图象图形学报》浏览原始摘要信息
点击此处可从《中国图象图形学报》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号