Current Issue Cover
融合残差上下文编码和路径增强的视杯视盘分割

梅华威1,2, 尚虹霖1,2, 苏攀1,3, 刘艳平1(1.华北电力大学控制与计算机工程学院, 保定 071003;2.复杂能源系统智能计算教育部工程研究中心, 保定 071003;3.河北省能源电力知识计算重点实验室, 保定 071000)

摘 要
目的 从眼底图像中分割视盘和视杯对于眼部疾病智能诊断来说是一项重要工作,U-Net及变体模型已经广泛应用在视杯盘分割任务中。由于连续的卷积与池化操作容易引起空间信息损失,导致视盘和视杯分割精度差且效率低。提出了融合残差上下文编码和路径增强的深度学习网络RCPA-Net,提升了分割结果的准确性与连续性。方法 采用限制对比度自适应直方图均衡方法处理输入图像,增强对比度并丰富图像信息。特征编码模块以ResNet34(residual neural network)为骨干网络,通过引入残差递归与注意力机制使模型更关注感兴趣区域,采用残差空洞卷积模块捕获更深层次的语义特征信息,使用路径增强模块在浅层特征中获得精确的定位信息来增强整个特征层次。本文还提出了一种新的多标签损失函数用于提高视盘视杯与背景区域的像素比例并生成最终的分割图。结果 在4个数据集上与多种分割方法进行比较,在ORIGA (online retinal fundus image database for glaucoma analysis)数据集中,本文方法对视盘分割的JC (Jaccard)指数为0.939 1,F-measure为0.968 6,视杯分割的JC和Fmeasure分别为0.794 8和0.885 5;在Drishti-GS1数据集中,视盘分割的JC和F-measure分别为0.951 3和0.975 0,视杯分割的JC和F-measure分别为0.863 3和0.926 6;在Refuge (retinal fundus glaucoma challenge)数据集中,视盘分割的JC和F-measure分别为0.929 8和0.963 6,视杯分割的JC和F-measure分别为0.828 8和0.906 3;在RIM-ONE (retinal image database for optic nerve evaluation)-R1数据集中,视盘分割的JC和F-measure分别为0.929 0和0.9628。在4个数据集上结果均优于对比算法,性能显著提升。此外,针对网络中提出的模块分别做了消融实验,验证了RCPA-Net中各个模块的有效性。结论 实验结果表明,RCPA-Net提升了视盘和视杯分割精度,预测图像更接近真实标签结果,同时跨数据集测试结果证明了RCPA-Net具有良好的泛化能力。
关键词
Optic disc and cup segmentation with combined residual context encoding and path augmentation

Mei Huawei1,2, Shang Honglin1,2, Su Pan1,3, Liu Yanping1(1.School of Control and Computer Engineering, North China Electric Power University, Baoding 071003, China;2.Engineering Research Center of Intelligent Computing for Complex Energy Systems, Ministry of Education, Baoding 071003, China;3.Hebei Key Laboratory of Knowledge Computing for Energy and Power, Baoding 071000, China)

Abstract
Objective Ophthalmic image segmentation is an important part of medical image analysis. Among these, optic disc(OD) and optic cup(OC) segmentation are crucial technologies for the intelligent diagnosis of glaucoma, which can cause irreversible damage to the eyes and is the second leading cause of blindness worldwide. The primary glaucoma screening method is the evaluation of OD and OC based on fundus images. The cup disc ratio(CDR) is one of the most representative glaucoma detection features. In general, eyes with CDR greater than 0. 65 are considered to have glaucoma. With the continuous development of deep learning, U-Net and its variant models, including superpixel classification and edge segmentation, have been widely used in OD and OC segmentation tasks. However, the segmentation accuracy of OD and OC is limited, and their efficiency is low during training due to the loss of spatial information caused by continuous convolution and pooling operations. To improve the accuracy and training efficiency of OD and OC segmentation, we proposed the residual context path augmentation U-Net(RCPA-Net), which can capture deeper semantic feature information and solve the problem of unclear edge localization. Method RCPA-Net includes three modules:feature coding module(FCM), residual atrous convolution(RAC) module, and path augmentation module(PAM). First, the FCM block adopts the ResNet34 network as the backbone network. By introducing the residual module and attention mechanism, the model is enabled to focus on the region of interest, and the efficient channel attention(ECA) is adopted to the squeeze and excitation(SE) module. The ECA module is an efficient channel attention module that avoids dimensionality reduction and captures cross-channel features effectively. Second, the RAC block is used to obtain the context feature information of a wider layer. Inspired by Inception-V4 and context encoder network (CE-Net), we fuse cavity convolution into the inception series network and stack convolution blocks. Traditional convolution is replaced with cavity convolution, such that the receptive field increases while the number of parameters remains the same. Finally, to shorten the information path between the lowlevel and top-level features, the PAM block uses an accurate low-level positioning signal and lateral connection to enhance the entire feature hierarchy. To solve the problem of extremely unbalanced pixels and generate the final segmentation map, we propose a new multi-label loss function based on the dice coefficient and focal loss. This function improves the pixel ratio between the OD/OC and background regions. In addition, we enhance the training data by flipping the image and adjusting the ratio of length and width. Then, the input images are processed using the contrast-limited adaptive histogram equalization method, and each resultant image is fused with its original one and then averaged to form a new three-channel image. This step aims to enhance image contrast and enrich image information. In the experimental stage, we use Adam optimization instead of the stochastic gradient descent method to optimize the model. The number of samples selected for each training stage is eight, and the weight decay is 0. 000 1. During training, the learning rate is adjusted adaptively in accordance with the number of samples selected each time. In outputting the prediction results, the maximum connected region in OD and OC is selected to obtain the final segmentation result. Result Four datasets(ORIGA, Drishti-GS1, Refuge, and RIM-ONE-R1) are employed to validate the performance of the proposed method. Then, the results are compared with various state-of-the-art methods, including U-Net, M-Net, and CE-Net. The ORIGA dataset contains 650 color fundus images of 3 072 × 2 048 pixels, and the ratio of the training set to the test set is 1:1 during the experiment. The Drishti-GS1 dataset contains 101 images, including 31 normal images and 70 diseased images. The fundus images are divided into two datasets, Groups A and B, which include 50 training samples and 51 testing samples, respectively. The 400 fundus images in the Refuge dataset are also divided into two datasets. Group A includes 320 training samples, while Group B includes 80 testing samples. The Jaccard index and F-measure score are used in the experimentation to evaluate the results of OD and OC segmentation. The results indicate that in the ORIGA dataset, the Jaccard index and F-measure of the proposed method in OD/OC segmentation are 0. 939 1/0. 794 8 and 0. 968 6/0. 885 5, respectively. In the DrishtiGS1 dataset, the results in OD/OC segmentation are 0. 951 3/0. 863 3 and 0. 975 0/0. 926 6, respectively. In the Refuge dataset, the results are 0. 929 8/0. 828 8 and 0. 963 6/0. 906 3, respectively. In the RIM-ONE-R1 dataset, the results of OD segmentation are 0. 929 0 and 0. 962 8. The results of the proposed method on the four datasets are all better than those of its counterparts, and the performance of the network is significantly improved. In addition, we conduct ablation experiments for the primary modules proposed in the network, where we perform comparative experiments with respect to the location of the modules, the parameters in the model, and other factors. The results of the ablation experiments demonstrate the effectiveness of each proposed module in RCPA-Net. Conclusion In this study, we propose RCPA-Net, which combines the advantages of deep segmentation models. The images predicted using RCPA-Net are closer to the real results, providing more accurate segmentation of OD and OC than several state-of-the-art methods. The experimentation demonstrates the high effectiveness and generalization ability of RCPA-Net.
Keywords

订阅号|日报