融合空洞卷积与注意力的胃癌组织切片分割
The fusing of dilated convolution and attention for segmentation of gastric cancer tissue sections
- 2021年26卷第9期 页码:2281-2292
收稿:2020-12-14,
修回:2021-2-19,
录用:2021-2-26,
纸质出版:2021-09-16
DOI: 10.11834/jig.200765
移动端阅览

浏览全部资源
扫码关注微信
收稿:2020-12-14,
修回:2021-2-19,
录用:2021-2-26,
纸质出版:2021-09-16
移动端阅览
目的
2
病理组织切片检查是诊断胃癌的金标准,准确发现切片中的病变区域有助于及时确诊并开展后续治疗。然而,由于病理切片图像的复杂性、病变细胞与正常细胞形态差异过小等问题,传统的语义分割模型并不能达到理想的分割效果。基于此,本文提出了一种针对病理切片的语义分割方法ADEU-Net(attention-dilated-efficient U-Net++),提高胃癌区域分割的精度,实现端到端分割。
方法
2
ADEU-Net使用经过迁移学习的EfficientNet作为编码器部分,增强图像特征提取能力。解码器采用了简化的U-Net++短连接方式,促进深浅层特征融合的同时减少网络参数量,并重新设计了其中的卷积模块提高梯度传递能力。中心模块使用空洞卷积对编码器输出结果进行多尺度的特征提取,增强模型对不同尺寸切片的鲁棒性。编码器与解码器的跳跃连接使用了注意力模块,以抑制背景信息的特征响应。
结果
2
在2020年“华录杯”江苏大数据开发与应用大赛(简称“SEED”大赛)数据集中与其他经典方法比较,验证了一些经典模型在该分割任务中难以拟合的问题,同时实验得出修改特征提取方式对结果有较大提升,本文方法在分割准确度上比原始U-Net提高了18.96%。在SEED数据集与2017年中国大数据人工智能创新创业大赛(brain of things,BOT)数据集中进行了消融实验,验证了本文方法中各个模块均有助于提高病理切片的分割效果。在SEED数据集中,本文方法ADEU-Net比基准模型在Dice系数、准确度、敏感度和精确度上分别提升了5.17%、2.7%、3.69%、4.08%;在BOT数据集中,本文方法的4项指标分别提升了0.47%、0.06%、4.30%、6.08%。
结论
2
提出的ADEU-Net提升了胃癌病理切片病灶点分割的精度,同时具有良好的泛化性能。
Objective
2
As the gold standard for the diagnosis of gastric cancer
pathological section has been a hotspot nowadays. The degree of precise detection of the lesion area in the section has rather beneficial to in-situ diagnosis and follow-up treatment. The pathologists have missed some subtle changes in cancerous cells in their practice. Automated gastric cancer cells segmentation has been aided to diagnose. Deep learning-based pathological section image of stomach have been obtained qualified classification via deep convolutional neural networks (DCNNs). Focused segmentation for pathological section has been challenged to some issues as below. First
the color and morphology between gastric cancer cells and normal cells to extract deep features. Second
the different magnifications in pathological section images have been hard to segment in different sizes. A semantic segmentation neural network called attention-dilated-efficient U-Net++ (ADEU-Net) has been demonstrated rather than original U-Net to facilitate the precision of gastric cancer cell segmentation.
Method
2
The illustrated framework is an encoder-decoder networks which can achieve end-to-end training. The capability of the encoder has affected the segmentation accuracy based on the deep features of pathological section images interpretation. The featured extraction part of EfficientNet has been adopted as the encoder via qualified classification. The initialized weights of EfficientNet have been pre-trained on ImageNet and its structure have been divided into five stages for the skipped cohesions. The decoder has been designed in terms of the structure of U-Net++. The encoder and decoder sub-networks have been via the integration of nested
dense skip pathways. An 8 GB GPU model training has been implemented based on the most skip cohesions in U-Net++. The convolution blocks in the decoder has been re-modified to the gradient transfer issues. An additional module called DBlock has been integrated to enhance the feature extraction capability for multi-sized pathological sections. Three multi-layers dilation rates convolution have been cascaded in DBlock to realize the features in receptive fields. The dilation rates of the stacked dilated convolution layers have been calculated to 1
2
5 and the receptive field of each layer has been realized 3
7
17 each in terms of the structure of hybrid dilated convolution (HDC). The featured maps have been concatenated by channel and fused via a 1×1 convolution layer to realize multi-scale features simultaneously. The attention mechanism has been used to replace the skip connection between the encoder and the decoder to suppress the feature correspondence of the background region effectively as well. The outputs of the encoder and the decoder of the upper layer have been conducted each based on a 1×1 convolution layer. The optional assigned weights to the parameters of the original feature map have been added together to form the attention gate. deep supervised learning can be altered to solve the low speed convergence in the training process.
Result
2
The experiments on two datasets called SEED and BOT have been conducted to verify the effectiveness of the method obtained from two gastric cancer cell section segmentation competition. The evaluation metrics of the models have Dice coefficient
sensitivity
pixel-wise accuracy and precision. Different segmentation results have also been calculated visually. First the baseline method has been compared to some classical models on SEED dataset in 18.96% accuracy higher than original U-Net and it has been found that the design of feature extraction has been crucial to the segmentation accuracy. Transfer-learning strategies of the encoder have been improved the results greatly. The further ablation experiments have been performed to each added module to confirm the results of segmentation.
The Dice coefficient
sensitivity
accuracy and precision has been increased by 5.17% and 0.47%
2.7% and 0.06%
3.69% and 4.30%
4.08% and 6.08% each compared with the baseline model's results with SEED and BOT. The results have demonstrated the effectiveness of each part of proposed algorithm. The visual segmentation results have more similar to the ground truth label.
Conclusion
2
A semantic segmentation model called ADEU-Net has been illustrated to the segmentation of pathological sections of gastric cancer task. The involvement of EfficientNet has beneficial to feature extraction
multi-scale features assembling cascade dilated convolution layers and the attention module in replace of the skip connection between the encoder and decoder.
Bian X W and Ping Y F. 2019. Pathology in China: challenges and opportunities. Journal of Third Military Medical University, 41(19): 1815-1817
卞修武, 平轶芳. 2019. 我国病理学科发展面临的挑战和机遇. 第三军医大学学报, 41(19): 1815-1817[DOI: 10.16016/j.1000-5404.201909212]
Clevert D A, Unterthiner T and Hochreiter S. 2016. Fast and accurate deep network learning by exponential linear units (ELUs)//Proceedings of the 4th International Conference on Learning Representations, ICLR 2016-Conference Track Proceedings. San Juan, Puerto Rico, USA: ICLR: 1-14
Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Franke U, Roth S and Schiele B. 2016. The cityscapes dataset for semantic urban scene understanding//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, USA: IEEE: 3213-3223[ DOI: 10.1109/CVPR.2016.350 http://dx.doi.org/10.1109/CVPR.2016.350 ]
Cosatto E, Laquerre P F, Malon C, Graf H P, Saito A, Kiyuna T, Marugame A and Kamijo K. 2013. Automated gastric cancer diagnosis on H and E-stained sections; ltraining a classifier on a large scale with multiple instance machine learning//Proceedings of Medical Imaging 2013: Digital Pathology. Lake Buena Vista (Orlando Area), USA: SPIE: 867605[ DOI: 10.1117/12.2007047 http://dx.doi.org/10.1117/12.2007047 ]
Deng J, Dong W, Socher R, Li L J, Li K and Li F F. 2009. Imagenet: a large-scale hierarchical image database//Proceedings of 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami, USA: IEEE: 248-255[ DOI: 10.1109/CVPR.2009.5206848 http://dx.doi.org/10.1109/CVPR.2009.5206848 ]
Everingham M, Van Gool L, Williams C K I, Winn J and Zisserman A. 2010. The pascal visual object classes (VOC) challenge. International Journal of Computer Vision, 88(2): 303-338[DOI: 10.1007/s11263-009-0275-4]
He K M, Zhang X Y, Ren S Q and Sun J. 2016. Deep residual learning for image recognition//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, USA: IEEE: 770-778[ DOI: 10.1109/CVPR.2016.90 http://dx.doi.org/10.1109/CVPR.2016.90 ]
Hu J, Shen L and Sun G. 2018. Squeeze-and-excitation networks//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE: 7132-7141[ DOI: 10.1109/CVPR.2018.00745 http://dx.doi.org/10.1109/CVPR.2018.00745 ]
Iizuka O, Kanavati F, Kato K, Rambeau M, Arihiro K and Tsuneki M. 2020. Deep learning models for histopathological classification of gastric and colonic epithelial tumours. Scientific Reports, 10(1): #1504[DOI: 10.1038/s41598-020-58467-9]
Jin P, Ji X Y, Kang W Z, Li Y, Liu H, Ma F H, Ma S, Hu H T, Li W K and Tian Y T. 2020. Artificial intelligence in gastric cancer: a systematic review. Journal of Cancer Research and Clinical Oncology, 146(9): 2339-2350[DOI: 10.1007/s00432-020-03304-9]
Khened M, Kollerathu V A and Krishnamurthi G. 2019. Fully convolutional multi-scale residual densenets for cardiac segmentation and automated cardiac diagnosis using ensemble of classifiers. Medical Image Analysis, 51: 21-45[DOI: 10.1016/j.media.2018.10.004]
Krizhevsky A, Sutskever I and Hinton G E. 2017. Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6): 84-90[DOI: 10.1145/3065386]
Lee C Y, Xie S N, Gallagher P, Zhang Z Y and Tu Z W. 2015. Deeply-supervised nets. Journal of Machine Learning Research, 38: 562-570
Li X M, Chen H, Qi X J, Dou Q, Fu C W and Heng P A. 2018. H-DenseUNet: hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Transactions on Medical Imaging, 37(12): 2663-2674[DOI: 10.1109/TMI.2018.2845918]
Liao F Z, Liang M, Li Z, Hu X L and Song S. 2019. Evaluate the malignancy of pulmonary nodules using the 3-D deep leaky noisy-OR network. IEEE Transactions on Neural Networks and Learning Systems, 30(11): 3484-3495[DOI: 10.1109/TNNLS.2019.2892409]
Liu B, Yao K, Huang M M, Zhang J H, Li Y and Li R. 2018. Gastric pathology image recognition based on deep residual networks//Proceedings of the 42nd IEEE Annual Computer Software and Applications Conference (COMPSAC). Tokyo, Japan: IEEE: 408-412[ DOI: 10.1109/COMPSAC.2018.10267 http://dx.doi.org/10.1109/COMPSAC.2018.10267 ]
Liu L Y, Jiang H M, He P C, Chen W Z, Liu X D, Gao J F and Han J W. 2019. On the variance of the adaptive learning rate and beyond[EB/OL]. [2020-05-17] . https://arxiv.org/pdf/1908.03265.pdf https://arxiv.org/pdf/1908.03265.pdf
Long J, Shelhamer E and Darrell T. 2015. Fully convolutional networks for semantic segmentation//Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston, USA: IEEE: 3431-3440[ DOI: 10.1109/CVPR.2015.7298965 http://dx.doi.org/10.1109/CVPR.2015.7298965 ]
Oktay O, Schlemper J, Folgoc L L, Lee M, Heinrich M, Misawa K, Mori K, McDonagh S, Hammerla N Y, Kainz B, Glocker B and Rueckert D. 2018. Attention U-net: learning where to look for the pancreas//Proceedings of the 1st Conference on Medical Imaging with Deep Learning. Amsterdam, The Netherlands: 1-10
Ronneberger O, Fischer P and Brox T. 2015. U-net: Convolutional networks for biomedical image segmentation//Proceedings of International Conference on Medical Image Computing and Computer-Assisted Intervention. Cham: Springer: 234-241[ DOI: 10.1007/978-3-319-24574-4_28 http://dx.doi.org/10.1007/978-3-319-24574-4_28 ]
Roth H R, Oda H, Hayashi Y, Oda M, Shimizu N, Fujiwara M, Misawa K and Mori K. 2017. Hierarchical 3D fully convolutional networks for multi-organ segmentation[EB/OL]. [2020-11-14] . https://arxiv.org/pdf/1704.06382.pdf https://arxiv.org/pdf/1704.06382.pdf
Sharma H, Zerbe N, Klempert I, Hellwich O and Hufnagl P. 2017. Deep convolutional neural networks for automatic classification of gastric carcinoma using whole slide images in digital histopathology. Computerized Medical Imaging and Graphics, 61: 2-13[DOI: 10.1016/j.compmedimag.2017.06.001]
Simonyan K and Zisserman A. 2015. Very deep convolutional networks for large-scale image recognition//Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015-Conference Track Proceedings. San Diego, USA: ICLR: 1-14
Tan M X and Le Q V. 2019. Efficientnet: rethinking model scaling for convolutional neural networks[EB/OL]. [2020-11-09] . https://arxiv.org/pdf/1905.11946v3.pdf https://arxiv.org/pdf/1905.11946v3.pdf
Wang P Q, Chen P F, Yuan Y, Liu D, Huang Z H, Hou X D and Cottrell G. 2018. Understanding convolution for semantic segmentation//Proceedings of 2018 IEEE Winter Conference on Applications of Computer Vision (WACV). Lake Tahoe, USA: IEEE: 1451-1460[ DOI: 10.1109/WACV.2018.00163 http://dx.doi.org/10.1109/WACV.2018.00163 ]
Xiao X, Lian S, Luo Z M and Li S Z. 2018. Weighted res-UNet for high-quality retina vessel segmentation//Proceedings of the 9th International Conference on Information Technology in Medicine and Education (ITME). Hangzhou, China: IEEE: 327-331[ DOI: 10.1109/ITME.2018.00080 http://dx.doi.org/10.1109/ITME.2018.00080 ]
Zhou L C, Zhang C and Wu M. 2018a. D-LinkNet: linknet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPR). Salt Lake City, USA: IEEE: 182-186[ DOI: 10.1109/CVPRW.2018.00034 http://dx.doi.org/10.1109/CVPRW.2018.00034 ]
Zhou Z, Siddiquee M M R, Tajbakhsh N and Liang J. 2018b. UNet++: a nested U-Net architecture for medical image segmentation//Stoyanov D, Taylor Z, Carneiro G, Syeda-Mahmood T, Martel A, Maier-Hein L, Tavares J M R S, Bradley A, Papa J P, Belagiannis V, Nascimento J C, Lu Z, Conjeti S, Moradi M, Greenspan H and Madabhushi A, eds. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. Cham: Springer: 3-11[ DOI: 10.1007/978-3-030-00889-5_1 http://dx.doi.org/10.1007/978-3-030-00889-5_1 ]
相关作者
相关机构
京公网安备11010802024621