2D级联CNN模型的放疗危及器官自动分割
Automatic segmentation of organs at risk in radiotherapy using 2D cascade-CNN model
- 2020年25卷第10期 页码:2110-2118
收稿:2020-06-10,
修回:2020-7-4,
录用:2020-7-11,
纸质出版:2020-10-16
DOI: 10.11834/jig.200281
移动端阅览

浏览全部资源
扫码关注微信
收稿:2020-06-10,
修回:2020-7-4,
录用:2020-7-11,
纸质出版:2020-10-16
移动端阅览
目的
2
精准的危及器官(organs at risk,OARs)勾画是肿瘤放射治疗过程中的关键步骤。依赖人工的勾画方式不仅耗费时力,且勾画精度容易受图像质量及医生主观经验等因素的影响。本文提出了一种2D级联卷积神经网络(convolutional neural network,CNN)模型,用于放疗危及器官的自动分割。
方法
2
模型主要包含分类器和分割网络两部分。分类器以VGG(visual geometry group)16为骨干结构,通过减少卷积层以及加入全局池化极大地降低了参数量和计算复杂度;分割网络则是以U-Net为基础,用双线性插值代替反卷积对特征图进行上采样,并引入Dropout层来缓解过拟合问题。在预测阶段,先利用分类器从输入图像中筛选出包含指定器官的切片,然后使用分割网络对选定切片进行分割,最后使用移除小连通域等方法对分割结果进一步优化。
结果
2
本文所用数据集共包含89例宫颈癌患者的腹盆腔CT(computed tomography)图像,并以中国科学技术大学附属第一医院多位放射医师提供的手工勾画结果作为评估的金标准。在实验部分,本文提出的分类器在6种危及器官(左右股骨、左右股骨头、膀胱和直肠)上的平均分类精度、查准率、召回率和F1-Score分别为98.36%、96.64%、94.1%和95.34%。基于上述分类性能,本文分割方法在测试集上的平均Dice系数为92.94%。
结论
2
与已有的CNN分割模型相比,本文方法获得了最佳的分割性能,先分类再分割的策略能够有效地避免标注稀疏问题并减少假阳性分割结果。此外,本文方法与专业放射医师在分割结果上具有良好的一致性,有助于在临床中实现更准确、快速的危及器官分割。
Objective
2
Accurate delineation of organs at risk (OARs) is an essential step in the radiation therapy for cancers. However
this procedure is frequently time consuming and error prone because of the large anatomical variation across patients
different experience of observers
and poor soft-tissue contrast in computed tomography (CT) scans. A computer-aided analysis system for OAR auto-segmentation from CT images will reduce the burden of doctors and the subjective errors and improve the effect of radiotherapy. In the early years
atlas-based methods are extremely popular and widely used in anatomy segmentation. However
the performance of atlas-based segmentation methods can be easily affected by various factors
such as the quality of atlas and registration methods. Recently
profits from the rapid growth of computing power and the amount of available data
deep learning
especially deep convolutional neural networks (CNNs)
has shown great potential in the field of image analysis. For most of the medical image segmentation tasks
the algorithms based on CNN outperform traditional methods. As a special fully CNN
U-Net adopts the design of encoder-decoder and fuses the high- and low-level features by skip connections to realize pixelwise segmentation. Given the outstanding performance of U-Net
numerous derivatives of the U-Net architecture have been gradually developed in various organ segmentation tasks. V-Net is proposed as an improvement scheme of U-Net to address the difficulties in processing 3D data. V-Net can fully utilize the 3D characteristics of images
although it is unsuitable for the 3D medical image datasets with few samples and the segmentation tasks of small organs. Therefore
a two-step 2D CNN model is proposed for the automatic and accurate segmentation of OARs in radiotherapy.
Method
2
In this study
we propose a novel cascade-CNN model that mainly includes a slice classifier and a 2D organ segmentation network. visual geometry group (VGG)16 is used as the backbone structure of the classifier and modified accordingly by considering the time overhead and the size of available data. To reduce the parameters and calculation complexity
three convolutional layers are removed
and an additional global max pooling layer is used as the bridging module between the feature extraction and the fully connected layer. The segmentation network is built upon the U-Net by replacing the deconvolutional layers with bilinear interpolation layers to perform the upsampling for features. The dropout layer and the data augmentation are utilized to avoid the overfitting problem. The classifier and the segmentation network for each organ are implemented in Keras and independently trained with the binary cross-entropy and dice losses
respectively. We use adaptive moment estimation to stabilize the gradient descent process during the training of the two models. In the inference stage
the slices containing the target organs are first selected by the classifier from the entire CT scans. These slices are then used as the input of the segmentation network to obtain the results
and some simple post-processing methods are applied to optimize the segmentation results.
Result
2
The dataset containing CT scans of 89 cases of cervical cancer patients and the manual segmentation results provided by multiple radiologists in The First Affiliated Hospital of University of Science and Technology of China (USTC) are used as the gold standard for evaluation. In the experimental part
the average classification accuracy
precision
recall
and F1-score of the classifier on the six organs (left femoral head
right femoral head
left femur
right femur
bladder
and rectum) are 98.36%
96.64%
94.1%
and 95.34%
respectively. On the basis of the performance of the above classifiers
the proposed method achieves high segmentation accuracy on the bladder
left femoral head
right femoral head
left femur
and right femur
with Dice coefficients of 94.16%
93.69%
95.09%
96.14%
and 96.57%
respectively. Compared with the single U-Net and cascaded V-Net
the Dice coefficient increases by 4.1% to 6.6%. For the rectum
all methods perform poorly because of the irregular shape and low contrast. The proposed method achieves a Dice coefficient of 72.99% on the rectum
which is approximately 20% higher than other methods. The comparison experiments demonstrate that the classifiers can effectively improve the overall segmentation accuracy.
Conclusion
2
In this work
we propose a novel 2D cascade-CNN model
which is composed of a classifier and a segmentation network
for the automatic segmentation of OARs in radiation therapy. The experiments demonstrate that the proposed method effectively alleviates the labeling sparse problem and reduces the false positive segmentation results. For the organs that immensely vary in shape and size
the proposed method obtains a significant improvement in segmentation accuracy. In comparison with existing neural network methods
the proposed method achieves state-of-the-art results in the segmentation task of OARs for cervical cancer and a better consistency with that of experienced radiologists.
Han M F, Yao G, Zhang W H, Mu G R, Zhan Y Q, Zhou X and Gao Y Z. 2019. Segmentation of CT thoracic organs by multi-resolution VB-nets[EB/OL].[2020-05-30] . http://ceur-ws.org/Vol-2349/SegTHOR2019_paper_1.pdf http://ceur-ws.org/Vol-2349/SegTHOR2019_paper_1.pdf
Han X, Hoogeman M S, Levendag P C, Hibbard L S, Teguh D N, Voet P, Cowen A C and Wolf T K. 2008. Atlas-based auto-segmentation of head and neck CT images//Proceedings of the 11th International Conference on Medical Image Computing and Computer-Assisted Intervention. New York: Springer: 434-441[ DOI:10.1007/978-3-540-85990-1_52 http://dx.doi.org/10.1007/978-3-540-85990-1_52 ]
He H and Chen S. 2020. Automatic tumor segmentation in PET by deep convolutional U-Net with pre-trained encoder. Journal of Image and Graphics, 25(1):171-179
何慧, 陈胜. 2020.改进预训练编码器U-Net模型的PET肿瘤自动分割.中国图象图形学报, 25(1):171-179 [DOI:10.11834/jig.190058]
He K M, Zhang X Y, Ren S Q and Sun J. 2016. Deep residual learning for image recognition//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas: IEEE: 770-778[ DOI:10.1109/CVPR.2016.90 http://dx.doi.org/10.1109/CVPR.2016.90 ]
Hu J, Shen L and Sun G. 2018. Squeeze-and-excitation networks//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE: 7132-7141[ DOI:10.1109/CVPR.2018.00745 http://dx.doi.org/10.1109/CVPR.2018.00745 ]
Ibragimov B and Xing L. 2017. Segmentation of organs-at-risks in head and neck CT images using convolutional neural networks. Medical Physics, 44(2):547-557[DOI:10.1002/mp.12045]
Krizhevsky A, Sutskever I and Hinton G E. 2012. ImageNet classification with deep convolutional neural networks//Proceedings of the 25th International Conference on Neural Information Processing Systems. Lake Tahoe: ACM: 1097-1105
Litjens G, Kooi T, Ehteshami Bejnordi B, Setio A A A, Ciompi F, Ghafoorian M, van der Laak J A W M, van Ginneken B and Sánchez C I. 2017. A survey on deep learning in medical image analysis. Medical Image Analysis, 42:60-88[DOI:10.1016/j.media.2017.07.005]
Men K, Dai J R and Li Y X. 2017. Automatic segmentation of the clinical target volume and organs at risk in the planning CT for rectal cancer using deep dilated convolutional neural networks. Medical Physics, 44(12):6377-6389[DOI:10.1002/mp.12602]
Miller K D, Nogueira L, Mariotto A B, Rowland J H, Yabroff K R, Alfano C M, Jemal A, Kramer J L and Siegel R L. 2019. Cancer treatment and survivorship statistics, 2019. CA:A Cancer Journal for Clinicians, 69(5):363-385[DOI:10.3322/caac.21565]
Milletari F, Navab N and Ahmadi S A. 2016. V-net: fully convolutional neural networks for volumetric medical image segmentation//Proceedings of the 4th International Conference on 3D Vision. Stanford: IEEE: 565-571[ DOI:10.1109/3DV.2016.79 http://dx.doi.org/10.1109/3DV.2016.79 ]
Ronneberger O, Fischer P and Brox T. 2015. U-net: convolutional networks for biomedical image segmentation//Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention. Munich: Springer: 234-241[ DOI:10.1007/978-3-319-24574-4_28 http://dx.doi.org/10.1007/978-3-319-24574-4_28 ]
Salehi S M, Erdogmus D and Gholipour A. 2017. Tversky loss function for image segmentation using 3D fully convolutional deep networks//Proceedings of the 8th International Workshop on Machine Learning in Medical Imaging. Quebec City: Springer: 379-387[ DOI:10.1007/978-3-319-67389-9_44 http://dx.doi.org/10.1007/978-3-319-67389-9_44 ]
Shelhamer E, Long J and Darrell T. 2017. Fully convolutional networks for semantic segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(4):640-651[DOI:10.1109/TPAMI.2016.2572683]
Simonyan K and Zisserman A. 2014. Very deep convolutional networks for large-scale image recognition[EB/OL].[2020-05-30] . https://arxiv.org/pdf/1409.1556v6.pdf https://arxiv.org/pdf/1409.1556v6.pdf
Tam C M, Yang X, Tian S, Jiang X, Beitler J J and Li S. 2018. Automated delineation of organs-at-risk in head and neck CT images using multi-output support vector regression//Proceedings of SPIE 10578, Medical Imaging 2018: Biomedical Applications in Molecular, Structural, and Functional Imaging. Houston: SPIE: 1-10[ DOI:10.1117/12.2292556 http://dx.doi.org/10.1117/12.2292556 ]
Trullo R, Petitjean C, Ruan S, Dubray B, Nie D and Shen D. 2017. Segmentation of organs at risk in thoracic CT images using a sharpmask architecture and conditional random fields//Proceedings of the 14th IEEE International Symposium on Biomedical Imaging. Melbourne: IEEE: 1003-1006[ DOI:10.1109/ISBI.2017.7950685 http://dx.doi.org/10.1109/ISBI.2017.7950685 ]
Vesal S, Ravikumar N and Maier A. 2019. A 2D dilated residual U-Net for multi-organ segmentation in thoracic CT.[EB/OL].[2020-05-30] . https://arxiv.org/pdf/1905.07710.pdf https://arxiv.org/pdf/1905.07710.pdf
Wang P Q, Chen P F, Yuan Y, Liu D, Huang Z H, Hou X D and Cottrell G. 2018. Understanding convolution for semantic segmentation//Proceedings of 2018 IEEE Winter Conference on Applications of Computer Vision. Lake Tahoe: IEEE: 1451-1460[ DOI:10.1109/WACV.2018.00163 http://dx.doi.org/10.1109/WACV.2018.00163 ]
Zheng S, Jayasumana S, Romera-Paredes B, Vineet V, Su Z Z, Du D L, Huang C and Torr P H S. 2015. Conditional random fields as recurrent neural networks//Proceedings of 2015 IEEE International Conference on Computer Vision. Santiago: IEEE: 1529-1537[ DOI:10.1109/ICCV.2015.179 http://dx.doi.org/10.1109/ICCV.2015.179 ]
Zhou Y X and Bai J. 2007. Multiple abdominal organ segmentation:an atlas-based fuzzy connectedness approach. IEEE Transactions on Information Technology in Biomedicine, 11(3):348-352[DOI:10.1109/titb.2007.892695]
Zhu W T, Huang Y F, Zeng L, Chen X M, Liu Y, Qian Z, Du N, Fan W and Xie X H. 2019. AnatomyNet:deep learning for fast and fully automated whole-volume segmentation of head and neck anatomy. Medical Physics, 46(2):576-589[DOI:10.1002/mp.13300]
相关作者
相关机构
京公网安备11010802024621