改进预训练编码器U-Net模型的PET肿瘤自动分割
Automatic tumor segmentation in PET by deep convolutional U-Net with pre-trained encoder
- 2020年25卷第1期 页码:171-179
收稿:2019-03-08,
修回:2019-7-14,
录用:2019-7-21,
纸质出版:2020-01-16
DOI: 10.11834/jig.190058
移动端阅览

浏览全部资源
扫码关注微信
收稿:2019-03-08,
修回:2019-7-14,
录用:2019-7-21,
纸质出版:2020-01-16
移动端阅览
目的
2
为制定放疗计划并评估放疗效果,精确的PET(positron emission tomography)肿瘤分割在临床中至关重要。由于PET图像存在低信噪比和有限的空间分辨率等特点,为此提出一种应用预训练编码器的深度卷积U-Net自动肿瘤分割方法。
方法
2
模型的编码器部分用ImageNet上预训练的VGG19编码器代替;引入基于Jaccard距离的损失函数满足对样本重新加权的需要;引入了DropBlock取代传统的正则化方法,有效避免过拟合。
结果
2
PET数据库共包含1 309幅图像,专业的放射科医师提供了肿瘤的掩模、肿瘤的轮廓和高斯平滑后的轮廓作为模型的金标准。实验结果表明,本文方法对PET图像中的肿瘤分割具有较高的性能。Dice系数、Hausdorff距离、Jaccard指数、灵敏度和正预测值分别为0.862、1.735、0.769、0.894和0.899。最后,给出基于分割结果的3维可视化,与金标准的3维可视化相对比,本文方法分割结果可以达到金标准的88.5%,这使得在PET图像中准确地自动识别和连续测量肿瘤体积成为可能。
结论
2
本文提出的肿瘤分割方法有助于实现更准确、稳定、快速的肿瘤分割。
Objective
2
Positron emission tomography (PET) is a crucial technique established for patient administration in neurology
oncology
and cardiology. Particularly in clinical oncology
fluorodeoxy glucose PET is usually applied in therapy monitoring
radiotherapy planning
staging
diagnosis
and follow-up. Adaptive radiation therapy assists in radiation treatment with the hope that specific therapies aimed at individual patients and target tumors can be developed to re-optimize the treatment plan as early as possible. The use of PET greatly benefits adaptive radiation therapy. Manual delineation is time-consuming and highly dependent on observers. Previous studies have shown that automatic computer-generated segmentations are more reproducible than manual delineations
especially for radiomic analysis. Therefore
automatic and accurate tumor delineation is highly demanded for subsequent determination of therapeutic options and achievement of an improved prognosis. Over the past decade
dozens of methods have been used
which rely on multiple image segmentation approaches or composed methods from broad categories
including thresholding
region-based
contour-based
and graph-based methods as well as clustering
statistical techniques
and machine learning. However
those methods depend on hand-crafted features and possess a limited capability to represent features. For medical image segmentation
convolutional neural networks (CNNs) have demonstrated competitive performance. Nevertheless
these methods show the image classification based on region
in which the integral input image is split up and turns into small regions. Then
whether the small region belongs to the target (foreground) or not can be predicted by the CNN model of each small region. Therefore
each region merely stands for a partial area of the image; thus
this algorithm merely involves limited contextual knowledge that belongs to the small region. A U-Net
which is considered an optimal segmentation network for medical imaging
is trained end to end. It includes a contractive path and an expensive path produced by a combination of convolutional
up-sampling
and pooling layers. This architecture certified itself to be highly effective in using limited amounts of data for segmentation problems. Affected by recent achievements in deep learning
we exploit an automatic tumor segmentation method by deep convolutional U-Net with a pre-trained encoder.
Method
2
In this paper
we present a fully automatic method for tumor segmentation by using a 14-layer U-Net model with two blocks of a VGG19 encoder pre-trained with ImageNet. The pre-trained VGG19 encoder contains 260 160 trainable parameters. The rest of our network consists of 14 layers with 1 605 961 trainable parameters. We fix the stride at 2. We propose three-step strategies to ensure effective and efficient learning with limited training data. First
we use the first two blocks of VGG19 as the contracting path and introduce rectified linear units (ReLUs) to each convolutional layer as the activation function. For the symmetrically expanding path
we arrange ReLUs and batch normalization after each convolutional layer. The loss of boundary pixels in each convolution layer necessitates cropping. For the last layer
we use a 1×1 convolution to map each 64-channel feature vector
and each component expresses the chance that the corresponding input pixel is within a target tumor. Second
a tumor holds a small portion within an entire PET image. Therefore
the pixel-wise classification tends to be biased to the outside of targets
leading to a high probability to partially segment or miss tumors. A loss function based on Jaccard distance is applied to eliminate the need for sample re-weighting
which is a typical procedure when using cross-entropy as the loss function for image segmentation due to a strong imbalance between the number of foreground and background pixels. Third
we import the DropBlock technique to replace the normal regularization dropout method because the former can help the U-Net efficiently avoid overfitting. This approach is a structured shape of dropout in which units in a successive region of a feature map are dropped together. In addition to the convolution layers
applying DropBlock in skip connections increases the accuracy.
Result
2
A database that contains 1 309 PET images is applied to train and test the proposed segmentation model. We split the database into a before-radiotherapy (BR) sub-database and an after-radiotherapy (AR) sub-database. We use the mask
contour
and smoothed contour of a tumor
which are provided by an expert radiologist
as truths for teaching the proposed model. Experimental results on the BR sub-database show that our method presented a relatively high performance of tumor segmentation in PET images. The Dice coefficient (DI)
Hausdorff distance
Jaccard index
sensitivity
and positive predicted value (PPV) are 0.862
1.735
0.769
0.894
and 0.899
respectively. In the test stage
the total processing time of the testing dataset of the BR sub-database needs an average of 1.39 s
which can meet clinical real-time requirements. Then
we fine-tune the weights of the model that we have selected on the BR sub-database by training the network further with the AR sub-database. Experimental results indicate a good segmentation performance with a DI of 0.852
SE of 0.840
and PPV of 0.893. Compared with the traditional U-Net
our method increased by 5.9%
15.1%
1.9%
respectively. Finally
the volume of the segmented tumors in the PET images is presented
enabling the accurate automated identification and serial measurement of tumor volumes in PET images.
Conclusion
2
This study uses a 14-layer U-Net architecture with a VGG19 pre-trained encoder for tumor segmentation in PET images. We demonstrate how to improve the performance of the U-Net by using a technique called fine-tuning in an encoder of network for initializing weights. Although fine-tuning has been widely applied in image classification tasks
it has not been applied to the like-U-Net-type architectures for medical image segmentation tasks. We use the Jaccard distance as the loss function to improve the segmentation performance. Overall results show that our approach is suitable for various tumors with minimum post-processing and without pre-processing. We believe that this method could be generalized effectively to other medical image segmentation tasks.
Bai B, Bading J and Conti P S. 2013. Tumor quantification in clinical positron emission tomography. Theranostics, 3(10):787-801[DOI:10.7150/thno.5629]
Bi X J and Xiao J. 2011. Application of DE algorithm and improved GVF Snake model in segmentation of PET image. Journal of Image and Graphics, 16(3):382-388
毕晓君, 肖婧. 2011.差分进化算法GVF Snake模型在PET图像分割中的应用.中国图象图形学报, 16(3):382-388)[DOI:10.11834/jig.20110320]
Cha K H, Hadjiiski L, Samala R K, Chan H P, Caoili E M and Cohan R H. 2016. Urinary bladder segmentation in CT urography using deep-learning convolutional neural network and level sets. Medical Physics, 43(4):1882-1896[DOI:10.1118/1.4944498]
Chen S, Zhong S, Yao L, Shang Y and Suzuki K. 2016. Enhancement of chest radiographs obtained in the intensive care unit through bone suppression and consistent processing. Physics in Medicine&Biology, 61(6):2283-2301[DOI:10.1088/0031-9155/61/6/2283]
Day E, Betler J, Parda D, Reitz B, Kirichenko A, Mohammadi S and Miften M. 2009. A region growing method for tumor volume segmentation on PET images for rectal and anal cancer patients. Medical Physics, 36(10):4349-4358[DOI:10.1118/1.3213099]
Foster B, Bagci U, Mansoor A, Xu Z and Mollura D J. 2014. A review on segmentation of positron emission tomography images. Computers in Biology and Medicine, 50:76-96[DOI:10.1016/j.compbiomed.2014.04.014]
Ghiasi G, Lin T Y and Le Q V. 2018. DropBlock: a regularization method for convolutional networks[EB/OL]. (2018-10-30)[2019-07-19] . https://arxiv.org/pdf/1810.12890.pdf https://arxiv.org/pdf/1810.12890.pdf
Hatt M, Lee J A, Schmidtlein C R, Naqa I E, Caldwell C, De Bernardi E, Lu W, Das S, Geets X, Gregoire V, Jeraj R, Macmanus M P, Mawlawi O R, Nestle U, Pugachev A B, Schöder H, Shepherd T, Spezi E, Visvikis D, Zaidi H and Kirov A S. 2017a. Classification and evaluation strategies of auto-segmentation approaches for PET: report of AAPM task group No. 211. Medical Physics, 44(6): e1-e42[ DOI: 10.1002/mp.12124 http://dx.doi.org/10.1002/mp.12124 ]
Hatt M, Tixier F, Pierce L, Kinahan P E, Le Rest C C and Visvikis D. 2017b. Characterization of PET/CT images using texture analysis:the past, the present… any future? European Journal of Nuclear Medicine and Molecular Imaging, 44(1):151-165[DOI:10.1007/s00259-016-3427-0]
Lecun Y, Bottou L, Bengio Y and Haffner P. 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278-2324[DOI:10.1109/5.726791]
Lelandais B, Ruan S, Denoeux T, Vera P and Gardin I. 2014. Fusion of multi-tracer PET images for dose painting. Medical Image Analysis, 18(7):1247-1259[DOI:10.1016/j.media.2014.06. 014]
Long J, Shelhamer E and Darrell T. 2015. Fully convolutional networks for semantic segmentation//Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition. Boston, MA, USA: IEEE, 3431-3440[ DOI: 10.1109/CVPR.2015.7298965 http://dx.doi.org/10.1109/CVPR.2015.7298965 ]
Moussallem M, Valette P J, Traverse-Glehen A, Houzard C, Jegou C and Giammarile F. 2012. New strategy for automatic tumor segmentation by adaptive thresholding on PET/CT images. Journal of Applied Clinical Medical Physics, 13(5):236-251[DOI:10.1120/jacmp.v13i5.3875]
Onoma D P, Ruan S, Gardin I, Monnehan G A, Modzelewski R and Vera P. 2012. 3D random walk based segmentation for lung tumor delineation in PET imaging//The 9th IEEE International Symposium on Biomedical Imaging. Barcelona, Spain: IEEE, 1260-1263[ DOI: 10.1109/ISBI.2012.6235791 http://dx.doi.org/10.1109/ISBI.2012.6235791 ]
Pereira S, Pinto A, Alves V and Silva C A. 2016. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Transactions on Medical Imaging, 35(5):1240-1251[DOI:10.1109/TMI.2016.2538465]
Ronneberger O, Fischer P and Brox T. 2015. U-Net: convolutional networks for biomedical image segmentation//Proceedings of the 18th InternationalConference on Medical Image Computing and Computer-Assisted Intervention. Munich, Germany: Springer, 234-241[ DOI: 10.1007/978-3-319-24574-4_28 http://dx.doi.org/10.1007/978-3-319-24574-4_28 ]
Thorwarth D, Geets X and Paiusco M. 2010. Physical radiotherapy treatment planning based on functional PET/CT data. Radiotherapy and Oncology, 96(3):317-324[DOI:10.1016/j.radonc.2010.07.012]
Yan D, Vicini F, Wong J and Martinez A. 1997. Adaptive radiation therapy. Physics in Medicine&Biology, 42(1):123-132[DOI:10.1088/0031-9155/42/1/008]
Yuan Y, Chao M and Lo Y C. 2017. Automatic skin lesion segmentation using deep fully convolutional networks with jaccard distance. IEEE Transactions on Medical Imaging, 36(9):1876-1886[DOI:10.1109/TMI.2017.2695227]
Zhong Z, Kim Y, Zhou L, Plichta K, Allen B, Buatti J and Wu X. 2018. 3D fully convolutional networks for co-segmentation of tumors on PET-CT images//Proceedings of the 15th IEEE International Symposium on Biomedical Imaging. Washington DC, USA: IEEE, 228-231[ DOI: 10.1109/ISBI.2018.8363561 http://dx.doi.org/10.1109/ISBI.2018.8363561 ]
相关作者
相关机构
京公网安备11010802024621