L-UNet:轻量化云遮挡道路提取网络
L-UNet: lightweight network for road extraction in cloud occlusion scene
- 2021年26卷第11期 页码:2670-2679
收稿:2020-08-24,
修回:2020-12-21,
录用:2020-12-28,
纸质出版:2021-11-16
DOI: 10.11834/jig.200461
移动端阅览

浏览全部资源
扫码关注微信
收稿:2020-08-24,
修回:2020-12-21,
录用:2020-12-28,
纸质出版:2021-11-16
移动端阅览
目的
2
道路提取是常见的遥感应用之一。现有的基于深度卷积网络的道路提取方法往往未考虑云遮挡给道路提取带来的影响,且提取网络模型较大,不利于在移动端部署,同时缺乏用于云遮挡场景下的道路提取数据集。对此,本文提出一种轻量化的UNet网络(lightweight UNet,L-UNet),高效地实现云遮挡下的道路提取。
方法
2
通过柏林噪声模拟云层以扩展现有道路提取数据集,进而训练L-UNet。使用移动翻转瓶颈卷积模块作为特征提取的主要结构,在深度可分离卷积的基础上加入扩展卷积和压缩激励模块,在减少参数量的同时大幅提升了分割效果。
结果
2
在DeepGlobe道路提取扩展数据集的测试中,与D-LinkNet相比,L-UNet的交并比(intersection over union,IoU)提升了1.97%,而参数量仅为D-LinkNet的1/5。在真实云遮挡遥感图像道路提取测试中,L-UNet的性能仍然最优,与D-LinkNet和UNet相比,IoU值分别提高19.47%和31.87%。
结论
2
L-UNet网络具有一定的云遮挡区域下道路标签生成能力,虽然在模拟云遮挡数据集下训练得到,但对于真实云遮挡仍具有较强的鲁棒性。L-UNet模型参数量很小,易于嵌入移动端。
Objective
2
Road extraction is one of the primary tasks in the field of remote sensing. It has been applied in many areas
such as urban planning
route optimization
and navigation. Especially in the event of disasters such as mudslides
floods
and earthquakes
road information will suddenly change. Embedding road extraction models on the mobile terminal has an essential application value for rapid rescue. In recent years
deep learning has provided new ideas for realizing road pixel-level extraction
such as the classic image segmentation network UNet and the improved road extraction networks based on UNet
such as Residual UNet
LinkNet
and D-LinkNet. They can achieve road extraction better than traditional methods based on primary image feature extraction. However
these methods that rely on deep convolutional networks still have two problems: 1) cloud occlusion seriously affects the retrieval of information about ground objects in remote sensing images. At present
these convolutional network models are all trained on clear remote sensing images and do not consider the effect of cloud occlusion on road extraction. Their road extraction performance on cloudy remote sensing images is substantially reduced. 2) Network lightweight design has been an engaging area of research for several years. None of the above models based on deep learning considers the lightweight design of deep convolutional networks
which adds considerable difficulty to the deployment of these models. To address these road extraction problems
a lightweight UNet (L-UNet) is proposed
and road extraction is implemented in an end-to-end manner in the cloud occlusion scene.
Method
2
1) To address the problem of cloud occlusion
the Perlin noise is used to simulate a cloud layer image
and then the artificial cloud layer image and an RGB remote sensing image merge through the alpha coefficient to simulate the cloud occlusion scene. This simulation method is used to extend the cloudless road extraction dataset. Specifically
20 000 artificial cloud layer images have been generated before the network training. During training
cloud layer images are randomly sampled with replacement. The selected cloud layer image is merged with the clear remote sensing image in the training dataset
thereby simulating the continually changing cloud occlusion scenes. 2) In terms of network lightweight
UNet
a fully convolutional neural network
is improved to obtain L-UNet. The main improvement is the use of mobile inverted bottleneck convolutional blocks (MBConv) in the encoder. The MBConv first uses depthwise separable convolution
which considerably reduces the number of network params. However
the performance of road extraction only using depthwise separable convolution is not ideal; thus
expand convolution is added. Expand convolution with several 1×1 convolution kernels can increase the number of feature channels for each layer in the encoder part. Therefore
each layer of the network can learn more abundant features. The MBConv also uses a squeeze-and-excitation block. The block consists of two parts: global pooling for squeeze and 1×1 convolution with swish function for excitation. The squeeze-and-excitation block rationalizes the relative weights between the output feature maps of each layer. It highlights the feature information related to roads and clouds
which is beneficial to the segmentation tasks. Moreover
the swish is selected as the activation function rather than the rectified linear unit (ReLU). The L-UNet model reduces the param of the original UNet model and achieves better results. 3) The training loss function is the sum of the binary cross-entropy loss and the dice coefficient loss. The optimizer for network training is Adam
which has an initial learning rate of 2E-4. The encoder parameters of L-UNet are initialized to "ImageNet" pretrained model parameters. Then
the training is finetuned. The PyTorch deep learning framework is selected to implement L-UNet model construction and experiment. L-UNet performs 233 epochs of training on two NVIDIA GTX 1080 TI GPUs and finally converges.
Result
2
Network training and comparison experiments are carried out on the DeepGlobe road extraction extended dataset. 1) In the trial of comparing the network structure
the baseline network is UNet# that only contains depthwise separable convolution. When the expand convolution and squeeze-and-excitation block are added separately
the corresponding intersection over union (IoU) values increase by 1.12% and 8.45%
respectively. Adding the expand convolution and squeeze-and-excitation block simultaneously increases the IoU index by 16.24%. 2) L-UNet is compared with other networks on the extended test dataset. The IoU index of L-UNet rises by 4.65% compared with UNet. The IoU index increases by 1.97% compared with D-LinkNet
which is the second most powerful. The L-UNet param is only 22.28 M
which is 1/7 of UNet and 1/5 of D-LinkNet. The Mask-IoU and Mask-P indices
which are used to measure the network's road prediction performance in the cloud occlusion area
are also higher than those of other networks. 3) For road extraction tests on several real cloudy remote sensing images from Sentinel-2 satellite
the performance of L-UNet remains the best. The average IoU of the detection results is higher than D-LinkNet's 19.47% and UNet's 31.87%.
Conclusion
2
This paper studies the problem of road extraction from remote sensing images in cloud occlusion scenes. Simulated cloud layers are added on existing datasets
and extended datasets are used to improve the robustness of existing deep learning-based methods against cloud occlusion interference. The proposed L-UNet network architecture dramatically reduces the param and has excellent performance for road extraction given cloud cover. It can even predict road labels under thick clouds through known visible road edges and trends; thus
its road detection results have a better consistency. Other tasks for extracting remotely sensed ground objects with cloud cover can also use our method in future work.
Cheng G L, Wang Y, Xu S B, Wang H Z, Xiang S M and Pan C H. 2017. Automatic road detection and centerline extraction via cascaded end-to-end convolutional neural network. IEEE Transactions on Geoscience and Remote Sensing, 55(6): 3322-3337[DOI:10.1109/TGRS.2017.2669341]
Cheng Y, Wang D, Zhou P and Zhang T. 2018. Model compression and acceleration for deep neural networks: the principles, progress, and challenges. IEEE Signal Processing Magazine, 35(1): 126-136[DOI:10.1109/MSP.2017.2765695]
Demir I, Koperski K, Lindenbaum D, Pang G, Huang J, Basu S, Hughes F, Tuia D and Raskar R. 2018. DeepGlobe 2018: a challenge to parse the earth through satellite images//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. Salt Lake City, USA: IEEE: 172-181[ DOI: 10.1109/CVPRW.2018.00031 http://dx.doi.org/10.1109/CVPRW.2018.00031 ]
Enomoto K, Sakurada K, Wang W M, Fukui H, Matsuoka M, Nakamura R and Kawaguchi N. 2017. Filmy cloud removal on satellite imagery with multispectral conditional generative adversarial nets//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops. Honolulu, USA: IEEE: 1533-1541[ DOI: 10.1109/CVPRW.2017.197 http://dx.doi.org/10.1109/CVPRW.2017.197 ]
Howard A G, Zhu M L, Chen B, Kalenichenko D, Wang W J, Weyand T, Andreetto M and Adam H. 2017. MobileNets: efficient convolutional neural networks for mobile vision applications[EB/OL]. [2020-08-11] . https://arxiv.org/pdf/1704.04861.pdf https://arxiv.org/pdf/1704.04861.pdf
Hu J, Shen L, Albanie S, Sun G and Wu E H. 2020. Squeeze-and-excitation networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(8): 2011-2023[DOI:10.1109/TPAMI.2019.2913372]
Lin C H, Lai K H, Chen Z B and Chen J Y. 2014. Patch-based information reconstruction ofcloud-contaminated multitemporal images. IEEE Transactions on Geoscience and Remote Sensing, 52(1): 163-174[DOI:10.1109/TGRS.2012.2237408]
Milletari F, Navab N and Ahmadi S A. 2016. V-Net: fully convolutional neural networks for volumetric medical image segmentation//Proceedings of the 4th International Conference on 3D Vision (3DV). Stanford, USA: IEEE: 565-571[ DOI: 10.1109/3DV.2016.79 http://dx.doi.org/10.1109/3DV.2016.79 ]
Perlin K. 2002. Improving noise//Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques. San Antonio, USA: ACM: 681-682[ DOI: 10.1145/566570.566636 http://dx.doi.org/10.1145/566570.566636 ]
Porter T and Duff T. 1984. Compositing digital images//Proceedings of the 11th Annual Conference on Computer Graphics and Interactive Techniques. New York, USA: ACM: 253-259[ DOI: 10.1145/800031.808606 http://dx.doi.org/10.1145/800031.808606 ]
Ronneberger O, Fischer P and Brox T. 2015. U-Net: convolutional networks for biomedical image segmentation//Proceedings of 2015 International Conference on Medical Image Computing and Computer-assisted Intervention. Munich, Germany: Springer: 234-241[ DOI: 10.1007/978-3-319-24574-4_28 http://dx.doi.org/10.1007/978-3-319-24574-4_28 ]
Sandler M, Howard A, Zhu M, Zhmoginov A and Chen L C. 2018. MobileNetV2: inverted residuals and linear bottlenecks//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE: 4510-4520[ DOI: 10.1109/CVPR.2018.00474 http://dx.doi.org/10.1109/CVPR.2018.00474 ]
Sarukkai V, Jain A, Uzkent B and Ermon S. 2020. Cloud removal in satellite images using spatiotemporal generative networks//Proceedings of 2020 IEEE Winter Conference on Applications of Computer Vision. Snowmass, USA: IEEE: 1785-1794[ DOI: 10.1109/WACV45572.2020.9093564 http://dx.doi.org/10.1109/WACV45572.2020.9093564 ]
Shi W Z, Zhu C Q and Wang Y. 2001. Road feature extraction from remotely sensed image: review and prospects. Acta Geodaetica et Cartographica Sinica, 30(3): 257-262
史文中, 朱长青, 王昱. 2001. 从遥感影像提取道路特征的方法综述与展望. 测绘学报, 30(3): 257-262[DOI:10.3321/j.issn:1001-1595.2001.03.014]
Tan M X and Le Q V. 2019. EfficientNet: rethinking model scaling for convolutional neural networks[EB/OL]. [2020-08-11] . https://arxiv.org/pdf/1905.11946.pdf https://arxiv.org/pdf/1905.11946.pdf
Woo S, Kim D, Cho D and Kweon I S. 2018. LinkNet: relational embedding for scene graph//Proceedings of the 32nd International Conference on Neural Information Processing Systems. Montréal, Canada: Curran Associates Inc.: 558-568[ DOI: 10.5555/3326943.3326995 http://dx.doi.org/10.5555/3326943.3326995 ]
Zhang J J, Wan G T, Zhang H Q, Li S S and Feng X X. 2020. Rapid road extraction from quick view imagery of high-resolution satellites with transfer learning. Journal of Image and Graphics, 25(7): 1501-1512
张军军, 万广通, 张洪群, 李山山, 冯旭祥. 2020. 迁移学习下高分快视数据道路快速提取. 中国图象图形学报, 25(7): 1501-1512[DOI:10.11834/jig.190441]
Zhang Z X, Liu Q J and Wang Y H. 2018. Road extraction by deep residual U-Net. IEEE Geoscience and Remote Sensing Letters, 15(5): 749-753[DOI:10.1109/LGRS.2018.2802944]
Zhou L C, Zhang C and Wu M. 2018. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. Salt Lake City, USA: IEEE: 182-186[ DOI: 10.1109/CVPRW.2018.00034 http://dx.doi.org/10.1109/CVPRW.2018.00034 ]
Zoph B and Le Q V. 2016. Neural architecture search with reinforcement learning[EB/OL]. [2020-08-11] . https://arxiv.org/pdf/1611.01578.pdf https://arxiv.org/pdf/1611.01578.pdf
相关作者
相关机构
京公网安备11010802024621