深度卷积网络卫星图像水泥厂目标检测
Cement plant detection on satellite images using deep convolution network
- 2019年24卷第4期 页码:550-561
收稿:2018-07-04,
修回:2018-9-14,
纸质出版:2019-04-16
DOI: 10.11834/jig.180424
移动端阅览

浏览全部资源
扫码关注微信
收稿:2018-07-04,
修回:2018-9-14,
纸质出版:2019-04-16
移动端阅览
目的
2
水泥厂作为重要的污染源企业需要对其进行统计和监管,近几年随着卫星遥感技术的发展和遥感影像分辨率的提高,使得基于卫星影像进行水泥厂目标检测成为可能。但是由于遥感图像中建筑目标的环境复杂多变,同时各个水泥厂在生产规模、设备构成、厂区结构、坐落方位上存在较大差异,图像表观上的形态各异和复杂环境干扰使得传统图像识别方法难以设计和提取有效特征。鉴于深度学习在视觉目标检测领域的成功应用,本文将研究应用深度卷积神经网络方法,实现在卫星图像上识别与定位水泥厂目标,为环保部门提供一种高效便捷的水泥厂目标检测和统计方法。
方法
2
基于面向目标检测与定位的Faster R-CNN深度学习框架,以准确检测与定位水泥厂区域为目的,以京津冀地区的水泥厂位置作为训练和测试数据集,选用3种结构不同的提取特征卷积神经网络模型进行了对比实验。并针对小样本训练容易出现的过拟合和误检问题,采用图像去雾预处理、数据扩充、引入负样本等技术进一步提升模型能力。
结果
2
测试集实验结果表明ResNet特征提取网络效果最好,准确率达到74%。为了进一步提高检出率并降低误检率,引入3种模型能力提升方法,在扩充检测数据集中的检出率达到94%,误检率降低到14%;在全球水泥厂数据集中的图像检出率达到96%,万幅随机图像的误检数量为30幅(0.3%)。对上海地区的卫星图像进行扫描检测,结果检测出11个已登记的水泥厂(共登记16个),另外还检测出17个未登记的水泥厂。
结论
2
对于卫星地图上水泥厂这种具有特殊建筑构造但也存在厂区几何形状各异、所处地理环境复杂、随季节性变化等特点,本文提出的基于深度卷积网络的卫星图像水泥厂检测方法,能够自动学习提取有效的图像特征并对目标进行准确检测。针对小样本训练问题,引入3种方法显著提高了模型的检测精度。在模型泛化能力测试中,经过优化后的模型在水泥厂建筑目标检测任务中表现良好,具有重要的应用价值。
Objective
2
The problem of environmental pollution in China has become increasingly serious with the rapid development of society and economy. Creating a beautiful ecological environment is becoming an important issue in the current national planning. Environmental pollution problems affect the sustainable development of the country and society. As an important source of pollution
cement plants must be effectively counted and monitored. With the development of satellite remote sensing
high-resolution and good quality images become available. At the same time
deep learning has made great progress in the field of target detection
and many excellent deep convolutional network models
such as Faster R-CNN
YOLO
SSD
and Mask R-CNN
have been proposed recently. In the object detection task on satellite images
huge differences lie in the plant area scales
the structures of equipment composition
and the orientations of each cement plant. Thus
various cement plants are presented with various appearances due to the complex natural geographical surroundings. Overcoming the problem of cement plant target detection and recognition using traditional artificial image feature methods is difficult. However
deep learning has achieved excellent performance in the field of image target recognition. The application of the deep convolutional network may be a brilliant method to locate cement plants on satellite images.
Method
2
A method of detecting and locating cement plant position using high-resolution satellite images was proposed based on the convolutional neural network of Faster R-CNN framework for image target detection. First
we used GoogleMap API to download Google Earth satellite images. We developed this high-resolution satellite image dataset of cement factory target using GoogleMap web API
which contains 464 cement plant locations
according to the Beijing-Tianjin-Hebei cement dataset given by the Satellite Environment Center
Ministry of Environmental protection. Through the training and testing datasets of the cement plant in Beijing-Tianjin-Hebei with three different feature extracting modules (namely
VGG(visual geometry group network)
ZF
and ResNet)
we compared the testing results among the three CNN models. Three methods
which include image haze removal using dark channel prior
data augmentation
and adding negative training samples
were introduced to solve the problems of overfitting and reduce high false positive rate because of insufficient amount of training data. We also verified the influence of different numbers of negative training samples on model training. We used these features to assist cement plant target detection
considering the characteristics of cement buildings with evident cylindrical cement tanks and heating reaction tower buildings.
Result
2
The visualized images of the convolutional feature map show that the identification of the deep convolution network for cement plant target detection is mainly based on special buildings in the plant area
such as cylindrical cement tanks
heating towers
and rectangular plants. The experimental results on the test set reveal that the ResNet achieves the best performance with the average accuracy rate of 74%. An optimization method by three methods is proposed to further enhance the detection of accurate rate and suppress the false positive rate. The precision of the promoted CNN model reached 94% in the augmented testing dataset
and the false positive rate was reduced to 14%. The true positive detection rate in the global cement plant dataset reached 96%
and the number of false detection of 10 000 random satellite images was reduced to 30 (0.3%). The actual cement plant target scan detection of the satellite images in the entire Shanghai area was conducted. To avoid dividing the cement plant into two image parts
we used the overlapping detection method
which also detected adjacent areas between satellite images. As a result
11 out of 16 registered cement plants
and 17 unregistered cement plants are detected.
Conclusion
2
The cement plant has special buildings of different shapes and background
even varying over time
thereby resulting in relatively difficult detection task. However
the method of cement plant detection on satellite images based on deep convolutional networks can automatically learn to extract effective features and can identify the position of the target in the image. In addition
some optimization methods
including image preprocessing
data augmentation
and adding negative samples to promote the performance of the model
are adopted to improve the model detection accuracy and solve the problem of few training data. The geographic latitude and longitude coordinates of cement targets are easy to obtain based on the result of the image object detection and its position information because the satellite image is geocoded. We can also estimate the cement plant area using information of detected boxes. In the model generalization capability test experiment
the proposed method achieves good performance in the detection and location tasks in the global cement dataset. The scanning results of the entire satellite image set in Shanghai indicate that the deep convolution network target detection method not only can detect most registered cement plants but also multiple unregistered cement plants. The method provides a reliable reference for monitoring environmental pollution sources. Furthermore
this model can be easily converted to detect other architectural targets using transfer learning techniques.
Richards J A, Jia X P. Remote Sensing Digital Image Analysis[M]. Berlin Heidelberg:Springer-Verlag, 2006:381-433.[DOI:10.1007/3-540-29711-1]
Zhang G M, Yin J P, Zhu E, et al. Hierarchical object detection in remote sensing image based on vision saliency[J]. Journal of Jilin University:Engineering and Technology Edition, 2007, 37(3):625-629.
张国敏, 殷建平, 祝恩, 等.遥感图像中基于视觉显著性的分层目标检测[J].吉林大学学报:工学版, 2007, 37(3):625-629. [DOI:10.13229/j.cnki.jdxbgxb2007.03.028]
Geng L L, Lin J, Long X X, et al. Research on feature matching algorithm for ZY-3 image[J]. Spacecraft Recovery&Remote Sensing, 2012, 33(3):93-99.
耿蕾蕾, 林军, 龙小祥, 等. "资源三号"卫星图像影像特征匹配方法研究[J].航天返回与遥感, 2012, 33(3):93-99. [DOI:10.3969/j.issn.1009-8518.2012.03.012]
Pal M. Random forest classifier for remote sensing classification[J]. International Journal of Remote Sensing, 2005, 26(1):217-222.[DOI:10.1080/01431160412331269698]
Krizhevsky A, Sutskever I, Hinton G E. ImageNet classification with deep convolutional neural networks[C]//Proceedings of the 25th International Conference on Neural Information Processing Systems. Lake Tahoe, Nevada: ACM, 2012: 1097-1105.
Zhou M F, Wang X L, Wang L, et al. Convolutional neural network models for high spatial resolution satellite imagery classification[J]. Journal of Image and Graphics, 2017, 22(7):996-1007.
周明非, 汪西莉, 王磊, 等.高分辨卫星图像卷积神经网络分类模型[J].中国图象图形学报, 2017, 22(7):996-1007. [DOI:10.11834/jig.160627]
Maggiori E, Tarabalka Y, Charpiat G, et al. Convolutionalneural networks for large-scale remote-sensing image classification[J]. IEEE Transactions on Geoscience and Remote Sensing, 2017, 55(2):645-657.[DOI:10.1109/TGRS.2016.2612821]
Chen Y S, Lin Z H, Zhao X, et al. Deep learning-based classification of hyperspectral data[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2014, 7(6):2094-2107.[DOI:10.1109/JSTARS.2014.2329330]
Girshick R. Fast R-CNN[C]//Proceedings of the IEEE International Conference on Computer Vision. 2015: 1440-1448.[ DOI:10.1109/ICCV.2015.169 http://dx.doi.org/10.1109/ICCV.2015.169 ]
Felzenszwalb P F, Girshick R B, McAllester D, et al. Object detection with discriminatively trained part-based models[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 32(9):1627-1645.[DOI:10.1109/TPAMI.2009.167]
Ren S Q, He K M, Girshick R, et al. Faster R-CNN:towards real-time object detection with region proposal networks[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017, 39(6):1137-1149.[DOI:10.1109/TPAMI.2016.2577031]
Liu W, Anguelov D, Erhan D, et al. SSD: single shot multibox detector[C]//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, The Netherlands: Springer, 2016: 21-37.[ DOI:10.1007/978-3-319-46448-0_2 http://dx.doi.org/10.1007/978-3-319-46448-0_2 ]
Redmon J, Divvala S, Girshick R, et al. You only look once: unified, real-time object detection[C]//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV, USA: IEEE, 2016: 779-788.[ DOI:10.1109/CVPR.2016.91 http://dx.doi.org/10.1109/CVPR.2016.91 ]
Zeiler M D, Fergus R. Visualizing and understanding convolutional networks[C]//Proceedings of the 13th European Conference on Computer Vision. Zurich, Switzerland: Springer, 2014: 818-833.[ DOI:010.1007/978-3-319-10590-1_53 http://dx.doi.org/010.1007/978-3-319-10590-1_53 ]
Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition[EB/OL]. 2015-04-10[2018-05-01] . https://arxiv.org/pdf/1409.1556.pdf https://arxiv.org/pdf/1409.1556.pdf .
He K M, Zhang X Y, Ren S Q, et al. Deep residual learning for image recognition[C]//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV, USA: IEEE, 2016: 770-778.[ DOI:10.1109/CVPR.2016.90 http://dx.doi.org/10.1109/CVPR.2016.90 ]
Xiang Y Y, Gao S, Chen Y H, et al. A Comparative study of Baidu map API, Lbs Amap API and Google maps API[J]. Software Guide, 2017, 16(9):19-21, 25.
向玉云, 高爽, 陈云红, 等.百度、高德及Google地图API比较研究[J].软件导刊, 2017, 16(9):19-21, 25. [DOI:10.11907/rjdk.171478]
Nasrabadi N M. Pattern recognition and machine learning[J]. Journal of Electronic Imaging, 2007, 16(4):049901.[DOI:10.1117/1.2819119]
Jia Y Q, Shelhamer E, Donahue J, et al. Caffe: Convolutional architecture for fast feature embedding[C]//Proceedings of the 22nd ACM International Conference on Multimedia. Orlando, Florida, USA: ACM, 2014: 675-678.[ DOI:10.1145/2647868.2654889 http://dx.doi.org/10.1145/2647868.2654889 ]
He K M, Sun J, Tang X O. Single image haze removal using dark channel prior[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2011, 33(12):2341-2353.[DOI:10.1109/TPAMI.2010.168]
Gonzalez R C, Woods R E, Eddins S L. Digital image processing using MATLAB[M]. Upper Saddle River, New Jersey:Pearson-Prentice-Hall, 2004:37-38.[DOI:10.5772/63028]
CemNet. The global cement report-online database of cement plants[EB/OL]. 2017-06-01[2017-12-01] . https://www.cemnet.com/global-cement-report/ https://www.cemnet.com/global-cement-report/
相关作者
相关机构
京公网安备11010802024621