自适应权重金字塔和分支强相关的SAR图像舰船检测
Ship detection in SAR images based on adaptive weight pyramid and branch strong correlation
- 2022年27卷第10期 页码:3127-3138
收稿:2021-05-25,
修回:2021-8-12,
录用:2021-8-19,
纸质出版:2022-10-16
DOI: 10.11834/jig.210373
移动端阅览

浏览全部资源
扫码关注微信
收稿:2021-05-25,
修回:2021-8-12,
录用:2021-8-19,
纸质出版:2022-10-16
移动端阅览
目的
2
合成孔径雷达(synthetic aperture radar
SAR)图像舰船目标检测在军事、民生领域发挥重要作用。由于SAR图像背景复杂且多为小尺度舰船目标,同时图像中的相干斑噪声导致舰船目标边缘模糊,现阶段目标检测模型无法快速高效地实现目标检测任务。为了提高模型检测精度,使模型具有更好的鲁棒性,提出了自适应权重金字塔和分支强相关的SAR图像检测模型。
方法
2
对特征提取网络提取的特征图经过采样、融合处理获得特征自适应权重,然后利用权重指导每层特征图充分融合空间位置信息和语义信息,更好地检测小尺度目标;分支强相关模块融合分类分支和回归分支的特征,对融合后的待检测特征分别采用1×1、3×3对称卷积核和1×3、3×1非对称卷积核捕获不同的舰船特征;构建IoU(intersection over union)分支,利用IoU分支作用于分类分支,避免高IoU低分类置信度的候选框被抑制,通过设置平衡因子平衡IoU分支和分类分支,使其能更好地指导回归分支优化候选框。
结果
2
在公开的遥感数据集SSDD(SAR ship detection dataset)上实验结果表明,本文模型的检测精度达到90.53%,F
1
值提升至94.35%,检测速度达到20.14帧/s。与其他SAR图像舰船目标检测算法相比有较好的检测效果。
结论
2
实验结果表明该模型相比原始模型具有更好的检测效果,满足实时性检测需求,在SAR图像实时检测中具有实际意义。
Objective
2
Synthetic aperture radar(SAR) imaging is interfered with the echo due to electromagnetic waves penetrate the rough object surface by emitting electromagnetic pulses and receiving reflected echo imaging
resulting in coherent spot noise. However
its multi-temporal and all-weather features are used in maritime applications like rescue
monitoring and vessel transportation supervision. Constant false alarm rate algorithm is use for SAR image target detection in common. It has the simple algorithm structure
but the complex and uneven distribution of ocean clutter types can cause missing detection. In the case of multiple targets
it is easy to appear that the reference unit contains target information
so the amplitude of target signal may be greater than the clutter amplitude
resulting in the estimated threshold is far greater than the real threshold. The constant false alarm rate algorithm cannot obtain the ideal detection effect. In recent years
with the rapid development of deep learning
object detection algorithms based on deep learning have achieved good results in optical image object detection. SAR images are different from optical images in that the targets in SAR images are mostly small-scale ships with bright spots and lack obvious details. Similar targets such as islands
buildings and so on will form similar ship shapes after electromagnetic wave reflection. Therefore
the application of optical detection models directly to SAR images cannot achieve ideal results. Feature pyramid network adopts top-down layer by layer fusion
so the high-level feature map lacks spatial location information
and the fusion of each level largely loses the semantic information contained in high-level features
while the semantic information of low-level features is insufficient. In order to improve the model detection accuracy and make the model more robust
a SAR image detection model with adaptive weight pyramid and branch strong correlation is proposed in this paper.
Method
2
The high-level features extracted by ResNet101 contain rich semantic information
and the underlying features contain rich spatial information. To have a better detection effect of target detection model
the feature map is required to contain semantic information and spatial location information both. Therefore
the featured information needs to be adopted based on a more effective fusion mechanism for fusion. It is necessary to use effective weights to guide the fusion mechanism for to the integration of more effective information. Our adaptive weight pyramid module is demonstrated
which first samples the feature map to make it have the same scale
then stitches concatenation the feature map after each layer of sampling into channel dimensions
adjusts the number of channels by 1×1 convolution
and generates the weight in the end through the Softmax function. To get a feature map of the layer
multiply each layer feature by the corresponding weight
and the other layers are analogy. Its fusion removes some redundant information under the weighting of machine learning. Multiple features are extracted through 1×1 and 3×3 symmetric convolution kernels and 1×3 and 3×1 asymmetric convolution kernels
following the feature fusion of classification branch and regression branch by the branch correlation module. The feature diagram contains the classification branch information in the regression branch based on the fused classification and regression features. In the detection phase
the candidate boxes are removed related to accurate positioning and low classification confidence
which reduces the detection accuracy of the model greatly. In order to solve this problem
the intersection over union(IoU) detection head is constructed in the regression branch
and IoU it is used for classification branch
thus avoiding the problem of the candidate box with high IoU low classification confidence being suppressed. In order to balance the IoU branch and the classification branch
the constrained balance factor is designed sand can guide the regression branch optimization candidate box better.
Result
2
Our model is evaluated on the public remote sensing dataset SSDD(SAR ship detection)
which had 1 160 pictures
including 2 456 vessel targets. The dataset is divided dataset into training sets and test sets by a ratio of 7 : 3
and enhanced the learning ability of the model by rotating
flipping
brightness and increasing Gaussian noise. All the experimental environments are based on ubuntu16.04 LTS operating system
equipped with CPU for lntel (R) Core (TM) i7-7700@3.6 GHz ×8
graphics card for NVIDIA GTX1080Ti memory for 11 GB
running under the Tensorflow framework
through CUDA8.0 and cuDNN5.0 accelerated training. Our model learning rate is set to 0.000 5
the learning rate is attenuated by 1/10 per 40 k iterations
and the network is convergent at 80 k iterations completely. The non-maximum suppression threshold is 0.5 and the predicted probability threshold is 0.6. The recall rate
accuracy rate
average accuracy
F1 value and detection speed are represented as the evaluation indexes of the model. The model realizes the best detection effect in terms of equilibrium factor takes 0.4. The final recall rate is increased from 89.17% to 93.63%
the accuracy rate is increased from 87.94% to 95.08%
the average accuracy is increased from 86.91% to 90.53%
F1 value is increased from 88.55% to 94.53%
and the detection speed reached 20.14 frame/s.
Conclusion
2
Our experimental results demonstrate that the feature map targets fused by adaptive weight pyramids are more prominent
and the strong correlated branches module enhances the correlation between the classification branch and the regression branch
and improves the candidate box regression ability. Our model has it potential to meet the real-time detection requirement of SAR imaging.
Abu A and Diamant R. 2020. CFAR detection algorithm for objects in sonar images. IET Radar, Sonar and Navigation, 14(11): 1757-1766 [DOI: 10.1049/iet-rsn.2020.0230]
Bodla N, Singh B, Chellappa R and Davis L S. 2017. Soft-NMS-improving object detection with one line of code//Proceedings of 2017 IEEE International Conference on Computer Vision. Venice, Italy: IEEE: 5562-5570 [ DOI:10.1109/iccv.2017.593 http://dx.doi.org/10.1109/iccv.2017.593 ]
Chang Y L, Anagaw A, Chang L, Wang Y C, Hsiao C Y and Lee W H. 2019. Ship detection based on YOLOv2 for SAR imagery. Remote Sensing, 11(7): #786 [DOI: 10.3390/rs11070786]
Chicco D and Jurman G. 2020. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genomics, 21(1): #6 [DOI: 10.1186/s12864-019-6413-7]
Girshick R. 2015. Fast R-CNN//Proceedings of 2015 IEEE International Conference on Computer Vision. Santiago, Chile: IEEE: 1440-1448 [ DOI:10.1109/ICCV.2015.169 http://dx.doi.org/10.1109/ICCV.2015.169 ]
Guo C X, Fan B, Zhang Q, Xiang S M and Pan C H. 2020. AugFPN: improving multi-scale feature learning for object detection//Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE: 12595-12601 [ DOI:10.1109/CVPR42600.2020.01261 http://dx.doi.org/10.1109/CVPR42600.2020.01261 ]
He K M, Zhang X Y, Ren S Q and Sun J. 2016. Deep residual learning for image recognition//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE: 770-778 [ DOI:10.1109/CVPR.2016.90 http://dx.doi.org/10.1109/CVPR.2016.90 ]
Huang G, Liu Z, Van Der Maaten L and Weinberger K Q. 2017. Densely connected convolutional networks//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE: 2261-2269 [ DOI:10.1109/cvpr.2017.243 http://dx.doi.org/10.1109/cvpr.2017.243 ]
Lee S H and Seo J K. 2005. Noise removal with Gauss curvature-driven diffusion. IEEE Transactions on Image Processing, 14(7): 904-909 [DOI: 10.1109/tip.2005.849294]
Lin T Y, Dollár P, Girshick R, He K M, Hariharan B and Belongie S. 2017a. Feature pyramid networks for object detection//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE: 936-944 [ DOI:10.1109/cvpr.2017.106 http://dx.doi.org/10.1109/cvpr.2017.106 ]
Lin T Y, Goyal P, Girshick R, He K M and Dollár P. 2017b. Focal loss for dense object detection//Proceedings of 2017 IEEE International Conference on Computer Vision. Venice, Italy: IEEE: 2999-3007 [ DOI:10.1109/ICCV.2017.324 http://dx.doi.org/10.1109/ICCV.2017.324 ]
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C Y and Berg A C. 2016. SSD: single shot MultiBox detector//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, the Netherlands: Springer:21-37 [ DOI:10.1007/978-3-319-46448-0_2 http://dx.doi.org/10.1007/978-3-319-46448-0_2 ]
Ma X S, Shen H F, Yang J and Zhang L P. 2015. Polarimetric SAR speckle filtering using a nonlocal weighted minimum mean squared error filter. Journal of Image and Graphics, 20(1): 140-150
马晓双, 沈焕锋, 杨杰, 张良培. 2015. 极化SAR相干斑抑制的非局部加权最小均方误差滤波算法. 中国图象图形学报, 20(1): 140-150 [DOI: 10.11834/jig.20150115]
Malhotra R, Shakya A, Ranjan R and Banshi R. 2021. Software defect prediction using binary particle swarm optimization with binary cross entropy as the fitness function. Journal of Physics: Conference Series, 1767: #012003 [DOI: 10.1088/1742-6596/1767/1/012003]
Redmon J and Farhadi A. 2018. Yolov3: an incremental improvement[EB/OL]. [2018-04-08] . https://arxiv.org/pdf/1804.02767.pdf https://arxiv.org/pdf/1804.02767.pdf
Ren S Q, He K M, Girshick R and Sun J. 2017. Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(6): 1137-1149 [DOI: 10.1109/TPAMI.2016.2577031]
Sugimoto M, Ouchi K and Nakamura Y. 2013. On the novel use of model-based decomposition in SAR polarimetry for target detectionon the sea. Remote Sensing Letters, 4(9): 843-852 [DOI: 10.1080/2150704X.2013.804220]
Tian Z, Shen C H, Chen H and He T. 2019. FCOS: fully convolutional one-stage object detection//Proceedings of 2019 IEEE/CVF International Conference on Computer Vision. Seoul, Korea(South): IEEE: 9626-9635 [ DOI:10.1109/iccv.2019.00972 http://dx.doi.org/10.1109/iccv.2019.00972 ]
Wang C C, Wang Y and Liao M S. 2012. Removal of azimuth ambiguities and detection of a ship: using polarimetric airborne C-band SAR images. International Journal of Remote Sensing, 33(10): 3197-3210 [doi: 10.1080/01431161.2011.633123]
Wang J Q, Chen K, Yang S, Loy C C and Lin D H. 2019a. Region proposal by guided anchoring//Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Long Beach, USA: IEEE: 2965-2974 [ DOI:10.1109/CVPR.2019.00308 http://dx.doi.org/10.1109/CVPR.2019.00308 ]
Wang Y Y, Wang C, Zhang H, Dong Y B and Wei S S. 2019b. Automatic ship detection based on RetinaNet using multi-resolution Gaofen-3 imagery. Remote Sensing, 11(5): #531 [DOI: 10.3390/rs11050531]
Zhang X H, Yao L B, Lyu Y F, Jian T, Zhao Z W and Zang J. 2020. Data-adaptive single-shot ship detector with a bidirectional feature fusion module for SAR images. Journal of Image and Graphics, 25(9): 1943-1952
张筱晗, 姚力波, 吕亚飞, 简涛, 赵志伟, 藏洁. 2020. 双向特征融合的数据自适应SAR图像舰船目标检测模型. 中国图象图形学报, 25(9): 1943-1952 [DOI: 10.11834/jig.190558]
Zhao Y Q, Rao Y, Dong S P and Zhang J Y. 2020. Survey on deep learning object detection. Journal of Image and Graphics, 25(4): 629-654
赵永强, 饶元, 董世鹏, 张君毅. 2020. 深度学习目标检测方法综述. 中国图象图形学报, 25(4): 629-654 [DOI: 10.11834/jig.190307]
相关作者
相关机构
京公网安备11010802024621