面向SAR目标识别的深度卷积神经网络结构设计
Architecture design of deep convolutional neural network for SAR target recognition
- 2018年23卷第6期 页码:928-936
收稿:2017-08-28,
修回:2017-12-8,
纸质出版:2018-06-16
DOI: 10.11834/jig.170473
移动端阅览

浏览全部资源
扫码关注微信
收稿:2017-08-28,
修回:2017-12-8,
纸质出版:2018-06-16
移动端阅览
目的
2
针对用于SAR(synthetic aperture radar)目标识别的深度卷积神经网络模型结构的优化设计难题,在分析卷积核宽度对分类性能影响基础上,设计了一种适用于SAR目标识别的深度卷积神经网络结构。
方法
2
首先基于二维随机卷积特征和具有单个隐层的神经网络模型-超限学习机分析了卷积核宽度对SAR图像目标分类性能的影响;然后,基于上述分析结果,在实现空间特征提取的卷积层中采用多个具有不同宽度的卷积核提取目标的多尺度局部特征,设计了一种适用于SAR图像目标识别的深度模型结构;最后,在对MSTAR(moving and stationary target acquisition and recognition)数据集中的训练样本进行样本扩充基础上,设定了深度模型训练的超参数,进行了深度模型参数训练与分类性能验证。
结果
2
实验结果表明,对于具有较强相干斑噪声的SAR图像而言,采用宽度更大的卷积核能够提取目标的局部特征,提出的模型因能从输入图像提取目标的多尺度局部特征,对于10类目标的分类结果(包含非变形目标和变形目标两种情况)接近或优于已知文献的最优分类结果,目标总体分类精度分别达到了98.39%和97.69%,验证了提出模型结构的有效性。
结论
2
对于SAR图像目标识别,由于与可见光图像具有不同的成像机理,应采用更大的卷积核来提取目标的空间特征用于分类,通过对深度模型进行优化设计能够提高SAR图像目标识别的精度。
Objective
2
To solve issues in the optimization design of deep convolutional neural network (DCNN) model architecture for synthetic aperture radar (SAR) target recognition
a DCNN model architecture for SAR target recognition is presented based on the analysis of the influence of convolution kernel size on classification performance.
Method
2
First
two-dimensional random convolution features and extreme learning machines (ELMs)
which are a single-hidden-layer neural network
are used to analyze the influence of convolution kernel size on SAR target recognition performance. Experimental results show that recognition performance increases as the kernel size increases although convolution kernels generate randomly and the convolution kernel with size 3×3 is unsuitable for SAR image recognition. Second
a DCNN architecture for SAR target recognition
in which the pixel resolution of input image is set to 88×88
is presented based on directed acyclic graph architecture. Multiple convolution kernels with different sizes
which are set as 5×5
7×7
9×9
and 11×11
are first adopted in the spatial-feature-extraction convolutional layer of DCNN to extract multi-scale local features from input images
and convolution kernels with large size
including 7×7
5×5
and 6×6
are then used in the last convolutional layers to extract semantic features. A fully connected layer is used as the classifier to recognize various types of targets and softmax loss function is used to train the parameters of the convolutional layers. The dropout strategy
which can improve regularization performance
is used between the fully connected layer and the output layer. Rectified linear units following behind each convolutional layer are used as activation functions
and pooling operations with width 3 and stride 2 are used to perform downsampling behind each activation function layer. Finally
MSTAR database
where the training samples are randomly augmented through sampling and adding speckle noises
is used to train the parameters of the proposed model architecture after setting proper training hyperparameters
and the recognition performances are tested in standard operating conditions
where target configurations with non-deformable and deformable conditions are considered.
Result
2
The MatConvNet toolbox is used to implement the proposed DCNN model architecture. In this task
90% of the augmented training samples are used to train the parameters of each convolutional layer
and other training samples are used to verify the trained parameters. The dropout rate is set as 0.1. The training procedure stops after 28 epochs
and the trained parameters are used to test recognition performance. The experimental results demonstrate that superior performance can be achieved for SAR image recognition because large-size kernels are used to extract spatial features from input image to overcome the influence of high-level speckle noise. This result is different from that of natural scene classification scenario using visible images
where small kernel sizes
such as 3×3
3×1
and 1×3
are used to achieve high recognition performance. The classification results based on the proposed architecture for 10 classes (including non-deformable and deformable target configurations) are compared with two DCNN models. The experimental results show that it can achieve comparable or better results than that of state-of-the-art deep model architectures
where the overall recognition performances reach 98.39% and 97.69%
respectively
for the two scenarios. The deep model using 3×3 convolutional layers can only achieve 93.16% recognition rate
which confirms our analysis on the influence of convolution kernel size on SAR image recognition performance. The recognition performance using the proposed DCNN model architecture is also better than that using random convolution features and ELM. This finding demonstrates that the DCNN model architecture can achieve satisfactory performance when deep architecture is carefully designed and more training samples are used to train these parameters.
Conclusion
2
A large convolution kernel size should be used to extract spatial features for SAR target recognition due to different imaging mechanisms compared with visible images
and better performance can be achieved through the optimization design of deep model architecture with augmented training samples.
El-Darymli K, Gill E W, Mcguire P, et al. Automatic target recognition in synthetic aperture radar imagery:a state-of-the-art review[J]. IEEE Access,2016, 4:6014-6058.[DOI:10.1109/ACCESS.2016.2611492]
He Z G, Lu J, Kuang G Y. A survey on feature extraction and selection of SAR images[J]. Signal Processing, 2008, 24(5):813-823.
贺志国, 陆军, 匡纲要. SAR图像特征提取与选择研究[J].信号处理, 2008, 24(5):813-823. [DOI:10.3969/j.issn.1003-0530.2008.05.024]
Amoon M, Rezai-Rad G A. Automatic target recognition of synthetic aperture radar (SAR) images based on optimal selection of Zernike moments features[J]. IET Computer Vision, 2013, 8(2):77-85.[DOI:10.1049/iet-cvi.2013.0027]
Du P J, Samat A, Gamba P, et al. Polarimetric SAR image classification by Boosted Multiple-Kernel Extreme Learning Machines with polarimetric and spatial features[J]. International Journal of Remote Sensing, 2014, 35(23):7978-7990.[DOI:10.1080/2150704X.2014.978952]
Zhao Q, Principe J C. Support vector machines for SAR automatic target recognition[J]. IEEE Transactions on Aerospace and Electronic Systems, 2001, 37(2):643-654.[DOI:10.1109/7.937475]
Sun Y J, LiuZ P, Todorovic S, et al. Adaptive boosting for SAR automatic target recognition[J]. IEEE Transactions on Aerospace and Electronic Systems, 2007, 43(1):112-125.[DOI:10.1109/TAES.2007.357120]
Song S L, Xu B, Yang J. SAR target recognition via supervised discriminative dictionary learning and sparse representation of the SAR-HOG feature[J]. Remote Sensing, 2016, 8(8):683.[DOI:10.3390/rs8080683]
Zhang H C, Nasrabadi N M, Zhang Y N, et al. Multi-view automatic target recognition using joint sparse representation[J]. IEEE Transactions on Aerospace and Electronic Systems, 2012, 48(3):2481-2497.[DOI:10.1109/TAES.2012.6237604]
Dong G G, Kuang G Y, Wang N, et al. SAR target recognition via joint sparse representation of monogenic signal[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2015, 8(7):3316-3328.[DOI:10.1109/JSTARS.2015.2436694]
Hinton G E, Salakhutdinov R R. Reducing the dimensionality of data with neural networks[J]. Science, 2006, 313(5786):504-507.[DOI:10.1126/science.1127647]
Lecun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998, 86(11):2278-2324.[DOI:10.1109/5.726791]
Krizhevsky A, Sutskever I, Hinton G E. ImageNet classification with deep convolutional neural networks[C]//Proceedings of 2012 Advances in Neural Information Processing Systems. Doha, Qatar: Curran Associates Inc, 2012: 1097-1105.
Szegedy C, Liu W, Jia Y Q, et al. Going Deeper with Convolutions[C]//Proceedings of 2015 IEEE Conference On Computer Vision and Pattern Recognition. Boston, USA: IEEE, 2015: 1-9. [ DOI:10.1109/CVPR.2015.7298594 http://dx.doi.org/10.1109/CVPR.2015.7298594 ]
Simonyan K, Zisserman A. Very deep convolutional networks for large-Scale image recognition[C]//Proceedings of 2005 International Conference on Learning Representations. San Diego, CA: Computer Science, 2015: 1-14.
He K M, Zhang X Y, Ren S Q, et al. Deep residual learning for image recognition[C]//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, USA: IEEE, 2016: 770-778. [ DOI:10.1109/CVPR.2016.90 http://dx.doi.org/10.1109/CVPR.2016.90 ]
Ding J, Chen B, Liu H W, et al. Convolutional neural network with data augmentation for SAR target recognition[J]. IEEE Geoscience and Remote Sensing Letters, 2016, 13(3):364-368.[DOI:10.1109/LGRS.2015.2513754]
Chen S Z, Wang H P, Xu F, et al. Target classification using the deep convolutional networks for SAR images[J]. IEEE Transactions on Geoscience and Remote Sensing, 2016, 54(8):4806-4817.[DOI:10.1109/TGRS.2016.2551720]
Goodfellow I, Bengio Y, Courville A. Deep Learning[M]. Massachusetts, USA:MIT Press, 2016.
Iandola F N, Han S, Moskewicz M W, et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and<0. 5MB model size[J/OL]. 2016-11-04, arXiv: 1602. 07360[cs. CV].
Saxe A M, Koh P W, Chen Z H, et al. On random weights and unsupervised feature learning[C]//Proceedings of the 28th International Conference on International Conference on Machine Learning. Bellevue, Washington, USA: Omni Press, 2011: 1089-1096.
Huang G, Huang G B, Song S J, et al. Trends in extreme learning machines:a review[J]. Neural Networks, 2015, 61:32-48.[DOI:10.1016/j.neunet.2014.10.001]
Huang G B, Bai Z, Kasun L L C, et al. Local receptive fields based extreme learning machine[J]. IEEE Computational Intelligence Magazine, 2015, 10(2):18-29.[DOI:10.1109/MCI.2015.2405316]
Goodfellow I J, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets[C]//Proceedings of the 27th International Conference on Neural Information Processing Systems. Cambridge, USA: MIT Press, 2014: 2672-2680.
Srivastava N, Hinton G, Krizhevsky A, et al. Dropout:a simple way to prevent neural networks from overfitting[J]. Journal of Machine Learning Research, 2014, 15(1):1929-1958.
Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift[C]//Proceedings of the 32nd International Conference on Machine Learning. Lille, France: PMLR, 2015: 448-456. https://www.researchgate.net/publication/286794765_Dropout_A_Simple_Way_to_Prevent_Neural_Networks_from_Overfitting .
相关文章
相关作者
相关机构
京公网安备11010802024621