样本优化选择的高光谱图像分类
Sample optimized selection of hyperspectral image classification
- 2019年24卷第1期 页码:135-148
收稿:2018-07-04,
修回:2018-8-8,
纸质出版:2019-01-16
DOI: 10.11834/jig.180437
移动端阅览

浏览全部资源
扫码关注微信
收稿:2018-07-04,
修回:2018-8-8,
纸质出版:2019-01-16
移动端阅览
目的
2
高光谱分类问题中,由于类内光谱特性存在差异性,导致常规的随机样本选择策略无法保证训练样本均匀覆盖样本空间。针对这一问题,提出基于类内再聚类的样本空间优化策略。同时为了进一步提高分类精度,针对低置信度分类结果,提出基于邻域高置信信息的修正策略。
方法
2
采用FCM(fuzzy C-means)聚类算法对每类样本进行类内再聚类,在所聚的每个子类内选择适当样本。利用两个简单分类器SVM(support vector machine)和SRC(sparse representation-based classifier),对分类结果进行一致性检测,确定高、低置信区域,对低置信区域,利用主成分图作为引导图对置信度图进行滤波,使得高置信信息向低置信区域传播,从而修正低置信区域分类结果。以上策略可以保证即便在较少的训练样本的情况下,也能够训练出较高的分类器,大幅度提高分类精度。
结果
2
使用3组实验数据,根据样本比例设置两组实验与经典以及最新分类算法进行对比。实验结果表明,本文算法均取得很大改进,尤其在样本比例较小的实验中效果显著。在小比例(一般样本选取比例的十分之一)训练样本实验中,对于India Pines数据集,OA(overall accuracy)值高达90.48%;在Salinas数据集上能达到99.68%;同样,PaviaU数据集的OA值为98.54%。3组数据集的OA值均比其他算法高出4%~6%。
结论
2
综上表明,本文算法通过样本空间优化策略选取有代表性、均衡性的样本,保证小比例样本下分类精度依然显著;基于邻域高置信信息的修正策略起到很好的优化效果。同时,本文算法适应多种数据集,具有很好的鲁棒性。
Objective
2
In recent years
an increasing number of applications have been realized for remote sensing images. Hyperspectral image classification is a widely used method for hyperspectral image processing. For the traditional high-spectral classification problem
the mainstream improvement methods were compared to optimize the classifier or the classification algorithm. This approach does not address the existing limitations; thus
we proposed new improvements on the basis of this problem. In other words
while improving the classifier
the sample space was optimized to obtain a group of representative training samples to ensure the overall classification accuracy. The traditional classification method considers only the improvement of the classification algorithm and the classifier and adopts various random selection methods for sample acquisition. The spectral properties of even the same kind of substances are different. In view of the differences in the in-class spectral characteristics
the conventional random sample selection strategy for random sample selection in a certain proportion of various substances cannot guarantee that the selected training samples can contain complete spectral features. To solve this problem
a sample space optimization strategy based on in-class reclustering was proposed. In this way
the selected training samples can be guaranteed to contain various spectral curves of each class
and the selected samples are uniformly distributed in each subclass of each class. Moreover
to further improve the classification accuracy
the classifier must also be improved. According to integrated learning
the accuracy of point classification with the same classification results for multiple classifiers is higher
whereas the error rate with deviations is higher. Therefore
for the low-confidence classification result with low classification accuracy
the high-confidence region with high accuracy was optimized again. In this paper
the optimization method was a correction strategy based on neighborhood high-confidence information
which optimized the classification results of the low-confidence region by obtaining the classification results of the high-confidence region to improve the accuracy of the results of low-confidence region and the overall classification accuracy. Given that the classification strategy used in this paper is point classification
the domain information was not considered. In fact
the category information at a certain point was the same as the category information in the same region affected by the neighborhood information. Therefore
we used the edge protection filter to smooth various types of information based on edge protection to ensure the similarity between the information of a certain point and the field information and to further improve the classification accuracy.
Method
2
In this paper
fuzzy C-means clustering algorithm was used to implement class-to-class clustering on each class of samples. As the spectral characteristics of the same kind of samples are different
the same kind of samples were grouped into several subclasses according to the difference in spectral characteristics. When selecting samples
we ensured that samples were selected from each subclass of each class to ensure that the sample covered the entire sample space. For the correction strategy of the neighborhood high-confidence information
this paper used edge protection filter to optimize the low-confidence information by using the high-confidence region information. First
two simple classifiers
namely
support vector machine (SVM) classifier and sparse representation-based classifier (SRC)
were used to test the consistency of the classification results of the two classifiers. The point set with consistent classification results was the high-confidence region
and the point set with inconsistent results was the low-confidence region. Then
the results of the low-confidence region were optimized by using the edge protection filter. First
the hyperspectral images were processed by principal component analysis to obtain the first principal component. Given that the first principal component contained most of the structure information of the image
the first principal component was used as the guide diagram. Then
the high-confidence region was filtered
and a small number of the low-confidence region point sets were propagated by the high-confidence information. In this way
the low-confidence region can obtain a new category information to replace the original low-confidence category information to modify the low-confidence region classification results. The edge-protecting filter has the feature of edge-preserving smoothness. The aforementioned strategies demonstrated that our classification effect was greatly improved. In addition
even when a small proportion of training samples was selected
the samples can be trained with a high classifier after the sample space was optimized to ensure the stability of the classification accuracy.
Result
2
The experiments used three sets of experimental data
namely
the India Pines
Salinas
and PaviaU datasets. Then
two sets of experiments were set up according to the different sample selection proportion to compare the classic and the proposed classification algorithm. In the first experiment
we selected 10%
1%
and 1% training samples. The experimental results revealed that the overall accuracy (OA) values of the three datasets reached up to 98.93%
99.78%
and 99.40% respectively
which was 1% higher than that of other optimal algorithms. In the second small-scale sample experiment
we set the sample proportions to 1%
0.3%
and 0.4%. The OA values for the India Pines
Salinas
and PaviaU datasets reached up to 90.48%
99.68%
and 98.54%.The OA values of the three groups of data were 4%-6% higher than that of other algorithms for the same proportions. The experimental results suggested that the proposed algorithm is superior to other algorithms
and that the classification accuracy was greatly improved
particularly in experiments with small sample proportions.
Conclusion
2
In this paper
a representative and balanced sample was selected through sample space optimization strategy to ensure that the classification accuracy remained significant for small-scale samples. The correction strategy based on neighborhood high-confidence information offered a good optimization effect. Moreover
this algorithm adapted to many datasets and achieved good robustness. In summary
the results showed that the reduction of sample proportion resulted in the rapid decline and instability of the classification effect of the traditional classification algorithm. The proposed algorithm offers obvious advantages
which ensures not only the high accuracy but also the stability of the classification results.
Fang L Y, Li S T, Kang X D, et al. Spectral-spatial hyperspectral image classification via multiscale adaptive sparse representation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2014, 52(12):7738-7749.[DOI:10.1109/TGRS.2014.2318058]
Kang X D, Li S T, Fang L Y, et al. Extended random walker-based classification of hyperspectral images[J]. IEEE Transactions on Geoscience and Remote Sensing, 2015, 53(1):144-153.[DOI:10.1109/TGRS.2014.2319373]
Chang C C, Lin C J. LIBSVM:a library for support vector machines[J]. ACM Transactions on Intelligent Systems and Technology, 2011, 2(3):#27.[DOI:10.1145/1961189.1961199]
Fauvel M, Benediktsson J A, Chanussot J, et al. Spectral and spatial classification of hyperspectral data using SVMs and morphological profiles[J]. IEEE Transactions on Geoscience and Remote Sensing, 2008, 46(11):3804-3814.[DOI:10.1109/TGRS.2008.922034]
Fang L Y, He N J, Li S T, et al. Extinction Profiles Fusion for Hyperspectral Images Classification[J]. IEEE Transactions on Geoscience&Remote Sensing, 2018, 56(3):1803-1815.[DOI:10.1109/TGRS.2017.2768479]
Ghamisi P, Souza R, Benediktsson J A, et al. Extinction profiles for the classification of remote sensing data[J]. IEEE Transactions on Geoscience and Remote Sensing, 2016, 54(10):5631-5645.[DOI:10.1109/TGRS.2016.2561842]
Fang L Y, Wang C, Li S T, et al. Hyperspectral image classification via multiple-feature-based adaptive sparse representation[J]. IEEE Transactions on Instrumentation and Measurement, 2017, 66(7):1646-1657.[DOI:10.1109/TIM.2017.2664480]
Fang L Y, Li S T, Duan W H, et al. Classification of hyperspectral images by exploiting spectral-spatial information of superpixel via multiple kernels[J]. IEEE Transactions on Geoscience and Remote Sensing, 2015, 53(12):6663-6674.[DOI:10.1109/TGRS.2015.2445767]
Lu T, Li S T, Fang L Y, et al. Set-to-set distance-based spectral-spatial classification of hyperspectral images[J]. IEEE Transactions on Geoscience and Remote Sensing, 2016, 54(12):7122-7134.[DOI:10.1109/TGRS.2016.2596260]
Li J, Bioucas-Dias J M, Plaza A. Hyperspectral image segmentation using a new Bayesian approach with active learning[J]. IEEE Transactions on Geoscience and Remote Sensing, 2011, 49(10):3947-3960.[DOI:10.1109/TGRS.2011.2128330]
Lou X L, Huang W G, Zhou C B, et al. A method for fast resampling of remote sensing imagery[J]. Journal of Remote Sensing, 2002, 6(2):96-101.
楼琇林, 黄韦艮, 周长宝, 等.遥感图像数据重采样的一种快速算法[J].遥感学报, 2002, 6(2):96-101. [DOI:10.11834/jrs.20020204]
Han H, Wang W Y, Mao B H. Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning[C] //Proceedings of 2005 International Conference on Advances in Intelligent Computing. Hefei, China: Springer, 2005: 878-887.[ DOI: 10.1007/11538059_91 http://dx.doi.org/10.1007/11538059_91 ]
Cai Z W, Fan Q F, Feris R S, et al. A unified multi-scale deep convolutional neural network for fast object detection[C]//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, The Netherlands: Springer, 2016: 354-370.[ DOI: 10.1007/978-3-319-46493-0_22 http://dx.doi.org/10.1007/978-3-319-46493-0_22 ]
Sun B, Kang X D, Li S T, et al. Random-walker-based collaborative learning for hyperspectral image classification[J]. IEEE Transactions on Geoscience and Remote Sensing, 2017, 55(1):212-222.[DOI:10.1109/TGRS.2016.2604290]
Qiao T, Ren J C, Wang Z, et al. Effective denoising and classification of hyperspectral images using curvelet transform and singular spectrum analysis[J]. IEEE Transactions on Geoscience and Remote Sensing, 2017, 55(1):119-133.[DOI:10.1109/TGRS.2016.2598065]
Shen J B, Yang X S, Li X L, et al. Intrinsic image decomposition using optimization and user scribbles[J]. IEEE Transactions on Cybernetics, 2013, 43(2):425-436.[DOI:10.1109/TSMCB.2012.2208744]
Kang X D, Li S T, Fang L Y, et al. Intrinsic image decomposition for feature extraction of hyperspectral images[J]. IEEE Transactions on Geoscience and Remote Sensing, 2015, 53(4):2241-2253.[DOI:10.1109/TGRS.2014.2358615]
Bezdek J C, Ehrlich R, Full W. FCM:the fuzzy c-means clustering algorithm[J]. Computers&Geosciences, 1984, 10(2-3):191-203.[DOI:10.1016/0098-3004(84)90020-7]
Li S T, Kang X D, Hu J W. Image fusion with guided filtering[J]. IEEE Transactions on Image Processing, 2013, 22(7):2864-2875.[DOI:10.1109/TIP.2013.2244222]
Villa A, Benediktsson J A, Chanussot J, et al. Hyperspectral image classification with independent component discriminant analysis[J]. IEEE Transactions on Geoscience and Remote Sensing, 2011, 49(12):4865-4876.[DOI:10.1109/TGRS.2011.2153861]
Fang L Y, He N J, Li S T, et al. A new spatial-spectral feature extraction method for hyperspectral images using local covariance matrix representation[J]. IEEE Transactions on Geoscience and Remote Sensing, 2018, 56(6):3534-3546.[DOI:10.1109/TGRS.2018.2801387]
Kang X D, Xiang X L, Li S T, et al. PCA-Based edge-preserving features for hyperspectral image classification[J]. IEEE Transactions on Geoscience and Remote Sensing, 2017, 55(12):7140-7151.[DOI:10.1109/TGRS.2017.2743102]
相关作者
相关机构
京公网安备11010802024621