Hyperspectral image classification evaluated across different datasets
- Vol. 26, Issue 8, Pages: 1969-1977(2021)
Published: 16 August 2021 ,
Accepted: 24 May 2021
DOI: 10.11834/jig.210123
移动端阅览
浏览全部资源
扫码关注微信
Published: 16 August 2021 ,
Accepted: 24 May 2021
移动端阅览
Erting Pan, Yong Ma, Jun Huang, Fan Fan, Hao Li, Jiayi Ma. Hyperspectral image classification evaluated across different datasets. [J]. Journal of Image and Graphics 26(8):1969-1977(2021)
目的
2
随着高光谱成像技术的飞速发展,高光谱数据的应用越来越广泛,各场景高光谱图像的应用对高精度详细标注的需求也越来越旺盛。现有高光谱分类模型的发展大多集中于有监督学习,大多数方法都在单个高光谱数据立方中进行训练和评估。由于不同高光谱数据采集场景不同且地物类别不一致,已训练好的模型并不能直接迁移至新的数据集得到可靠标注,这也限制了高光谱图像分类模型的进一步发展。本文提出跨数据集对高光谱分类模型进行训练和评估的模式。
方法
2
受零样本学习的启发,本文引入高光谱类别标签的语义信息,拟通过将不同数据集的原始数据及标签信息分别映射至同一特征空间以建立已知类别和未知类别的关联,再通过将训练数据集的两部分特征映射至统一的嵌入空间学习高光谱图像视觉特征和类别标签语义特征的对应关系,即可将该对应关系应用于测试数据集进行标签推理。
结果
2
实验在一对同传感器采集的数据集上完成,比较分析了语义—视觉特征映射和视觉—语义特征映射方向,对比了5种基于零样本学习的特征映射方法,在高光谱图像分类任务中实现了对分类模型在不同数据集上的训练和评估。
结论
2
实验结果表明,本文提出的基于零样本学习的高光谱分类模型可以实现跨数据集对分类模型进行训练和评估,在高光谱图像分类任务中具有一定的发展潜力。
Objective
2
Hyperspectral sensors are evolving towards miniaturization and portability with the rapid development of hyperspectral imaging technology. The acquisition of hyperspectral data has become easier and less expensive as a result of this breakthrough. The broad applications of hyperspectral images in various scenes have arisen an increasing demand for high-precision and detailed annotations. The growth of existing hyperspectral classification models mainly focuses on supervised learning
and many of them have reached an almost perfect performance. However
nearly all of these models are trained and evaluated in a single hyperspectral data cube. On this condition
the trained classification model cannot be directly transferred to a new hyperspectral dataset to obtain reliable annotations. The main reason is that different hyperspectral datasets are collected in irrelevant scenes and have covered inconsistent object categories. Accordingly
existing hyperspectral classification models have poor generalization capacity. The further development of the hyperspectral image classification model is also constrained. Consequently
a hyperspectral classification model with a generalization capability and ability to adapt to new and unseen classes across the different datasets must be developed. In this study
we propose a specific unique paradigm
which is to train and evaluate a hyperspectral classification model across different datasets. As previously mentioned
some new and unseen categories might be encountered when evaluating the classification model on another hyperspectral dataset. We introduce zero-shot learning into hyperspectral image classification to address this problem.
Method
2
Zero-shot learning can distinguish data from unseen categories
except for identifying data categories that have appeared in the training set. Zero-shot learning is based on the principle of allowing the model learn to understand the semantics of categories at first. Specifically
this mechanism employs auxiliary knowledge (such as attributes) to embed category labels into the semantic space and uses the data in the training set to learn the mapping relationship from images to semantics. On the basis of zero-shot learning
we introduce hyperspectral category label semantics as side information to address the unseen categories in the classification across different datasets. The overall workflow of our model can be divided into three steps. The first step is feature extraction
including hyperspectral image visual feature extraction and label semantics extraction. Hyperspectral image feature extraction aims to obtain high-level hyperspectral features with great capability to distinguish
and we refer to them as visual features. Most existing hyperspectral classification models are designed to extract robust hyperspectral features and have fine classification performance. Hence
we directly fine-tune current models to embed hyperspectral images into visual feature space. In the label semantic extraction
word2vec can map each word to a vector
representing the relationship between words. We employ word2vec model trained on a large external corpus to obtain hyperspectral label vectors. This model embeds a hyperspectral category label into a label semantic space. The second step is feature mapping. According to the choice of embedding space
this mechanism can be divided into visual to semantic feature mapping and semantic to visual feature mapping. Feature mapping is used to map these two features into the same feature space
and the model learns and optimizes the mapping on the basis of the correspondence between the hyperspectral data and annotations in the training set. The learned mapping establishes the correspondence between the hyperspectral visual features and the label semantic features. The third step is to employ this learned mapping to perform label reasoning on the testing set. Specifically
the mapping is applied to the hyperspectral visual features and label semantic features of the testing data
and the corresponding labels for the test set data are inferred by measuring the similarity of the two features.
Result
2
We selected a pair of datasets collected by the same type of hyperspectral sensor for comparative experiments
namely
the Salinas and Indian Pines datasets. This step is conducted to avoid the issue of differences in the physical representation of the object spectra resulting from hyperspectral data collection by different hyperspectral sensors. The workflow of our method is divided into three steps; thus
we adopt different models in each step for comparative experiments. We compared two visual feature extraction models
different feature mapping models
including visual to semantic feature mapping models
and semantic to visual feature mapping models. The quantitative evaluation metric is top-
k
accuracy. We listed the results of top-1 accuracy and top-3 accuracy. Experimental results show that the method that employs spatial-spectral united network (SSUN)as the visual feature extraction model and relation network (RN) as the zero-shot learning model reaches the best performance. Comparative experiments in different visual feature extraction models demonstrate that SSUN can obtain more distinguishable features because it considered the features in the spatial and spectral domains. When compared with the results from all models related to feature mapping
the semantic to visual feature mapping models outperform the other approaches. This result indicates that choosing visual feature space as the embedding space is a preferable alternative. We also analyze the reason for the current unsatisfied classification performances in detail.
Conclusion
2
In this study
we propose a specific pattern to train and evaluate the classification model across different datasets to improve the poor migration ability of the existing classification model. We introduce the semantic description of category labels in hyperspectral classification to cope with new and unseen categories in a new hyperspectral dataset and establish the association between seen and new datasets. The experimental results show that the feature extraction model SSUN improves the performance
and the semantic to visual feature mapping model RN outperforms several approaches. In conclusion
the proposed hyperspectral classification model based on zero-shot learning can be trained and evaluated across different datasets. The experimental results indicate that the proposed model has certain development potential in hyperspectral image classification tasks.
高光谱图像分类深度学习特征提取零样本学习高光谱语义特征
hyperspectral image classificationdeep learningfeature extractionzero-shot learningsemantic features of hyperspectral data
Frome A, Corrado G S, Shlens J, Bengio S, Dean J, Ranzato M A and Mikolov T. 2013. DeViSE: a deep visual-semantic embedding model//Proceedings of the 26th International Conference on Neural Information Processing Systems. Lake Tahoe, USA: Curran Associates Inc.: 2121-2129
Jiang J J, Ma J Y, Wang Z, Chen C and Liu X M. 2019. Hyperspectral image classification in the presence of noisy labels. IEEE Transactions on Geoscience and Remote Sensing, 57(2): 851-865[DOI:10.1109/TGRS.2018.2861992]
Kodirov E, Xiang T and Gong S G. 2017. Semantic autoencoder for zero-shot learning//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, USA: IEEE: 4447-4456[DOI: 10.1109/CVPR.2017.473http://dx.doi.org/10.1109/CVPR.2017.473]
Lampert C H, Nickisch H and Harmeling S. 2014. Attribute-based classification for zero-shot visual object categorization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 36(3): 453-465[DOI:10.1109/TPAMI.2013.140]
Li C, Ma Y, Mei X G, Liu C Y and Ma J Y. 2016. Hyperspectral image classification with robust sparse representation. IEEE Geoscience and Remote Sensing Letters, 13(5): 641-645[DOI:10.1109/LGRS.2016.2532380]
Li G D, Zhang C J, Gao F and Zhang X Y. 2019. Doubleconvpool-structured 3D-CNN for hyperspectral remote sensing image classification. Journal of Image and Graphics, 24(4): 639-654
李冠东, 张春菊, 高飞, 张雪英. 2019. 双卷积池化结构的3D-CNN高光谱遥感影像分类方法. 中国图象图形学报, 24(4): 639-654[DOI:10.11834/jig.180422]
Li S T, Song W W, Fang L Y, Chen Y S, Ghamisi P and Benediktsson J A. 2019. Deep learning for hyperspectral image classification: an overview. IEEE Transactions on Geoscience and Remote Sensing, 57(9): 6690-6709[DOI:10.1109/TGRS.2019.2907932]
Mei X G, Pan E T, Ma Y, Dai X B, Huang J, Fan F, Du Q L, Zheng H and Ma J Y. 2019. Spectral-spatial attention networks for hyperspectral image classification. Remote Sensing, 11(8): #963[DOI:10.3390/rs11080963]
Mikolov T, Sutskever I, Chen K, Corrado G and Dean J. 2013. Distributed representations of words and phrases and their compositionality[EB/OL]. [2021-02-11].https://proceedings.neurips.cc/paper/2013/file/9aa42b31882ec039965f3c4923ce901b-Paper.pdfhttps://proceedings.neurips.cc/paper/2013/file/9aa42b31882ec039965f3c4923ce901b-Paper.pdf
Mou L C, Ghamisi P and Zhu X X. 2017. Deep recurrent neural networks for hyperspectral image classification. IEEE Transactions on Geoscience and Remote Sensing, 55(7): 3639-3655[DOI:10.1109/TGRS.2016.2636241]
Paoletti M E, Haut J M, Plaza J and Plaza A. 2019. Deep learning classifiers for hyperspectral imaging: a review. ISPRS Journal of Photogrammetry and Remote Sensing, 158: 279-317[DOI:10.1016/j.isprsjprs.2019.09.006].
Ran Q, Yu H Y, Gao L R, Li W and Zhang B. 2018. Superpixel and subspace projection-based support vector machines for hyperspectral image classification. Journal of Image and Graphics, 23(1): 95-105
冉琼, 于浩洋, 高连如, 李伟, 张兵. 2018. 结合超像元和子空间投影支持向量机的高光谱图像分类. 中国图象图形学报, 23(1): 95-105[DOI:10.11834/jig.170201]
Sung F, Yang Y X, Zhang L, Tao X, Torr P H S and Hospedales T M. 2018. Learning to compare: relation network for few-shot learning//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE: 1199-1208[DOI: 10.1109/CVPR.2018.00131http://dx.doi.org/10.1109/CVPR.2018.00131]
Tong Q X, Zhang B and Zhang L F. 2016. Current progress of hyperspectral remote sensing in China. Journal of Remote Sensing, 20(5): 689-707
童庆禧, 张兵, 张立福. 2016. 中国高光谱遥感的前沿进展. 遥感学报, 20(5): 689-707[DOI:10.11834/jrs.20166264]
Xian Y Q, Lampert C H, Schiele B and Akata Z. 2019. Zero-shot learning-a comprehensive evaluation of the good, the bad and the ugly. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(9): 2251-2265[DOI:10.1109/TPAMI.2018.2857768]
Xu Y H, Zhang L P, Du B and Zhang F. 2018. Spectral-spatial unified networks for hyperspectral image classification. IEEE Transactions on Geoscience and Remote Sensing, 56(10): 5893-5909[DOI:10.1109/TGRS.2018.2827407]
Zhang B. 2016. Advancement of hyperspectral image processing and information extraction. Journal of Remote Sensing, 20(5): 1062-1090
张兵. 2016. 高光谱图像处理与信息提取前沿. 遥感学报, 20(5): 1062-1090[DOI:10.11834/jrs.20166179]
Zhang L, Xiang T and Gong S G. 2017. Learning a deep embedding model for zero-shot learning//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, USA: IEEE: 3010-3019[DOI: 10.1109/CVPR.2017.321http://dx.doi.org/10.1109/CVPR.2017.321]
相关文章
相关作者
相关机构