嵌入式深度神经网络高光谱图像聚类
Embedded deep neural network hyperspectral image clustering
- 2020年25卷第1期 页码:193-205
收稿:2019-04-09,
修回:2019-6-17,
录用:2019-6-24,
纸质出版:2020-01-16
DOI: 10.11834/jig.190129
移动端阅览

浏览全部资源
扫码关注微信
收稿:2019-04-09,
修回:2019-6-17,
录用:2019-6-24,
纸质出版:2020-01-16
移动端阅览
目的
2
高光谱图像的高维特性和非线性结构给聚类任务带来了"维数灾难"和线性不可分问题,以往的工作将特征提取过程与聚类过程互相剥离,难以同时优化。为了解决上述问题,提出了一种新的嵌入式深度神经网络模糊C均值聚类方法(EDFCC)。
方法
2
EDFCC算法为了提取更加有效的深层特征,联合优化高光谱图像的特征提取和聚类过程,将模糊C均值聚类算法嵌入至深度自编码器网络中,可以保持两任务联合优化的优势,同时利用深度自编码器网络降维以及逼近任意非线性函数的能力,逐步将原始数据映射到潜在特征空间,提取数据的深层特征。所提方法采用模糊C均值聚类算法约束特征提取过程,学习适用于聚类的高光谱数据深层特征,动态调整聚类指示矩阵。
结果
2
实验结果表明,EDFCC算法在Indian Pines和Pavia University两个高光谱数据集上的聚类精度分别达到了42.95%和60.59%,与当前流行的低秩子空间聚类算法(LRSC)相比分别提高了3%和4%,相比于基于自编码器的数据聚类算法(AEKM)分别提高了2%和3%。
结论
2
EDFCC算法能够从高光谱图像的高维光谱信息中提取更加有效的深层特征,提升聚类精度,并且由于EDFCC算法不需要额外的训练过程,大大提升了聚类效率。
Objective
2
Hyperspectral remote sensing
which is also called imaging spectral remote sensing
is a combined imaging and spectroscopy of multi-dimensional information retrieval technology. It carries abundant spectral information and is widely used in earth observation. A hyperspectral image is a kind of nonlinear structured data with a high dimension
and it poses a great challenge to the clustering task. If direct processing of the spectral information of hyperspectral images requires a large amount of computation
then appropriate dimensionality reduction methods for the nonlinear structure of hyperspectral data must be adopted. Although many clustering methods have been proposed
these traditional methods involve shallow linear models
the efficiency of the similarity measure is low
and the clustering effect is often poor for high-dimensional or hyperspectral data with a nonlinear structure. Traditional clustering algorithms encounter difficulties when clustering high-dimensional data. The concept of subspace clustering has been proposed to solve the problem of high-dimensional data clustering. Subspace clustering can solve the clustering problem of high-dimensional data. However
existing subspace clustering algorithms typically employ shallow models to estimate the underlying subspaces of unlabeled data points and cluster them into corresponding clusters. They have several limitations. First
the clustering effect of these subspace clustering methods depends on the quality of the affinity matrix. Second
due to the linear assumption of the data
these methods cannot deal with data with a nonlinear structure. Several nuclear methods have been proposed to overcome these shortcomings. These methods map the data to a predefined kernel space where they perform subspace clustering. A disadvantage of these nuclear space clustering methods is that their performance depends heavily on the kernel functions used. Existing data transformation methods include linear transformation
such as principal component analysis (PCA)
and nonlinear transformation
such as the kernel method. However
data with a highly complex potential structure is still a huge challenge to the effectiveness of existing clustering methods
and most clustering algorithms
such as shallow models
can only extract shallow features. Owing to the limited representation capacity of the employed shallow models
the algorithms may fail in handling realistic data with high-dimensional nonlinear structures. Moreover
most learning approaches treat feature extraction and clustering separately
train the feature extraction model well
and only use the clustering algorithm once in the feature representation of data to obtain clustering results.
Method
2
To solve these problems
the use of spectral information is maximized
and a new subspace clustering algorithm
that is
embedded deep neural network fuzzy c-means clustering (EDFCC)
is proposed in this study. The EDFCC algorithm can effectively extract the spectral information of hyperspectral images and be used for hyperspectral image clustering. The fuzzy c-means clustering algorithm is embedded into the deep autoencoder network
and the joint learning deep autoencoder network and fuzzy c-means clustering algorithm are used. Optimizing the two tasks jointly can substantially improve the performance of both. First
the feature extraction process of data is assumed to be an unknown transformation
which may be a nonlinear function. To preserve the local structure
the representation of each data point is learned by minimizing the reconstruction error
that is
the feature extraction process is completed by learning the deep autoencoder network. Data should be clustered in an effective manner to learn the representation of the potential features of data suitable for clustering. The fuzzy c-means clustering algorithm is used to constrain the feature extraction process and make the generated features suitable for clustering. The motivation for designing the EDFCC algorithm is to maintain the advantage of the joint optimization of the two tasks while using the capability of the deep autoencoder network to approximate any nonlinear function
gradually map the input data points to the potential nonlinear space
and adjust the clustering indicator matrix dynamically with the model training.
Result
2
Two hyperspectral data sets
namely
Indian Pines and Pavia University
are used to test the validity of the EDFCC algorithm. The quantitative evaluation metrics include accuracy and normalized mutual information. The Indian Pines dataset contains data acquired by the airborne visible infrared imaging spectrometer with a spectral range of 0.41~2.45 m
spatial resolution of 25 m
spectral resolution of 10 nm
and a total of 145×145 sample points. A total of 220 original bands are available
but the water vapor absorption band and bands with a low signal-to-noise ratio are excluded. The remaining 200 bands are used as research objects. The Indian Pines dataset has 16 different feature categories. Indian Pines shows that the overall clustering accuracy of the EDFCC algorithm is 42.95%
which is 3% higher than that of the best LRSC algorithm. The Pavia University dataset was obtained by the airborne reflector optical spectral imager in Germany. Its spectral range is 0.43~0.86 m
and its spatial resolution is 1.3 m. The dataset contains 610×340 sample points. A total of 115 original bands exist
but the noise bands are removed. The 103 remaining bands are used as research objects. The Pavia University dataset has nine types of ground objects. The dataset shows that the overall clustering accuracy of the EDFCC algorithm is 60.59%
which is 4% higher than that of the best LRSC algorithm. When compared with the AEKM algorithm for deep clustering
the AEKM algorithm is improved by 2% and 3%.
Conclusion
2
The EDFCC algorithm is proposed in this study. The algorithm is first applied in hyperspectral image clustering as a joint learning framework. The indicator matrix can be dynamically adjusted because of joint learning
and no additional training process is required
which greatly improves the training efficiency. Experimentalresults show that the EDFCC algorithm can extract many effective deep features from the high-dimensional spectral information of hyperspectral images and improve clustering accuracy.
Bezdek J C. 1981. Pattern recognition with fuzzy objective function algorithms. Boston, MA: Springer[ DOI: 10.1007/978-1-4757-0450-1 http://dx.doi.org/10.1007/978-1-4757-0450-1 ]
Chang J L, Wang L F, Meng G F, Xiang S and Pan C. 2017. Deep adaptive image clustering//Proceedings of IEEE International Conference on Computer Vision. Venice, Italy: IEEE, 5880-5888[ DOI: 10.1109/ICCV.2017.626 http://dx.doi.org/10.1109/ICCV.2017.626 ]
Dunn J C. 1973. A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. Journal of Cybernetics, 3(3):32-57[DOI:10.1080/01969727308546046]
Elhamifar E and Vidal R. 2013. Sparse subspace clustering:algorithm, theory, and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(11):2765-2781[DOI:10.1109/TPAMI.2013.57]
Elhamifar E and Vidal R. 2009. Sparse subspace clustering//Proceedings of 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami, FL, USA: IEEE, 2790-2797[ DOI: 10.1109/CVPR.2009.5206547 http://dx.doi.org/10.1109/CVPR.2009.5206547 ]
Ester M, Kriegel H P, Sander J and Xu X. 1996. A density-based algorithm for discovering clusters in large spatial databases with noise//Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining. Portland, Oregon: ACM, 226-231
Fang S, Zhu F J, Dong Z Y and Zhang J. 2019. Sample optimized selection of hyperspectral image classification. Journal of Image and Graphics, 24(1):135-148
方帅, 祝凤娟, 董张玉, 张晶. 2019.样本优化选择的高光谱图像分类.中国图象图形学报, 24(1):135-148)[DOI:10.11834/jig.180437]
Foody G M and Mathur A. 2004. A relative evaluation of multiclass image classification by support vector machines. IEEE Transactions on Geoscience and Remote Sensing, 42(6):1335-1343[DOI:10.1109/TGRS.2004.827257]
Hartigan J A and Wong M A. 1979. Algorithm AS 136:a k-means clustering algorithm. Journal of the Royal Statistical Society. Series C (Applied Statistics), 28(1):100-108[DOI:10.2307/2346830]
Hinton G E andSalakhutdinov R R. 2006. Reducing the dimensionality of data with neural networks. Science, 313(5786):504-507[DOI:10.1126/science.1127647]
Hofmann T, Schölkopf B and Smola A J. 2008. Kernel methods in machine learning. The Annals of Statistics, 36(3):1171-1220[DOI:10.1214/009053607000000677]
Kohonen T. 1990. The self-organizing map.Proceedings of 1990 IEEE, 78(9):1464-1480[DOI:10.1109/5.58325]
Landgrebe D. 2002. Hyperspectral image data analysis. IEEE Signal Processing Magazine, 19(1):17-28[DOI:10.1109/79.974718]
Li C G and Vidal R. 2016. A structured sparse plus structured low-rank framework for subspace clustering and completion. IEEE Transactions on Signal Processing, 64(24):6557-6570[DOI:10.1109/TSP.2016.2613070]
Li X, Lyu J and Yi Z. 2018. An efficient representation-based method for boundary point and outlier detection. IEEE Transactions on Neural Networks and Learning Systems, 29(1):51-62[DOI:10.1109/TNNLS.2016.2614896]
Liu G C, Lin Z C, Yan S C, Sun J, Yu Y and Ma Y. 2012. Robustrecovery of subspace structures by low-rank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1):171-184[DOI:10.1109/tpami.2012.88]
Nie F P, Wang H, Huang H and Ding C. 2011. Unsupervised and semi-supervised learning via $\ell$ 1 -norm graph//Proceedings of International Conference on Computer Vision. Barcelona, Spain: IEEE, 2268-2273[ DOI: 10.1109/ICCV.2011.6126506 http://dx.doi.org/10.1109/ICCV.2011.6126506 ].
Patel V M and Vidal R. 2014. Kernel sparse subspace clustering//Proceedings of 2014 IEEE International Conference on Image Processing. Paris, France: IEEE, 2849-2853[ DOI: 10.1109/ICIP.2014.7025576 http://dx.doi.org/10.1109/ICIP.2014.7025576 ]
Peng X, Yi Z and Tang H. 2015. Robust subspace clustering via thresholding ridge regression//Proceedings of the 29th AAAI Conference on Artificial Intelligence. Austin, Texas: ACM, 3827-3833
Peng X, Yu Z, Yi Z and Tang H. 2017. Constructing the L2-graph for robust subspace learning and subspace clustering. IEEE Transactions on Cybernetics, 47(4):1053-1066[DOI:10.1109/TCYB.2016.2536752]
Ran Q, Yu H Y, Gao L R, Li W and Zhang B. 2018. Superpixel and subspace projection-based support vector machines for hyperspectral image classification. Journal of Image and Graphics, 23(1):95-105
冉琼, 于浩洋, 高连如, 李伟, 张兵. 2018.结合超像元和子空间投影支持向量机的高光谱图像分类.中国图象图形学报, 23(1):95-105)[DOI:10.11834/jig.170201]
Reynolds D. 2015. Gaussian Mixture Models//Li S Z, Jain A K. Encyclopedia of Biometrics. 2nd ed. New York: Springer, 827-832[ DOI: 10.1007/978-1-4899-7488-4_196 http://dx.doi.org/10.1007/978-1-4899-7488-4_196 ]
Schmidhuber J. 2015. Deep learning in neural networks:an overview. Neural Networks, 61:85-117[DOI:10.1016/j.neunet.2014.09.003]
Song C F, Liu F, Huang Y Z, Wang L and Tan T. 2013. Auto-encoder based data clustering//Proceedings of the 18th Iberoamerican Congress on Pattern Recognition. Havana, Cuba: Springer, 117-124[ DOI: 10.1007/978-3-642-41822-8_15 http://dx.doi.org/10.1007/978-3-642-41822-8_15 ]
Vidal R. 2011. Subspace clustering. IEEE Signal Processing Magazine, 28(2):52-68[DOI:10.1109/MSP.2010.939739]
von Luxburg U. 2007. A tutorial on spectral clustering. Statistics and Computing, 17(4):395-416[DOI:10.1007/s11222-007-9033-z]
Wold S, Esbensen K and Geladi P. 1987. Principal component analysis. Chemometrics and Intelligent Laboratory Systems, 2(1-3):37-52[DOI:10.1016/0169-7439(87)80084-9]
Yang Y Z, Feng J S, Jojic N, Yang J and Huang T S. 2016. l 0 -sparse subspace clustering//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, The Netherlands: Springer, 731-747[ DOI: 10.1007/978-3-319-46475-6_45 http://dx.doi.org/10.1007/978-3-319-46475-6_45 ].
Yin M, Guo Y, Gao J B, He Z and Xie S. 2016. Kernel sparse subspace clustering on symmetric positive definite manifolds//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV, USA: IEEE, 5157-5164[ DOI: 10.1109/CVPR.2016.557 http://dx.doi.org/10.1109/CVPR.2016.557 ]
Yu W B, Wang Z Y, Li S S and Sun X. 2016. Hyperspectral image clustering based on density peaks and superpixel segmentation. Journal of Image and Graphics, 21(10):1402-1410
于文博, 王忠勇, 李山山, 孙旭. 2016.整合超像元分割和峰值密度的高光谱图像聚类.中国图象图形学报, 21(10):1402-1410)[DOI:10.11834/jig.20161015]
Yuan Y, Lin J and Wang Q. 2016. Dual-clustering-based hyperspectral band selection by contextual analysis. IEEE Transactions on Geoscience and Remote Sensing, 54(3):1431-1445[DOI:10.1109/tgrs.2015.2480866]
Zhang R, Nie F P and Li X H. 2017a. Embedded clustering via robust orthogonal least square discriminant analysis//Proceedings of 2017 IEEE International Conference on Acoustics, Speech and Signal Processing. New Orleans, LA, USA: IEEE, 2332-2336[ DOI: 10.1109/ICASSP.2017.7952573 http://dx.doi.org/10.1109/ICASSP.2017.7952573 ]
Zhang R, Nie F P and Li X L. 2017b. Self-weighted spectral clustering with parameter-free constraint. Neurocomputing, 241:164-170[DOI:10.1016/j.neucom.2017.01.085]
相关作者
相关机构
京公网安备11010802024621