Hybrid Collaborative Representation for Remote-Sensing Image Scene Classification
Abstract
:1. Introduction
- We propose a novel hybrid collaborative representation-based classification method that considers both conventional collaborative representation and class-specific collaborative representation.
- We extend our proposed hybrid collaborative representation-based classification method to arbitrary kernel space to find the nonlinear structures hidden in the image features.
- The proposed hybrid collaborative representation-based classification method is evaluated on four benchmark remote-sensing image datasets and achieves state-of-the-art performance.
2. Proposed Method
2.1. Overview of CRC
2.2. Class-Specific Collaborative Representation
2.3. Hybrid Collaborative Representation
2.4. Hybrid Collaborative Representation with Kernels
2.5. Optimization of the Objective Function
2.6. Hybrid Collaborative Representation-Based Classification with Kernels
Algorithm 1 Algorithm for hybrid collaborative representation-based classification with kernels |
Require: Training samples , , and test sample y
|
3. Experimental Results
3.1. Experimental Settings
3.2. Experiment on UC Merced Land Use Dataset
3.2.1. Parameter tuning on UC Merced Land Use Dataset
3.2.2. Comparison with Several Classical Classifier Methods on UC Merced Land Use Dataset
3.2.3. Confusion Matrix on UC Merced Land Use Dataset
3.2.4. Comparison with State-of-the-Art Approaches
3.3. Experiment on the RSSCN7 Dataset
3.3.1. Comparison with Several Classical Classifier Methods on the RSSCN7 Dataset
3.3.2. Confusion Matrix on the RSSCN7 Dataset
3.4. Experiment on the WHU-RS19 Dataset
3.5. Experiment on the AID Dataset
4. Discussion
- For RS image classification, both shared attributes and class-specific attributes are vital to the representation of testing samples with training samples. So, based on CRC, we propose a hybrid collaborative representation-based classification method that can decrease the reconstruction error and improve the classification rate. Through comparison with several state-of-the-art methods for RS image classification, we can see that our proposed method is capable of efficiently promoting the performance of classifying remote-sensing images.
- Because of the existence of a nonlinear structure in image features, we extended our method into Reproducing Kernel Hilbert Space to further improve the performance of our method with kernel functions. From the experimental results of comparing with several classical classification methods, we can see that the classification rates of the Hybrid-KCRC method on these four datasets are all higher than that with the NN, LIBLINEAR, SOFTMAX, SLRC-L2, CRC and CS-CRC methods. Obviously, our proposed Hybrid-KCRC method achieved superior performance to these methods.
- We took the UC-Merced dataset as an example and evaluated the performance of our proposed Hybrid-KCRC method per class with a confusion matrix. From the confusion matrix, we can see that the Hybrid-KCRC method is better than other methods in most categories.
- It is true that there are several pretrained models to extract features and the performance of the resnet model outperforms the performance of the vgg model. In our paper, however, we paid more attention to the design of the classifier and not feature extraction. We only extracted the features of remote-sensing images to complete the classification task. The vgg is also very popular candidate models for extracting CNN activations of images. That are the reasons why we chose vgg. As a matter of fact, our method could be further improved by using other, better feature-extraction pretrained models. To demonstrate this, we also extracted CNN features with the pretrained Resnet model [44], and layer pool5 was utilized. The feature was a 2048-dimensional vector for each image. The final feature of each image was -normalized. The experimental results are shown in Table 6. For fair comparison on each dataset, we fixed the ratio of the number of training sets of the UC-Merced dataset, the WHU-RS19 dataset, the RSSCN7 dataset, and the AID dataset to , , and , respectively. From Table 6, we can see that the features extracted via the pretrained resnet model performed better than the features extracted via the pretrained vgg model. Our proposed Hybrid-KCRC methods also achieved superior performance to classical methods, such as SVM and CRC.
- Figures about classification rates with a different number of training samples clearly illustrate that RBF and polynomial kernel functions suit RS images better. Notably, classification rate increased by 1% with the Hybrid-KCRC method from linear kernel to polynomial kernel for the AID dataset. For the RBF kernel function, the metric was different from the linear kernel. The distance between two points x and y from the linear kernel space would be closer in the RBF kernel space if x and y were close, with the contrary conclusion if x and y were far away. This makes representation more discriminative and achieves higher classification accuracy. For the polynomial kernel, the linear kernel can be a special case of the polynomial kernel (). Note that kernel function can be approximated by [45] to save time for the learning algorithm. We will adopt the approximation of kernel to save time in future works.
- In the literature [23], the author proposed a joint collaborative representation (CR) classification method that uses several complementary features to represent images, including spectral values and spectral gradient features, Gabor texture features, and DMP features. This multifeature fusion can also be implemented via our proposed method. In the literature [24], the author proposed a spatial-aware CR for hyperspectral image classification that utilized both spatial and spectral features to represent an image. The penalty term can also be added into the objective function of our proposed method.
5. Conclusions
Author Contributions
Funding
Conflicts of Interest
Abbreviations
CRC | Collaborative Representation-Based Classification |
CS-CRC | Class-Specific Collaborative Representation-Based Classification |
RS | Remote Sensing |
EO | Earth Observation |
NS | Nearest Subspace |
BovW | Bag-of-Visual-Words |
CNNs | Convolutional Neural Networks |
LRC | Linear Regression-Based Classification |
POLY | Polynomial Kernel |
RBF | Radial Basis Function Kernel |
NN | Nearest Neighbor |
SLRC | Superposed Linear Representation Classifier |
References
- Navalgund, R.R.; Jayaraman, V.; Roy, P.S. Remote sensing applications: An overview. Curr. Sci. 2007, 93, 1747–1766. [Google Scholar]
- Liu, W.; Ma, X.; Zhou, Y.; Tao, D.; Cheng, J. p-Laplacian Regularization for Scene Recognition. IEEE Trans. Cybern. 2018, 1–14. [Google Scholar] [CrossRef] [PubMed]
- Ma, X.; Liu, W.; Li, S.; Tao, D.; Zhou, Y. Hypergraph p-Laplacian Regularization for Remotely Sensed Image Recognition. IEEE Trans. Geosci. Remote Sens. 2018, PP, 1–11. [Google Scholar] [CrossRef]
- Yang, X.; Liu, W.; Tao, D.; Cheng, J.; Li, S. Multiview Canonical Correlation Analysis Networks for Remote Sensing Image Recognition. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1855–1859. [Google Scholar] [CrossRef]
- Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the Sigspatial International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; ACM: New York, NY, USA, 2010; pp. 270–279. [Google Scholar]
- Cheng, G.; Han, J.; Guo, L.; Liu, Z.; Bu, S.; Ren, J. Effective and Efficient Midlevel Visual Elements-Oriented Land-Use Classification Using VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4238–4249. [Google Scholar] [CrossRef] [Green Version]
- Zou, J.; Li, W.; Chen, C.; Du, Q. Scene Classification Using Local and Global Features with Collaborative Representation Fusion. Inf. Sci. 2016, 348, 209–226. [Google Scholar] [CrossRef]
- Zhang, F.; Du, B.; Zhang, L. Saliency-guided unsupervised feature learning for scene classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2175–2184. [Google Scholar] [CrossRef]
- Fu, M.; Yuan, Y.; Lu, X. Unsupervised feature learning for scene classification of high resolution remote sensing image. In Proceedings of the IEEE China Summit and International Conference on Signal and Information Processing, Chengdu, China, 12–15 July 2015; pp. 206–210. [Google Scholar]
- Han, J.; Zhang, D.; Cheng, G.; Guo, L.; Ren, J. Object Detection in Optical Remote Sensing Images Based on Weakly Supervised Learning and High-Level Feature Learning. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3325–3337. [Google Scholar] [CrossRef] [Green Version]
- Li, J.; Huang, X.; Gamba, P.; Bioucas-Dias, J.M.; Zhang, L.; Benediktsson, J.A.; Plaza, A. Multiple Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1592–1606. [Google Scholar] [CrossRef]
- Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef]
- Li, Y.; Huang, X.; Liu, H. Unsupervised Deep Feature Learning for Urban Village Detection from High-Resolution Remote Sensing Images. Photogramm. Eng. Remote Sens. 2017, 83, 567–579. [Google Scholar] [CrossRef]
- Yu, Y.; Zhong, P.; Gong, Z. Balanced data driven sparsity for unsupervised deep feature learning in remote sensing images classification. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Fort Worth, TX, USA, 23–28 July 2017; pp. 668–671. [Google Scholar]
- Chaib, S.; Yao, H.; Gu, Y.; Amrani, M. Deep feature extraction and combination for remote sensing image classification based on pre-trained CNN models. In Proceedings of the International Conference on Digital Image Processing, Hong Kong, China, 19–22 May 2017; International Society for Optics and Photonics: Bellingham, WA, USA, 2017; p. 104203D. [Google Scholar]
- Li, W.; Du, Q. A survey on representation-based classification and detection in hyperspectral remote sensing imagery. Pattern Recognit. Lett. 2016, 83, 115–123. [Google Scholar] [CrossRef]
- Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
- Baraniuk, R.G. Compressive sensing [lecture notes]. IEEE Signal Process. Mag. 2007, 24, 118–121. [Google Scholar] [CrossRef]
- John, W.; Allen, Y.Y.; Arvind, G.; Shankar, S.S.; Yi, M. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar]
- Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the IEEE International Conference on Computer vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 471–478. [Google Scholar]
- Wu, S.; Chen, H.; Bai, Y.; Zhu, G. A remote sensing image classification method based on sparse representation. Multimed. Tools Appl. 2016, 75, 12137–12154. [Google Scholar] [CrossRef]
- Tang, X.; Liu, Y.; Chen, J. Improvement of Remote Sensing Image Classification Method Based on Sparse Representation. Comput. Eng. 2016, 42, 254–258, 265. [Google Scholar] [CrossRef]
- Li, J.; Zhang, H.; Zhang, L.; Huang, X.; Zhang, L. Joint Collaborative Representation With Multitask Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5923–5936. [Google Scholar] [CrossRef]
- Jiang, J.; Chen, C.; Yu, Y.; Jiang, X.; Ma, J. Spatial-Aware Collaborative Representation for Hyperspectral Remote Sensing Image Classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 404–408. [Google Scholar] [CrossRef]
- Deng, W.; Hu, J.; Guo, J. Face recognition via collaborative representation: Its discriminant nature and superposed representation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 2513–2521. [Google Scholar] [CrossRef]
- Liu, B.D.; Shen, B.; Wang, Y.X. Class specific dictionary learning for face recognition. In Proceedings of the IEEE International Conference on Security, Pattern Analysis, and Cybernetics (ICSPAC), Wuhan, China, 18–19 October 2014; pp. 229–234. [Google Scholar]
- Liu, B.D.; Shen, B.; Gui, L.; Wang, Y.X.; Li, X.; Yan, F.; Wang, Y.J. Face recognition using class specific dictionary learning for sparse representation and collaborative representation. Neurocomputing 2016, 204, 198–210. [Google Scholar] [CrossRef]
- Wang, W.; Yan, Y.; Winkler, S.; Sebe, N. Category specific dictionary learning for attribute specific feature selection. IEEE Trans. Image Process. 2016, 25, 1465–1478. [Google Scholar] [CrossRef] [PubMed]
- Zorzi, M.; Sepulchre, R. AR Identification of Latent-Variable Graphical Models. IEEE Trans. Autom. Control 2016, 61, 2327–2340. [Google Scholar] [CrossRef] [Green Version]
- Zou, Q.; Ni, L.; Zhang, T.; Wang, Q. Deep Learning Based Feature Selection for Remote Sensing Scene Classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2321–2325. [Google Scholar] [CrossRef]
- Sheng, G.; Yang, W.; Xu, T.; Sun, H. High-resolution satellite scene classification using a sparse coding based multiple feature combination. Int. J. Remote Sens. 2012, 33, 2395–2412. [Google Scholar] [CrossRef]
- Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
- Fan, R.E.; Chang, K.W.; Hsieh, C.J.; Wang, X.R.; Lin, C.J. LIBLINEAR: A library for large linear classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
- Yu, Y.; Gong, Z.; Wang, C.; Zhong, P. An Unsupervised Convolutional Feature Fusion Network for Deep Representation of Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2018, 15, 23–27. [Google Scholar] [CrossRef]
- Lu, X.; Zheng, X.; Yuan, Y. Remote sensing scene classification by unsupervised representation learning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5148–5157. [Google Scholar] [CrossRef]
- Lazebnik, S.; Schmid, C.; Ponce, J. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition (CVPR), New York, NY, USA, 17–22 June 2006; pp. 2169–2178. [Google Scholar]
- Vaduva, C.; Gavat, I.; Datcu, M. Latent Dirichlet allocation for spatial analysis of satellite images. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2770–2786. [Google Scholar] [CrossRef]
- Cheriyadat, A.M. Unsupervised feature learning for aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 439–451. [Google Scholar] [CrossRef]
- Penatti, O.A.; Nogueira, K.; dos Santos, J.A. Do deep features generalize from everyday objects to remote sensing and aerial scenes domains? In Proceedings of the IEEE International Conference on Computer vision (CVPR) Workshops, Boston, MA, USA, 7–12 June 2015; pp. 44–51. [Google Scholar]
- Hu, F.; Xia, G.S.; Hu, J.; Zhang, L. Transferring deep convolutional neural networks for the scene classification of high-resolution remote sensing imagery. Remote Sens. 2015, 7, 14680–14707. [Google Scholar] [CrossRef]
- Lin, D.; Fu, K.; Wang, Y.; Xu, G.; Sun, X. MARTA GANs: Unsupervised representation learning for remote sensing image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2092–2096. [Google Scholar] [CrossRef]
- Li, P.; Ren, P.; Zhang, X.; Wang, Q.; Zhu, X.; Wang, L. Region-Wise Deep Feature Representation for Remote Sensing Images. Remote Sens. 2018, 10, 871. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Zorzi, M.; Chiuso, A. The Harmonic Analysis of Kernel Functions. Automatica 2018, 94, 125–137. [Google Scholar] [CrossRef]
Methods\Datasets | UC-Merced |
---|---|
NN | 81.88 |
LIBLINEAR | 89.57 |
SOFTMAX | 88.00 |
SLRC-L2 | 89.79 |
CRC | 90.40 |
CS-CRC | 89.10 |
Hybrid-KCRC (linear) | 90.67 |
Hybrid-KCRC (POLY) | 91.43 |
Hybrid-KCRC (RBF) | 91.43 |
Hybrid-KCRC (Hellinger) | 90.90 |
Methods | Year | Accuracy |
---|---|---|
SPMK [37] | 2006 | |
LDA-SVM [38] | 2013 | |
SIFT + SC [39] | 2013 | |
Saliency + SC [8] | 2014 | |
CaffeNet [40] (without fine-tuning) | 2015 | |
CaffeNet [41] + VLAD | 2015 | |
DCGANs [42] (without augmentation) | 2017 | |
MAGANs [42] (without augmentation) | 2017 | |
WDM [36] | 2017 | |
UCFFN [35] | 2018 | |
CNN-W + VLAD with SVM [43] | 2018 | |
CNN-R + VLAD with SVM [43] | 2018 | |
VGG19 + liblinear | ||
VGG19 + CRC | ||
VGG19 + CS-CRC | ||
VGG19 + Hybrid-KCRC (linear) | ||
VGG19 + Hybrid-KCRC (POLY) | ||
VGG19 + Hybrid-KCRC (RBF) | ||
VGG19 + Hybrid-KCRC (Hellinger) |
Methods\Datasets | RSSCN7 |
---|---|
NN | 76.44 |
LIBLINEAR | 84.84 |
SOFTMAX | 82.14 |
SLRC-L2 | 81.99 |
CRC | 85.77 |
CS-CRC | 84.23 |
Hybrid-KCRC (linear) | 86.39 |
Hybrid-KCRC (POLY) | 87.34 |
Hybrid-KCRC (RBF) | 87.29 |
Hybrid-KCRC (Hellinger) | 86.71 |
Methods\Datasets | WHU-RS19 |
---|---|
NN | 87.74 |
LIBLINEAR | 94.42 |
SOFTMAX | 93.29 |
SLRC-L2 | 94.18 |
CRC | 94.58 |
CS-CRC | 93.95 |
Hybrid-KCRC (linear) | 94.76 |
Hybrid-KCRC (POLY) | 95.34 |
Hybrid-KCRC (RBF) | 95.34 |
Hybrid-KCRC (Hellinger) | 95.39 |
Methods\Datasets | AID |
---|---|
NN | 65.32 |
LIBLINEAR | 79.93 |
SOFTMAX | 76.13 |
SLRC-L2 | 79.27 |
CRC | 80.73 |
CS-CRC | 77.92 |
Hybrid-KCRC (linear) | 81.07 |
Hybrid-KCRC (POLY) | 82.07 |
Hybrid-KCRC (RBF) | 82.05 |
Hybrid-KCRC (Hellinger) | 81.28 |
Models\Datasets | UC-Merced () | RSSCN7 () | WHU-RS19 () | AID () |
---|---|---|---|---|
CaffeNet + SVM [32] | 95.02 | 96.24 | 88.25 | 89.53 |
VGG16 + SVM [32] | 95.21 | 96.05 | 87.18 | 89.64 |
GoogleNet + SVM [32] | 94.31 | 94.71 | 85.84 | 86.39 |
VGG19 + SVM | 94.67 | 95.42 | 85.99 | 90.35 |
VGG19 + CRC | 95.05 | 95.63 | 86.97 | 89.58 |
VGG19 + Hybrid-KCRC (linear) | 96.17 | 95.68 | 88.16 | 89.93 |
VGG19 + Hybrid-KCRC (POLY) | 96.29 | 96.42 | 89.21 | 91.75 |
VGG19 + Hybrid-KCRC (RBF) | 96.26 | 96.5 | 89.17 | 91.82 |
VGG19 + Hybrid-KCRC (Hellinger) | 96.33 | 95.82 | 88.47 | 90.35 |
Resnet + SVM | 96.90 | 97.74 | 91.5 | 92.97 |
Resnet + CRC | 97.00 | 98.03 | 92.47 | 92.85 |
Resnet + Hybrid-KCRC (linear) | 97.29 | 98.05 | 92.89 | 92.87 |
Resnet + Hybrid-KCRC (POLY) | 97.40 | 98.16 | 93.11 | 93.98 |
Resnet + Hybrid-KCRC (RBF) | 97.43 | 98.13 | 93.07 | 94.00 |
Resnet + Hybrid-KCRC (Hellinger) | 97.36 | 98.37 | 92.87 | 93.15 |
© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, B.-D.; Xie, W.-Y.; Meng, J.; Li, Y.; Wang, Y. Hybrid Collaborative Representation for Remote-Sensing Image Scene Classification. Remote Sens. 2018, 10, 1934. https://doi.org/10.3390/rs10121934
Liu B-D, Xie W-Y, Meng J, Li Y, Wang Y. Hybrid Collaborative Representation for Remote-Sensing Image Scene Classification. Remote Sensing. 2018; 10(12):1934. https://doi.org/10.3390/rs10121934
Chicago/Turabian StyleLiu, Bao-Di, Wen-Yang Xie, Jie Meng, Ye Li, and Yanjiang Wang. 2018. "Hybrid Collaborative Representation for Remote-Sensing Image Scene Classification" Remote Sensing 10, no. 12: 1934. https://doi.org/10.3390/rs10121934
APA StyleLiu, B. -D., Xie, W. -Y., Meng, J., Li, Y., & Wang, Y. (2018). Hybrid Collaborative Representation for Remote-Sensing Image Scene Classification. Remote Sensing, 10(12), 1934. https://doi.org/10.3390/rs10121934