Next Article in Journal
Comparing Winds near Tropical Oceanic Precipitation Systems with and without Lightning
Next Article in Special Issue
Data Augmentation and Spectral Structure Features for Limited Samples Hyperspectral Classification
Previous Article in Journal
Hyperspectral Anomaly Detection via Graph Dictionary-Based Low Rank Decomposition with Texture Feature Extraction
Previous Article in Special Issue
Pixel-Wise Classification of High-Resolution Ground-Based Urban Hyperspectral Images with Convolutional Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Spectral Feature Extraction Framework for Hyperspectral Images

1
Beijing Key Laboratory of Embedded Real-Time Information Processing Technology, School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, China
2
School of Electronics Engineering and Computer Science, Peking University, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Current address: School of Automation, Beijing Institute of Technology, 5 Zhongguancun Nandajie, Haidian District, Beijing 100081, China.
Remote Sens. 2020, 12(23), 3967; https://doi.org/10.3390/rs12233967
Submission received: 3 November 2020 / Revised: 25 November 2020 / Accepted: 28 November 2020 / Published: 4 December 2020
(This article belongs to the Special Issue Feature Extraction and Data Classification in Hyperspectral Imaging)

Abstract

:
Extracting diverse spectral features from hyperspectral images has become a hot topic in recent years. However, these models are time consuming for training and test and suffer from a poor discriminative ability, resulting in low classification accuracy. In this paper, we design an effective feature extracting framework for the spectra of hyperspectral data. We construct a structured dictionary to encode spectral information and apply learning machine to map coding coefficients. To reduce training and testing time, the sparsity constraint is replaced by a block-diagonal constraint to accelerate the iteration, and an efficient extreme learning machine is employed to fit the spectral characteristics. To optimize the discriminative ability of our model, we first add spectral convolution to extract abundant spectral information. Then, we design shared constraints for subdictionaries so that the common features of subdictionaries can be expressed more effectively, and the discriminative and reconstructive ability of dictionary will be improved. The experimental results on diverse databases show that the proposed feature extraction framework can not only greatly reduce the training and testing time, but also lead to very competitive accuracy performance compared with deep learning models.

Graphical Abstract

1. Introduction

Feature extraction of hyperspectral images (HSIs) is a significant topic at present and is widely applied in different HSI applications [1,2], including hyperspectral classification [3], target detection [4], and image fusion [5]. However, the variability and redundancy of spectra make it challenging to extract valid features from HSIs. A large number of feature learning techniques have been developed to describe spectral characteristics, which can be roughly categorized into two types: linear and nonlinear algorithms. Linear models exploit the original spectral information or linearly derive various features from such information. These kinds of features have been widely used to represent the linear separability of certain classes [6]. The common linear models are independent component analysis [7], principal component analysis [8], and linear discriminant analysis [9]. Although these models are simple and compact, they suffer from poor representation ability and cannot cope with intricate HSI data.
The nonlinear models are more effective for class discrimination due to the existence of nonlinear class boundaries. These approaches adopt nonlinear transformations to better represent spectral features of HSIs. The kernel-based method [10] is a common nonlinear model that maps samples into higher dimensional space. Support vector machine (SVM) [11,12,13] is a representative kernel-based method and has been proven to be effective for HSI classification. In [14], Bruzzone proposed a transductive SVM that can simultaneously utilize labeled and unlabeled data. Nonetheless, kernel-based algorithms usually lack a theoretical basis for the selection of the corresponding parameters and are not scalable to large datasets. Another widely used nonlinear model is the deep learning method with strong potential for feature learning. Chen et al. [15] verified the eligibility of the stacked autoencoder (SAE) by classical spectral information-based classification. A similar model was proposed by Chen et al. [16], who applied deep belief networks (DBNs) to extract features in practice. In [17,18,19,20], multiple dimension convolutional neural networks (CNNs) were adopted for HSI classification. Rasti et al. [21] provided a technical overview of the state-of-the-art techniques for HSI classification, especially the deep learning models. However, deep learning models require numerous labeled data points, strictly limiting their application domain. Moreover, the trained models are inflexible, and their parameters are difficult to adjust.
Recently, dictionary-based methods have been introduced into HSI recognition. Compared with deep learning models, dictionary-based methods can represent spectral characteristics more effectively with less HSI data. Regarding sparse representation-based classification (SRC), References [22,23] constructed an unsupervised dictionary that often engendered unstable sparse coding. References [24,25] combined the kernel model with sparse coding to make samples more separable. Li et al. [26] designed a robust sparse representation algorithm against outliers in practice. To obtain a compact and discriminative dictionary, Zhang and Li [27] absorbed label information and constructed a k-singular-value decomposition (K-SVD) dictionary for feature learning. Moreover, Reference [28] optimized the discriminative dictionary and applied it to process HSIs. In [29,30], learning vector quantization was adopted for dictionary-based models for hyperspectral classification. In general, dictionary-based methods show great potential for HSI feature representation. However, these dictionaries are time consuming, and their discriminative ability is poor.
To address the aforementioned drawbacks, we propose an efficient framework that trains a discriminative structure dictionary to describe HSIs. The main novelties of the proposed model are threefold:
(1)
We design an efficient feature learning framework that calculates the structured dictionary to encode spectral information and adopts machine learning to map the coding coefficients. The block-diagonal constraint is applied to increase the efficiency of coding, and an effective extreme learning machine (ELM) is employed to complete the mapping.
(2)
We apply spectral convolution to extract the mean value and local variation of the spectra of HSIs. Then, the dictionary learning is carried out to capture more local spectral characteristics of HSI data.
(3)
We devise a new shared constraint for all of the subdictionaries. In this way, the common and specific features of HSI samples will be learned separately to achieve a more discriminative representation.

2. Materials and Methods

In this section, we first introduce the experimental datasets and then elaborate the proposed feature extracting framework for HSIs.

2.1. The Study Datasets

The experimental datasets include three well-known HSI datasets, and we randomly select 10 % of each dataset for training and the rest for testing. The detailed information is presented as follows.
Center of Pavia [31]: The HSI data were collected by the airborne sensor of the reflective optics system imaging spectrometer (ROSIS) located in the urban area of Pavia, Northern Italy. The image consisted of 1096 × 492 pixels at a ground sampling distance (GSD) of 1.3 m with 102 spectral bands in the range of 430 nm to 860 nm. In this dataset, nine main categories are investigated for the land cover classification task. The number of training and testing samples is specifically listed in Table 1.
Botswana [32]: This dataset was collected by the Hyperion sensors on the NASA Earth Observing 1 (EO-1) satellite over the Okavango Delta, Botswana. It has 1476 × 256 pixels at a GSD of 30 m with 145 spectral channels ranging from 400 nm to 2500 nm. There are 14 challenging classes for the land cover classification task. Table 2 lists the scene categories and the number of training and testing samples used in the classification task.
Houston University 2013 [21]: The dataset was collected by the compact airborne spectrographic imager (CASI) sensor over the campus of the University of Houston and its surrounding areas, in Houston, TX, USA. It contains 349 × 1905 pixels at a GSD of 1 m with 144 spectral channels ranging from 364 nm to 1046 nm. The specific training and test information for the data is detailed in Table 3.

2.2. Related Works

Recently, dictionary learning has led to promising results in HSI classification recognition. Dictionary learning aims to learn a set of atoms, also called visual words in the computer vision community, in which a few atoms can be linearly combined to well approximate a given signal [33]. Here, we briefly introduce several mainstream dictionary-based approaches.

2.2.1. Review of Sparse Representation-Based Classification

Wright et al. [22] proposed the sparse representation-based classification (SRC) model, which is widely applied in HSI classification [30]. Suppose there are C classes of HSIs. Let X = [ X 1 , , X i , , X C ] be the set of original training samples, where X i is the subset of training samples from class i. Then, sparse coding vector a corresponding to dictionary D is obtained by the l p -norm minimization constraint as follows:
a = arg min a X D a 2 2 + λ a p ,
where λ is a positive scalar and p is usually zero or one. The test samples can be classified via the following:
arg min i X D a i 2 2 ,
where a i is the coefficient vector associated with class i. SRC has impressive performance in face recognition and is robust to different noises [33]. It acts as a leading methodtoward classification with the help of dictionary coding. Nevertheless, it is obvious that the SRC model naively employs all the training samples as one dictionary. The dictionary of SRC suffers from redundant atoms and a disordered structure, making is unsuitable for complex HSI classification.

2.2.2. Review of Class-Specific Dictionary Learning

As discussed in [34], the pre-defined dictionary of the SRC model incorporates much redundancy, as well as noise and trivial information. To solve this problem, Yang et al. [34] constructed a class-specific dictionary, in which sub-dictionary D i of learned dictionary D = [ D 1 , , D i , , D C ] corresponds to class i. The sub-dictionary could be learned class-by-class as follows:
D i = arg min D i X i D i A i 2 2 + λ A i p ,
where A i is the coding result of samples X i on sub-dictionary D i . Equation (3) can be seen as the basic model of the class-specific dictionary learning model since each D i is trained separately from the samples of a specific class. We can apply reconstruction error X D i A i 2 to classify HSI data. However, Equation (3) does not consider the discriminative ability between different coefficients, resulting in low classification accuracy.

2.2.3. Review of Fisher Discriminant Dictionary Learning

Yang et al. [35] proposed a complex model named Fisher discriminant dictionary learning (FDDL), which adopts the Fisher criterion to learn a structured dictionary. Suppose that X = [ X 1 , , X i , , X C ] R ( L × N ) refers to all N training HSI samples from C classes with L band number. The coding matrix A = [ A 1 , , A i , , A C ] R ( N A × N ) is the corresponding coefficient over dictionary D containing N A atoms. The ith training sample can be computed as X i = D i A i , and the objective function is shown as follows:
L o s s D , A = arg min D , A L R + λ 1 L S + λ 2 L D ,
where λ 1 and λ 2 are the regularization parameters. L R , L S , and L D denote reconstructive loss, sparse constraint loss, and discriminative loss, respectively:
L R = X i D A i F 2 + X i D i A i i F 2 + j = 1 , j i C D j A i j F 2 ,
L S = A 1 ,
L D = t r S W A t r S B A + η A F 2 ,
where · F is the Frobenius norm. In Equation (5), the first term X i D A i F 2 guarantees reconstruction fidelity, while the rest of the terms are designed for the discriminative ability of dictionary D. As for Equation (6), A 1 is a sparsity constraint and can be calculated by lasso [35]. Equation (7) based on the Fisher criterion [35] can be completed by minimizing the within-class scatter of A, denoted by S W ( A ) , and maximizing the between-class scatter of S B ( A ) . The last elastic term of Equation (7) is applied to solve the non-convex problem.
The atoms of the structured dictionary in FDDL are strongly correlated with specific classes, which will improve the representation ability of D. However, the FDDL model is time consuming and unsuitable for practical application. More importantly, the structure of the FDDL model needs improvement to enhance the reconstructive ability.

2.3. Proposed Framework

Figure 1 shows the workflow of the proposed framework in which we construct a structured dictionary to extract spectral features for classification application. Spectral convolution is first introduced into our model to extract the abundant information. Following the convolution, the corresponding coding representations are built for the test spectral data. We design the shared constraint for all of subdictionaries to enhance the discriminative ability of the structured dictionary. Finally, the ELM model is adopted to map the coding coefficients to the corresponding labels.

2.3.1. Spectral Convolution

The HSI data contain a massive amount of spectral characteristics, such as reflection peaks and valleys, which play important roles in spectral classification. To extract this spectral information, we design different convolution masks for the original samples. The masks are as follows:
M 1 = 0 1 0 , M 2 = 1 / 3 1 / 3 1 / 3 , M 3 = 1 / 4 1 / 2 1 / 4 .
To achieve stable classification performance, we apply M 1 to preserve the original data. Inspired by the wave transform, we design mask M 2 to extract the main structure (mean values) of spectral samples and mask M 3 to capture the detailed information (local variation) of the spectra. As shown in Figure 2, the results of M 2 capture the main signal of spectra ( M 1 ) and the values of M 3 change with the local variation in the spectra ( M 1 ). Mask M 2 can be adopted to describe the main structure of spectral samples, while mask M 3 can be applied to describe the local reflection valleys and peaks of spectral data. However, the running time is closely related to the number of masks. In this work, we only employ three convolutional masks to extract the spectral information, and there are other possible masks that can be applied to extract the spectral characteristics.

2.3.2. Structured Dictionary

To encode the spectral information, most of the dictionary-based methods [34,35] are based on the sparsity constraint under the following framework:
arg min D , A X D A F 2 + λ A p + φ D i , A i ,
where λ 0 is a scalar constant. The first term X D A F 2 is the fidelity constraint to ensure the representation ability of trained dictionary. The second term A p is the sparsity constraint, and the remaining term φ D i , A i is the additional constraint for some discrimination promotion function. These models will train a structured dictionary to represent signals, which will promote discrimination between classes. However, the sparsity constraint is time consuming on the coding coefficients, making the model inefficient. More importantly, the role of sparse coding in classification is still an open problem [36,37,38], and some experts have argued that sparse coding may not be crucial for dictionary classification.
As described in [38], the block-diagonal constraint is an efficient way to calculate coding coefficients. Here, we built the structured dictionary model as follows:
{ A , D } = arg min A , D i = 1 C X i D i A i F 2 + j = 1 , j i C A i j F 2 ,
where the coefficient matrix A will be nearly block diagonal. The objective function in Equation (10) is generally non-convex. We introduce a variable matrix P to calculate the coefficient matrix A. Matrix P R N A × L is an encoder, and code A can be calculated as A = P X . With the encoder P = [ P 1 ; ; P j ; ; P C ] , we want the encoder P j to be able project the samples X i ( j i ) to a nearly null space, i.e., P j X i 0 , j i . Therefore, Equation (10) can be relaxed to the following problem:
{ A , D , P } = arg min A , D , P i = 1 C X i D i A i F 2 + τ P i X i A i F 2 + λ P i X i ¯ F 2 ,
where τ and λ are scalar constants, P i X i = A i , and X i ¯ denotes the complementary data matrix of subset X i in the whole training set X. Equation (11) can be implemented via a two-stage iterative algorithm: updating A with fixed D and P and updating D and P with fixed A.
(1) Suppose that D and P are fixed, and A are updated as follows:
{ A } = arg min A i = 1 C X i D i A i F 2 + τ P i X i A i F 2 .
Equation (12) is a standard least squares problem, and we achieve the closed-form solution:
A i ( k + 1 ) = D i ( k ) T D i ( k ) + τ I 1 τ P i ( k ) X i + D i ( k ) T D i ( k ) ,
where I is the unit matrix.
(2) Fixing A, D and P are updated as follows:
{ P } = arg min P i = 1 C τ P i X i A i F 2 + λ P i X i ¯ F 2 { D } = arg min D i = 1 C X i D i A i F 2 , s . t . d i 2 2 1 ,
where d i is the atom of the structured dictionary and d i 2 2 1 is to make the dictionary more stable. The closed-form solution of P can be obtained as:
P i k + 1 = τ A i ( k ) X i T τ X i X i T + λ X i ¯ X i ¯ T + γ I 1 ,
where γ is a small number. D can be calculated by introducing a variable S:
{ D , S } = arg min D , S i = 1 C X i D i A i F 2 s . t . D = S ,   d i 2 2 1 .
The optimal solution of Equation (16) can be achieved by the alternating direction method of multipliers (ADMM) algorithm [39]:
D k + 1 = arg min D i = 1 C X i D i ( k ) A i ( k ) F 2 + ρ D i ( k ) S i ( k ) + T i ( k ) F 2 S k + 1 = arg min S i = 1 C ρ D i ( k + 1 ) S i ( k ) + T i ( k ) F 2 T k + 1 = T k + D ( k + 1 ) S ( k + 1 ) ,
where ρ is an ever-changing value with a fixed ratio and T is a temp matrix. All these closed-form solutions converge rapidly, and a balance between the discrimination and representation power of the model can be achieved.

2.3.3. Shared Constraint

To improve the representation and reconstructive ability of the subdictionaries, we design the shared constraint for subdictionaries. As shown in Figure 3, the test samples contain the shared features, and our shared constraint (the c o m subdictionary) is added to describe duplicated information (shared features). Then, the discriminative features will be “amplified” relative to the original ones, and constructing a new structured dictionary is easier than ever.
Here, we design a subdictionary D c o m to calculate the class-shared characteristics as follows:
D = { D 1 , D 2 , , D C , D c o m } ,
where D c o m denotes the shared subdictionary. The corresponding objective function is modified as follows:
{ A , D } = arg min A , D i = 1 C X i D i A i F 2 + X i D c o m A c o m F 2 + j = 1 , j i C A i j F 2 , = arg min A , D i = 1 C X i D i c o m A i c o m F 2 + j = 1 , j i C A i j F 2 ,
where D i c o m = [ D i , D c o m ] and A i c o m = [ A i , A c o m ] . The introduction of D c o m will not affect the solution procedure. With the calculation of term X i D c o m A c o m F 2 , the results of term i = 1 C j = 1 , j i C A i j F 2 tend to be closer to zero, and the corresponding reconstructive ability of the structured dictionary will be improved.

2.3.4. Feature Extraction Framework

We construct the structured dictionary and encode the spectral information of HSIs. The coding coefficients A will be fed into the learning classifier to achieve better performance than directly using the minimum reconstruction error for classification. Different learning classifiers, such as SVM [12] and neural networks (NNs), can be employed to map the coding coefficients. However, these tools are often time consuming. Therefore, we employ an efficient machine technique, i.e., the extreme learning machine, to classify the HSIs.
In [40], Huang et al. proposed an ELM for generalized single-hidden-layer feed-forward neural networks (SLFNs), which has been widely applied in various application [41,42]. The ELM tries to learn an approximation function based on the training data. Suppose that SLFNs with K hidden nodes can be represented as follows:
f L ( x i ) = j = 1 K g ( x i , a i j , b i j ) β j ,
where a i j is the input weight connecting the input x i to the j-th hidden node, b i j is the bias connecting the input x i with the j-th hidden node, g ( · ) is the activation function, and β j is the output weight of the j-th hidden node. The activation function g ( · ) can be any nonlinear piecewise continuous function as follows:
g ( x ; θ ) = 1 1 + e x p ( ( a T X + b ) ) ,
g ( x ; θ ) = e x p ( b X a 2 ) ,
where Equations (21) and (22) are the sigmoid and radial basis function (RBF), θ = ( a , b ) are the parameters of the mapping function, and · 2 denotes the Euclidean norm.
Huang et al. [43] proved that SLFNs can approximate any continuous target function over any compact subset X with the above sigmoid and RBF functions. Training ELMs is equivalent to settling a regularized least-squares problem, which is considerately more efficient than training an SVM or learning with back-propagation. Therefore, in our model, an ELM is adopted for mapping the coding coefficients into different classes of HSIs.

3. Experimental Results and Discussion

In this section, we compare the performance of our proposed method with other feature extracting models, including SVM [12], FDDL [35], DPL [38], ResNet [44], RNN [21], and CNN [21] for HSI classification. We report the overall accuracy (OA), average accuracy (AA), and kappa coefficient of the different datasets and present the corresponding classification maps. The proposed method is evaluated, and relevant results are summarized and discussed in detail as follows.

3.1. Compared Methods and Evaluation Indexes

The SVM model (the codes for SVM were otained from https://www.csie.ntu.edu.tw/~cjlin/libsvm/) is a representative kernel-based method and has shown effective performance in HSI classification [12,13,45]. Yang et al. [35] proposed a complicated model named FDDL (the codes of FDDL were from http://www4.comp.polyu.edu.hk/~cslzhang/papers.htm), which was applied in HSI classification in [46]. The DPL [38] method (http://www4.comp.polyu.edu.hk/~cslzhang/papers.htm) is constructed to reduce the running time of learning the dictionary model. Convolutional neural networks (CNNs) [21] (all the CNNs models were downloaded from https://github.com/BehnoodRasti/HyFTech-Hyperspectral-Shallow-Deep-Feature-Extraction-Toolbox) are the most popularly adopted deep model for hyperspectral classification. Compared to traditional deep fully connected networks, CNNs possess weight-sharing and local-connection characteristics, making their training processes more efficient and effective. ResNet [44] adopts a residual networks to address the degradation problem and enhances the convergence rate of the CNN model, which is employed in HSI classification [47]. Recurrent neural networks (RNNs) [48,49] process all the spectral bands as a sequence and adopt a flexible network structure to classify HSIs. All experiments were repeated 10 times with the average classification results reported for comparison.
We used the following criteria to evaluate the performance of the different methods for HSI classification used in this paper, which include:
Overall accuracy (OA): the number of correctly classified HSI pixels divided by the total number of tests [50];
Average accuracy (AA): the average value of the classification accuracies of all classes [50];
Kappa coefficient: A statistical measurement of agreement between the final classification and the ground-truth map [50].

3.2. Discussions of Different Datasets

(1) Center of Pavia: Table 4 lists the classification results of the compared algorithms, and Figure 4 shows the confusion matrix of our model (only to one decimal place). In Table 4, one can observe that all the CNN-based models have a good performance. The best performance is achieved by the proposed framework whose OA, AA, and kappa coefficients are 98.39 % , 95.83 % , and 97.23 % , respectively. Compared with the dictionary learning- and deep learning-based models, our model gains significant classification accuracy for this dataset, especially for Class No. 2; see Figure 4. The confusion matrix for our model is shown in Figure 4, indicating that our algorithm distinguishes surface regions quite effectively.
For illustrative purposes, Figure 5 shows the obtained classification maps of the compared methods on the Center of Pavia dataset. Figure 5a,b is the RGB image and ground truth map, and Figure 5c–h is the corresponding classification results of SVM, FDDL, DPL, ResNet, RNN, CNN, and the proposed model. We employ yellow and red rectangles to highlight the interesting regions. We can observe from Figure 5 that the classification maps obtained by the proposed feature extractor are smoother in the regions sharing the same materials and sharper on the edges between different materials. The classification map produced from our model is the closest one compared with the results from other approaches. Our method is capable of extracting the intrinsic invariant feature representation from the HSI, achieving a more effective feature extraction.
(2) Botswana: The class-specific classification accuracies for the Botswana dataset and corresponding confusion matrix of our model are provided in Table 5 and Figure 6, respectively. From the results, one can see that the proposed algorithm outperforms the other algorithms in terms of OA, AA, and kappa, especially for Class Nos. 10 and 13. The proposed method significantly improves the results with a very high accuracy when tested with the Botswana dataset. From the illustrative results in the confusion matrix map, our model shows more discriminative ability between different classes. The confusion matrix can also confirm the class-specific classification accuracies presented in Table 5.
Figure 7 shows the classification maps for the Botswana dataset where Figure 7a,b is the RGB image and ground truth map and Figure 7c–h is the corresponding classification results of SVM, FDDL, DPL, ResNet, RNN, CNN, and the proposed model. We employ yellow and red rectangles to highlight the interesting regions. From the illustrative presentation in the classification maps, the compared algorithms show more noisy scattered point in the maps. The proposed method can remove them and lead to smoother classification results without blurring the boundaries. The result of our model is the closest one compared with the state-of-the-art methods. It demonstrates the effectiveness of the proposed structured dictionary learning model.
(3) Houston University 2013: Table 6 lists the classification result of the compared methods on the Houston University 2013 dataset, and Figure 8 shows the corresponding confusion matrix of our model. In Table 6, it is obvious that our model achieves slightly better performance than CNN-based models. The OA, AA, and kappa coefficients of our framework are 86.82 % , 86.44 % , and 85.74 % , respectively. Compared with the dictionary learning- and deep learning-based models, our model gains significant classification accuracy over this dataset, especially for Class Nos. 8, 9, and 12. The confusion matrix for our model is shown in Figure 8, indicating that our algorithm distinguishes surface regions quite effectively.
For illustrative purposes, Figure 9 shows the obtained classification maps of the compared methods on the Houston University 2013 dataset. Figure 9a,b is the RGB image and ground truth map, and Figure 9c–h is the corresponding classification results of SVM, FDDL, DPL, ResNet, RNN, CNN, and the proposed model. We employ yellow and red rectangles to highlight the interesting regions. As shown in Figure 9, our model removes the effects of salt-and-pepper noise from the classification maps effectively and simultaneously preserves the meaningful structure or objects. Owing to the robustness in local changes of the spectra, our model obtains more accurate classification maps in the area in and around the parking lot. Generally speaking, our model clearly shows superior performance in effective classification of HSIs.

3.3. Small Training Samples

The impact of the sample size for HSI classification has been reported in many research studies [23,24,28]. To confirm the effectiveness of our framework on small training samples, we randomly selected 5 % of the Botswana dataset for training and the rest for testing. As shown in Table 7, the classification performance is extremely susceptible to the number of training data. The reduction to 5 % of the training samples leads to a decrease of about 2 % 4 % in classification accuracy. The OA, AA, and kappa of our model are 88.42%, 88.95%, and 87.46%, beating all other compared methods. This result suggests that our model has the potential to achieve higher level accuracy with a limited sample size.

3.4. Time Cost

All the experiments in this paper were implemented with MATLAB 2018b and Python on a Windows 10 operation system and conducted on an Intel Core i7-8700 CPU 3.20 GHz desktop with 16GB memory. The training and testing time of different models are listed in Table 8. Overall, the training and testing time of our model are far less than the SVM- and CNN-based models, which clearly shows the superior efficiency of our approach in classification application.

4. Conclusions

In this work, we propose an efficient spectral feature extraction framework for HSI data. This algorithm is more suitable for low spatial resolution HSIs with a lack of spatial features. To improve the efficiency of our framework, we replace the sparsity constraint with the block-diagonal constraint to reduce the coding computation and employ an ELM model to map the coding coefficients. More importantly, we design spectral convolution and perform the dictionary learning on these features to capture more local spectral characteristics of the data. We also design a new shared constraint to construct a discriminative dictionary in the learning. Extensive experiments are conducted on three HSI datasets, and both qualitative and quantitative results demonstrate the effectiveness of the proposed feature learning model. Furthermore, the proposed approach consistently achieves higher classification accuracy even under a small number of training samples. In comparison to the SVM- and CNN-based models, our framework requires much less computation time, which demonstrates its potential and superiority in the HSI classification task. In the future, we will continue to incorporate the spatial information into the model to further strengthen the feature representation ability.

Author Contributions

Funding acquisition, B.Z.; Methodology, Z.L.; Supervision, B.Z.; Visualization, W.W.; Writing—original draft, Z.L. and W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (NSFC) under Grant 91738302 and in part by the National Natural Science Foundation of China (NSFC) under Grant 31727901.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dou, P.; Zeng, C. Hyperspectral Image Classification Using Feature Relations Map Learning. Remote Sens. 2020, 12, 2956. [Google Scholar] [CrossRef]
  2. Santos-Rufo, A.; Mesas-Carrascosa, F.-J.; García-Ferrer, A.; Meroño-Larriva, J.E. Wavelength Selection Method Based on Partial Least Square from Hyperspectral Unmanned Aerial Vehicle Orthomosaic of Irrigated Olive Orchards. Remote Sens. 2020, 12, 3426. [Google Scholar] [CrossRef]
  3. Li, J.; Huang, X.; Gamba, P.; Bioucas-Dias, J.M.; Zhang, L.; Benediktsson, J.A.; Plaza, A. Multiple Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1592–1606. [Google Scholar] [CrossRef] [Green Version]
  4. Yang, S.; Shi, Z. Hyperspectral Image Target Detection Improvement Based on Total Variation. IEEE Trans. Image Process. 2016, 25, 2249–2258. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, X.; Deng, C.; Chanussot, J.; Hong, D.; Zhao, B. Stfnet: A two-stream convolutional neural network for spatiotemporal image fusion. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6552–6564. [Google Scholar] [CrossRef]
  6. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  7. Bayliss, J.; Gualtieri, J.; Cromp, R. Analysing hyperspectral data with independent component analysis. Proc. Int. Soc. Opt. Eng. 1997, 3240, 133–143. [Google Scholar]
  8. Rodarmel, C.; Shan, J. Principal Component Analysis for Hyperspectral Image Classification. Surv. Land Inf. Syst. 2002, 62, 115–122. [Google Scholar]
  9. Ji, S.; Ye, J. Generalized linear discriminant analysis: A unified framework and efficient model selection. IEEE Trans. Neural Netw. 2008, 19, 1768–1782. [Google Scholar]
  10. Scholkopf, B.; Smola, A. Learning with Kernels? Support Vector Machines, Regularization, Optimization and Beyond; MIT Press: Cambridge, MA, USA, 2002; pp. 1768–1782. [Google Scholar]
  11. Camps-Valls, G.; Bruzzone, L. Kernel-based methods for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1351–1362. [Google Scholar] [CrossRef]
  12. Archibald, R.; Fann, G. Feature selection and classification of hyperspectral images with support vector machines. IEEE Geosci. Remote Sens. Lett. 2007, 4, 674–677. [Google Scholar] [CrossRef]
  13. Bahria, S.; Essoussi, N.; Limam, M. Hyperspectral data classification using geostatistics and support vector machines. Remote Sens. Lett. 2011, 2, 99–106. [Google Scholar] [CrossRef]
  14. Bruzzone, L.; Chi, M.; Marconcini, M. A novel transductive svm for semisupervised classification of remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3363–3373. [Google Scholar] [CrossRef] [Green Version]
  15. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  16. Chen, Y.; Zhao, X.; Jia, X. Spectralcspatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  17. Romero, A.; Gatta, C.; Camps-Valls, G. Unsupervised deep feature extraction for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 1349–1362. [Google Scholar] [CrossRef] [Green Version]
  18. Zhao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  19. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  20. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  21. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature Extraction for Hyperspectral Imagery: The Evolution from Shallow to Deep (Overview and Toolbox). arXiv 2020, arXiv:2003.02822. [Google Scholar] [CrossRef]
  22. Wright, J.; Yang, A.; Ganesh, A.; Sastry, S.; Ma, Y. Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 210–227. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Chen, Y.; Nasrabadi, N.; Tran, T. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  24. Chen, Y.; Nasrabadi, N.; Tran, T. Hyperspectral image classification via kernel sparse representation. IEEE Trans. Geosci. Remote Sens. 2013, 51, 217–231. [Google Scholar] [CrossRef] [Green Version]
  25. Gao, S.; Tsang, I.; Chia, L. Sparse representation with kernels. IEEE Trans. Image Process. 2013, 22, 423–434. [Google Scholar]
  26. Li, C.; Ma, Y.; Mei, X.; Liu, C.; Ma, J. Hyperspectral image classification with robust sparse representation. IEEE Geosci. Remote Sens. Lett. 2016, 13, 641–645. [Google Scholar] [CrossRef]
  27. Zhang, Q.; Li, B. Discriminative k-svd for dictionary learning in face recognition. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2691–2698. [Google Scholar]
  28. Du, P.; Xue, Z.; Li, J.; Plaza, A. Learning discriminative sparse representations for hyperspectral image classification. IEEE J. Sel. Top. Signal Process. 2015, 9, 1089–1104. [Google Scholar] [CrossRef]
  29. Wang, Z.; Nasrabadi, N.; Huang, T. Spatial-spectral classification of hyperspectral images using discriminative dictionary designed by learning vector quantization. IEEE Trans. Geosci. Remote Sens. 2014, 52, 4808–4822. [Google Scholar] [CrossRef]
  30. Gao, L.; Yu, H.; Zhang, B.; Li, Q. Locality-preserving sparse representation-based classification in hyperspectral imagery. J. Appl. Remote Sens. 2016, 10, 1–15. [Google Scholar] [CrossRef]
  31. Mei, X.; Pan, E.; Ma, Y.; Dai, X.; Huang, J.; Fan, F.; Du, Q.; Zheng, H.; Ma, J. Spectral-Spatial Attention Networks for Hyperspectral Image Classification. Remote Sens. 2019, 11, 963. [Google Scholar] [CrossRef] [Green Version]
  32. Yang, X.; Zhang, X.; Ye, Y.; Lau, R.; Lu, S.; Li, X.; Huang, X. Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification. Remote Sens. 2020, 12, 2033. [Google Scholar] [CrossRef]
  33. Shu, K.; Wang, D. A Brief Summary of Dictionary Learning Based Approach for Classification. arXiv 2012, arXiv:1205.6544. [Google Scholar]
  34. Yang, M.; Zhang, L.; Yang, J.; Zhang, D. Metaface learning for sparse representation based face recognition. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010. [Google Scholar]
  35. Yang, M.; Zhang, L.; Feng, X.; Zhang, D. Fisher discrimination dictionary learning for sparse representation. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  36. Coates, A.; Ng, A. The importance of encoding versus training with sparse coding and vector quantization. In Proceedings of the International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011. [Google Scholar]
  37. Zhang, L.; Yang, M.; Feng, X. Sparse representation or collaborative representation: Which helps face recognition? In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  38. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Projective Dictionary Pair Learning for Pattern Classification. Adv. Neural Inf. Process. Syst. 2014, 27, 793–801. [Google Scholar]
  39. Boyd, S.; Parikh, N.; Chu, E. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Found. Trends Mach. Learn. 2010, 3, 1–122. [Google Scholar] [CrossRef]
  40. Huang, G.; Zhu, Q.; Siew, C. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  41. Liu, X.; Deng, C.; Wang, S.; Huang, G.; Zhao, B.; Lauren, P. Fast and Accurate Spatiotemporal Fusion Based Upon Extreme Learning Machine. IEEE Geosci. Remote Sens. Lett. 2016, 13, 2039–2043. [Google Scholar] [CrossRef]
  42. Zhou, S.; Deng, C.; Wang, W.; Huang, G.; Zhao, B. GenELM: Generative Extreme Learning Machine feature representation. Neurocomputing 2019, 362, 41–50. [Google Scholar] [CrossRef]
  43. Huang, G.; Chen, L.; Siew, C. Universal Approximation Using Incremental Constructive Feedforward Networks With Random Hidden Nodes. IEEE Trans. Neural Netw. 2006, 17, 879–892. [Google Scholar] [CrossRef] [Green Version]
  44. Zhong, Z.; Li, J.; Ma, L.; Jiang, H.; Zhao, H. Deep residual networks for hyperspectral image classification. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; Volume 142–149, pp. 1824–1827. [Google Scholar]
  45. Zhao, C.; Liu, W.; Xu, Y.; Wen, J. A spectral-spatial SVM-based multi-layer learning algorithm for hyperspectral image classification. Remote Sens. Lett. 2018, 9, 218–227. [Google Scholar] [CrossRef]
  46. Yuan, Z.; Sun, H.; Ji, K.; Zhou, H. Hyperspectral Image Classification Using Fisher Dictionary Learning based Sparse Representation. Remote Sens. Technol. Appl. 2014, 29, 646–652. [Google Scholar]
  47. Meng, Z.; Li, L.; Tang, X.; Feng, Z.; Jiao, L.; Liang, M. Multipath Residual Network for Spectral-Spatial Hyperspectral Image Classification. Remote Sens. 2019, 11, 1896. [Google Scholar] [CrossRef] [Green Version]
  48. Shi, C.; Pun, C. Multi-scale hierarchical recurrent neural networks for hyperspectral image classification. Neurocomputing 2018, 294, 82–93. [Google Scholar] [CrossRef]
  49. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef] [Green Version]
  50. Mou, L.; Ghamisi, P.; Zhu, X. Deep Recurrent Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Workflow of the proposed feature extraction model.
Figure 1. Workflow of the proposed feature extraction model.
Remotesensing 12 03967 g001
Figure 2. Examples of spectral data with different convolution masks.
Figure 2. Examples of spectral data with different convolution masks.
Remotesensing 12 03967 g002
Figure 3. Overview of the built dictionary for different models. Shared constraints are applied for structured dictionaries to represent the shared features between subdictionaries, and unique features can acquire effective expressions.
Figure 3. Overview of the built dictionary for different models. Shared constraints are applied for structured dictionaries to represent the shared features between subdictionaries, and unique features can acquire effective expressions.
Remotesensing 12 03967 g003
Figure 4. The confusion matrix of our model on the Pavia of Center dataset.
Figure 4. The confusion matrix of our model on the Pavia of Center dataset.
Remotesensing 12 03967 g004
Figure 5. Classification maps of the Center of Pavia dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) ours. The yellow and red rectangles correspond to building and water areas.
Figure 5. Classification maps of the Center of Pavia dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) ours. The yellow and red rectangles correspond to building and water areas.
Remotesensing 12 03967 g005
Figure 6. The confusion matrix of our model on the Botswana dataset.
Figure 6. The confusion matrix of our model on the Botswana dataset.
Remotesensing 12 03967 g006
Figure 7. Classification maps of the Botswana dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) O = ours. The yellow and red rectangles correspond to grassland and mountain areas.
Figure 7. Classification maps of the Botswana dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) O = ours. The yellow and red rectangles correspond to grassland and mountain areas.
Remotesensing 12 03967 g007
Figure 8. The confusion matrix of our model on the Houston University 2013 dataset.
Figure 8. The confusion matrix of our model on the Houston University 2013 dataset.
Remotesensing 12 03967 g008
Figure 9. Classification maps of the Houston University 2013 dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) ours. The yellow and red rectangles correspond to building areas and the parking lot.
Figure 9. Classification maps of the Houston University 2013 dataset with the compared methods: (a) RGB image; (b) ground truth; (c) FDDL; (d) DPL; (e) ResNet; (f) RNN; (g) CNN; (h) ours. The yellow and red rectangles correspond to building areas and the parking lot.
Remotesensing 12 03967 g009
Table 1. Scene categories of the Center of Pavia dataset with the number of training and testing samples shown for each class.
Table 1. Scene categories of the Center of Pavia dataset with the number of training and testing samples shown for each class.
Class No.Class NameTrainingTest
1Water652758,751
2Trees6505858
3Asphalt2902615
4Self-Blocking Bricks2141926
5Bitumen6545895
6Tiles7586827
7Shadows7286559
8Meadows3122810
9Bare Soil2161949
Table 2. Scene categories of the Botswana dataset with the number of training and testing samples shown for each class.
Table 2. Scene categories of the Botswana dataset with the number of training and testing samples shown for each class.
Class No.Class NameTrainingTest
1Water27243
2Hippo grass1091
3Floodplain grasses 125226
4Floodplain grassed 221194
5Reeds26243
6Riparian26243
7Fire scar25234
8Island interior20183
9Acacia woodlands31283
10Acacia shrublands24224
11Acacia grasslands30275
12Short mopane18163
13Mixed mopane26242
14Exposed soils986
Table 3. Scene categories of the Houston University 2013 dataset with the number of training and testing samples shown for each class.
Table 3. Scene categories of the Houston University 2013 dataset with the number of training and testing samples shown for each class.
Class No.Class NameTrainingTest
1Healthy grass1251126
2Stressed grass1251129
3Synthetic grass69628
4Tree1241120
5Soil1241118
6Water32293
7Residential1261142
8Commercial1241120
9Road1251127
10Highway1221105
11Railway1231112
12Parking Lot 11231110
13Parking Lot 246423
14Tennis court42386
15Running track66594
Table 4. Classification accuracy for the Center of Pavia dataset. FDDL, Fisher discriminant dictionary learning.
Table 4. Classification accuracy for the Center of Pavia dataset. FDDL, Fisher discriminant dictionary learning.
Class No.SVMFDDLDPLResNetRNNCNNOurs
10.98660.98820.98560.98450.98360.99660.9998
20.63020.23190.37430.66410.41180.74960.9507
30.97080.98510.96820.96440.99020.96690.9667
40.50550.37600.25680.48770.46460.52560.8728
50.99690.98480.97290.98350.99240.99050.9732
60.66590.69440.85760.70350.83350.93310.9534
70.91630.88110.91430.93630.94650.95030.9547
80.94160.95950.97110.95040.97940.99040.9922
90.99650.96430.98250.98950.99300.98740.9616
OA0.92340.90570.92440.92890.93310.96630.9839
AA0.84560.78500.80930.85150.84390.89890.9583
kappa0.89270.86770.89370.90040.90600.95240.9723
Table 5. Classification accuracy for the Botswana dataset.
Table 5. Classification accuracy for the Botswana dataset.
Class No.SVMFDDLDPLResNetRNNCNNOurs
10.94650.97120.97940.98350.93460.94920.9342
21.00000.85710.93410.98900.91890.83330.8132
30.84510.79200.84960.82740.83660.92640.9735
40.89180.78870.91750.89180.78460.93230.9433
50.70370.68310.72840.75720.77040.82190.8724
60.68310.64610.63790.62140.62500.78610.8025
70.96150.74790.93160.90170.92340.96070.9573
80.88520.91260.98360.97810.82140.90050.9781
90.72790.70320.67840.77390.76510.76510.9435
100.73210.47770.83480.85270.77040.80710.9688
110.74180.75640.89450.88360.84040.85170.8691
120.90800.80370.88340.98160.77460.85800.8221
130.57850.78100.85540.73970.73710.89660.9256
140.90700.66280.79070.79070.74040.89010.9302
OA0.80170.75150.84200.84440.80170.86760.9130
AA0.82230.75600.85000.85520.80310.86990.9095
kappa0.78540.73110.82890.83160.78500.85660.9057
Table 6. Classification accuracy for the Houston University 2013 dataset.
Table 6. Classification accuracy for the Houston University 2013 dataset.
Class No.SVMFDDLDPLResNetRNNCNNOur
10.8890.90760.98310.93870.95380.92240.9645
20.93530.94770.98140.97520.96280.98240.9779
30.95860.99840.98250.99040.98570.98880.9809
40.88750.94460.86340.95980.97140.94350.9652
50.92840.97760.99020.97230.97850.96630.9821
60.87030.98290.96930.95900.92490.96910.8635
70.62610.78810.69960.79770.78200.85670.8862
80.7250.51880.65710.56340.42230.79450.8429
90.5510.65570.73290.70630.70450.72690.7995
100.63890.42440.84620.77470.77380.78080.7086
110.51170.43170.59260.77520.83540.78890.7707
120.53960.53150.65950.60360.74500.73480.7550
130.27660.54140.28840.64300.57450.48790.5012
140.96890.99480.98960.98960.97930.99080.9870
150.95450.98820.98480.95620.97810.93510.9815
OA0.74090.74760.81030.82550.82800.85490.8682
AA0.75080.77560.81470.84040.83810.85790.8644
kappa0.71990.72710.79490.81140.81420.84310.8574
Table 7. The classification results with 5% of the Botswana dataset for training the models.
Table 7. The classification results with 5% of the Botswana dataset for training the models.
Class No.SVMFDDLDPLResNetRNNCNNOurs
10.96890.98050.87551.00000.63140.93121.0000
20.98960.79170.98960.98960.23700.79340.9063
30.65270.68620.87450.84520.77620.87790.9791
40.91220.61950.92200.84390.07140.88460.9073
50.50780.60940.70700.78910.76190.83330.8242
60.53910.52340.70700.66410.43560.71940.7500
70.81780.94740.78140.88660.82910.95510.9393
80.90160.78240.95850.97930.35910.83960.9482
90.50170.57860.78930.75250.66810.75230.8528
100.60170.72030.91100.77120.81250.70790.7839
110.81720.62760.72760.87930.86710.75950.9483
120.72090.55230.60470.93600.44090.76610.9419
130.56470.73330.94900.57650.77880.87180.7490
140.82420.92310.81320.81320.22220.79380.9231
OA0.71250.70670.82150.82500.61220.81920.8842
AA0.73550.71880.82930.83680.56370.82040.8895
kappa0.68890.68270.80670.81070.58150.80420.8746
Table 8. Training and testing time of the HSI classification algorithms on the three datasets.
Table 8. Training and testing time of the HSI classification algorithms on the three datasets.
DatasetTime (s)SVMCNNRNNOurs
CodingELM
Pavia ofTraining286.14404.16800.680.431.48
CenterTesting6.788.219.334.50 × 10 4 0.63
BotswanaTraining51.4470.03296.240.030.05
Testing1.771.93.64 2.90 × 10 4 0.13
HoustonTraining62.50106.04256.010.150.25
University 2013Testing2.112.663.12 3.20 × 10 5 0.17
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Z.; Zhao, B.; Wang, W. An Efficient Spectral Feature Extraction Framework for Hyperspectral Images. Remote Sens. 2020, 12, 3967. https://doi.org/10.3390/rs12233967

AMA Style

Li Z, Zhao B, Wang W. An Efficient Spectral Feature Extraction Framework for Hyperspectral Images. Remote Sensing. 2020; 12(23):3967. https://doi.org/10.3390/rs12233967

Chicago/Turabian Style

Li, Zhen, Baojun Zhao, and Wenzheng Wang. 2020. "An Efficient Spectral Feature Extraction Framework for Hyperspectral Images" Remote Sensing 12, no. 23: 3967. https://doi.org/10.3390/rs12233967

APA Style

Li, Z., Zhao, B., & Wang, W. (2020). An Efficient Spectral Feature Extraction Framework for Hyperspectral Images. Remote Sensing, 12(23), 3967. https://doi.org/10.3390/rs12233967

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop