Next Article in Journal
Precise Orbit Determination for Climate Applications of GNSS Radio Occultation including Uncertainty Estimation
Previous Article in Journal
European Radiometry Buoy and Infrastructure (EURYBIA): A Contribution to the Design of the European Copernicus Infrastructure for Ocean Colour System Vicarious Calibration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised and Supervised Feature Extraction Methods for Hyperspectral Images Based on Mixtures of Factor Analyzers

by
Bin Zhao
1,
Magnus O. Ulfarsson
1,
Johannes R. Sveinsson
1,* and
Jocelyn Chanussot
1,2
1
Faculty of Electrical and Computer Engineering, University of Iceland, 101 Reykjavik, Iceland
2
Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(7), 1179; https://doi.org/10.3390/rs12071179
Submission received: 24 February 2020 / Revised: 28 March 2020 / Accepted: 31 March 2020 / Published: 7 April 2020
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
This paper proposes three feature extraction (FE) methods based on density estimation for hyperspectral images (HSIs). The methods are a mixture of factor analyzers (MFA), deep MFA (DMFA), and supervised MFA (SMFA). The MFA extends the Gaussian mixture model to allow a low-dimensionality representation of the Gaussians. DMFA is a deep version of MFA and consists of a two-layer MFA, i.e, samples from the posterior distribution at the first layer are input to an MFA model at the second layer. SMFA consists of single-layer MFA and exploits labeled information to extract features of HSI effectively. Based on these three FE methods, the paper also proposes a framework that automatically extracts the most important features for classification from an HSI. The overall accuracy of a classifier is used to automatically choose the optimal number of features and hence performs dimensionality reduction (DR) before HSI classification. The performance of MFA, DMFA, and SMFA FE methods are evaluated and compared to five different types of unsupervised and supervised FE methods by using four real HSIs datasets.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) provide abundant spectral information about a scene [1]. In general, an HSI contains hundreds of spectral bands with high spectral resolution [2,3,4]. Having sufficient spectral information makes it possible to discriminate different materials within a scene by using a classifier [5,6,7,8]. However, the high dimensionality of HSIs makes the processing computationally and memory costly. To achieve an acceptable classification accuracy for an image of high dimensionality many conventional HSI processing require many training samples [9,10,11]. This is known as the Hughes phenomenon or the curse of dimensionality [12]. Thus when we have a limited number of training samples, we have a trade-off between classification accuracy and the number of spectral bands [13,14,15,16,17,18,19,20,21]. Dimensionality reduction (DR) is a very effective way to solve this problem [22,23,24,25,26,27,28,29,30,31,32]. Dimensionality reduced data should be a good representation of the original data. In addition, both the computing time and the number of training samples required will become less when the data dimensionality is lower. Therefore, DR is a very important pre-processing step for HSI classification [33,34,35,36,37,38,39]. In general, DR can be divided into feature selection (FS) and feature extraction (FE). In this paper, we focus on FE. There exist several classical and novel statistical FE methods in the literature that have been used in HSI processing. FE methods are either unsupervised or supervised. Principal component analysis (PCA) [40] is a classical unsupervised FE method. PCA projects the original data onto a lower dimensional linear subspace of the original data space and can also be expressed as the maximum likelihood solution of a probabilistic latent variable model [41]. This reformulation of PCA is called probabilistic principal component analysis (PPCA) [42] and is an example of a linear-Gaussian framework, in which all of the marginal and conditional distributions are assumed to be Gaussian. Factor analysis (FA) [41,43] is also a linear Gaussian latent variable model closely related to PPCA. For FA, the conditional distribution of the observed variables given the latent variable have diagonal rather than an isotropic covariance matrix. In addition to these classical unsupervised FE methods, there are several novel unsupervised FE methods in the literature, such as orthogonal total variation component analysis (OTVCA) [44], edge-preserving filtering [45], Gaussian pyramid based multi-scale feature extraction (MSFE) [46], sparse and smooth low-rank analysis (SSLRA) [47], etc. For supervised FE methods, the new features should contain most discriminative information based on the labeled samples. There exist several supervised FE methods, such as linear discriminant analysis (LDA) [48], nonparametric weighted feature extraction (NWFE) [49], manifold-learning based HSI feature extraction [50], low-rank representation with the ability to preserve the local pairwise constraints information (LRLPC) [51], etc. Supervised methods are usually better than unsupervised methods for HSI classification [52,53,54], since they have access to labeled data. However, the effectiveness depends on how well the labeled dataset represents the whole original dataset.
For both PPCA and FA, all the marginal and conditional distributions of the HSI are assumed to be Gaussian. However, in practice, most HSIs cannot be assumed to obey a Gaussian distribution. To overcome this problem, we propose mixtures of factor analyzers (MFA), deep MFA (DMFA), and supervised MFA (SMFA) FE methods for HSI. We also propose an image segmentation method based on the Gaussian mixture model for MFA, DMFA, and SMFA to solve the problem of a non-normal distribution. MFA assumes a low-dimensionality representation of the Gaussians in the Gaussian mixture model. DMFA consists of a two-layer MFA, which inputs the samples from the posterior distribution at the first layer to an MFA model at the second layer. SMFA is a supervised FE method that uses labeled samples to extract features of HSI. Based on these three FE methods, a framework for HSI classification is also proposed in this paper. While the dimensionality of the desired features needs to be selected by the user in conventional DR methods, the proposed framework automatically determines the dimensionality of features according to classification accuracy without prior supervision by the user. The contribution of the paper are summarized as follows:
  • Two unsupervised FE methods, MFA and DMFA, are proposed for HSI. MFA and DMFA are particularly suitable for DR of HSI with a non-normal distribution and unlabeled samples.
  • A supervised FE method, SMFA, is proposed for HSI. SMFA can be effectively used for DR of HSI with a non-normal distribution and labeled samples.
  • An image segmentation method based on the Gaussian mixture model is proposed for MFA, DMFA, and SMFA to solve the problem of a non-normal distribution.
  • Frameworks for extracting the most useful features for HSI classification based on the MFA, DMFA, and SMFA DR methods are proposed.
The paper is organized as follows. Section 2 briefly describes the three FE methods and a framework which automatically extracts optimal features for HSI classification. Section 3 presents experimental results and analysis of the results. Finally, Section 4 concludes the paper.

2. Proposed FE Methods and Framework

2.1. MFA

Let x denote a D-dimensional spectral vector, z denote a d-dimensional latent vector, and  m { 1 , . . . , M } denote the component indicator variable of the M factor analyzers in MFA. The MFA model can be defined as
p ( m ) = π m , m = 1 M π m = 1 ,
p ( z | m ) = p ( z ) = N ( z ; 0 , I ) ,
p ( x | z , m ) = N ( x ; W m z + μ m , Ψ ) ,
where N ( z ; 0 , I ) means that z is Gaussian vector with zero mean and d × d identity matrix I as the covariance matrix. The parameters of the m-th factor analyzer include a mixing proportion π m , mean μ m , a  D × d factor loading matrix W m , and a D × D diagonal matrix Ψ which represents the independent noise variances for each band.
The parameters z , μ m , W m , and  Ψ of MFA are estimated (trained) by using an expectation maximization (EM) algorithm [55]. An example demonstrating how MFA works is shown in Figure 1a,b. The schematic of the MFA is shown in Figure 2.
The performance of MFA for classification can be improved by increasing the dimensionality d of the latent factors per component of the mixture of factor analyzers or the number M of mixture components. However, for high dimensionality data, this approach quickly leads to overfitting. Below we discuss a cross-validation scheme to select d while avoiding overfitting.

2.2. DMFA

Figure 1c shows a case where the posteriors have non-normal distribution, to solve this problem the DMFA model was proposed. Instead of a simple standard normal prior, the DMFA model uses a more powerful MFA prior:
p ( z | m ) = MFA ( θ m ( 2 ) ) ,
where θ m ( 2 ) is the model parameter in the second layer and it emphasizes that the new MFA’s parameters are at the second layer and specific to component m of the first layer MFA, while holding the first layer parameters fixed. Thus, the DMFA is equivalent to fitting component-specific second layer MFAs with vectors drawn from p ( z , m , | x ; θ m ( 1 ) ) as data, where θ m ( 1 ) is the model parameter in the first layer.
Using p ( k m | m ) = π k m ( 2 ) to denote the second layer mixing proportion of mixture component k m , and  K m denote the total number of factor analyzers in the second layer for specific m of the first layer, so
m : k m = 1 K m π k m ( 2 ) = 1 ,
p D M F A ( z , m ) = p ( m ) p ( z | m ) = p ( m ) p ( k m | m ) p ( z | k m ) .
For convenience, denote all possible second layer mixture components with s = 1 , . . . , S , where  S = m = 1 M K m . The mixing proportions are π s ( 2 ) = p ( m ( s ) ) p ( k m ( s ) | m ( s ) ) , where m ( s ) and k m ( s ) are the first and second layer mixture components m and k m to which s corresponds.
Therefore, the DMFA model is
p ( s ) = π s ( 2 ) ,
p ( z ( 2 ) | s ) = N ( z ( 2 ) ; 0 , I ) ,
p ( z ( 1 ) | z ( 2 ) , s ) = N ( z ( 1 ) ; W s ( 2 ) z ( 2 ) + μ s ( 2 ) , Ψ ( 2 ) ) ,
m m ( s ) , ( deterministic ) ,
p ( x | z ( 1 ) , m ) = N ( x ; W m ( 1 ) z ( 1 ) + μ m ( 1 ) , Ψ ( 1 ) ) ,
where (10) is fully deterministic as each s belongs to one and only one m. z ( 1 ) R d ( 1 ) , z ( 2 ) R d ( 2 ) , W m ( 1 ) R D × d ( 1 ) , W s ( 2 ) R d ( 1 ) × d ( 2 ) , μ m ( 1 ) R d ( 1 ) , μ s ( 2 ) R d ( 2 ) , Ψ ( 1 ) and Ψ ( 2 ) are D × D and d ( 1 ) × d ( 1 ) diagonal matrices of the first and second layers, respectively.
For the DMFA algorithm, the same scheme can be extended to training third-layer MFAs, but in this paper, we only consider the two-layer DMFA model.
The DMFA model can be trained by using a greedy layers-wise algorithm. The first layer of DMFA is trained as described above in Section 2.1, when training the second layer of DMFA, freezing the first layer parameters and treating the sampled first layer factor values for each mixture component { z n ( 1 ) } m as training data for the second layer of DMFA. The DMFA model is summarized in Algorithm 1, and an illustration of the DMFA are shown in Figure 1c and Figure 3, respectively.
Algorithm 1 DMFA algorithm
  • Step 1: Input HSI X = { x 1 , x 2 , . . . , x N } , the maximum number of EM iteration (default=1000).   
  • Step 2: Train the first layer of DMFA on X with M mixture components and d ( 1 ) dimensional latent factors using the EM algorithm.  
  • Step 3: Use the first layer latent factor dataset Y m = { z n ( 1 ) } m for each of the M mixture components as training data for the second layer of DMFA.  
  • Step 4: Train the second layer of DMFA on Y m with d ( 2 ) dimensional latent factors and K m mixture components using the EM algorithm.  
  • Step 5: Output DR results Z = { z 1 ( 2 ) , z 2 ( 2 ) , . . . , z N ( 2 ) } .

2.3. SMFA

SMFA is a supervised FE method, let y denote an output value (label) for each D-dimensional labeled spectral vector x . The SMFA model can be defined as
p ( x | z , m ) = N ( x ; W m z + μ m , Ψ ) ,
p ( y | z , m ) = N ( y ; W y m z + μ y m , Ψ y ) ,
where the parameters of the m-th factor analyzer include mean μ m , a  D × d factor loading matrix W m , and a D × D diagonal matrix Ψ which represents the independent noise variances for each band. W y m , μ y m , and  Ψ y are similar defined.
The parameters z , μ m , W m , Ψ , W y m , μ y m , and  Ψ y of SMFA are also estimated by using the EM algorithm [55]. The schematic of the SMFA is shown in Figure 4.

2.4. Framework

Traditionally, in DR, the dimensionality of desired features has to be initialized by the user. In this paper, we propose a framework that automatically selects the optimal dimensionality of desired features for HSI. We use the classification accuracy of a classifier on validation samples to automatically determine the dimensionality of the features. Different classifiers such as maximum likelihood (ML), support vector machine (SVM), and random forest (RF), can be used in this framework. The framework based on MFA, SMFA, and DMFA for HSI classification are summarized in Algorithms 2 and 3, respectively.
Algorithm 2 Framework based on MFA and SMFA
  • Step 1: Input HSI X , training samples;   
  • Step 2: Automatically select the optimal number of features d and mixture components M:
             for M = 2 : M c
                   for d = 3 : D 2
                         Run MFA (or SMFA);
                         Five-fold cross-validation of SVM (ML, or RF) on training samples;                      Save the cross-validation (CV) score CV M , d ;
                   end
             end
             Return M ^ and d ^ corresponding to the best CV;
  • Step 3: Run MFA (or SMFA) with M ^ and d ^ ;  
  • Step 4: Run SVM (ML or RF) classification;  
  • Step 5: Output Classification results.
Algorithm 3 Framework based on DMFA
  • Step 1: Input HSI X , training samples;   
  • Step 2: Automatically select the optimal number of features in the first layer d ( 1 ) , in the second layer d ( 2 ) , mixture components in the first layer M, and in the second layer K m :
             for M = 2 : M c
                   for d 1 = 3 : D 2
                         for K m = 2 : M
                               for d 2 = 3 : d 1
                                     Run DMFA;
                                     Five-fold cross-validation of SVM (ML, or RF) on training samples;
                                     Save the cross-validation accuracy (CVA) CVA M , d 1 , K m , d 2 ;
                               end
                         end
                   end
             end
             Return M ^ , d 1 ^ , K m ^ , and  d 2 ^ according to the best CVA;
  • Step 3: Run DMFA with M = M ^ , d ( 1 ) = d 1 ^ , K m = K m ^ , and  d ( 2 ) = d 2 ^ ;  
  • Step 4: Run SVM (ML or RF) classification;  
  • Step 5: Output Classification results.

3. Experiments and Results

The experiments were done using ML, RF, and SVM classifiers, but since ML and RF gave inferior or slightly inferior results compared to SVM results, only the results of the SVM classifier are reported.

3.1. Experimental Datasets

The Indian Pines dataset was collected by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor at Indian Pines. The image contains 145 × 145 pixels with a spatial resolution of 20 m and 220 spectral bands from 400 nm to 2500 nm. In the experiments, noisy bands and atmospheric vapor absorption bands are excluded leaving 200 spectral bands. Figure 5 shows the false-color composite of the Indian Pines image and the corresponding ground reference map, respectively. The nine largest classes are considered for classification [24,56]. For the Indian Pines, the University of Pavia, and the Salinas datasets, 10% of the labeled samples for each class are randomly selected as training samples, and the remaining 90% are used as the test set, respectively. Table 1, Table 2 and Table 3 provide information on the number of training and test samples for each class of interest, respectively.
The Houston dataset was provided by the IEEE Geoscience and Remote Sensing Society (GRSS) for the Data Fusion Contest in 2013. This image is of the University of Houston campus and the neighboring urban area. The dataset has 349 × 1905 pixels with a spatial resolution of 2.5 m and 114 spectral bands coverage ranging from 380 nm to 1050 nm. This HSI contains fifteen classes of interest. Figure 6 shows the false-color composite of the Houston image and the corresponding ground reference map, respectively. The training and test samples were given according to the IEEE GRSS Data Fusion Contest in 2013. The spatial positions and the number of training and test samples for each class of interest is fixed by the IEEE GRSS Data Fusion Contest. Table 4 provides information on the number of training and test samples for each class of interest. It is important to note that the standard sets of training and test samples were used for the dataset to make the results entirely comparable with most of the methods available in the literature.
The University of Pavia dataset was captured by the Reflective Optics System Imaging Spectrometer sensor over the city of Pavia, Italy. This image has 610 × 340 pixels with a spatial resolution of 1.3 m and 115 spectral bands coverage ranging from 0.43 μ m to 0.86 μ m. In the experiments, this data contains nine classes of interest and has 103 spectral bands after removing 12 noisy bands. Figure 7 shows the false-color composite of the University of Pavia image and the corresponding ground reference map.
The Salinas dataset was acquired by the AVIRIS sensor over Salinas Valley, California. This image has 512 × 217 pixels with a spatial resolution of 3.7 m and 204 spectral bands after removing 20 water absorption bands. This image contains sixteen classes of interest. Figure 8 shows the false-color composite of the Salinas image and the corresponding ground reference map.

3.2. Experimental Setup

An SVM classifier is used to evaluate the performance of the proposed methods. The SVM classifier is a supervised classification method that uses a kernel method to map the data with a non-linear transformation to a higher dimensional space and in that space tries to find a linear separating hyperplane from different classes. In the experiments, for SVM, the LibSVM Toolbox for MATLAB was applied with a radial basis function (RBF) kernel [57,58]. The five-fold cross-validation is used to find the best parameters, i.e., the kernel parameter and regularization parameter, in SVM. The evaluation metrics used are overall accuracy (OA), average accuracy (AA), and Kappa coefficient (KC), as well as standard deviation (STD). To further evaluate the performance of the proposed algorithms, the following statistical and DR methods: PCA, PPCA, FA, LDA, NWFE, MFA, DMFA, and SMFA are used for comparison. Each experiment is run ten times, and the average of these ten experiments is reported.

3.3. Tuning Parameter Estimation and Assessment

For the MFA and SMFA algorithms, we need to estimate the number of mixture components M, and the dimensionality of latent factors d. In the experiments, M { 2 , 3 , . . . , M c } , where M c is the maximal number of classes considered, d { 3 , 4 , . . . , D / 2 } , where D is the input dimensionality of original datasets, we use five-fold cross-validation to obtain the optimal parameters M and d.
The assessment of the effect of the tuning parameters (M and d) on the performance of the proposed methods was of interest. Since we were interested in the classification accuracy, we investigated the effect of the number of mixture components M and the dimensionality of latent factors d on OA. Figure 9a,b shows the 3-dimensional surface of the OA of MFA and SMFA with respect to the values of parameters M and d for the SVM classifier, respectively. It can be seen that the OA gradually increases in the beginning as the d increases, and then keeps stable with slight fluctuation as the d increases, but decreases slightly when the d reaches some value. It can also be observed that the OA is insensitive to M.
For the DMFA algorithm, we need to estimate the number of mixture components M and K m , and the dimensionality of latent factors d ( 1 ) and d ( 2 ) , in the first and second layer of DMFA, respectively. In the experiments, M is the same as in the MFA, K m { 2 , . . . , M } , we set K m = 2 for all m in our experiments. d ( 1 ) is the same as d, d ( 2 ) { 3 , 4 , . . . , d ( 1 ) } . Five-fold cross-validation can be used to obtain the optimal parameters. To analyze the impact of the number of mixture components M and the dimensionality of latent factors d ( 1 ) in the first layer on the performance of DMFA, the output results of the first layer of DMFA were used for HSI classification. Figure 9c shows the OAs with respect to the values of parameters M and d ( 1 ) . Figure 9a,b show similar things for MFA and SMFA, respectively.

3.4. Classification

The first experiment was performed on the Indian Pines dataset. Figure 10 shows the classification maps obtained by different methods. From the figures, it can be seen that the classification maps obtained by PCA, PPCA, FA, LDA, and NWFE are not very satisfactory since they have lots of visible noise. By contrast, MFA, DMFA, and SMFA give much better classification maps, all of them have a smoother appearance and preserve more details on edges. Besides visual comparison, Table 5 presents the quantitative classification results for all the methods, and there it can be observed that the MFA, DMFA, and SMFA achieve much higher classification accuracies than PCA, PPCA, and FA, respectively. These results imply that the performance of DR could be improved by considering the Gaussian mixture model. Moreover, DMFA and SMFA clearly outperform MFA, and the performance of DMFA and SMFA is similar, both of them present the highest OA, AA, and KC and achieves most of the top classification accuracy values for individual classes. This indicates that MFA, DMFA, and SMFA could extract more useful information for classification from a complicated HSI. Moreover, SMFA is better than LDA and NWFE, this means that SMFA is an effective supervised DR method. MFA, DMFA, and SMFA improve the OA by 11.11%, 13.38%, and 13.42% using SVM compared to other methods in the experiment, respectively. It is interesting to note that all DR methods based on FA (SMFA, DMFA, MFA, and FA) gave a better performance than PCA and PPCA. The reason for this is that noise could be distributed inconsistently for different components in real HSI. Table 5 also gives the STDs of classification results for different DR methods for the Indian Pines dataset. It can be seen that all the methods give similar and stable classification results. Table 6 compares the CPU processing time (in seconds) used by different DR methods for the Indian Pines dataset. All methods were implemented in Matlab R2019a on a computer having Intel(R) Core(TM) i7-6700 processor (3.40 GHz), 8.00 GB of memory, and 64-bit Windows 10 Operating System. It can be seen that the running times for MFA, DMFA, and SMFA were 3.85, 18.07, and 0.12 s, respectively. It is worth noting that the running time for the supervised methods (LDA, NWFE, and SMFA) is affected considerably by the number of labeled (training) samples used, and the unsupervised methods are affected by the total size of the dataset.
The second experiment was performed on the Houston dataset. Figure 11 shows the classification maps obtained by different methods. From the figures, we can see that the proposed MFA, DMFA, and SMFA algorithms also outperform the other algorithms. Table 7 presents the quantitative classification results of the different DR methods. As shown in Table 7, the performance of DMFA and SMFA is better than MFA and much better than PCA, PPCA, FA, LDA, and NWFE. MFA, DMFA, and SMFA improve the OA by 4.60%, 6.67%, and 7.35% compared to other methods in the experiment, respectively. Table 7 also presents the STDs of classification results. It can be seen that FA and LDA present the most stable results. MFA, DMFA, and SMFA have slight fluctuation for each experiment and give relatively stable results.
The third and fourth experiments were performed on the University of Pavia and Salinas datasets. It should be noted that the NWFE method does not work for the Salinas dataset. Therefore, in the experiments of the Salinas dataset, there are no experimental results for NWFE. Figure 12 and Figure 13 show the classification maps obtained by different methods on the University of Pavia and Salinas datasets, respectively. From Figure 12 and Figure 13, it can be seen that the classification maps obtained by MFA, DMFA, and SMFA are much better than PCA, PPCA, FA, LDA, and NWFE, respectively. Table 8 and Table 9 present the quantitative classification results. As shown in Table 8 and Table 9, the classification accuracies of the proposed MFA, DMFA, and SMFA methods are much better than PCA, PPCA, and FA methods. These results further demonstrate that, instead of using a single Gaussian distribution model, the performance of DR could be improved by considering the Gaussian mixture model. For the University of Pavia dataset, MFA, DMFA, and SMFA improve the OA by 6.05%, 7.02%, and 7.02% compared to other methods used in the experiment, respectively. For the Salinas dataset, MFA, DMFA, and SMFA improve the OA by 5.11%, 5.49%, and 5.62% compared to other methods used in the experiment, respectively. Moreover, in the experiments, DMFA and SMFA are clearly better than MFA, and DMFA and SMFA give similar and highest OA, AA, and KC and also achieve most of the top classification accuracy values for the individual classes. This also further indicates that MFA, DMFA, and SMFA could extract more effective information for classification from a complicated HSI. Table 8 and Table 9 also give the STDs of classification results for different DR methods for the University of Pavia and Salinas datasets, respectively. It can be seen that all the methods give similar and relatively stable classification results. This also further demonstrates that MFA, DMFA, and SMFA are stable and effective DR methods.

4. Conclusions

In this paper, MFA, DMFA, and SMFA were proposed for feature extraction of HSIs and were then used for classification of them. MFA, DMFA, and SMFA are probabilistic DR methods, instead of assuming that a whole HSI obeys a Gaussian distribution, the methods use a Gaussian mixture model to extract more effective information for DR. The Gaussian mixture model is used for MFA to allow a low-dimensionality representation of the Gaussian. A two-layer MFA, DMFA, utilizes the samples from the posterior at the first layer to an MFA model at the second layer. MFA and DMFA are two unsupervised DR method. The methods are particularly suitable for DR of HSI with a non-normal distribution and unlabeled samples. SMFA is a supervised DR method and uses labeled samples to extract features. SMFA can be effectively used to DR of HSI with a non-normal distribution and labeled samples.
Based on the three DR methods, we also proposed a framework for HSI classification, the overall accuracy of a classifier on validation samples is used to automatically determine the optimal number of features of DR for HSI classification. This framework can automatically extract the most effective feature for HSI classification. To validate the performance of DR, we conduct experiments in terms of SVM classification based on four real HSIs. The experimental results show that MFA, DMFA, and SMFA can give better results than statistical DR comparison methods.
In the future, more validations on other applications, e.g., hyperspectral unmixing, target detection, will be incorporated in our future work.

Author Contributions

B.Z. wrote the manuscript and performed all the experiments. M.O.U., J.R.S. and J.C. revised the manuscript and improved its presentation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Doctoral Grants of the University of Iceland Research Fund and the Icelandic Research Fund under Grant 174075-05.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Harsanyi, J.C.; Chang, C.I. Hyperspectral image classification and dimensionality reduction: An orthogonal subspace projection approach. IEEE Trans. Geosci. Remote Sens. 1994, 32, 779–785. [Google Scholar] [CrossRef] [Green Version]
  2. Jia, X.; Richards, J.A. Segmented principal components transformation for efficient hyperspectral remote-sensing image display and classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 538–542. [Google Scholar]
  3. Camps-Valls, G.; Bruzzone, L. Kernel-based methods for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1351–1362. [Google Scholar] [CrossRef]
  4. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  5. Chang, C.; Du, Q.; Sun, T.; Althouse, M.L. A joint band prioritization and band-decorrelation approach to band selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2631–2641. [Google Scholar] [CrossRef] [Green Version]
  6. Feng, F.; Li, W.; Du, Q.; Zhang, B. Dimensionality reduction of hyperspectral image with graph-based discriminant analysis considering spectral similarity. Remote Sens. 2017, 9, 323. [Google Scholar] [CrossRef] [Green Version]
  7. Xu, X.; Li, J.; Huang, X.; Dalla-Mura, M.; Plaza, A. Multiple morphological component analysis based decomposition for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3083–3102. [Google Scholar] [CrossRef]
  8. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2011, 50, 879–893. [Google Scholar] [CrossRef]
  9. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of machine-learning classification in remote sensing: An applied review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef] [Green Version]
  10. Wieland, M.; Pittore, M. Performance evaluation of machine learning algorithms for urban pattern recognition from multi-spectral satellite images. Remote Sens. 2014, 6, 2912–2939. [Google Scholar] [CrossRef] [Green Version]
  11. Pal, M.; Foody, G.M. Feature selection for classification of hyperspectral data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef] [Green Version]
  12. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef] [Green Version]
  13. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, L.; Zhong, Y.; Huang, B.; Gong, J.; Li, P. Dimensionality reduction based on clonal selection for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2007, 45, 4172–4186. [Google Scholar] [CrossRef]
  15. Luo, F.; Huang, H.; Duan, Y.; Liu, J.; Liao, Y. Local geometric structure feature for dimensionality reduction of hyperspectral imagery. Remote Sens. 2017, 9, 790. [Google Scholar] [CrossRef] [Green Version]
  16. Huang, H.; Yang, M. Dimensionality reduction of hyperspectral images with sparse discriminant embedding. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5160–5169. [Google Scholar] [CrossRef]
  17. Ulfarsson, M.O.; Palsson, F.; Sigurdsson, J.; Sveinsson, J.R. Classification of big data with application to imaging genetics. Proc. IEEE 2016, 104, 2137–2154. [Google Scholar] [CrossRef] [Green Version]
  18. Gormus, E.T.; Canagarajah, N.; Achim, A. Dimensionality reduction of hyperspectral images using empirical mode decompositions and wavelets. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1821–1830. [Google Scholar] [CrossRef]
  19. Esser, E.; Moller, M.; Osher, S.; Sapiro, G.; Xin, J. A convex model for nonnegative matrix factorization and dimensionality reduction on physical space. IEEE Trans. Image Process. 2012, 21, 3239–3252. [Google Scholar] [CrossRef] [Green Version]
  20. Zhao, B.; Gao, L.; Liao, W.; Zhang, B. A new kernel method for hyperspectral image feature extraction. Geo-Spat. Inf. Sci. 2017, 20, 309–318. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, P.; Jiao, L.; Liu, F.; Gou, S.; Zhao, J.; Zhao, Z. Dimensionality reduction of hyperspectral imagery using sparse graph learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 1165–1181. [Google Scholar] [CrossRef]
  22. Plaza, A.; Martinez, P.; Plaza, J.; Perez, R. Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations. IEEE Trans. Geosci. Remote Sens. 2005, 43, 466–479. [Google Scholar] [CrossRef] [Green Version]
  23. Dong, Y.; Du, B.; Zhang, L.; Zhang, L. Dimensionality reduction and classification of hyperspectral images using ensemble discriminative local metric learning. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2509–2524. [Google Scholar] [CrossRef]
  24. Gao, L.; Zhao, B.; Jia, X.; Liao, W.; Zhang, B. Optimized kernel minimum noise fraction transformation for hyperspectral image classification. Remote Sens. 2017, 9, 548. [Google Scholar] [CrossRef] [Green Version]
  25. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–spatial classification of hyperspectral data using loopy belief propagation and active learning. IEEE Trans. Geosci. Remote Sens. 2012, 51, 844–856. [Google Scholar] [CrossRef]
  26. Bruce, L.M.; Koger, C.H.; Li, J. Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2331–2338. [Google Scholar] [CrossRef]
  27. Mojaradi, B.; Abrishami-Moghaddam, H.; Zoej, M.J.V.; Duin, R.P. Dimensionality reduction of hyperspectral data via spectral feature extraction. IEEE Trans. Geosci. Remote Sens. 2009, 47, 2091–2105. [Google Scholar] [CrossRef]
  28. Wu, Z.; Li, Y.; Plaza, A.; Li, J.; Xiao, F.; Wei, Z. Parallel and distributed dimensionality reduction of hyperspectral data on cloud computing architectures. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2270–2278. [Google Scholar] [CrossRef]
  29. Chen, G.; Qian, S. Denoising and dimensionality reduction of hyperspectral imagery using wavelet packets, neighbour shrinking and principal component analysis. Int. J. Remote Sens. 2009, 30, 4889–4895. [Google Scholar] [CrossRef]
  30. Du, H.; Qi, H. An FPGA implementation of parallel ICA for dimensionality reduction in hyperspectral images. In Proceedings of the 2004 IEEE International Geoscience and Remote Sensing Symposium, Anchorage, AK, USA, 20–24 September 2004; IEEE: Hoboken, NJ, USA, 2004; Volume 5, pp. 3257–3260. [Google Scholar]
  31. Feng, Z.; Yang, S.; Wang, S.; Jiao, L. Discriminative spectral–spatial margin-based semisupervised dimensionality reduction of hyperspectral data. IEEE Geosci. Remote Sens. Lett. 2014, 12, 224–228. [Google Scholar] [CrossRef]
  32. Chen, S.; Zhang, D. Semisupervised dimensionality reduction with pairwise constraints for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2010, 8, 369–373. [Google Scholar] [CrossRef]
  33. Kang, X.; Li, S.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification with edge-preserving filtering. IEEE Trans. Geosci. Remote Sens. 2013, 52, 2666–2677. [Google Scholar] [CrossRef]
  34. Li, H.; Xiao, G.; Xia, T.; Tang, Y.Y.; Li, L. Hyperspectral image classification using functional data analysis. IEEE Trans. Cybern. 2013, 44, 1544–1555. [Google Scholar]
  35. Ghamisi, P.; Benediktsson, J.A.; Ulfarsson, M.O. Spectral–spatial classification of hyperspectral images based on hidden Markov random fields. IEEE Trans. Geosci. Remote Sens. 2013, 52, 2565–2574. [Google Scholar] [CrossRef] [Green Version]
  36. Liu, X.; Bo, Y. Object-based crop species classification based on the combination of airborne hyperspectral images and LiDAR data. Remote Sens. 2015, 7, 922–950. [Google Scholar] [CrossRef] [Green Version]
  37. Möckel, T.; Dalmayne, J.; Prentice, H.; Eklundh, L.; Purschke, O.; Schmidtlein, S.; Hall, K. Classification of grassland successional stages using airborne hyperspectral imagery. Remote Sens. 2014, 6, 7732–7761. [Google Scholar]
  38. Pan, L.; Li, H.; Deng, Y.; Zhang, F.; Chen, X.; Du, Q. Hyperspectral dimensionality reduction by tensor sparse and low-rank graph-based discriminant analysis. Remote Sens. 2017, 9, 452. [Google Scholar] [CrossRef] [Green Version]
  39. Licciardi, G.; Chanussot, J. Spectral transformation based on nonlinear principal component analysis for dimensionality reduction of hyperspectral images. Eur. J. Remote Sens. 2018, 51, 375–390. [Google Scholar] [CrossRef] [Green Version]
  40. Roger, R.E. Principal components transform with simple, automatic noise adjustment. Int. J. Remote Sens. 1996, 17, 2719–2727. [Google Scholar] [CrossRef]
  41. Lawley, D.N. A modified method of estimation in factor analysis and some large sample results. Upps. Symp. Psychol. Factor Anal. 1953, 17, 35–42. [Google Scholar]
  42. Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 1999, 61, 611–622. [Google Scholar] [CrossRef]
  43. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
  44. Rasti, B.; Ulfarsson, M.O.; Sveinsson, J.R. Hyperspectral feature extraction using total variation component analysis. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6976–6985. [Google Scholar] [CrossRef]
  45. Zhang, Y.; Kang, X.; Li, S.; Duan, P.; Benediktsson, J.A. Feature extraction from hyperspectral images using learned edge structures. Remote Sens. Lett. 2019, 10, 244–253. [Google Scholar] [CrossRef]
  46. Tu, B.; Li, N.; Fang, L.; He, D.; Ghamisi, P. Hyperspectral Image Classification with Multi-Scale Feature Extraction. Remote Sens. 2019, 11, 534. [Google Scholar] [CrossRef] [Green Version]
  47. Rasti, B.; Ghamisi, P.; Ulfarsson, M.O. Hyperspectral Feature Extraction Using Sparse and Smooth Low-Rank Analysis. Remote Sens. 2019, 11, 121. [Google Scholar] [CrossRef] [Green Version]
  48. Li, M.; Yuan, B. 2D-LDA: A statistical linear discriminant analysis for image matrix. Pattern Recognit. Lett. 2005, 26, 527–532. [Google Scholar] [CrossRef]
  49. Kuo, B.C.; Landgrebe, D.A. Nonparametric weighted feature extraction for classification. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1096–1105. [Google Scholar]
  50. Zhang, P.; He, H.; Gao, L. A nonlinear and explicit framework of supervised manifold-feature extraction for hyperspectral image classification. Neurocomputing 2019, 337, 315–324. [Google Scholar] [CrossRef]
  51. Ahmadi, S.A.; Mehrshad, N.; Razavi, S.M. Supervised feature extraction method based on low-rank representation with preserving local pairwise constraints for hyperspectral images. Signal Image Video Process. 2019, 13, 583–590. [Google Scholar] [CrossRef]
  52. Zhang, X.; He, Y.; Zhou, N.; Zheng, Y. Semisupervised dimensionality reduction of hyperspectral images via local scaling cut criterion. IEEE Geosci. Remote Sens. Lett. 2013, 10, 1547–1551. [Google Scholar] [CrossRef]
  53. Yang, S.; Jin, P.; Li, B.; Yang, L.; Xu, W.; Jiao, L. Semisupervised dual-geometric subspace projection for dimensionality reduction of hyperspectral image data. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3587–3593. [Google Scholar] [CrossRef]
  54. Wu, H.; Prasad, S. Semi-supervised dimensionality reduction of hyperspectral imagery using pseudo-labels. Pattern Recognit. 2018, 74, 212–224. [Google Scholar] [CrossRef]
  55. Moon, T.K. The expectation-maximization algorithm. IEEE Signal Process. Mag. 1996, 13, 47–60. [Google Scholar] [CrossRef]
  56. Richards, J.A.; Jia, X. Remote Sensing Digital Image Analysis; Springer: New York, NY, USA, 1999; Volume 3. [Google Scholar]
  57. Gualtieri, J.A.; Cromp, R.F. Support vector machines for hyperspectral remote sensing classification. In Proceedings of the 27th AIPR Workshop: Advances in Computer-Assisted Recognition, Washington, DC, USA, 14–16 October 1998; International Society for Optics and Photonics: Bellingham, WA, USA, 1999; Volume 3584, pp. 221–232. [Google Scholar]
  58. Chang, C.; Lin, C. LIBSVM: A Library for Support Vector Machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
Figure 1. ( a ) A scatterplot of HSI samples over two spectral bands. ( b ) Illustration of the MFA model with each ellipse representing a Gaussian component. MFA has three components colored red ( m = 1 ), green ( m = 2 ) and blue ( m = 3 ). Their mixing proportions are given by π m . ( c ) Illustration of DMFA model with each ellipse representing a Gaussian component. The number of components and mixing proportions of the first layer of DMFA are the same as MFA. For the red component, we further learn a second layer of DMFA with three components. For the green and blue components, both of them are learned a second layer of DMFA with two components, respectively. We also introduce the second layer component indicator variable k m = 1 , 2 , . . . , K m and mixing proportions π m ( 2 ) , where K m is the total number of the second layer components associated with the first layer component m. K m is specific to the first layer component and need not be the same for all m. In this example, K 1 = 3 , K 2 = 2 and K 3 = 2 .
Figure 1. ( a ) A scatterplot of HSI samples over two spectral bands. ( b ) Illustration of the MFA model with each ellipse representing a Gaussian component. MFA has three components colored red ( m = 1 ), green ( m = 2 ) and blue ( m = 3 ). Their mixing proportions are given by π m . ( c ) Illustration of DMFA model with each ellipse representing a Gaussian component. The number of components and mixing proportions of the first layer of DMFA are the same as MFA. For the red component, we further learn a second layer of DMFA with three components. For the green and blue components, both of them are learned a second layer of DMFA with two components, respectively. We also introduce the second layer component indicator variable k m = 1 , 2 , . . . , K m and mixing proportions π m ( 2 ) , where K m is the total number of the second layer components associated with the first layer component m. K m is specific to the first layer component and need not be the same for all m. In this example, K 1 = 3 , K 2 = 2 and K 3 = 2 .
Remotesensing 12 01179 g001
Figure 2. The schematic of the MFA corresponding to Figure 1b.
Figure 2. The schematic of the MFA corresponding to Figure 1b.
Remotesensing 12 01179 g002
Figure 3. The schematic of the DMFA corresponding to Figure 1 c .
Figure 3. The schematic of the DMFA corresponding to Figure 1 c .
Remotesensing 12 01179 g003
Figure 4. The schematic of the SMFA corresponding to Figure 1b.
Figure 4. The schematic of the SMFA corresponding to Figure 1b.
Remotesensing 12 01179 g004
Figure 5. Indian Pines dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map containing nine land-cover classes.
Figure 5. Indian Pines dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map containing nine land-cover classes.
Remotesensing 12 01179 g005
Figure 6. Houston dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Figure 6. Houston dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Remotesensing 12 01179 g006
Figure 7. University of Pavia dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Figure 7. University of Pavia dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Remotesensing 12 01179 g007
Figure 8. Salinas dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Figure 8. Salinas dataset. ( a ) Three-band false-color image. ( b ) Ground truth-map reference.
Remotesensing 12 01179 g008
Figure 9. OAs versus the reduced dimensionality d and the number of mixture components M in the proposed methods with SVM classifier on the Indian Pines dataset. ( a ) MFA ( b ) SMFA, and ( c ) DMFA in the first layer.
Figure 9. OAs versus the reduced dimensionality d and the number of mixture components M in the proposed methods with SVM classifier on the Indian Pines dataset. ( a ) MFA ( b ) SMFA, and ( c ) DMFA in the first layer.
Remotesensing 12 01179 g009
Figure 10. Classification maps for the Indian Pines dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Figure 10. Classification maps for the Indian Pines dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Remotesensing 12 01179 g010
Figure 11. Classification maps for the Houston dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Figure 11. Classification maps for the Houston dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Remotesensing 12 01179 g011
Figure 12. Classification maps for the University of Pavia dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Figure 12. Classification maps for the University of Pavia dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) NWFE, ( f ) MFA, ( g ) DMFA, and ( h ) SMFA DR methods.
Remotesensing 12 01179 g012
Figure 13. Classification maps for the Salinas dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) MFA, ( f ) DMFA, and ( g ) SMFA DR methods.
Figure 13. Classification maps for the Salinas dataset obtained by SVM classification after using ( a ) PCA, ( b ) PPCA, ( c ) FA, ( d ) LDA, ( e ) MFA, ( f ) DMFA, and ( g ) SMFA DR methods.
Remotesensing 12 01179 g013
Table 1. Indian Pines HSI: Number of training and test samples.
Table 1. Indian Pines HSI: Number of training and test samples.
Class Number Class Name Training Samples Test Samples
1Corn-no till1431291
2Corn-min till83751
3Grass/Pasture50447
4Grass/Trees75672
5Hay-windrowed49440
6Soybean-no till97871
7Soybean-min till2472221
8Soybean-clean till61553
9Woods1291165
Total9348411
Table 2. University of Pavia dataset: Number of training and test samples.
Table 2. University of Pavia dataset: Number of training and test samples.
Class Number Class Name Training Samples Test Samples
1Asphalt6635968
2Meadows186516,784
3Gravel2101889
4Trees3062758
5Painted metal sheets1351210
6Bare Soil5034526
7Bitumen1331197
8Self-Blocking Bricks3683314
9Shadows95852
Total427838,498
Table 3. Salinas dataset: Number of training and test samples.
Table 3. Salinas dataset: Number of training and test samples.
Class Number Class Name Training Samples Test Samples
1Brocoli_green_weeds_12011808
2Brocoli_green_weeds_23733353
3Fallow1981778
4Fallow_rough_plow1391255
5Fallow_smooth2682410
6Stubble3963563
7Celery3583221
8Grapes_untrained112710,144
9Soil_vinyard_develop6205583
10Corn_senesced_green_weeds3282950
11Lettuce_romaine_4wk107961
12Lettuce_romaine_5wk1931734
13Lettuce_romaine_6wk92824
14Lettuce_romaine_7wk107963
15Vinyard_untrained7276541
16Vinyard_vertical_trellis1811626
Total541548,714
Table 4. Houston dataset: Number of training and test samples.
Table 4. Houston dataset: Number of training and test samples.
Class Number Class Name Training Samples Test Samples
1Healthy grass1981053
2Stressed grass1901064
3Synthetic grass192505
4Trees1881056
5Soil1861056
6Water182143
7Residential1961072
8Commercial1911053
9Road1931059
10Highway1911036
11Railway1811054
12Parking Lot 11921041
13Parking Lot 2184285
14Tennis Court181247
15Running Track187473
Total283212,197
Table 5. The classification results (%) of different DR methods on the Indian Pines dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs. The best classification results are given in bold typeface.
Table 5. The classification results (%) of different DR methods on the Indian Pines dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs. The best classification results are given in bold typeface.
CN PCA PPCA FA LDA NWFE MFA DMFA SMFA
175.13 ± 2.3981.22 ± 1.3986.48 ± 1.9389.31 ± 0.9884.51 ± 2.2593.00 ± 1.8797.49 ± 1.5196.34 ± 1.32
283.33 ± 2.1684.00 ± 4.2189.30 ± 3.4972.44 ± 3.9782.82 ± 2.1691.71 ± 1.0599.17 ± 0.8995.61 ± 1.04
394.51 ± 1.6294.77 ± 2.2797.48 ± 0.8992.84 ± 2.6291.50 ± 1.7594.42 ± 1.0897.35 ± 1.6695.08 ± 1.51
496.09 ± 1.0295.95 ± 0.9997.65 ± 0.9897.32 ± 1.6298.51 ± 1.4998.53 ± 0.4098.39 ± 0.4798.81 ± 1.08
599.55 ± 0.1499.77 ± 0.1899.84 ± 0.1699.55 ± 0.2399.32 ± 0.1699.83 ± 0.1899.86 ± 0.1499.84 ± 0.18
675.03 ± 1.9982.06 ± 2.2782.50 ± 2.1480.37 ± 2.8982.43 ± 3.0090.53 ± 0.6895.48 ± 1.4097.36 ± 0.62
778.22 ± 1.8577.62 ± 1.4588.40 ± 1.7082.85 ± 2.0588.74 ± 2.1195.56 ± 1.5896.91 ± 1.2298.24 ± 1.42
883.54 ± 2.6883.00 ± 2.3489.21 ± 2.6877.76 ± 2.7479.57 ± 3.5696.75 ± 1.3498.74 ± 1.0399.28 ± 1.06
999.40 ± 0.3299.83 ± 0.4199.66 ± 0.2699.57 ± 0.3699.49 ± 0.5599.83 ± 0.3099.74 ± 0.5499.91 ± 0.46
AA87.20 ± 0.3588.69 ± 0.7392.30 ± 0.6788.00 ± 0.6489.65 ± 0.8895.59 ± 1.0198.14 ± 0.8697.85 ± 0.51
OA84.48 ± 0.3585.98 ± 0.3990.96 ± 0.6087.20 ± 0.5689.28 ± 0.8295.59 ± 0.8697.86 ± 0.7297.90 ± 0.60
KC0.8173 ± 0.00400.8345 ± 0.00480.8940 ± 0.00720.8495 ± 0.00650.8738 ± 0.00960.9459 ± 0.01020.9749 ± 0.00850.9753 ± 0.0070
Table 6. CPU processing times in seconds by different DR methods applied to the Indian Pines (INPS), Houston (HSN), University of Pavia (UPA), and Salinas (SAS) datasets (the number of features = 20).
Table 6. CPU processing times in seconds by different DR methods applied to the Indian Pines (INPS), Houston (HSN), University of Pavia (UPA), and Salinas (SAS) datasets (the number of features = 20).
DatasetsPCAPPCAFALDANWFEMFADMFASMFA
INPS0.1163.6670.890.121.053.8518.070.12
HSN2.911335.0921.450.895.02103.77151.914.54
UPA0.62285.014.010.210.9128.1079.301.55
SAS0.56356.4422.150.3818.9787.851.62
Table 7. The classification results (%) of different DR methods on the Houston dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
Table 7. The classification results (%) of different DR methods on the Houston dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
CN PCA PPCA FA LDA NWFE MFA DMFA SMFA
198.0596.20 ± 0.2996.3780.5380.72 ± 0.4598.98 ± 0.8999.31 ± 1.6679.49 ± 1.37
296.0196.36 ± 0.1395.9780.9282.99 ± 0.1895.33 ± 1.6395.24 ± 1.5883.46 ± 1.57
310099.97 ± 0.2710010099.42 ± 0.5899.76 ± 0.2499.82 ± 0.1999.48 ± 0.34
498.0597.35 ± 1.3697.7292.9089.49 ± 0.9499.89 ± 0.3299.47 ± 1.7889.30 ± 2.09
597.4197.45 ± 0.1697.1196.9798.86 ± 0.3898.42 ± 0.8298.11 ± 1.0999.96 ± 0.11
699.9495.07 ± 0.4695.3093.7195.11 ± 0.2147.24 ± 0.8099.97 ± 0.2792.31 ± 0.35
781.3482.05 ± 0.3187.9184.8079.76 ± 2.4466.22 ± 1.3889.53 ± 3.0282.37 ± 1.23
881.4568.87 ± 0.1784.1871.1372.46 ± 4.7384.06 ± 1.1481.82 ± 2.6490.03 ± 0.11
983.4078.34 ± 0.6078.0872.5275.54 ± 1.1987.65 ± 1.0491.66 ± 0.6683.95 ± 1.08
1070.0383.58 ± 0.4865.3273.1781.27 ± 3.2186.21 ± 0.2466.67 ± 4.7697.97 ± 1.59
1167.8669.57 ± 1.8657.0485.3993.55 ± 2.1068.21 ± 0.1674.51 ± 0.5988.99 ± 4.01
1285.5089.15 ± 5.1282.6882.5285.11 ± 1.8695.80 ± 2.2894.07 ± 2.6496.25 ± 1.19
1330.6930.06 ± 1.3343.8570.5370.88 ± 1.7492.52 ± 3.2689.56 ± 2.3165.26 ± 0.89
1499.1898.39 ± 0.2489.0599.1999.60 ± 0.3194.64 ± 0.8397.24 ± 1.1399.96 ± 0.22
1510099.41 ± 0.5910095.9898.73 ± 0.5099.54 ± 0.5499.88 ± 0.1491.97 ± 1.61
AA85.9385.50 ± 0.5084.7085.3586.94 ± 0.4787.68 ± 0.3691.81 ± 0.3989.42 ± 0.69
OA83.4583.75 ± 0.5982.0583.5985.35 ± 0.5686.65 ± 0.4288.72 ± 0.4989.40 ± 0.67
KC0.82070.8240 ± 0.00650.80530.82230.8412 ± 0.00590.8552 ± 0.00460.8775 ± 0.00530.8848 ± 0.0075
Table 8. The classification results (%) of different DR methods on the University of Pavia dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
Table 8. The classification results (%) of different DR methods on the University of Pavia dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
CN PCA PPCA FA LDA NWFE MFA DMFA SMFA
192.00 ± 1.0193.16 ± 0.8294.65 ± 0.5893.20 ± 0.5594.10 ± 0.4998.07 ± 0.4598.39 ± 0.6698.69 ± 0.27
295.71 ± 0.5396.12 ± 0.2596.43 ± 0.2496.86 ± 0.4197.52 ± 0.3899.68 ± 0.1099.02 ± 0.4299.79 ± 0.08
378.98 ± 2.0682.23 ± 1.8680.16 ± 1.7276.13 ± 2.8174.01 ± 4.0090.15 ± 2.2998.39 ± 1.4494.28 ± 1.93
494.24 ± 0.9594.61 ± 0.8293.86 ± 0.7892.57 ± 0.9089.63 ± 1.2695.65 ± 0.3698.72 ± 0.9596.27 ± 0.61
599.41 ± 2.2798.99 ± 2.0199.21 ± 1.2399.17 ± 0.1699.59 ± 0.1399.89 ± 0.1299.93± 0.3599.81 ± 0.09
690.96 ± 2.0191.80 ± 0.7593.54 ± 1.0981.11 ± 1.0690.39 ± 1.7298.63 ± 0.3099.80 ± 0.8899.98 ± 0.30
785.40 ± 4.3488.69 ± 1.9890.08 ± 2.1980.95 ± 1.3583.12 ± 4.1794.32 ± 0.7699.75 ± 1.0699.92 ± 0.68
877.35 ± 2.1483.85 ± 0.9784.40 ± 1.6186.33 ± 2.0384.91 ± 2.3094.63 ± 0.8197.15 ± 1.1297.31 ± 1.54
991.55 ± 1.7899.89± 0.5899.31± 0.8799.65 ± 0.1899.88 ± 0.1197.07 ± 0.9399.64 ± 1.0797.77 ± 0.58
AA89.57 ± 0.5392.27 ± 0.3992.57 ± 0.4889.55 ± 0.1990.35 ± 0.6696.47 ± 0.3098.98 ± 0.4998.22 ± 0.25
OA91.78 ± 0.2993.32 ± 0.2793.80 ± 0.2991.85 ± 0.1893.02 ± 0.3297.90 ± 0.1498.87 ± 0.4198.87 ± 0.14
KC0.8908 ± 0.00380.9112 ± 0.00360.9177 ± 0.00380.8314 ± 0.00230.9071 ± 0.00430.9722 ± 0.00190.9850 ± 0.00550.9850 ± 0.0019
Table 9. The classification results (%) of different DR methods on the Salinas dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
Table 9. The classification results (%) of different DR methods on the Salinas dataset, the best results are in bold typeface. The row of each class number (CN) is the mean accuracy ± standard deviation based on ten runs.
CN PCA PPCA FA LDA MFA DMFA SMFA
199.91± 0.1799.32± 0.0299.29± 0.3899.95 ± 0.0599.94± 0.0999.82 ± 0.0899.89± 0.18
299.88 ± 0.2399.97 ± 0.0999.82 ± 0.1499.79 ± 0.0399.79 ± 0.1299.99± 0.2199.99± 0.12
398.88 ± 0.3699.04 ± 0.2999.21 ± 0.1699.78 ± 0.1299.83 ± 0.0799.72 ± 0.2199.96± 0.22
499.13 ± 0.4899.76 ± 0.4299.76 ± 0.3099.12 ± 0.5398.96 ± 0.5097.93 ± 0.4698.41 ± 0.18
599.05 ± 0.6498.19 ± 0.5699.13 ± 0.1598.96 ± 0.3299.50 ± 0.3499.92 ± 0.4299.17 ± 0.20
699.94 ± 0.2099.92± 0.0999.03± 0.0699.83 ± 0.0899.23 ± 0.2099.98± 0.1499.97 ± 0.05
799.94 ± 0.1999.89± 0.1499.95± 0.2099.94 ± 0.0999.81 ± 0.1099.91 ± 0.1399.91 ± 0.12
888.88 ± 0.9287.55 ± 1.0184.53 ± 1.1290.09 ± 1.2997.73 ± 0.9199.25 ± 0.0499.71 ± 0.60
999.44 ± 0.1699.38 ± 0.1499.34 ± 0.1599.95 ± 0.2099.91 ± 0.0799.93 ± 0.2999.99± 0.03
1096.27 ± 0.5398.19 ± 0.7297.80 ± 0.8298.61 ± 0.4598.99 ± 0.4499.49 ± 0.8399.76 ± 0.87
1199.03 ± 0.8099.04 ± 0.7499.68 ± 0.4798.86 ± 0.4199.97 ± 0.1999.90 ± 1.0899.97± 0.28
1298.36 ± 0.1498.86 ± 0.0799.26 ± 0.1599.97± 0.3299.94 ± 0.4999.71 ± 0.1599.89 ± 0.13
1399.52 ± 0.5199.76 ± 0.2599.76 ± 0.2299.39 ± 0.2999.43± 0.0999.52 ± 0.6099.96± 0.45
1499.79 ± 1.0298.43 ± 0.8399.16 ± 0.5896.78 ± 0.9399.93 ± 1.2399.27 ± 1.6099.99± 0.15
1584.51 ± 1.2085.68 ± 1.5483.23 ± 1.9672.62 ± 2.4497.38 ± 1.1598.03 ± 1.6497.75 ± 2.26
1699.38 ± 0.3099.75 ± 0.7499.58± 1.1598.65 ± 0.4499.76± 0.9799.08 ± 0.6399.92± 0.80
AA97.62 ± 0.1587.73 ± 0.1297.54 ± 0.0997.02 ± 0.1299.49 ± 0.0899.47 ± 0.2699.66 ± 0.14
OA95.09 ± 0.1795.08 ± 0.0994.19 ± 0.1293.91 ± 0.1099.02 ± 0.2099.40 ± 0.4099.53 ± 0.34
KC0.9453 ± 0.00190.9452 ± 0.00100.9353 ± 0.00130.9321 ± 0.00110.9890 ± 0.00220.9933 ± 0.00450.9948 ± 0.0038

Share and Cite

MDPI and ACS Style

Zhao, B.; Ulfarsson, M.O.; Sveinsson, J.R.; Chanussot, J. Unsupervised and Supervised Feature Extraction Methods for Hyperspectral Images Based on Mixtures of Factor Analyzers. Remote Sens. 2020, 12, 1179. https://doi.org/10.3390/rs12071179

AMA Style

Zhao B, Ulfarsson MO, Sveinsson JR, Chanussot J. Unsupervised and Supervised Feature Extraction Methods for Hyperspectral Images Based on Mixtures of Factor Analyzers. Remote Sensing. 2020; 12(7):1179. https://doi.org/10.3390/rs12071179

Chicago/Turabian Style

Zhao, Bin, Magnus O. Ulfarsson, Johannes R. Sveinsson, and Jocelyn Chanussot. 2020. "Unsupervised and Supervised Feature Extraction Methods for Hyperspectral Images Based on Mixtures of Factor Analyzers" Remote Sensing 12, no. 7: 1179. https://doi.org/10.3390/rs12071179

APA Style

Zhao, B., Ulfarsson, M. O., Sveinsson, J. R., & Chanussot, J. (2020). Unsupervised and Supervised Feature Extraction Methods for Hyperspectral Images Based on Mixtures of Factor Analyzers. Remote Sensing, 12(7), 1179. https://doi.org/10.3390/rs12071179

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop