Next Article in Journal
SAR Imaging Distortions Induced by Topography: A Compact Analytical Formulation for Radiometric Calibration
Next Article in Special Issue
Transferable Deep Learning from Time Series of Landsat Data for National Land-Cover Mapping with Noisy Labels: A Case Study of China
Previous Article in Journal
InSAR Coherence Analysis for Wetlands in Alberta, Canada Using Time-Series Sentinel-1 Data
Previous Article in Special Issue
Remote Sensing Image Scene Classification via Label Augmentation and Intra-Class Constraint
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
Engineering Research Center of Intelligent Perception and Autonomous Control, Ministry of Education, Beijing 100124, China
3
Traffic Management Research Institute of the Ministry of Public Security, Wuxi 214151, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(16), 3316; https://doi.org/10.3390/rs13163316
Submission received: 24 June 2021 / Revised: 16 August 2021 / Accepted: 18 August 2021 / Published: 21 August 2021
(This article belongs to the Special Issue Deep Learning for Remote Sensing Data)

Abstract

:
Hyperspectral classification is an important technique for remote sensing image analysis. For the current classification methods, limited training data affect the classification results. Recently, Conditional Variational Autoencoder Generative Adversarial Network (CVAEGAN) has been used to generate virtual samples to augment the training data, which could improve the classification performance. To further improve the classification performance, based on the CVAEGAN, we propose a Self-Attention-Based Conditional Variational Autoencoder Generative Adversarial Network (SACVAEGAN). Compared with CVAEGAN, we first use random latent vectors to obtain more enhanced virtual samples, which can improve the generalization performance. Then, we introduce the self-attention mechanism into our model to force the training process to pay more attention to global information, which can achieve better classification accuracy. Moreover, we explore model stability by incorporating the WGAN-GP loss function into our model to reduce the mode collapse probability. Experiments on three data sets and a comparison of the state-of-art methods show that SACVAEGAN has great advantages in accuracy compared with state-of-the-art HSI classification methods.

1. Introduction

With the development of remote sensing technology, hyperspectral images (HSI) have made great breakthroughs in earth observation. Different from three-channel color images, HSI can simultaneously collect images in hundreds of spectral bands, providing rich spectral information. Therefore, HSI are widely used in many fields [1,2,3,4,5,6], such as satellite remote sensing, agriculture observation, and mineral exploration.
For images classification problems, many algorithms [7,8,9,10] have be used, such as kernel extreme learning machine (KELM) and kernel principal component (KPCK) [7], wavelet coefficients [8], and neural network [9]. Many traditional methods have also been used to solve HSI classification problems, such as support vector machine (SVM) [11,12,13], K nearest neighbor (KNN) [14,15], maximum likelihood [16], neural network [17], and logistic regression [18,19]. However, these algorithms usually require manual designing of spectral and spatial features. Therefore, the classification results may be affected by manual features instead of algorithms. Moreover, traditional HSI classification algorithms usually encounter the Huges phenomenon [20], which severely affects the classification results.
Recently, deep learning methods [21,22,23,24,25,26,27,28,29,30,31] have received attention in HSI classification. Chen et al. [21] first applied the convolution network to solve the HSI classification problem. To augment the training data, CNN-PPF [26] was proposed by pairing selected samples. In this way, the training data can be expanded to obtain better classification results. In Reference [31], a two-channel neural network was proposed, which extracts features from spectral and spatial domains. In Reference [30], a novel local spatial sequence method based on recurrent neural network (RNN) was proposed, which can extract local information and semantic information for HSI classification. On the other hand, Hamida et al. [22] proposed a 3D convolution network to extract the spatial-spectral features of hyperspectral data for the classification and has achieved excellent results. However, most CNN-based HSI classification methods use fixed-size patches to train the model, which cannot reflect the complex spatial structure information in HSI. Therefore, MS-DenseNet [28] uses multi-scale patches to solve this problem.
Although deep learning methods have made great progress in HSI classification, they still face the problem of limited labeled data. To solve the problem, some scholars have proposed Generative Adversarial Networks (GAN) [32] and other models [33,34,35,36,37,38,39,40] to generate virtual samples to augment the training data. Zhu et al. [36] proposed a GAN-based [32] classification method for HSI and discussed its practicability and effectiveness in HSI classification. This method generates virtual samples through the GAN [32] network to alleviate the lack of labeled data in HSI and to improve the accuracy of the model. Multi-class spatial-spectral GAN (MSGAN) [34] was also proposed; it generates samples of spatial and spectral information by designing two generators and then uses a discriminator to extract spatial-spectral features and to output the final category. DropBlock structure (DBGAN) [41] was also applied to a GAN [40], which allowed the CNN to learn features by discarding continuous regions in the feature map to improve performance. Progressive growing GAN (PG-GAN) [42] and Wasserstein GAN gradient penalty (WGAN-GP) [43] were also proposed; they gradually increase the depth of the network, making the training process faster. Capsule network (CapsNet) [44] and convolutional long short-term memory (ConvLSTM) [45] were combined in Reference [39]. It designed a new discriminator that extracts low-level features and combines them with local spatial sequence information to form high-level context features. CVAEGAN [37] was also proposed for hyperspectral classification; it makes the classification process more stable and obtains better performance.
In this paper, we propose a new HSI classification model called SACVAEGAN. Compared with the original CVAEGAN [37] model, our proposed SACVAEGAN can generate virtual samples from the generated and random latent vector and can generate more high-quality training data. We also apply the self-attention mechanism [46] to the VAE and discriminator so that the model focuses more on global features rather than local features, thus making the virtual samples generated by the model more realistic. In the discriminator, the loss function in WGAN-GP [43] is applied to the optimization process, which makes the model training more stable. Moreover, in the encoder and classifier, we extract the features of HSI data from the spectral and spatial dimensions, which can effectively enhance the performance.
The rest of this paper is organized as follows. Section 2 presents the CVAEGAN [47] model and self-attention method [46]. Section 3 presents each module of SACVAEGAN in detail. Section 4 carries out a comprehensive evaluation of the proposed method. Section 5 analyzes each module of the self-attention mechanism, the WGAN-GP loss function, and the virtual samples. Finally, conclusions are drawn in Section 6.

2. Related Work

In this section, the CVAEGAN [47] and the self-attention mechanism [46,48] are introduced, which is the basis of our proposed method.

2.1. Conditional Variational Auto-Encoder Generative Adversarial Networks

In recent years, generative models have been roughly divided into VAE and GAN. Both have their strengths and weaknesses. The image generated by the VAE is normal but blurry. The GAN generates clear images, but the GAN model is hard to converge in the training stage and the samples generated from GAN are far from natural. The CVAEGAN network combines the advantages of the two to improve the effect of generating images.
CVAEGAN [47] is a network structure that combines VAE [49] and GAN [32]. It consists of four parts: (1) an encoder network E, used to learn the relationship between a latent vector space and the real image space; (2) the generator network G, which generates the corresponding virtual sample through the given latent vector; (3) the discriminator network D, used to judge whether a given sample is a real sample or a virtual sample; and (4) the classifier network C, used to classify a given sample.
The encoder network E transfers the data x to a latent vector space with a Gaussian distribution N ( 0 , 1 ) through the learned distribution. It outputs the mean value μ and covariance ϵ of the latent vector corresponding to the data sample x. Then, through z = μ + r e x p ( ϵ ) , r N ( 0 , 1 ) , the latent vector z can be obtained. At the same time, the distance between p ( z ) and q ( z | x ) can be reduced through K L divergence, where p represents the true latent vector distribution and q represents the latent vector distribution predicted by the model. The following is the equation:
K L ( p | | q ) = ( p ( x ) l o g ( p ( x ) q ( x ) ) )
L K L = 1 2 ( μ T μ + ( exp ( ϵ ) ϵ 1 ) )
The classifier takes the data x as input and outputs the posterior probability P ( c | x ) . Classifier C minimizes the following loss function as follows:
L C = E x P d a t a ( x ) [ l o g P ( c | x ) ]
Suppose the generator G generates data samples G ( z ) corresponding to the latent vector by receiving the latent vector z. The discriminator is a binary classifier to judge the true or false of a given data sample. The loss function of the discriminator is as follows:
L D = E x P d a t a ( x ) [ l o g D ( x ) ] E x P z ( z ) [ l o g ( 1 D ( G ( z , c ) ) ) ]
In the generator, to solve the problem of unstable training and to synthesize more realistic data samples, CVAEGAN [47] adds a feature matching method to the generator. The generator G minimizes the following loss function:
L G = 1 2 ( x G ( z , c ) 2 2 + f D ( x ) f D ( G ( z , c ) ) 2 2 + f C ( x ) f C ( G ( z , c ) ) 2 2 )
L G D = 1 2 ( E x P r f D ( x ) E z r P z r f D ( G ( z r , c ) ) 2 2
L G C = 1 2 ( E x P r f C ( x ) E z r P z r f C ( G ( z r , c ) ) 2 2
where f D and f C represent the middle layer features of the discriminator and classifier, respectively; G ( z , c ) represents the sample generated by the latent vector generated by the encoder; and G ( z r , c ) represents the sample generated by the latent vector generated randomly.

2.2. Self-Attention

The purpose of the self-attention [46,48] mechanism is to improve the classification performance of the classifier. Most GAN-based methods [32] used self-attention in their encoder or generator to enhance the performance. Figure 1 shows the structure of self-attention. f ( x ) , g ( x ) , and h ( x ) are all ordinary 1 × 1 convolutions, and the difference lies in the different output channels. The output of f ( x ) is transposed and multiplied by the output of g ( x ) , and the attention map is obtained after the softmax activation function. After multiplying the obtained attention map by the output of h ( x ) pixel by pixel, the final attention feature map can be obtained. The specific process is as follows.
Suppose that the image features x R C × N in the previous hidden layer are transformed into two features spaces f and g to calculate attention. Among them, f ( x ) = W f x , g ( x ) = W g x . β j , i represents the degree of participation in the i area when the model synthesizes the image content of the j area, that is, the correlation.
β j , i = e x p ( s i j ) i = 1 N e x p ( s i j ) , s i j = f ( x i ) T g ( x j )
The output of the attention layer is o = ( o 1 , o 2 , , o j , , o N R C × N ) , C is the number of channels, and N is the number of feature positions of the previous hidden layer feature.
o j = v ( i = 1 N β j , i h ( x i ) ) , h ( x i ) = W h x i , v ( x i ) = W v x i
Finally, the output of the attention layer is multiplied by a scale parameter, and the input feature map is added. Therefore, the final output is as follows:
y i = γ o i + x i
To allow the network to pay attention to the neighborhood information, γ is a scalar that can be learned in the self-attention training process with the initial value of 0, and then, the weight is slowly assigned to other long-distance features.

3. Methodology

In this section, we introduce the details of the proposed SACVAEGAN method. As shown in Figure 2, the network structure consists of three modules: discriminator, VAE, and classifier. The module of the discriminator is used to determine whether the input samples are real samples or virtual samples. The VAE module is divided into two parts: encoder and generator. The encoder transfers real samples to the latent vector space, and the generator uses the latent vectors generated by the encoder and the random latent vectors to generate virtual hyperspectral samples. The classifier module classifies the input real samples and virtual samples.

3.1. Discriminator

As shown in Figure 3, it is the module of the discriminator. It consists of four convolution layers. For each layer, the kernel size of each layer is 3 × 3 . The self-attention mentioned in Section 2 is applied after the first and second convolution layers. After the last layer, we reshape the data into a feature vector. According to Reference [50], the label information should also be input to the discriminator to make the model more stable. Therefore, we reshape the label to a vector through a full connection layer and concatenate it with the feature vector. Then, a full connection layer is applied to reduce the dimension. At last, the sigmoid function is used to determine whether the data are real.
The loss function of the discriminator D is as follows:
L D = L D G P + L D G ( z r a n d o m | y )
The first term is the loss in WGAN-GP between x and G ( z | y ) , which can make the model more stable, and the second term is the loss of the discriminator, which judges whether G ( z r a n d o m | y ) is false.
Among them:   
L D G P = E z P z [ D ( G ( z | y ) | y ) ] E x P ( x r e a l ) [ D ( x | y ) ] + λ E z P z [ ( G ( z | y ) D ( G ( z | y ) | y ) 2 1 ) 2 ]
L D x z r a n d o m = E z p z r a n d o m [ l o g D ( G ( z r a n d o m | y ) | y ) ]
where z represents the latent vector generated by the encoder,  z r a n d o m represents the latent vector generated randomly, x r e a l represents real samples, G ( z | y ) represents the virtual samples generated by the generator according to z and the corresponding label, G ( z r a n d o m | y ) represents the virtual sample generated by the generator according to z r a n d o m and the corresponding label, and y represents the label.

3.2. Variational Auto-Encoder

The module of the Variational Auto-Encoder (VAE) is shown in Figure 4. The VAE consists of two parts: the encoder and the generator. The encoder E is used to transfer the real samples to the latent vector space, while the generator G uses the latent vector to generate the virtual hyperspectral sample. We can see from Figure 4 that the encoder E is divided into two spectral-spatial feature extraction networks: one is used to get the mean vector μ , and the other is for the covariance ϵ of the latent vector space. For each feature extraction network, the network structure is the same. It consists of spectral and spatial feature extraction networks. For the spectral feature extraction network, it consists of 4 1-D convolution layers with 5 × 1 kernel. Self-attention is also applied to the first and second layers. For the spatial feature network, there are four 2-D convolution layers with 3 × 3 kernel and self-attention in the first and second layers. After the spectral and spatial feature extraction network, we concatenate the spectral-spatial feature together and use a full connection layer for dimension reduction. After we obtain the mean vector μ and the covariance ϵ , we use the following equation to obtain the latent vector:
z = μ + r e x p ( ϵ )
The goal of generator G is to learn the distribution of training data and to generate virtual hyperspectral samples. For the generator G, it consists of two full connection layers and four transposed convolution layers. After obtaining the latent vector and the corresponding label, it uses two full connection layers to reshape the vector. Then, the vector is reshaped into a 3-D data cube, which is sent to the transposed convolution layers. The kernel size of the transposed convolution layer is 3 × 3 . At last, we can obtain a virtual hyperspectral sample.
The loss function of VAE is as follows:
L V A E = L k l + L G + L v a e D + L v a e C
The first term is the KL divergence, which is used to reduce the difference between the distribution of the obtained latent vector and the assumed latent vector distribution. The second term is the l 2 reconstruction loss between x and G ( z | y ) . The third term is the sum of the pair-wise feature matching loss between x and G(z|y) in the discriminator D and the loss of the discriminator in judging whether G ( z r a n d o m | y ) is true. The last term is the sum of the pair-wise feature matching loss between x and G ( z | y ) in the classifier C and the loss of the classification result of the classifier.
Among them:
L k l = 1 2 ( μ T μ + s u m ( e x p ( ϵ ) ϵ 1 ) )
L G = 1 2 ( x G ( z | y ) 2 2 )
L v a e D = f D ( x ) f D ( G ( z | y ) ) 2 2 + E [ l o g ( 1 D ( G ( z r a n d o m | y ) ) ) ]
L v a e C = f C ( x ) f C ( G ( z | y ) ) 2 2 E [ l o g P ( y | G ( z r a n d o m | y ) ) ]
where μ and ϵ represent the mean value and covariance generated by the encoder, respectively; f D represents the features of the middle layer in the discriminator D; f C represents the features of the middle layer of the classifier; the real samples is represented as x r e a l ; the virtual sample generated according to the latent vector generated by the encoder is G ( z | y ) ; the virtual sample generated according to the randomly generated latent vector is G ( z r a n d o m | y ) ; and y represents the label.

3.3. Classifier

The module of classifier C is shown in Figure 5. The purpose of classifier C is to obtain the classification results. For classifier C, it also consists of spectral-spatial feature extraction networks. For the spectral feature extraction network, there are five 1-D convolution layers with the kernel of 1 × 5 , and for the spatial feature extraction network, it also consists of five 2-D convolution layers with 3 × 3 kernel. Finally, we concatenate the spectral and spatial features together and then send them to two full connection layers to obtain the final results. The loss function L C is as follows:
L C = L C x r e a l + λ 1 L C x z + λ 2 L C x z r a n d o m
where the first term is the loss of the result obtained by classifying x. The second term is the sum of the pair-wise feature matching loss between x and G ( z | y ) . The last term is the loss of the result obtained by classifying G ( z r a n d o m | y ) . λ 1 and λ 2 are the weights of L C x z and L C x x r a n d o m loss, respectively.
L C x r e a l = E [ l o g P ( c | x r e a l ) ]
L C x z = f C ( x r e a l ) f C ( x z ) 2 2
L C z r a n d o m = E [ l o g P ( y | x z r a n d o m ) ]
where f C represents the characteristics of the middle layer of the classifier, x r e a l represents the real samples, x z represents the virtual sample generated by inputting the latent vector z generated by the encoder into the generator, x z r a n d o m represents the virtual sample generated by inputting the randomly generated latent vector z r a n d o m into the generator, and y represents the label.
The procedure of the proposed SACVAEGAN is summarized in Algorithm 1.
Algorithm 1: Optimization procedure of SACVAEGAN.
Input: the training samples x t r a i n = ( x i ) i = 1 N and test samples x t e s t = ( x i ) i = 1 S from C
    classes, the onehot lables ( y i ) i = 1 N of training samples, batch size n,
    the number of training epochs E, noise dimension d, the updating times k
    of the discriminator
1 Initialize all the weight matrices and biases
2while every epoch do
3 sample d dimensional noises z r a n d o m = ( z j ) j = 1 d ;
4 generate n latent vectors z = ( z j ) j = 1 d by Encoder E
5 generate n virtual samples G ( z | y ) through the generator according to the
    latent vector z;
6 generate n virtual samples G ( z r a n d o m | y ) through the generator according to the
    latent vector z r a n d o m ;
7for k steps do
8  updating parameters of the D by minimizing L D
9  end
10 updating parameters of the VAE by minimizing L V A E
11  updating parameters of C by minimizing L C
12end
Output: the labels of the test samples x t e s t classified by the trained sample
     classifier in SACVAEGAN

4. Experiments

To evaluate the accuracy of our proposed SACVAEGAN, we conducted training and evaluation on three data sets, namely the Indian Pines data set, the PaviaU data set, and the Salinas data set. We used three performance metrics: O A , A A , and K. O A is the overall classification accuracy, which represents the ratio between the categories that are properly classified and the total number of categories. A A is the average classification accuracy, which represents the average accuracy between categories. K represents the kappa coefficient for different weights in the confusion matrix.

4.1. Hyperspectral Data Sets

4.1.1. Indian Pines

The Indian Pines data set was collected by airborne visible infrared imaging spectrometer (AVRIS) sensors in Indian Pines testing in northwest Indiana. The AVIRIS sensor captured images in 0.4 to 2.5 microns. The data consist of 145 × 145 pixels and 224 spectral reflection bands. By deleting the bands covering the water absorption area, the number of spectral was reduced to 200. There are 16 categories in total, and the distribution of samples in each category is extremely unbalanced.

4.1.2. Paviau

The PaviaU data set was acquired by Reflective Optics System Imaging Spectrometer (ROSIS) sensors in Pavia region of northern Italy. The data consist of 610 × 340 pixels and 115 spectral reflection bands. By deleting 12 bands affected by noise, 103 spectral signatures remained. The size of the data is 610 × 340. However, most of them are background pixels, with only 42,776 pixels being foreground pixels in nine categories, including trees, Asphalt, Bricks, Meadows, etc.

4.1.3. Salinas

The Salinas data set was also collected by the AVIRIS sensor in the Salinas Valley, California, with a spatial resolution of 3.7 m. The data consist of 512 × 217 pixels and 224 spectral reflection bands. By deleting the bands covering the water absorption area, the spectrum was reduced to 204. The image size is 512 × 217 with 16 classes.
We divided the labeled samples in the three data sets into two parts: the training set and the testing set. Among them, for the Indian Pines data set, we chose 5 % data for training. For the PaviaU data set, we chose 2 % points as the training samples. For the Salinas data set, we chose 1 % points as the training samples. Table 1, Table 2 and Table 3 show the number of training samples and testing samples for each category on the Indian Pines data set, PaviaU data set, and Salinas data set, respectively.

4.2. Parameter Analysis

Before training the model, some important factors should be analyzed, which may affect the results. Due to the introduction of WGAN-GP, we used RMSprop as the optimizer in the VAE module and the discriminator module. The Adam optimizer was used in the classifier module. We mainly discuss three factors: the different size of patch sizes, the network regularization method, and the parameters of λ 1 and λ 2 .

4.2.1. Analysis of the Size of Patches

To analyze the influence of the size of patches, we calculated the overall accuracy (OA) on three data sets using different sizes of patches. As shown in the Table 4, the size of the patches is 19 × 19 , 23 × 23 , 27 × 27 , 31 × 31 . It can be seen from Table 4 for the Indian Pines and the PaviaU data sets. When the patch size was 27, the proposed method obtained the best performance, while for the Salinas data set, when the size was set to 31, it obtained the best results. Therefore, in the experiments, we set the patch size as 27, with better overall accuracy (OA).

4.2.2. Analysis of the Network Regularization Method

To analysis the regularization methods, we used different methods to regularize our proposed method, including batch normalization (BN) and dropout. Batch normalization can normalize the input data and can solve the problem of the gradient disappearing and the gradient exploding. It can also accelerate the convergence speed. Dropout enhances the generalization of the model by making the activation value have a certain probability p during forward propagation. As shown in Table 5, when the proposed method uses dropout and batch normalization, it obtains the best performance. Therefore, we applied dropout and Batch Normalization (BN) in our proposed method.

4.2.3. Analysis of the λ 1 and λ 2 in the Classifier Loss Function

In order to analyze the influence of λ 1 and λ 2 in Equation (17), we conducted experiments with different values of λ 1 and λ 2 . We set λ 1 from 0 to 0.5 and λ 2 from 0.5 to 1 with the interval of 0.1. We calculated the performance on three data sets. As can be seen in Table 6, Table 7 and Table 8, when λ 1 = 0.8 and λ 2 = 0.2 , our model obtains the best average classification accuracy. Therefore, we set the value of λ 1 to 0.8, and λ 2 to 0.2 in our experiments.

4.3. Classification Results

To illustrate the effectiveness of our proposed method, SACVAEGAN was compared with several different hyperspectral image classification methods, including SVM-Radial Basis Function ( R B F ) , Two-CNN [31], 3D-CNN [22], DCGAN [36], DBGAN [41], and CVAEGAN [37]. SVM-Radial Basis Function uses a Gaussian kernel function, and the parameter gamma of the kernel function was set to 0.125. The hyperparameters of the deep learning method were all set to the parameters mentioned in the corresponding paper. We used the classifier module for the final classification. Table 9, Table 10 and Table 11 show the comparison of the methods we proposed on the Indian Pines data set, the PaviaU data set, and the Salinas data set. All of the classification methods were tuned to the best settings. The training set and test set are shown in Table 1, Table 2 and Table 3.
Table 9, Table 10 and Table 11 show the results of the proposed methods on the Indian Pines data set, the PaviaU data set, and the Salinas data set, respectively. From the tables, we can see that the performance of the deep learning methods is better than the traditional method. This is because the deep learning method can extract better features. Moreover, for the GAN-based method, the performance is better than other deep learning methods because the GAN-based method could generate more training data, which could help the training. From these tables, we can see that our proposed method obtained the best performance. This may be due to the following reasons: First, compared with DCGAN and DBGAN, the VAE module is applied to our proposed method, which makes it higher. Second, compared with CVAEGAN, the conditional GAN is also applied to our proposed method, which could generate more high-quality training data and can enhance the performance.
Figure 6, Figure 7 and Figure 8 show the false color images of the three HSI data sets, the corresponding ground truth maps, and the classification maps of each method. It can be seen that the classification result diagram is consistent with the classification results shown in Table 9, Table 10 and Table 11. From these figures, we can see that the classification maps of our proposed method are clearer, less noisy, and closer to the ground truth. This indicates that our proposed method has a better classification capability than other methods.

4.4. Analysis of the Size of the Training Set

To verify the robustness of SACVAEGAN for different sizes of training data, we selected different proportions of samples for training. For the three data sets, we selected 1 to 5% data for training. We show the results in Figure 9. From Figure 9, we can see that our proposed method obtains a better performance than other methods. This demonstrates that our proposed method is more stable compared with other methods and robust to different sizes of training data.

4.5. Investigation on the Run Time

To evaluate the effectiveness of our proposed method, we show the training and testing time of seven different methods in Table 12. The experiments were run with an NVIDIA GTX 1660 GPU and an Inter i5-9300H 2.40-GHz CPU with 16 GB of RAM. It can be seen from Table 12 that the traditional method is much faster than the deep learning method. The deep learning method using the generative adversarial network is slower in training speed than and similar in test speed to the other deep learning methods. The method proposed in this paper requires a relatively long training time on the three data sets. The reason is that our proposed method extracts spectral and spatial features for training to enhance the performance, so more training time is required.

5. Discussion

To analyze the influence of the self-attention mechanism, WGAN-GP loss, and additional generated samples on the classification accuracy of the proposed method, we conducted several ablation experiments to analyze the spatial features, spectral features, and overall accuracy under different conditions.

5.1. Spatial Feature Analysis

To illustrate the advantages of the spatial features of the generated samples, we draw the corresponding virtual samples under different conditions. As shown in Figure 10, it can be seen that the samples generated by our method are closer to the real sample distribution. Thus, the detailed features captured by the virtual sample can improve the classification performance of the model in return. To analyze the difference between the spatial features of the generated virtual sample and the real sample, we calculated the mean square error (MSE) between the generated sample and the real sample. The mean square error can be calculated using Equation (24). As shown in Table 13, the sample generated by our proposed method is closer to the real sample compared with other methods.
M S E ( x , y ) = i = 1 n ( x i y i ) 2 n

5.2. Spectral Feature Analysis

To illustrate the advantages of spectral feature, we plot the spectral feature map corresponding to the generated virtual sample under different conditions in Figure 11. It can be seen that the spectral feature distribution of the virtual sample generated by our method is more consistent with the spectral feature distribution of real samples. In order to analyze the difference in spectral features between the virtual sample and the real sample, we calculated the spectral information divergence (SID) between the virtual sample and the real sample. SID is based on the theory of information theory to measure the difference between two spectral calculated by Equation (25). The smaller the SID value, the more similar the spectral. As shown in Table 14, the proposed method obtains the optimal SID on the three data sets.
S I D ( x , y ) = i = 1 l p i l o g ( p i q i ) + i = 1 l q i l o g ( q i p i )

5.3. Overall Accuracy Analysis

Self-attention enables the model to better extract global features. For the WGAN-GP, it can make the model more stable and can improve performance. Additional virtual samples can improve the generalization performance and classification accuracy. To analyze the advantage of each module, we run the experiments with or without each module for 10 times. As shown in Table 15, each module contributes to the improvement of classification accuracy. When the proposed method uses all three strategies, it achieves the best performance. Therefore, we apply all three strategies in our proposed method.
To analyze the impact of WGAN-GP loss on model performance, we calculate the Frechet Inception Distance (FID) of the proposed method with and without WGAN-GP. FID is a criterion to evaluate the performance of GAN. The basic idea is to input the training samples and generated samples into the classifier. Then, it extracts the features of the middle layer of the classifier. Assuming that the data conforms to the multivariate Gaussian distribution, it estimates the mean values of μ t r a i n and μ g e n , and the variance of σ t r a i n and σ g e n of the Gaussian distribution of the generated sample and the training sample. Then, it calculates the Freche distance of two Gaussian distributions using Equation (26). The smaller the value of FID, the better the performance of GAN. We run experiments 10 times on each data set and obtain the average values. It can be seen from Table 16 that, when the WGAN-GP loss is added, the FID of our model is lower. This proves that adding WGAN-GP can improve the performance of the SACVAEGAN model to a certain extent.
F I D = μ t r a i n μ g e n + t r ( σ t r a i n + σ g e n 2 ( σ t r a i n σ g e n ) 1 2 )

6. Conclusions

In this paper, a self-attention-based conditional variational autoencoder generative adversarial network (SACVAEGAN) is proposed for hyperspectral classification. We combine the Conditional GAN with CVAEGAN, which can generate more high-quality training data to enhance the performance. Moreover, the self-attention mechanism is also applied to our proposed SACVAEGAN to extract better features. A novel loss function is used to make the whole training process more stable. Compared with the GAN-based methods, SACVAEGAN achieved a better classification performance than the state-of-the-art methods on three commonly used hyperspectral image data sets by incorporating an extra self-attention mechanism and the WGAN-GP loss. In the future, we will explore more GAN-based models for HSI classification.

Author Contributions

All authors contributed to this manuscript: Conceptualization, L.T. and Z.C.; Methodology, L.T. and Z.C.; Supervision, L.T. and C.X.; Validation, Z.C and B.Q.; Resources, Z.C. and J.Y.; Writing original draft, Z.C.; Writing review and editing, Z.C., L.T. and B.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the National Natural Science Foundation of China under Grant 61701009, 61801479 and Beijing Municipal Education Commission Science and Technology Program under Grant KM202010005016.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. An Augmented Linear Mixing Model to Address Spectral Variability for Hyperspectral Unmixing. IEEE Trans. Image Process. 2019, 28, 1923–1938. [Google Scholar] [CrossRef] [Green Version]
  2. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of Spectral–Temporal Response Surfaces by Combining Multispectral Satellite and Hyperspectral UAV Imagery for Precision Agriculture Applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  3. Hunt, E.R.; Daughtry, C.S.T.; Mirsky, S.B.; Hively, W.D. Remote Sensing With Simulated Unmanned Aircraft Imagery for Precision Agriculture Applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4566–4571. [Google Scholar] [CrossRef]
  4. Tao, C.; Wang, Y.J.; Zou, B.; Tu, Y.L.; Jiang, X.L. Assessment and Analysis of Migrations of Heavy Metal Lead and Zinc in Soil with Hyperspectral Inversion Model. Spectrosc. Spectr. Anal. 2018, 38, 1850. [Google Scholar]
  5. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A General End-to-End 2-D CNN Framework for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3–13. [Google Scholar] [CrossRef] [Green Version]
  6. Salah Bourennane, C.F. Robust Denoising Method based on Tensor Models Decomposition for Hyperspectral Imagery. WSEAS Trans. Signal Process. 2019, 15, 92–98. [Google Scholar]
  7. Sawssen, B.; Okba, T.; Noureeddine, L. A Mammographic Images Classification Technique via the Gaussian Radial Basis Kernel ELM and KPCA. Int. J. Appl. Math. Comput. Sci. Syst. Eng. 2020, 2, 92–98. [Google Scholar]
  8. Vetova, S. A Comparative Study of Image Classification Models using NN and Similarity Distance. Int. J. Electr. Eng. Comput. Sci. (EEACS) 2021, 3, 109–113. [Google Scholar]
  9. Vetova, S. Covid Image Classification using Wavelet Feature Vectors and NN. Eng. World 2021, 3, 38–42. [Google Scholar]
  10. Luqman Hakim, M.I.Z. Implementation of Discrete Wavelet Transform on Movement Images and Recognition by Artificial Neural Network Algorithm. WSEAS Trans. Signal Process. 2019, 15, 149–154. [Google Scholar]
  11. Zhong, S.; Chang, C.I.; Zhang, Y. Iterative Support Vector Machine for Hyperspectral Image Classification. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3309–3312. [Google Scholar]
  12. Sun, S.; Zhong, P.; Xiao, H.; Liu, F.; Wang, R. An active learning method based on SVM classifier for hyperspectral images classification. In Proceedings of the 2015 7th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Tokyo, Japan, 2–5 June 2015; pp. 1–4. [Google Scholar]
  13. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geoence Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  14. Song, W.; Li, S.; Kang, X.; Huang, K. Hyperspectral image classification based on KNN sparse representation. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 2411–2414. [Google Scholar]
  15. Guo, Y.; Cao, H.; Han, S.; Sun, Y.; Bai, Y. Spectral–Spatial HyperspectralImage Classification with K-Nearest Neighbor and Guided Filter. IEEE Access 2018, 6, 18582–18591. [Google Scholar] [CrossRef]
  16. Jia, X. Block-based maximum likelihood classification for hyperspectral remote sensing data. In Proceedings of the IGARSS’97. 1997 IEEE International Geoscience and Remote Sensing Symposium Proceedings: Remote Sensing—A Scientific Vision for Sustainable Development, Singapore, 3–8 August 1997; Volume 2, pp. 778–780. [Google Scholar]
  17. Ratle, F.; Camps-Valls, G.; Weston, J. Semisupervised Neural Networks for Efficient Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2271–2282. [Google Scholar] [CrossRef]
  18. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised Hyperspectral Image Classification Using Soft Sparse Multinomial Logistic Regression. IEEE Geosci. Remote Sens. Lett. 2013, 10, 318–322. [Google Scholar]
  19. Khodadadzadeh, M.; Li, J.; Plaza, A.; Bioucas-Dias, J.M. A Subspace-Based Multinomial Logistic Regression for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2105–2109. [Google Scholar] [CrossRef]
  20. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 14, 55–63. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  22. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D Deep Learning Approach for Remote Sensing Image Classification. IEEE Trans. Geoence Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, Y.; Wang, Y.; Gu, Y.; He, X.; Ghamisi, P.; Jia, X. Deep Learning Ensemble for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1882–1897. [Google Scholar] [CrossRef]
  24. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 7, 2094–2107. [Google Scholar] [CrossRef]
  25. Zhang, M.; Li, W.; Du, Q. Diverse Region-Based CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2018, 27, 2623–2634. [Google Scholar] [CrossRef]
  26. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  27. Xu, X.; Li, W.; Ran, Q.; Du, Q.; Gao, L.; Zhang, B. Multisource Remote Sensing Data Classification Based on Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2018, 56, 937–949. [Google Scholar] [CrossRef]
  28. Xie, J.; He, N.; Fang, L.; Ghamisi, P. Multiscale Densely-Connected Fusion Networks for Hyperspectral Images Classification. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 246–259. [Google Scholar] [CrossRef]
  29. Ying, L.; Haokui, Z.; Qiang, S. Spectral-Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar]
  30. Zhang, X.; Sun, Y.; Jiang, K.; Li, C.; Jiao, L.; Zhou, H. Spatial Sequential Recurrent Neural Network for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4141–4155. [Google Scholar] [CrossRef] [Green Version]
  31. Yang, J.; Zhao, Y.; Chan, J.C.W.; Yi, C. Hyperspectral image classification using two-channel deep convolutional neural network. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5079–5082. [Google Scholar]
  32. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  33. Wang, X.; Tan, K.; Du, Q.; Chen, Y.; Du, P. Caps-TripleGAN: GAN-Assisted CapsNet for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7232–7245. [Google Scholar] [CrossRef]
  34. Feng, J.; Yu, H.; Wang, L.; Cao, X.; Zhang, X.; Jiao, L. Classification of Hyperspectral Images Based on Multiclass Spatial–Spectral Generative Adversarial Networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5329–5343. [Google Scholar] [CrossRef]
  35. Zhong, Z.; Li, J.; Clausi, D.A.; Wong, A. Generative Adversarial Networks and Conditional Random Fields for Hyperspectral Image Classification. IEEE Trans. Cybern. 2020, 50, 3318–3329. [Google Scholar]
  36. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  37. Wang, H.; Tao, C.; Qi, J.; Li, H.; Tang, Y. Semi-Supervised Variational Generative Adversarial Networks for Hyperspectral Image Classification. In Proceedings of the IGARSS 2019—2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 9792–9794. [Google Scholar]
  38. Zhang, F.; Bai, J.; Zhang, J.; Xiao, Z.; Pei, C. An Optimized Training Method for GAN-Based Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 1–5. [Google Scholar] [CrossRef]
  39. Wang, W.Y.; Li, H.C.; Deng, Y.J.; Shao, L.Y.; Lu, X.Q.; Du, Q. Generative Adversarial Capsule Network With ConvLSTM for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 523–527. [Google Scholar] [CrossRef]
  40. Yin, J.; Li, W.; Han, B. Hyperspectral Image Classification Based on Generative Adversarial Network with Dropblock. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 405–409. [Google Scholar]
  41. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  42. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive growing of gans for improved quality, stability, and variation. arXiv 2017, arXiv:1710.10196. [Google Scholar]
  43. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, C.A. Improved Training of Wasserstein GANs. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5767–5777. [Google Scholar]
  44. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic Routing Between Capsules. arXiv 2017, arXiv:1710.09829. [Google Scholar]
  45. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  46. Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  47. Bao, J.; Chen, D.; Wen, F.; Li, H.; Hua, G. CVAE-GAN: Fine-Grained Image Generation through Asymmetric Training. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2764–2773. [Google Scholar]
  48. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning (PMLR 2019), Beach, CA, USA, 9–15 June 2019; pp. 7354–7363. [Google Scholar]
  49. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2014, arXiv:1312.6114. [Google Scholar]
  50. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
Figure 1. The self-attention module for SACVAEGAN. The ⊗ denotes matrix multiplication.
Figure 1. The self-attention module for SACVAEGAN. The ⊗ denotes matrix multiplication.
Remotesensing 13 03316 g001
Figure 2. The structure of the proposed SACVAEGAN.
Figure 2. The structure of the proposed SACVAEGAN.
Remotesensing 13 03316 g002
Figure 3. The structure of the discriminator D.
Figure 3. The structure of the discriminator D.
Remotesensing 13 03316 g003
Figure 4. The structure of the V A E .
Figure 4. The structure of the V A E .
Remotesensing 13 03316 g004
Figure 5. The structure of the classifier C.
Figure 5. The structure of the classifier C.
Remotesensing 13 03316 g005
Figure 6. Illustration of the classification results on the Indian Pines data set. (a) Ground truth, (b) SVM-RBF (51.69%), (c) Two-CNN (90.10%), (d) 3D-CNN (77.80%), (e) DCGAN (91.11%), (f) DBGAN (84.42%), (g) CVAEGAN (93.47%), and (h) SACVAEGAN (95.98%).
Figure 6. Illustration of the classification results on the Indian Pines data set. (a) Ground truth, (b) SVM-RBF (51.69%), (c) Two-CNN (90.10%), (d) 3D-CNN (77.80%), (e) DCGAN (91.11%), (f) DBGAN (84.42%), (g) CVAEGAN (93.47%), and (h) SACVAEGAN (95.98%).
Remotesensing 13 03316 g006
Figure 7. Illustration of the classification results on the PaviaU data set. (a) Ground truth, (b) SVM-RBF (76.00%), (c) Two-CNN (95.80%), (d) 3D-CNN (96.55%), (e) DCGAN (93.88%), (f) DBGAN (96.99%), (g) CVAEGAN (98.07%), and (h) SACVAEGAN (98.30%).
Figure 7. Illustration of the classification results on the PaviaU data set. (a) Ground truth, (b) SVM-RBF (76.00%), (c) Two-CNN (95.80%), (d) 3D-CNN (96.55%), (e) DCGAN (93.88%), (f) DBGAN (96.99%), (g) CVAEGAN (98.07%), and (h) SACVAEGAN (98.30%).
Remotesensing 13 03316 g007
Figure 8. Illustration of the classification results on the PaviaU data set. (a) Ground truth, (b) SVM-RBF (75.25%), (c) Two-CNN (95.79%), (d) 3D-CNN (91.05%), (e) DCGAN (90.39%), (f) DBGAN (94.13%), (g) CVAEGAN (98.49%), and (h) SACVAEGAN (98.92%).
Figure 8. Illustration of the classification results on the PaviaU data set. (a) Ground truth, (b) SVM-RBF (75.25%), (c) Two-CNN (95.79%), (d) 3D-CNN (91.05%), (e) DCGAN (90.39%), (f) DBGAN (94.13%), (g) CVAEGAN (98.49%), and (h) SACVAEGAN (98.92%).
Remotesensing 13 03316 g008
Figure 9. OA(%) of different methods with different training percentages of samples. (a) Indian Pines. (b) PaviaU. (c) Salinas.
Figure 9. OA(%) of different methods with different training percentages of samples. (a) Indian Pines. (b) PaviaU. (c) Salinas.
Remotesensing 13 03316 g009
Figure 10. Illustration of the spatial features on the Indian Pines (ae), PaviaU (fj), and Salinas (ko) data sets.
Figure 10. Illustration of the spatial features on the Indian Pines (ae), PaviaU (fj), and Salinas (ko) data sets.
Remotesensing 13 03316 g010
Figure 11. Illustration of the spectral features on the Indian Pines (ae), PaviaU (fj), Salinas (ko) data sets. The orange line represents the mean, the green line represents the mean add the variance, and the blue line represents the mean minus the variance.
Figure 11. Illustration of the spectral features on the Indian Pines (ae), PaviaU (fj), Salinas (ko) data sets. The orange line represents the mean, the green line represents the mean add the variance, and the blue line represents the mean minus the variance.
Remotesensing 13 03316 g011
Table 1. Training and testing data in the Indian Pines data set.
Table 1. Training and testing data in the Indian Pines data set.
No.ClassTrainingTest
1Alfalfa244
2Corn-notill711357
3Corn-mintill41789
4Corn12225
5Grass-pasture24459
6Grass-trees37693
7Grass-pasture-mowed127
8Hay-windrowed24454
9Oats119
10Soybean-notill49923
11Soybean-mintill1232332
12Soybean-clean30563
13Wheat10195
14Woods631202
15Buildings-Grass-Trees-Drives19367
16Stone-Steel-Towers588
Total 5129737
Table 2. Training and testing data in the PaviaU data set.
Table 2. Training and testing data in the PaviaU data set.
No.ClassTrainingTest
1Asphalt1326499
2Meadows37318,276
3Gravel422057
4Trees613003
5Painted metal sheets271318
6Bare Soil1004929
7Bitumen271303
8Self-Blocking Bricks743608
9Shadows19928
Total 85541,921
Table 3. Training and testing data in the Salinas data set.
Table 3. Training and testing data in the Salinas data set.
No.ClassTrainingTest
1Brocoli_green_weeds_1201989
2Brocoli_green_weeds_2373689
3Fallow201956
4Fallow_rough_plow141380
5Fallow_smooth272651
6Stubble393920
7Celery363543
8Grapes_untrained11311,158
9Soil_vinyard_develop626141
10Corn_senesced_green_weeds333245
11Lettuce_romaine_4wk111057
12Lettuce_romaine_5wk191908
13Lettuce_romaine_6wk9907
14Lettuce_romaine_7wk111059
15Vinyard_untrained727196
16Vinyard_vertical_trellis181789
Total 54153,588
Table 4. Overall accuracy of different patch sizes in the Indian pines, PaviaU, and Salinas data sets.
Table 4. Overall accuracy of different patch sizes in the Indian pines, PaviaU, and Salinas data sets.
Patch SizeIndian PinesPaviaUSalinas
19 × 1992.7898.9197.05
23 × 2394.7899.0197.80
27 × 2795.9499.3298.82
31 × 3195.0299.1099.19
Table 5. Network regularization method in the Indian Pines, PaviaU, and Salinas data sets.
Table 5. Network regularization method in the Indian Pines, PaviaU, and Salinas data sets.
ProposeedIndian PinesPaviaUSalinas
None94.0198.3797.70
Dropout95.2698.2698.00
BN95.5098.6398.21
Both95.9499.3298.82
Table 6. Overall accuracy in the Indian Pines data set with different λ 1 and λ 2 .
Table 6. Overall accuracy in the Indian Pines data set with different λ 1 and λ 2 .
λ 1 0.50.60.70.80.91
λ 2
094.5094.8295.2395.7395.6795.39
0.194.5194.7195.3595.3295.7195.32
0.293.5694.7595.4695.9495.7695.44
0.393.7293.3294.6795.5595.1095.59
0.493.8594.1094.4394.5794.6694.91
0.593.4393.8794.2294.0194.6294.68
Table 7. Overall accuracy in the PaviaU data set with different λ 1 and λ 2 .
Table 7. Overall accuracy in the PaviaU data set with different λ 1 and λ 2 .
λ 1 0.50.60.70.80.91
λ 2
098.9999.0199.0599.2499.1098.82
0.198.8398.7499.2199.2899.1499.07
0.298.9199.1299.2799.3299.2999.19
0.398.6198.6799.0199.1399.0798.91
0.498.6998.4398.8798.5798.7698.59
0.598.3498.3197.8998.4698.3998.21
Table 8. Overall accuracy in the Salinas data set with different λ 1 and λ 2 .
Table 8. Overall accuracy in the Salinas data set with different λ 1 and λ 2 .
λ 1 0.50.60.70.80.91
λ 2
098.4798.5198.5798.6598.4198.61
0.198.3898.4298.5898.6698.7298.67
0.298.4698.3298.7898.8298.8198.79
0.398.1298.0998.4198.3198.5398.37
0.497.7598.1898.4398.2998.0598.17
0.597.3797.5198.2597.9398.1198.03
Table 9. Classification results on the Indian Pines data set.
Table 9. Classification results on the Indian Pines data set.
ClassSVM−RBFTwo−CNN3D−CNNDCGANDBGANCVAEGANProposed
1069.78 ± 1.9817.70 ± 1.1175.97 ± 0.0437.83 ± 7.4471.01 ± 2.3766.67 ± 3.29
228.07 ± 2.6384.78 ± 0.2668.96 ± 0.8485.60 ± 0.0177.23 ± 1.3890.13 ± 0.2797.15 ± 0.38
30.33 ± 0.9988.39 ± 0.5968.11 ± 1.4690.72 ± 0.0181.16 ± 1.2688.76 ± 0.9191.04 ± 1.63
41.11 ± 2.9094.68 ± 0.4267.75 ± 3.6795.36 ± 0.0189.37 ± 3.9590.44 ± 2.8892.55 ± 1.80
52.11 ± 4.0880.68 ± 0.7587.52 ± 0.8481.78 ± 0.0224.43 ± 5.8884.82 ± 2.0691.58 ± 1.05
692.58 ± 2.3595.04 ± 0.3288.26 ± 0.6595.62 ± 0.0191.84 ± 0.9799.45 ± 0.1197.99 ± 0.31
7090.00 ± 2.6665.31 ± 6.2881.07 ± 0.0514.29 ± 9.1679.76 ± 1.9442.86 ± 2.92
899.21 ± 1.2797.05 ± 0.2299.16 ± 0.2197.47 ± 0.0098.42 ± 0.32100 ± 098.60 ± 0.82
9036.00 ± 0.6360.00 ± 6.0680.00 ± 0.0429.50 ± 6.6545.00 ± 4.0838.33 ± 5.93
100.35 ± 1.0484.44 ± 0.4768.90 ± 0.7787.14 ± 0.0183.70 ± 0.6989.27 ± 0.9396.02 ± 1.3
1194.99 ± 1.7693.18 ± 0.2780.14 ± 0.7296.99 ± 0.0094.12 ± 0.7597.65 ± 0.6498.56 ± 0.31
120.23 ± 0.5482.48 ± 0.3854.85 ± 2.3683.98 ± 0.0173.88 ± 2.4485.83 ± 1.3389.71 ± 0.68
130.29 ± 30.7994.44 ± 0.4698.47 ± 0.5996.59 ± 0.0195.41 ± 0.4899.02 ± 0.6199.84 ± 0.13
1498.66 ± 0.3896.17 ± 0.1992.94 ± 0.3698.18 ± 0.0197.53 ± 0.4099.31 ± 0.0498.58 ± 0.78
152.70 ± 4.2495.52 ± 0.6269.73 ± 1.1897.93 ± 0.0287.42 ± 2.3092.57 ± 2.2095.51 ± 0.78
1682.61 ± 4.3784.95 ± 1.8480.49 ± 2.1088.17 ± 0.0250.86 ± 8.5178.49 ± 0.5192.47 ± 2.63
OA (%)51.69 ± 0.7390.10 ± 0.2177.80 ± 0.4691.11 ± 0.4584.42 ± 0.3793.47 ± 0.1195.94 ± 0.20
AA (%)33.27 ± 1.7585.41 ± 0.3573.02 ± 0.7387.53 ± 0.5567.00 ± 0.2986.97 ± 0.2486.71 ± 0.69
Kappa (%)41.46 ± 0.9288.73 ± 0.2474.68 ± 0.5289.87 ± 0.5182.14 ± 0.4282.54 ± 0.1395.36 ± 0.22
Table 10. Classification results on the PaviaU data set.
Table 10. Classification results on the PaviaU data set.
ClassSVM−RBFTwo−CNN3D−CNNDCGANDBGANCVAEGANProposed
191.49 ± 1.0094.81 ± 0.1696.91 ± 0.1892.70 ± 0.3596.03 ± 0.3796.31 ± 0.5399.60 ± 0.15
296.80 ± 2.7699.51 ± 0.0799.15 ± 0.0899.40 ± 0.1699.67 ± 0.0999.98 ± 0.0299.99 ± 0.01
31.89 ± 3.5079.15 ± 0.2885.88 ± 0.8572.87 ± 3.1080.98 ± 3.0990.41 ± 0.6897.09 ± 1.01
463.57 ± 12.5994.14 ± 0.2594.73 ± 0.4591.33 ± 1.5297.90 ± 0.3395.96 ± 0.2297.96 ± 0.44
598.87 ± 0.1698.74 ± 0.3499.72 ± 0.0699.48 ± 0.1698.23 ± 0.64100 ± 0100 ± 0
614.84 ± 1.7996.88 ± 0.3895.32 ± 0.4996.34 ± 0.9198.90 ± 0.3199.25 ± 0.37100 ± 0
7084.09 ± 0.2986.46 ± 0.9985.41 ± 3.6290.79 ± 1.2994.29 ± 2.9496.54 ± 1.95
891.90 ± 2.5992.15 ± 0.2494.41 ± 0.3383.13 ± 0.2093.83 ± 1.0498.75 ± 0.2698.23 ± 0.68
999.82 ± 0.0892.73 ± 0.7397.02 ± 0.5486.91 ± 0.3792.67 ± 1.7990.11 ± 1.0297.04 ± 0.85
OA (%)76.00 ± 0.4995.80 ± 0.0996.55 ± 0.1793.88 ± 0.3996.99 ± 0.3498.07 ± 0.0499.32 ± 0.06
AA (%)62.13 ± 1.1692.47 ± 0.1494.40 ± 0.3389.19 ± 0.2395.33 ± 0.2196.12 ± 0.3798.49 ± 0.14
Kappa (%)66.28 ± 0.4894.41 ± 0.1295.42 ± 0.2391.86 ± 0.5196.00 ± 0.4597.43 ± 0.0599.10 ± 0.08
Table 11. Classification results on the Salinas data set.
Table 11. Classification results on the Salinas data set.
ClassSVM−RBFTwo−CNN3D−CNNDCGANDBGANCVAEGANProposed
194.92 ± 1.4393.16 ± 0.6197.58 ± 0.1496.33 ± 1.1023.48 ± 0.1999.93 ± 0.0599.30 ± 0.31
295.36 ± 1.4398.79 ± 0.1599.84 ± 0.0492.39 ± 1.6597.82 ± 1.49100 ± 098.60 ± 0.57
360.68 ± 4.3399.65 ± 0.1698.46 ± 0.3192.02 ± 1.3698.50 ± 0.8399.19 ± 0.4599.85 ± 0.10
498.44 ± 0.3696.90 ± 0.3798.76 ± 0.2497.20 ± 0.4097.30 ± 0.5897.87 ± 0.5998.02 ± 0.77
596.83 ± 0.5696.15 ± 0.2087.56 ± 0.9397.53 ± 0.5096.03 ± 1.2598.32 ± 0.3399.22 ± 0.20
698.25 ± 0.7998.65 ± 0.1898.86 ± 0.2491.02 ± 0.2099.78 ± 0.1199.99 ± 0.0199.92 ± 0.03
799.05 ± 0.1796.12 ± 0.3298.10 ± 0.3196.68 ± 0.5898.70 ± 0.32100 ± 099.95 ± 0.03
896.77 ± 2.8794.22 ± 0.2883.66 ± 0.6981.53 ± 0.9696.06 ± 0.5897.23 ± 0.8997.79 ± 0.36
998.65 ± 0.3698.36 ± 0.1397.29 ± 0.1599.18 ± 0.3398.82 ± 0.5099.80 ± 0.1799.44 ± 0.29
1061.48 ± 11.9099.78 ± 0.0992.75 ± 0.5397.87 ± 0.4698.45 ± 0.3699.30 ± 0.3099.82 ± 0.13
110.11 ± 0.3490.69 ± 0.2489.96 ± 0.9792.08 ± 1.6363.55 ± 0.3093.07 ± 0.5496.47 ± 0.57
1260.40 ± 7.6891.99 ± 0.1498.71 ± 0.2691.26 ± 0.8295.99 ± 0.8097.61 ± 1.1599.12 ± 0.47
1338.82 ± 39.9790.97 ± 0.5498.76 ± 0.1495.82 ± 0.9294.07 ± 0.4895.49 ± 1.2896.98 ± 0.49
1487.97 ± 1.7193.90 ± 0.2994.90 ± 0.3596.50 ± 0.9695.95 ± 0.7897.69 ± 0.4598.57 ± 0.45
152.91 ± 5.8892.06 ± 0.1975.98 ± 1.0578.98 ± 1.6897.93 ± 0.2897.67 ± 0.3098.81 ± 0.55
1653.00 ± 3.1699.23 ± 0.1594.23 ± 1.1481.08 ± 2.1097.34 ± 0.8098.73 ± 0.6397.79 ± 0.47
OA (%)75.25 ± 1.2395.79 ± 0.0791.05 ± 0.2490.39 ± 0.3594.13 ± 0.9498.49 ± 0.0198.82 ± 0.11
AA (%)71.48 ± 3.0195.66 ± 0.0694.09 ± 0.1992.80 ± 0.3090.61 ± 2.3498.24 ± 0.2998.72 ± 0.08
Kappa (%)71.92 ± 1.4695.31 ± 0.0790.04 ± 0.2789.30 ± 0.3993.46 ± 1.0498.69 ± 0.1298.80 ± 0.01
Table 12. Training and testing time on the three data sets.
Table 12. Training and testing time on the three data sets.
AlgorithmsTimeIndian PinesPaviaUSalinas
SVM−RBFTrain(s)0.00980.00800.0043
Test(s)0.13660.40650.5430
Two−CNNTrain(s)2023.644306.7810,957.61
Test(s)22.4748.75140.63
3D−CNNTrain(s)448.001565.822559.34
Test(s)5.2715.1221.38
DCGANTrain(s)7019.6914,053.5110,273.40
Test(s)7.2825.6331.62
DBGANTrain(s)3485.747922.476551.87
Test(s)4.5713.7318.06
CVAEGANTrain(s)4394.088514.976988.92
Test(s)3.7016.4719.72
ProposedTrain(s)14,101.2825,665.4217,524.97
Test(s)10.2643.10102.79
Table 13. MSE when self-attention, WGAN-GP, and additional virtual samples are removed.
Table 13. MSE when self-attention, WGAN-GP, and additional virtual samples are removed.
AlgorithmsIndian PinesPaviaUSalinas
SACVAEGAN without self-attention0.0210.0480.013
SACVAEGAN without wgan-gp0.0220.0520.023
SACVAEGAN without additional virtual samples0.0200.0460.021
SACVAEGAN0.0170.0430.011
Table 14. SID when self-attention, WGAN-GP, and additional virtual samples are removed.
Table 14. SID when self-attention, WGAN-GP, and additional virtual samples are removed.
AlgorithmsIndian PinesPaviaUSalinas
SACVAEGAN without self-attention0.0190.0410.003
SACVAEGAN without wgan-gp0.0120.0210.028
SACVAEGAN without additional virtual samples0.0110.0350.011
SACVAEGAN0.0110.0140.002
Table 15. OA when self-attention, WGAN-GP, and additional virtual samples are removed.
Table 15. OA when self-attention, WGAN-GP, and additional virtual samples are removed.
AlgorithmsIndian PinesPaviaUSalinas
SACVAEGAN without self-attention95.3798.5298.38
SACVAEGAN without wgan-gp95.6199.0198.67
SACVAEGAN without additional virtual samples95.4398.9998.64
SACVAEGAN95.9499.3298.82
Table 16. FID value of the model with or without WGAN-GP.
Table 16. FID value of the model with or without WGAN-GP.
AlgorithmsIndian PinesPaviaUSalinas
SACVAEGAN10.211.782.37
SACVAEGAN without wgan-gp12.3512.6211.83
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, Z.; Tong, L.; Qian, B.; Yu, J.; Xiao, C. Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification. Remote Sens. 2021, 13, 3316. https://doi.org/10.3390/rs13163316

AMA Style

Chen Z, Tong L, Qian B, Yu J, Xiao C. Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification. Remote Sensing. 2021; 13(16):3316. https://doi.org/10.3390/rs13163316

Chicago/Turabian Style

Chen, Zhitao, Lei Tong, Bin Qian, Jing Yu, and Chuangbai Xiao. 2021. "Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification" Remote Sensing 13, no. 16: 3316. https://doi.org/10.3390/rs13163316

APA Style

Chen, Z., Tong, L., Qian, B., Yu, J., & Xiao, C. (2021). Self-Attention-Based Conditional Variational Auto-Encoder Generative Adversarial Networks for Hyperspectral Classification. Remote Sensing, 13(16), 3316. https://doi.org/10.3390/rs13163316

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop