Next Article in Journal
SDGH-Net: Ship Detection in Optical Remote Sensing Images Based on Gaussian Heatmap Regression
Next Article in Special Issue
Hyperspectral Image Classification with Localized Graph Convolutional Filtering
Previous Article in Journal
Vegetation Cover Change and Its Attribution in China from 2001 to 2018
Previous Article in Special Issue
Multiscale Adjacent Superpixel-Based Extended Multi-Attribute Profiles Embedded Multiple Kernel Learning Method for Hyperspectral Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial-Spectral Transformer for Hyperspectral Image Classification

1
School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin 150001, China
2
Shanghai Jiao Tong University, Shanghai 201100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(3), 498; https://doi.org/10.3390/rs13030498
Submission received: 3 December 2020 / Revised: 23 January 2021 / Accepted: 27 January 2021 / Published: 30 January 2021

Abstract

:
Recently, a great many deep convolutional neural network (CNN)-based methods have been proposed for hyperspectral image (HSI) classification. Although the proposed CNN-based methods have the advantages of spatial feature extraction, they are difficult to handle the sequential data with and CNNs are not good at modeling the long-range dependencies. However, the spectra of HSI are a kind of sequential data, and HSI usually contains hundreds of bands. Therefore, it is difficult for CNNs to handle HSI processing well. On the other hand, the Transformer model, which is based on an attention mechanism, has proved its advantages in processing sequential data. To address the issue of capturing relationships of sequential spectra in HSI in a long distance, in this study, Transformer is investigated for HSI classification. Specifically, in this study, a new classification framework titled spatial-spectral Transformer (SST) is proposed for HSI classification. In the proposed SST, a well-designed CNN is used to extract the spatial features, and a modified Transformer (a Transformer with dense connection, i.e., DenseTransformer) is proposed to capture sequential spectra relationships, and multilayer perceptron is used to finish the final classification task. Furthermore, dynamic feature augmentation, which aims to alleviate the overfitting problem and therefore generalize the model well, is proposed and added to the SST (SST-FA). In addition, to address the issue of limited training samples in HSI classification, transfer learning is combined with SST, and another classification framework titled transferring-SST (T-SST) is proposed. At last, to mitigate the overfitting problem and improve the classification accuracy, label smoothing is introduced for the T-SST-based classification framework (T-SST-L). The proposed SST, SST-FA, T-SST, and T-SST-L are tested on three widely used hyperspectral datasets. The obtained results reveal that the proposed models provide competitive results compared to the state-of-the-art methods, which shows that the concept of Transformer opens a new window for HSI classification.

Graphical Abstract

1. Introduction

Due to the advances in imaging spectrometry, hyperspectral sensors tend to capture the intensity of reflectance of a given scene with increasingly higher spatial and spectral resolution [1]. The obtained hyperspectral image (HSI) contains both spatial features and a continuous diagnostic spectrum of different objects at the same time [2]. Thus, the obtained abundant information makes HSI useful in many areas including effective measurement of agricultural performance [3], plant diseases detection [4], identification of minerals [5], disease diagnosis and image-guided surgery [6], ecosystem measurement [7], and earth monitoring [8]. To fully use the obtained HSI, many data processing techniques have been explored, such as unmixing, detection, and classification [8].
HSI classification aims to categorize the content of each pixel in the scene [9], which is a basic procedure in applications such as identifying the type of land-cover classes in earth monitoring [10].
A great many supervised methods have been proposed for HSI classification in the last two decades [11]. In the early stage of HSI classification, the HSI classification methods used spectral information only. A typical spectral classifier was introduced in [12], which was based on the support vector machine (SVM). SVM shows its low sensitivity to high dimensionality [13]; therefore, many SVM-based classifiers have been proposed to handle the spectral classification of HSI [14]. Hyperspectral sensors can provide abundant spatial information of the observing scene with the development of imaging technology. It is reasonable to develop spectral-spatial classifiers. Numerous morphological operations have been developed to extract the spatial features of HSI for following spatial-spectral classification, such as morphological profiles (MPs) [15], extended MPs (EMPs) [16], extended multi-attribute profile (EMAP) [17], and extinction profiles (EPs) [18]. However, the aforementioned HSI classifiers are not deep models [11].
In recent years, deep learning techniques, especially the deep convolutional neural network (CNN), have revolutionized the means of remote sensing data processing. The task of HSI classification is not an exception. In [19], stacked auto-encoder was introduced as a deep model for HSI feature extraction and classification. After that, several deep learning models such as the deep belief network [20], CNN [21,22], recurrent neural network [23,24], generative adversarial network [25,26], and capsule network [27,28] were investigated for HSI classification and obtained good classification performance.
Because of its local connection and shared weights, which makes it effective to capture local correlations, CNN is quite useful for image processing, including HSI classification. According to the input information of models, CNN-based HSI classification methods can be divided into three types: spectral CNN, spatial CNN, and spectral-spatial CNN. Spectral CNN-based HSI classification receives the pixel vector as input and uses CNN to classify the HSI only in the spectral domain. For example, Hu et al. proposed 1-D CNN with five convolutional layers to extract the spectral features of HSI [29]. Moreover, an interesting work was proposed in [30], which used CNN to extract pixel-pair features for HSI classification and obtained good classification performance.
Spatial CNN-based methods are the second type of CNN-based HSI classification methods. In addition to spectral information, the obtained HSI contains abundant spatial information; therefore, it is reasonable to use spatial CNN (2-D CNN) to extract the spatial features of HSI. Most of existing spatial CNN-based HSI classification methods were conducted on one or several principal components. For example, in [31], the cropped spatial patches of pixel-centered neighbors, which belong to the first principal component of his, were used to train a 2-D CNN for HSI classification.
Spectral-spatial CNN-based methods are the third type of CNN-based HSI classification methods, which aim for joint exploitation of spectral and spatial HSI features in a unified framework. Since the input of HSI is a cubic tensor, 3-D convolution was used for HSI classification [32]. For example, in [33], He et al. proposed a 3D deep CNN to jointly extracted spatial and spectral features by computing multiscale features. In [34], the 3-D convolutional layer and batch normalization layer were utilized to extract spectral-spatial information and regularize the model, respectively. Due to the good classification performance obtained by CNN-based methods, CNN has become the de-facto standard for HSI classification in recent years.
Existing CNN models for HSI classification have achieved state-of-the-art performance; however, there are still several limitations. First, some information of input HSI is ignored and is not well explored in CNN-based methods. CNN is a vector-based method, which considers the inputs to be a collection of pixel vectors [35]. For HSI, it intrinsically has a sequence-based data structure in the spectral domain. Therefore, using CNN can lead to information loss when dealing with hyperspectral pixel vectors [36]. Second, learning long-range sequential dependence back and forth between distant positions of bands is difficult. Since convolutional operations process a local neighborhood, the receptive field of CNN is strictly restricted by its kernel size and the number of layers, which has made it less advantageous in capturing long-range dependencies of input data [37]. Therefore, it is difficult to learn the long-range dependencies of HSI, which usually contain hundreds of spectral bands.
Very recently, a model called Transformer [38], which is based on the self-attention mechanism [39], has been proposed for natural language processing. Transformer uses attention to draw global dependency within a sequence of input. For deep learning models including Transformer, there is a common problem of vanishing-gradient, which hampers the convergence in the training procedure [40]. To alleviate the vanishing-gradient problem, a new type of Transformer, which uses dense connection to strengthen feature propagation, titled DenseTransformer, is proposed in this study.
Furthermore, two classification frameworks based on DenseTransformer are proposed for HSI classification. The first classification framework combines CNN, DenseTransformer, and multilayer perceptron. In the second classification framework, transfer learning strategy is combined with Transformer to improve the HSI classification performance with limited training samples.
The main contributions of this study are summarized as follows.
(1) A modified Transformer titled DenseTransformer is proposed, which uses dense connection to alleviate the vanishing-gradient problem in Transformer.
(2) A new classification framework, i.e., spatial-spectral Transformer (SST), is proposed for HSI classification, which combines CNN, DenseTransformer, and multilayer perceptron (MLP). In the proposed SST, a well-designed CNN is used to extract the spatial features of HSI, and the proposed DenseTransformer is used to capture sequential spectra relationships of HSI, and the MLP is used to finish the classification task.
(3) Furthermore, dynamic feature augmentation, which aims to alleviate the overfitting problem and therefore generalize the model well, is proposed and added to the SST to form a new HSI classification method (i.e., SST-FA).
(4) Another new classification framework, i.e., transferring spatial-spectral Transformer (T-SST), is proposed to further improve the performance of HSI classification. The proposed T-SST uses the pre-trained VGG-like model on a large dataset as the initialization of the used CNN in SST; therefore, it enhanced the HSI classification accuracy with limited training samples.
(5) At last, label smoothing is introduced into Transformer-based classification. Label smoothing is combined with T-SST to formulate a new HSI classification method titled T-SST-L.
The rest of this paper is organized as follows. The proposed SST and transferring SST for HSI classification are presented in Section 2 and Section 3, respectively. The experimental results and discussions are reported in Section 4. Section 5 presents the conclusion of this study.

2. Spatial-Spectral Transformer for Hyperspectral Image Classification

The framework of the proposed SST for HSI classification is shown in Figure 1. In general, there are three parts in the classification method: CNN-based spatial feature extraction, modified Transformer-based spatial-spectral feature extraction, and MLP-based classification.
Firstly, for each band of HSI, a 2D patch, which contains the neighboring pixels of the pixel to be classified, is selected as input. There are n (i.e., the number of bands of HSI) patches for a training sample. After that, a well-designed CNN is used to extract the features of each 2D patch, and then the extracted features are sent to Transformer. Then, the modified Transformer is used to obtain the relationship of the sequential spatial features. At last, the obtained spatial-spectral features are used to get the classification result.

2.1. CNN-Based HSI Spatial Feature Extraction

CNN has powerful capability to extract spatial features of image, and it is widely used for image processing such as classification, detection, and segmentation. For HSI, it contains abundant spatial information. CNN is used in this study to effectively extract the spatial features of HSI.
CNN contains a wide range of different architectures. How to choose a proper architecture is important. Although HSI is a 3-D cube, a 3-D CNN is not used in this study. Instead, a 2-D CNN is used in this classification framework. Furthermore, we use 2-D CNN separately to extract the features of each band in his, and the extracted features are fed into a Transformer.
VGGNet is a simple but effective model, which considers the depth of appropriate layers and does not increase the total number of parameters compared to previous AlexNet [41]. Therefore, we used VGG-like architecture. The original VGG contains 16 layers, which includes 13 convolutional layers and three fully connected layers. Each convolutional layer is followed by BN layer and ReLU operation, and the max pooling layer is added after the second, fourth, seventh, tenth, and 13th convolutional layer. Possibly, the usage of the whole 16 layers is not a good choice for HSI spatial feature extraction. How to design a proper CNN architecture is a key point for a successful HSI classifier. In the experimental part, we designed a VGG-like deep CNN for spatial feature extraction of HSI.

2.2. Spectral-Spatial Transformer for HSI Classification

CNN uses local connection to extract neighboring features of inputs. HSI usually contains hundreds of bands; therefore, it is difficult for CNN to obtain spectral relationships in a long distance. The self-attention mechanism can obtain the relationship of every two bands. For example, Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) contains 224 bands. Using self-attention, a matrix with the shape of 224 × 224 can be obtained through the learning procedure. Each element in the matrix represents the relationship between the two bands.
As shown in Figure 1, the extracted features by CNN in the last part are then sent to the Transformer to learn long-range dependencies, which mainly contains three elements.
The first element is called the position embedding, which aims to capture the positional information of the different bands. This element modifies the output features of the last part, which depends on its positions without changing these full features. In this paper, one dimensional position embedding is utilized, which considers the input features as a sequence of different bands. These generated positional embeddings are added to the features, then sent together to the next element. In addition, a learnable position embedding is prepared (i.e., number zero), whose state serves as the whole representations of the band. This learnable position embedding combines with the third part to finish the classification task.
The second element is the Transformer encoder, which is the core part of our model. The Transformer encoder contains a total of d encoder blocks, and each encoder block consists of a multi-head attention and a MLP layer, coupled with layer normalization and residual connection. In each encoder block, a normalization layer is added before each multi-head attention and an MLP layer, and residual connections are designed after each multi-head attention and MLP layer.
Let denote the number of n bands of HSI ( b 1 ,   b 2 , b n ) by B R n × d m o d e l , where d m o d e l indicates the dimension of the extracted features by CNN. The Transformer encoder aims to capture the interaction among all   n bands of HSI by encoding each band in terms of the global contextual information. Specifically, three learnable weight matrices including queries (i.e., Q ), keys (i.e., K ) of dimension d k , and values (i.e., V ) of dimension d v are defined. The dot products are applied to compute the query with all keys, and then the softmax function is used to compute the weights on the values. The output of attention is defined as follows:
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V ,  
where d k is the dimension of K .
It is beneficial to project the queries, keys, and values several times (i.e., h times) with different and learned projections, and then these results are concatenated. This process is named the multi-head attention. Each result of those parallel computations of attention is called a head.
M u l t i H e a d ( Q , K , V ) = c o n c a t ( h e a d 1 , , h e a d h ) W O ,  
where h e a d i =   Attention ( Q W i Q , K W i K , V W i V ) , W i Q R d m o d e l × d q , W i K R d m o d e l × d k , W i V R d m o d e l × d v , and W O R h × d v × d m o d e l are parameter matrices.
After that, the weights extracted by the multi-head attention mechanism are sent to the MLP layer, whose output features are 512 dimensions. Here, MLP is constituted by two fully connected layers with a nonlinearity named the Gaussian error linear unit (GELU) activation between. Here, GELU is the variant of the ReLU, which can be defined as follows [42]:
G E L U = x Φ ( x ) = x · 1 2 [ 1 + erf ( x / 2 ) ] ,  
where Φ ( x ) indicates the standard Gaussian cumulative distribution function, erf ( x ) = 0 x e t 2 d t .
Before the MLP layer, there is always a normalization layer [43], which not only reduces the training time by normalizing neurons, but also alleviates the vanishing or exploding gradient problem. For the i th summed input at the   l -th layer a i l , the normalization layer represents as follows:
a i l = g l σ l · ( a i l μ l ) + b ,
where a i l is normalized summed input, and μ l and σ l represent expectation and variance at l th layer, respectively. g l and b indicate the learned scale parameter and shift parameter, respectively.
For a deep learning model, there is a common problem titled vanishing-gradient, which hampers the convergence in the training of the deep Transformer model [40]. To alleviate the vanishing-gradient and strengthen feature propagation, short-cut connection is used to form a DenseTransformer. Specially, each layer in DenseTransformer has connections of the previous layers in the DenseTransformer. For a traditional Transformer H L ( . ) with L layers, there are L connections, and a DenseTransformer has L ( L + 1 ) 2 connections. The DenseTransformer encourages feature reuse and therefore mitigates the vanishing-gradient problem. Figure 2 shows the proposed DenseTransformer when L = 3.
The proposed DenseTransformer consists of L layers, considering a single e L indicates the output of the traditional Transformer H L ( . ) at the L -th layer. Consequently, the L -th layer of the proposed DenseTransformer receives the weights produced by the previous preceding layers e 0 , e 1 , , e L 1 , which can be defined as follows:
e L = H L ( [ e 0 , e 1 , , e L 1 ] ) ,  
D e n s e M u l t i H e a d ( Q , K , V ,   e L ) = c o n c a t ( h e a d 1 , , h e a d h ) W O ,
where h e a d i =   Attention ( Q W i Q , K W i K , V W i V ,   ( e L ) i ) , W i Q R d m o d e l × d q , W i K R d m o d e l × d k , W i V R d m o d e l × d v , and W O R h × d v × d m o d e l are parameter matrices.
The third part of SST is MLP. The architecture of MLP includes two fully connected layers with a GELU operation, where the last fully connected layer (i.e., the softmax layer) aims to generate the final results for HSI classification. In softmax, for an input vector R , the probability that the input belongs to category i can be estimated as follows:
P ( Y = i | R ,   W ,   b ) = s ( W R + b ) = e W i R + b i j e W j R + b j ,  
where W   and b are weights and biases of the Softmax layer, respectively.
In the MLP, the size of the input-layer is set to be the same as the size of the output-layer of the Transformer, and the size of the output-layer is set to be the same as the total number of classes. Softmax ensures the activation of each output unit sums to 1. Therefore, the output can be deemed as a set of conditional probabilities.

2.3. Dynamic Feature Augmentation

Due to the proposed, SST is often susceptible to overfitting and therefore requires proper regularization to generalize well. In this subsection, a simple regularization technique called dynamic feature augmentation is proposed, which is implemented by randomly masking out features during training. Then, SST is combined with feature augmentation to form a new HSI classifier (i.e., SST-FA), which improves the robustness and overall classification performance of SST.
Specially, the dimension of the spatial features extracted by the VGG is high (i.e., 512-dimension), which is easy to overfit for the Transformer model. Here, a coordinate is first randomly selected in the features, then, a mask is placed around the coordinate, which decides how many features are set to zero. Note that the coordinate dynamically changes w.r.t. epochs during training, which ensures the Transformer model receives different features. The proposed SST-FA is not only easy to implement, but also able to further improve the Transformer model performance.

3. Heterogeneous Transferring Spatial-Spectral Transformer for Hyperspectral Image Classification

The collection of training samples is not only expensive but also time-consuming. Therefore, limited training samples are a common issue in HSI classification. To address this issue, transfer learning is combined with SST in this study. Transfer learning is a technique that extracts the knowledge from the source domain and transfers it to the target domain [44]. For example, in CNN-based transfer learning, the learned weights on the source domain can be used to initialize the net of the target domain. Therefore, when it is properly used, transfer learning can improve the classification performance of the target task if the number of training samples is limited.
To further improve classification performance of the proposed SST, transferring SST (T-SST) is proposed in this section. Figure 3 shows the framework of the proposed T-SST for HSI classification. In general, there are three parts in the classification method: transferring CNN-based spatial feature extraction, Transformer-based spatial-spectral feature extraction, and MLP-based classification.

3.1. Heterogeneous Mapping of Two Datasets

There is a problem of simply using transfer learning for HSI classification, due to the fact that the large-scale dataset (i.e., the source dataset) has three channels, but HSI (i.e., the target dataset) contains hundreds of channels. To solve the problem caused by heterogeneous transfer learning, a mapping layer is used to handle the issue of different number of channels (i.e., bands) of the two datasets.
The pre-trained model on the large-scale ImageNet dataset has three channels of input (i.e., R, G, and B), but CNN in T-SST receives one band as input.
Let O R W × H be the input of CNN, in which W × H represents the weight and height of a 2D patch. O R W × H × 3 is the mapped data for subsequent processing, and α R 3 × 1 . Therefore,
O = O × α .
There are three learnable parameters in the heterogeneous mapping. The mapping operation is combined with subsequent CNN to form an end-to-end learning system.

3.2. The Proposed T-SST for HSI Classification

Transfer learning is a technique that aims at extracting knowledge from the source domain and applying it to the target domain [44]. The learned knowledge from the source task is used to improve the performance of the target task. In deep learning-based transfer learning, deep models can learn a lot of knowledge from a large dataset such as ImageNet, and the learnt knowledge can be transferred to a new task such as HSI classification. Therefore, the proper usage of transfer learning may reduce the number of necessary training samples.
Many previous studies proved that the learnt weights in CNN of the original domain can be re-used in the new task [45]. For an image classification task, the first several layers usually extract low-level features (i.e., blobs, corners, and edges), and the low-level features are usually common in image classification tasks. Due to the similarity tasks between the ImageNet and HSI classification, the transfer learning step can be facilitated by fine-tuning on HSI classification task. Specifically, the learned weights of VGGNet on the ImageNet dataset can be utilized to initialize the network of the HSI classification and then fine-tune the weights on an HSI classification task.
Here, a new classification framework titled T-SST is proposed for HSI classification, which is a combination of transferred VGGNet, modified Transformer (i.e., DenseTransformer), and MLP. In T-SST, VGGNet with 16 layers is used, which was trained on the ImageNet dataset, and the well-trained weights of all the convolutional layers from the source task are transferred to our target task. Then, these initialized weights are fine-tuned on the HSI dataset.
Using transferred VGGNet, more robust and discriminant features can be extracted compared with the original VGGNet, which is useful for the following processing. The obtained features using transferred VGGNet are used as inputs of Transformer. Specially, a 2D patch, which contains the neighboring pixels in a band of HSI, is an input of transferred VGGNet. VGGNet uses all the convolutional layers to extract the features of the input, then the obtained features are fed into the DenseTransformer. The following MLP is used to obtain the final classification results.

3.3. The Proposed T-SST-L for HSI Classification

Without sufficient training samples, the model faces a problem of “overfitting”, which means that the classification accuracy on test data will be low. This problem is expected when T-SST is applied to HSI classification, because it is a common issue that there are only limited training samples in real application. To address the overfitting issue in T-SST, label smoothing is introduced.
In classification, each training sample x has the corresponding label y { 1 , 2 , , C } . C is the number of classes. Here, we use a C -dimensional one-hot vector y to represent the label of training sample x ,
y k = δ k , y ,  
where k = 1 ,   2 , , C , δ k , y represents the discrete Dirac delta function, which equals 1 for k = y and 0 otherwise.
However, work in [46] has shown that, if we assign all ground truth labels as “hard labels” (i.e., the δ k , y ), the model will struggle with many efforts to push the predicted distribution of labels towards the hard label. Moreover, this can be effectively relieved if the labels are properly smoothed, i.e., assigned tiny probability mass on the zeros in δ k , y . Intuitively, this happens because the model becomes too confident about its predictions. Thus, in this paper, a mechanism called label smoothing for encouraging the model to be less confident is introduced in this paper to achieve better performance. Label smoothing changes the original label y k to y k , which can be defined as follows:
y k = ( 1 ε ) δ k , y + ( 1 δ k , y ) ε C 1 ,
where y k mixes the label of training sample x and the fixed uniform distribution of the number of C 1 classes; ε is the smoothing factor [47].
By reducing the model to learn the full probability label of each training sample, the label smoothing mechanism can mitigate the overfitting problem and increase the generalization ability of the model in a simple form.

4. Experimental Results

4.1. Hyperspectral Datasets

In this study, the performance of proposed methods is evaluated on three public datasets, including the Salinas, Pavia University (Pavia), and Indian Pines datasets. Table 1 reports the information of all the datasets including the sensor, number of bands, spatial resolution, pixel size dimension, number of classes, and year of data acquisition. The descriptions of all the datasets are summarized below.
For the Salinas dataset, it was collected by the AVIRIS over Salinas Valley, CA, USA, in 1998. After removing 20 bands of low signal to noise ratio (SNR), 204 bands were used in the experiments.
There are 512 × 217 pixels with 3.7-m spatial resolution included in this hyperspectral image. The common 16 classes are labeled in the ground truth. The false-color composite image, the available ground-truth map, and the Scale bar are shown in Figure 4.
For the Pavia dataset, it was obtained by the Reflective Optics Spectrographic Imaging System (ROSIS) sensor over an urban scene by Pavia University, Italy, in 2001. This dataset has a size of 610 × 340 pixels with 1.3 m spatial resolution, and the spectral ranges from 0.43 to 0.86 μm. Due to the SNR and water absorption, 12 bands were removed, and the remaining 103 bands were used in the experiments. There are nine urban classes to discriminate. Figure 5 shows the false-color composite image, the available ground-truth map, and the scale bar.
For the Indian Pines dataset, it was captured by the AVIRIS sensor over the Indian Pines region in Northwestern Indiana, 1992. The spatial size of it is 145 × 145 with the spatial resolution of 20 m. The number of spectral bands is 200 with the wavelengths from 0.4 to 2.5 μm after discarding 20 water absorption bands. Sixteen ground truth classes are labeled in the available ground truth. The false-color composite image, the available ground-truth map, and the scale bar are shown in Figure 6.
Each dataset is divided into three subsets: training set, validate set, and test set. The training set consists of 200 labeled samples for model training, which are randomly selected from all the labeled samples, the validate set includes 50 labeled samples for guiding the model, and the remains are used as the test set.
The input HSI datasets are normalized into [−0.5, 0.5]. According to Figure 1, to capture the relationships in a long distance in the spectral domain, the HSI data cube consists of n bands, where the order of bands follows the spectral order. The neighborhood pixels of each sample are set to 33 × 33, and then these samples are fed into VGGNet.

4.2. Training Details

The VGGNet with 16 layers are adopted in the experiments, which contains 13 convolutional layers and three fully connected layers. The characteristic of VGGNet is mainly adopted small convolutional filters with 3 × 3 size. Moreover, the 13 convolutional layers can be divided into five groups, and each group contains two or three convolutional layers, which are followed by a max pooling layer. The architecture of VGGNet in SST is similar to VGGNet; to reduce the overfitting for HSI classification, several convolutional layers are ignored, the first three convolutional layer groups reduce one convolutional layer, and the fourth convolutional layer reduces the first two convolutional layers. In addition, for the T-SST, the architecture of VGGNet adopts all the convolutional layers, which are used as initialized weights. For the Pavia and Indian pines datasets, dropout is added. Then, HSI training samples are exploited to fine-tune the VGGNet by the back-propagation algorithm. Due to the input of VGGNet in each band, a mapping layer is designed, whose input is each band and output is three features. Then these features are sent to the VGGNet to extract the discriminative features.
During the training procedure, the mini-batch algorithm is adopted, which is set to 128 for all the datasets [48]. For the SST, the initial learning rate is set to 8 × 10−5 for the Salinas dataset, and 9 × 10−5 for the Pavia and Indian Pines datasets, and the learning rate is reduced by 0.9 for each epoch. In the experiments, the small learning rate is found to be suitable for the SST for HSI classification. Additionally, for the T-SST, the learning rate is set to 3 × 10−4, 9 × 10−5, and 1 × 10−4 for the Salinas, Pavia, and Indian Pines datasets, respectively. The learning rate is reduced by 0.7, 0.9, and 0.9 for each epoch for the Salinas, Pavia, and Indian Pines datasets, respectively. Additionally, the training epoch is set to 150 for the Salinas dataset, and for the Pavia and Indian Pines datasets, the training epoch is set to 80. Furthermore, the overall accuracies (OA), average (AA) accuracies, and kappa coefficient (K) are considered to evaluate the performance of different methods.

4.3. Parameter Analysis

To give a comprehensive study of the spatial-spectral Transformer, some key parameters involved in the Transformer are analyzed in this section, including the number of attention heads, the depth of the Transformer encoder, and the smoothing factor ε of the proposed T-SST-L. For the number of attention heads and the depth of the Transformer encoder, they not only influence the robustness of the model, but also affect the complexity of the model. With the increment of the model depth, it is easy for the model to encounter the overfitting problem. For the smoothing factor ε , the value of ε could influence the performance of the model. Thus, the optimal parameter settings of these parameters are needed to investigate.
To analyze the influences of these parameters to the model, other parameters are fixed; 200 training samples are used for searching optimal parameters. Figure 7 shows the analysis results evaluated by OA (%) on the Salinas, Pavia, and Indian Pines datasets, respectively. For searching for the optimal number of attention heads, one, two, four, eight, and 16 of attention heads are chosen. This result is shown in Figure 7a: it can be seen that the best number of attention heads is two for all the datasets. For the depth of the Transformer encoder, the depths ranging from one to five are searched. Figure 7b shows that the best depth is two for all the datasets: it can be concluded that the lacking depths may cause incomplete information, while for the too deep models, the accuracies are decreased, due to the large amounts of parameters that are needed to train. In the experiments, according to these results, the number of attention heads and depths of the Transformer encoder are set to two for all the datasets to lead to better classification results.
ε is the smoothing factor of T-SST-L; to validate the influences of T-SST-L with different values of ε , the grid search method is utilized to search the optimal value of 𝜀 varying from 0.2 to 0.9. The OA of different values of   ε is shown in Figure 8. As can be seen, the OAs of different values of 𝜀 are fluctuant, but the proposed T-SST-L obtains the best result when the value of 𝜀 is set to 0.9 on the three datasets. Therefore, in all experiments, the value of 𝜀 is set to 0.9 for all datasets to obtain the best performance for HSI classification.

4.4. The Classification Results of SST and SST-FA for HSI Classification

In this section, the proposed SST and SST-FA are verified by using several comparison experiments, including the traditional methods (i.e., RBF-SVM and EMP-SVM) and the classical CNN related methods (i.e., CNN, SSRN, and VGG). For RBF-SVM, the radial basis function is adopted as the kernel, and a grid search method is used for finding the best value of C and γ , which are in the exponentially growing sequence { 10 3 , 10 2 ,…, 10 3 }. The best parameters   C and γ are obtained using five-fold cross-validation. EMP-SVM combines EMP with SVM; for EMP, a disk-shape structure element with an increasing size from two to eight is designed for the opening and closing operations in EMP to extract features. The architectures of CNN and SSRN are implemented following the settings described in [34,49], respectively.
The experimental results of SST and SST-FA are reported in Table 2, Table 3 and Table 4. As can be seen, the values of OA, AA, and kappa achieve by the proposed SST-FA are the best, which reach 94.94%, 93.37%, and 88.98% on the Salinas, Pavia, Indian Pines, respectively.
All the experimental results demonstrate that SST-FA reaches the best performance on all the HSI datasets, which has advantages in alleviating the overfitting. For the proposed SST, take the Salinas dataset as an example, compared to the traditional methods, the OA is 11.33% and 6.83% points higher than that of RBF-SVM and EMP-SVM, respectively; the AA is 7.65% and 4.46% points better, respectively, and the kappa is 12.66% and 7.58% points higher. Besides, compared to CNN, OA of SST is improved by 6.02%, 1.33%, and 1.96% on the Salinas, Pavia, and Indian Pines datasets, respectively. In addition, compared to CNN-based methods including SSRN and VGG, for the Indian Pines dataset, the accuracy of the proposed SST achieves 88.77%, which increases of 5.56% and 1.97%, respectively. SST also offers improvement on the Salinas and Pavia datasets. Figure 9 shows learning curves of SST including loss, accuracy of training, and validate samples on the three datasets. The experimental results demonstrate that the proposed SST has the advantages in extracting sequential information of HSI.

4.5. The Classification Results of the Proposed T-SST and T-SST-L for HSI Classification

In this section, the experimental results of the proposed T-SST and T-SST-L are presented to test the performance of HSI classification, which use the pre-trained VGGNet on a large dataset as the initialized weights of VGGNet. To further verify that the proposed T-SST and T-SST-L are superior to other methods for HSI classification, EMP-random forest (RF), EMP-CNN, VGG, and T-CNN are selected for comparison. For EMP-RF, the detailed information about EMP is the same as the previous settings. Then, the features extracted by EMP are fed into the RF classifier with 200 decision trees [50]. Additionally, EMP-CNN combines EMP with CNN and is implemented for spectral-spatial classification. Specifically, the architecture design of EMP-CNN is similar to CNN [49]. Moreover, we adopt a comparison method named VGG, whose architecture follows all the convolutional layers in VGGNet; after that, a fully connected layer is added for HSI classification. In addition, to demonstrate the proposed Transformer method with transfer learning is effective, CNN with transfer learning (T-CNN) is also utilized. Three bands are randomly selected from all bands, then all the VGGNet weights of the first seven convolutional layers are used to initialize T-CNN to finish HSI classification task.
The results of the proposed T-SST and T-SST-L are reported in Table 5, Table 6 and Table 7. The proposed T-SST-L achieves competitive results as compared to state-of-the-art well-designed networks. The OA of the proposed T-SST-L reaches 96.83%, 93.73%, and 91.20% on Salinas, Pavia, and Indian Pines datasets, respectively. In addition, it can be observed that the proposed T-SST is superior to other existing methods on all the datasets. Specifically, the T-SST outperforms the EMP-RF by 4.16%, 5.08%, and 4.96% in terms of OA on Salinas, Pavia, and Indian Pines datasets, respectively. Additionally, the accuracy obtained by the proposed T-SST on the Salinas dataset is 3.09% better than that of the EMP-CNN, while, for the Indian Pines dataset, it is 2.84%. In addition, compared to T-CNN, the proposed T-SST achieves about 2% improvements on the Salinas dataset in terms of OA and K . Furthermore, compared to the proposed T-SST, the proposed T-SST-L increases accuracies by 1.03% and 1.14% on Salinas and Indian Pines datasets, respectively. It demonstrates that label smoothing is an effective method to prevent the overfitting problem.

4.6. Analysis of Transformer Encoder Representation of the Proposed T-SST

In this subsection, to see how the proposed T-SST captures long-distance dependency relations, we analyze the Transformer encoder representation by visualizing the normalized attention weights. Figure 10, Figure 11 and Figure 12 display normalized attention weights between bands on the three hyperspectral datasets.
Since there are hundreds of bands in HSI, it is difficult to make complete visualization for each of them. Therefore, the first band, the middle band, and the last band are chosen to illustrate the long-range dependency relations between the chosen band and another band. Specially, band 1, band 100, and band 204 are chosen on the Salinas dataset; band 1, band 50, and band 103 are chosen on the Pavia dataset; and band 1, band 100, and band 200 are chosen on the Indian Pines dataset.
As shown in Figure 10, Figure 11 and Figure 12, the values of attention weight strongly fluctuate on the three hyperspectral datasets, and the value of attention weight can be high even if the two bands have a long distance.
Take the Salinas dataset as an example: in Figure 10a, the value of normalized attention weight between band 50 and band 203 is very high, although distance of the two bands is far. The results demonstrate that T-SST tends to capture long-range dependency relations.

4.7. Classification Maps

Here, to fully evaluate the classification results, Figure 13, Figure 14 and Figure 15 display classification maps of different methods from a visual perspective on all the datasets; the methods include EMP-SVM, CNN, SSRN, VGG, and our proposed methods (i.e., SST-FA and T-SST-L). Through comparison, it can be observed that classification maps of EMP-SVM produce more errors for all the datasets, while for the proposed SST and T-SST, there exist fewer noise points. In addition, in Figure 15, compared to other CNN-based methods, for example, in comparison of CNN, SSRN, and VGG (see Figure 15b–d), many pixels are misclassified on the boundary among different classes on the Indian Pines dataset, while the proposed methods are able to classify more classes correctly (i.e., Soybean-clean) and have a clearer distinction. Obviously, SST-FA and T-SST-L produce classification maps with the highest quality compared to other approaches, which demonstrates that the proposed SST-FA and T-SST-L are effective in enhancing the performance of model, respectively.

4.8. Time Consumption

The execution time of different methods for the three HSI datasets with 200 training samples is reported in Table 8. All the experiments are conducted on a computer with an Intel Core i7-10700F processor with 2.9 GHz, 64 GB of DDR4 RAM, an NVIDIA GeForce RTX 3070 graphical processing unit (GPU). For the traditional methods including RBF-SVM, EMP-SVM, and EMP-RF, the processing time is short, but these methods achieve poor performance. In addition, for the CNN and T-CNN, since CNN includes fewer parameters than other competitive deep-learning-based methods and T-CNN only contains three bands for transfer learning, the running time is short. Additionally, compared to CNN, SSRN and VGG take longer time, because SSRN needs more epochs to train the network and VGG contains many 3 × 3 convolutional kernels. For the proposed methods (i.e., SST, T-SST, and T-SST-L), since the proposed methods consider the model of Transformer, the processing times are long.

5. Conclusions

In this study, Transformer is investigated for HSI classification. Specifically, DenseTransformer is proposed, which uses dense connection to alleviate the vanishing-gradient problem in the training of a Transformer.
Moreover, two classification frameworks (i.e., SST and T-SST) have been proposed to handle the task of HSI classification. The proposed methods obtained superior performance in terms of classification accuracy on the three popular HSI datasets.
For the proposed SST-based HSI classification method, it took full advantage of CNN to capture spatial features of a 2D patch and made best of DenseTransformer to capture relationships in a long distance in spectral domain. The used self-attention mechanism considered the intrinsic sequential data structure of a pixel vector of his, and the combination of CNN and DenseTransformer obtained the spectral-spatial discriminate features, which are useful for the following HSI classification.
In addition, DenseTransformer combined with dynamic feature augmentation (i.e., SST-FA) is proposed for alleviating the overfitting problem, and thus it enhances the accuracy of the model in a simple form.
Furthermore, the effectiveness of T-SST has been tested. The proposed T-SST combined transfer learning and SST to further improve the classification performance. To use the pre-trained model on the ImageNet dataset, a heterogeneous mapping layer was designed, which was used to map the model from the source domain (i.e., ImageNet dataset) to target domain (i.e., HSI). The obtained experimental results have shown the usefulness of T-SST for HSI classification.
At last, label smoothing has been proved as a useful regularization technique in Transformer-based HSI classification. The proposed T-SST-L led to high performance compared to SST, SST-FA, T-SST, and other methods.
The proposed SST and T-SST have shown the potential of the proposed DenseTransformer for HSI classification. However, it is in the early stage of Transformer-based HSI classification. In our future work, various improvements of Transformer can be used to open a new widow for HSI accurate classification.

Author Contributions

Conceptualization: Y.C.; methodology: X.H., Y.C., and Z.L.; writing—original draft preparation: X.H., Y.C., and Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of China under the Grant 61971164.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  2. Della, C.; Bekit, A.; Lampe, B.; Chang, C.-I. Hyperspectral image classification via compressive sensing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8290–8303. [Google Scholar] [CrossRef]
  3. Datt, B.; McVicar, T.; Van Niel, T.; Jupp, D.; Pearlman, J. Preprocessing EO-1 Hyperion hyperspectral data to support the application of agricultural indexes. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1246–1259. [Google Scholar] [CrossRef] [Green Version]
  4. Mahlein, A.; Oerke, E.; Steiner, U.; Dehne, H. Recent advances in sensing plant diseases for precision crop protection. Eur. J. Plant Pathol. 2012, 133, 197–209. [Google Scholar] [CrossRef]
  5. Murphy, R.; Schneider, S.; Monteiro, S. Consistency of measurements of wavelength position from hyperspectral imagery: Use of the ferric iron crystal field absorption at ∼900 nm as an indicator of mineralogy. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2843–2857. [Google Scholar] [CrossRef]
  6. Lu, G.; Fei, B. Medical hyperspectral imaging: A review. J. Biomed. Opt. 2014, 19, 010901. [Google Scholar] [CrossRef] [PubMed]
  7. Hestir, E.; Brando, V.; Bresciani, M.; Giardino, C.; Matta, E.; Villa, P.; Dekker, A. Measuring freshwater aquatic ecosystems: The need for a hyperspectral global mapping satellite mission. Remote Sens. Environ. 2015, 167, 181–195. [Google Scholar] [CrossRef] [Green Version]
  8. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in Hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  9. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; Giorgi, A.D.; Fang, L.; Chen, Y.; Chi, M.; et al. New frontiers in spectral-spatial hyperspectral image classification: The latest advances based on mathematical morphology, Markov random fields, segmentation, sparse representation, and deep learning. IEEE Geosci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  10. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature extraction for hyperspectral imagery: The evolution from shallow to deep: Overview and toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  11. Audebert, N.; Le, B.; Lefevre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef] [Green Version]
  12. Melgani, F.; Lorenzo, B. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  13. Gualtieri, J.; Chettri, S. Support vector machines for classification of hyperspectral data. In Proceedings of the IGARSS, Honolulu, HI, USA, 24–28 July 2000; pp. 813–815. [Google Scholar]
  14. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef] [Green Version]
  15. Fauvel, M.; Benediktsson, J.A.; Chanussot, J.; Sveinsson, J.R. Spectral and spatial classification of hyperspectral data using SVMs and morphological profiles. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3804–3814. [Google Scholar] [CrossRef] [Green Version]
  16. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  17. Mura, M.D.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  18. Fang, L.; He, N.; Li, S.; Ghamisi, P.; Benediktsson, J.A. Extinction profiles fusion for hyperspectral images classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1803–1815. [Google Scholar] [CrossRef]
  19. Ma, X.; Wang, H.; Geng, J. Spectral-spatial classification of hyperspectral image based on deep auto-encoder. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  20. Zhong, P.; Gong, Z.; Li, S.; Schönlieb, C.-B. Learning to diversify deep belief networks for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2017, 55, 3516–3530. [Google Scholar] [CrossRef]
  21. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  22. Gong, Z.; Zhong, P.; Yu, Y.; Hu, W.; Li, S. A CNN with multiscale convolution and diversified metric for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3599–3618. [Google Scholar] [CrossRef]
  23. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  24. Zhang, X.; Sun, Y.; Jiang, K.; Li, C.; Jiao, L.; Zhou, H. Spatial sequential recurrent neural networks for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2018, 11, 4141–4155. [Google Scholar] [CrossRef] [Green Version]
  25. Feng, J.; Yu, H.; Wang, L.; Zhang, X.; Jiao, L. Classification of hyperspectral images based on multiclass spatial–spectral generative adversarial networks. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5329–5343. [Google Scholar] [CrossRef]
  26. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative adversarial networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  27. Zhu, K.; Chen, Y.; Ghamisi, P.; Jia, X.; Benediktsson, J.A. Deep convolutional capsule network for hyperspectral image spectral and spectral-spatial classification. Remote Sens. 2019, 11, 223. [Google Scholar] [CrossRef] [Green Version]
  28. Paoletti, M.; Haut, J.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.; Li, J.; Pla, F. Capsule networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2145–2160. [Google Scholar] [CrossRef]
  29. Hu, W.; Huang, Y.; Li, W.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef] [Green Version]
  30. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral image classification using deep pixel-pair features. IEEE Trans. Geosci. Remote Sens. 2016, 55, 844–853. [Google Scholar] [CrossRef]
  31. Haut, J.; Paoletti, M.; Plaza, J.; Plaza, A.; Li, J. Visual attention-driven hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  32. Li, Y.; Zhang, H.; Shen, Q. Spectral-spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef] [Green Version]
  33. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  34. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  35. Tal, L.; Emmanuel, D.; Yoav, G. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies. Trans. Assoc. Comput. Linguist. 2016, 4, 521–535. [Google Scholar]
  36. Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef] [Green Version]
  37. Tang, G.; Mathias, M.; Rio, A.; Sennrich, R. Why self-attention? A targeted evaluation of neural machine translation architectures. arXiv 2018, arXiv:1808.08946v3. Available online: https://arxiv.org/ pdf/1808.08946v3.pdf (accessed on 11 November 2020).
  38. Lin, Z.; Feng, M.; Santos, C.; Yu, M.; Xiang, B.; Zhou, B.; Bengio, Y. A structured self-attentive sentence embedding. arXiv 2017, arXiv:1703.03130. Available online: https://arxiv.org/ pdf/1703.03130.pdf (accessed on 9 March 2017).
  39. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  40. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. J. Mach. Learn. Res. 2010, 9, 249–256. [Google Scholar]
  41. Windrim, L.; Melkumyan, A.; Murphy, R.J.; Chlingaryan, A.; Ramakrishnan, R. Pretraining for hyperspectral convolutional neural network classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2798–2810. [Google Scholar] [CrossRef]
  42. Hendrycks, D.; Gimpel, K. Gaussian error linear units (GELUs). arXiv 2020, arXiv:1606.08415. Available online: https://arxiv.org/ pdf/1606.08415.pdf (accessed on 8 July 2020).
  43. Ba, J.; Kiros, J.; Hinton, G. Layer normalization. arXiv 2016, arXiv:1607.06450. Available online: https://arxiv.org/ pdf/1607.06450v1.pdf (accessed on 21 July 2016).
  44. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? In Proceedings of the Advance in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 3320–3328. [Google Scholar]
  45. Oquab, M.; Bottou, L.; Laptev, I.; Sivic, J. Learning and transferring mid-level image representations using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 20–23 June 2014; pp. 1717–1724. [Google Scholar]
  46. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. Available online: https://arxiv.org/abs/1503.02531 (accessed on 9 March 2015).
  47. Muller, R.; Kornblith, S.; Hinton, G. When does label smoothing help? In Proceedings of the NeurIPS, Vancouver, BC, Canada, 8–14 December 2019; pp. 4696–4705. [Google Scholar]
  48. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2017, arXiv:1412.6980. Available online: https://arxiv.org/abs/1412.6980 (accessed on 30 January 2017).
  49. Chen, Y.; Zhu, L.; Ghamisi, P.; Jia, X.; Li, G.; Tang, L. Hyperspectral images classification with gabor filtering and convolutional neural network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2355–2359. [Google Scholar] [CrossRef]
  50. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The framework of SST for HSI classification.
Figure 1. The framework of SST for HSI classification.
Remotesensing 13 00498 g001
Figure 2. A DenseTransformer.
Figure 2. A DenseTransformer.
Remotesensing 13 00498 g002
Figure 3. The framework of T-SST for HSI classification.
Figure 3. The framework of T-SST for HSI classification.
Remotesensing 13 00498 g003
Figure 4. The Salinas dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Figure 4. The Salinas dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Remotesensing 13 00498 g004
Figure 5. The Pavia dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Figure 5. The Pavia dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Remotesensing 13 00498 g005
Figure 6. The Indian Pines dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Figure 6. The Indian Pines dataset. (a) False-color composite image; (b) ground truth map; (c) scale bar.
Remotesensing 13 00498 g006
Figure 7. Results of SST varying (a) the number of attention heads and (b) the model depth.
Figure 7. Results of SST varying (a) the number of attention heads and (b) the model depth.
Remotesensing 13 00498 g007
Figure 8. Analysis of the performing results of the parameters ε .
Figure 8. Analysis of the performing results of the parameters ε .
Remotesensing 13 00498 g008
Figure 9. Learning curves of SST on the three HSI datasets. (a) Salinas; (b) Pavia; (c) Indian Pines.
Figure 9. Learning curves of SST on the three HSI datasets. (a) Salinas; (b) Pavia; (c) Indian Pines.
Remotesensing 13 00498 g009
Figure 10. Visualization of attention weights for the head 2, layer 2 of T-SST on the Salinas dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 100 and another band; (c) dependency relations between the band 204 and another band.
Figure 10. Visualization of attention weights for the head 2, layer 2 of T-SST on the Salinas dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 100 and another band; (c) dependency relations between the band 204 and another band.
Remotesensing 13 00498 g010
Figure 11. Visualization of attention weights for the head 2, layer 2 of T-SST on the Pavia dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 50 and another band; (c) dependency relations between the band 103 and another band.
Figure 11. Visualization of attention weights for the head 2, layer 2 of T-SST on the Pavia dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 50 and another band; (c) dependency relations between the band 103 and another band.
Remotesensing 13 00498 g011
Figure 12. Visualization of attention weights for the head 2, layer 2 of T-SST on the Indian Pines dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 100 and another band; (c) dependency relations between the band 200 and another band.
Figure 12. Visualization of attention weights for the head 2, layer 2 of T-SST on the Indian Pines dataset. (a) Dependency relations between the band 1 and another band; (b) dependency relations between the band 100 and another band; (c) dependency relations between the band 200 and another band.
Remotesensing 13 00498 g012
Figure 13. Salinas. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Figure 13. Salinas. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Remotesensing 13 00498 g013
Figure 14. Pavia. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Figure 14. Pavia. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Remotesensing 13 00498 g014
Figure 15. Indian Pines. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Figure 15. Indian Pines. (a) False-color composite image. The classification maps using (b) EMP-SVM; (c) CNN; (d) SSRN; (e) VGG; (f) SST-FA; (g) T-SST-L.
Remotesensing 13 00498 g015
Table 1. Details of each HIS dataset.
Table 1. Details of each HIS dataset.
DatasetSensorNumber of BandsSpatial ResolutionSize Number of ClassesYear of Data Acquisition
SalinasAVIRIS2043.7 m512 × 217161998
PaviaROSIS1031.3 m610 × 34092001
Indian PinesAVIRIS20020 m145 × 145161992
Table 2. Test accuracy with different preprocessing methods on the Salinas dataset.
Table 2. Test accuracy with different preprocessing methods on the Salinas dataset.
MethodRBF-SVMEMP-SVMCNNSSRNVGGSSTSST-FA
OA(%)83.09 ± 1.0887.59 ± 2.3988.40 ± 2.1388.73 ± 2.0689.25 ± 5.2094.42 ± 1.5994.94 ± 1.31
AA(%)85.46 ± 2.0688.65 ± 1.9392.48 ± 2.7992.63 ± 1.4790.16 ± 4.8193.11 ± 1.8593.05 ± 0.92
K×10081.07 ± 1.1986.15 ± 2.7588.27 ± 2.7387.51 ± 2.2688.05 ± 5.8293.73 ± 1.8094.32 ± 1.48
Brocoli_green_weeds_194.15 ± 0.5095.28 ± 4.2181.75 ± 5.36100.00 ± 0.0082.46 ± 17.9492.58 ± 11.9092.44 ± 13.63
Brocoli_green_weeds_298.57 ± 0.8998.49 ± 0.2888.89 ± 7.69100.00 ± 0.0088.85 ± 2.1197.50 ± 4.0697.47 ± 3.37
Fallow90.56 ± 0.5070.53 ± 22.5889.35 ± 3.8572.85 ± 13.9892.04 ± 1.6191.47 ± 9.5588.47 ± 12.43
Fallow_rough_plow98.93 ± 0.4099.80 ± 0.1280.01 ± 5.8699.78 ± 2.1589.86 ± 0.3791.98 ± 11.8492.60 ± 7.79
Fallow_smooth95.23 ± 0.6393.34 ± 4.4088.40 ± 2.5798.68 ± 0.5498.32 ± 3.1895.59 ± 8.2393.09 ± 13.63
Stubble99.25 ± 0.9199.36 ± 0.2290.45 ± 6.3899.85 ± 0.8999.97 ± 1.5399.53 ± 0.8499.46 ± 0.90
Celery98.82 ± 0.3397.48 ± 1.8497.96 ± 2.7999.43 ± 0.4396.98 ± 3.3698.65 ± 2.1497.89 ± 3.43
Grapes_untrained78.50 ± 0.5790.09 ± 7.8369.63 ± 9.6262.90 ± 14.9575.66 ± 11.5295.22 ± 2.9095.09 ± 2.78
Soil_vinyard_develop94.11 ± 0.5098.89 ± 0.0889.33 ± 9.7997.87 ± 3.9698.44 ± 0.1399.46 ± 1.1299.70 ± 0.44
Corn_senesced_green_weeds85.56 ± 0.3690.74 ± 2.2385.75 ± 9.0888.08 ± 5.2096.76 ± 13.2998.28 ± 1.8899.69 ± 0.60
Lettuce_romaine_4wk90.63 ± 0.7893.06 ± 2.0288.92 ± 9.7286.54 ± 7.1795.72 ± 28.2599.87 ± 0.1598.47 ± 2.80
Lettuce_romaine_5wk99.48 ± 0.0399.98 ± 0.0482.07 ± 9.3799.41 ± 5.7597.41 ± 1.0189.80 ± 10.9596.08 ± 3.15
Lettuce_romaine_6wk20.08 ± 2.4777.91 ± 39.0382.65 ± 8.3084.74 ± 7.6996.82 ± 37.3797.32 ± 4.9897.47 ± 3.02
Lettuce_romaine_7wk66.29 ± 1.6798.34 ± 0.9985.41 ± 9.2599.33 ± 1.1898.23 ± 1.4096.37 ± 3.8293.81 ± 2.07
Vinyard_untrained59.14 ± 1.0635.73 ± 30.4076.80 ± 18.4893.40 ± 12.1784.13 ± 35.1381.42 ± 16.3786.72 ± 15.08
Vinyard_vertical_trellis66.96 ± 0.7879.35 ± 6.2256.06 ± 12.0298.71 ± 2.5850.94 ± 41.6465.07 ± 14.6960.33 ± 21.01
Table 3. Test accuracy with different preprocessing methods on the Pavia dataset.
Table 3. Test accuracy with different preprocessing methods on the Pavia dataset.
MethodRBF-SVMEMP-SVMCNNSSRNVGGSSTSST-FA
OA(%)80.06 ± 1.5289.43 ± 1.2091.41 ± 1.4491.59 ± 3.5791.72 ± 2.1292.74 ± 1.0893.37 ± 1.96
AA(%)69.48 ± 3.0380.37 ± 3.6081.03 ± 4.9987.56 ± 3.5784.13 ± 5.4183.60 ± 2.3585.01 ± 3.78
K×10074.19 ± 2.0285.81 ± 1.6489.12 ± 1.8388.96 ± 1.5689.53 ± 2.7090.80 ± 1.4191.65 ± 2.49
Asphalt88.75 ± 2.7091.84 ± 1.7592.36 ± 5.6299.62 ± 2.5789.85 ± 2.0594.38 ± 5.0898.36 ± 5.16
Meadows94.94 ± 1.4798.02 ± 0.6898.85 ± 1.0498.85 ± 4.9198.68 ± 1.2399.9 ± 0.2797.49 ± 0.94
Gravel34.89 ± 15.4666.63 ± 14.1042.94 ± 22.8890.99 ± 25.4168.99 ± 11.7072.86 ± 6.2948.88 ± 14.58
Trees64.11 ± 9.7692.10 ± 5.1493.52 ± 3.5394.13 ± 8.5586.72 ± 2.1877.73 ± 9.4489.7 ± 9.26
Metal sheets89.35 ± 8.2086.26 ± 21.9299.33 ± 0.7299.63 ± 2.7499.71 ± 0.2179.02 ± 7.6483.21 ± 9.83
Bare soil66.54 ± 6.1578.66 ± 7.1298.62 ± 1.5477.72 ± 8.9191.36 ± 3.2399.88 ± 3.9398.32 ± 1.52
Bitumen55.77 ± 22.7974.43 ± 10.1245.56 ± 27.3059.63 ± 15.3475.24 ± 22.8759.78 ± 15.7784.97 ± 13.92
Bricks73.39 ± 7.3190.92 ± 3.7090.60 ± 5.6867.52 ± 23.2897.24 ± 1.5997.02 ± 1.6198.43 ± 2.03
Shadow57.61 ± 21.5174.11 ± 22.2167.52 ± 33.5199.90 ± 2.8949.41 ± 26.6971.84 ± 14.8565.78 ± 8.79
Table 4. Test accuracy with different preprocessing methods on the Indian Pines dataset.
Table 4. Test accuracy with different preprocessing methods on the Indian Pines dataset.
MethodRBF-SVMEMP-SVMCNNSSRNVGGSSTSST-FA
OA(%)79.75 ± 1.8981.53 ± 2.3986.81 ± 1.2883.21 ± 1.2586.80 ± 1.0888.77 ± 1.0788.98 ± 0.66
AA(%)60.11 ± 2.5672.43 ± 6.3568.30 ± 4.8262.88 ± 7.1263.18 ± 1.9866.75 ± 6.4468.15 ± 1.06
K×10075.82 ± 2.3578.87 ± 2.7884.20 ± 1.5780.86 ± 1.3984.10 ± 1.2986.48 ± 1.2786.70 ± 0.78
Alfalfa38.67 ± 26.6650.27 ± 41.3738.70 ± 32.3925.00 ± 35.0835.56 ± 9.0735.40 ± 33.2462.91 ± 29.44
Corn-notill80.01 ± 9.1772.43 ± 7.5184.64 ± 6.0787.16 ± 12.4490.74 ± 1.9890.52 ± 3.5191.56 ± 3.23
Corn-min52.50 ± 11.8380.77 ± 5.4263.15 ± 22.1095.61 ± 20.2538.80 ± 20.6650.81 ± 15.6663.28 ± 13.48
Corn23.85 ± 22.7174.44 ± 19.2855.37 ± 45.7920.60 ± 17.8255.32 ± 27.3022.13 ± 36.0613.83 ± 27.66
Grass-pasture61.00 ± 16.1077.04 ± 8.1056.41 ± 10.4084.93 ± 21.9151.24 ± 11.6266.00 ± 10.4664.64 ± 10.76
Grass-trees91.76 ± 4.4995.27 ± 3.5496.90 ± 3.9499.71 ± 6.2294.21 ± 2.6093.50 ± 2.8394.40 ± 1.18
Grass-pasture-mowed16.58 ± 23.1249.63 ± 49.6442.54 ± 38.1628.70 ± 48.0413.58 ± 19.2142.96 ± 43.8729.63 ± 22.83
Hay-windrowed87.89 ± 8.5697.43 ± 2.9085.84 ± 27.2897.65 ± 2.1573.44 ± 18.4282.15 ± 17.7870.66 ± 21.10
Oats42.21 ± 37.6428.95 ± 39.6812.75 ± 18.339.28 ± 43.2224.07 ± 34.055.88 ± 13.158.82 ± 17.65
Soybean-notill74.57 ± 8.6974.92 ± 7.7290.33 ± 6.4880.77 ± 8.6185.78 ± 4.0891.54 ± 3.9192.57 ± 2.70
Soybean-mintill92.74 ± 3.2385.30 ± 4.8594.21 ± 3.3283.09 ± 6.5795.70 ± 1.8297.29 ± 1.8297.37 ± 2.38
Soybean-clean48.60 ± 13.6160.49 ± 10.0158.07 ± 12.4770.77 ± 21.7269.35 ± 8.9170.19 ± 11.4765.73 ± 8.25
Wheat61.59 ± 25.5075.54 ± 37.8885.99 ± 14.6298.48 ± 12.9696.75 ± 2.7281.24 ± 17.7158.94 ± 22.58
Woods94.26 ± 2.1596.56 ± 3.2496.63 ± 2.5396.99 ± 3.8698.02 ± 1.9397.16 ± 3.3698.88 ± 1.80
Buildings-Grass-Trees52.78 ± 28.2373.77 ± 9.5567.38 ± 28.3838.76 ± 18.1758.04 ± 42.3880.15 ± 26.4298.55 ± 2.20
Stone-Steel-Towers42.78 ± 34.4266.09 ± 43.0363.96 ± 33.5953.85 ± 44.5130.22 ± 12.7361.08 ± 37.6778.69 ± 38.57
Table 5. Test accuracy with different preprocessing methods on the Salinas dataset.
Table 5. Test accuracy with different preprocessing methods on the Salinas dataset.
MethodEMP-RFEMP-CNNVGGT-CNNT-SSTT-SST-L
OA(%)91.64 ± 0.8292.71 ± 0.6689.25 ± 5.2093.17 ± 1.4695.80 ± 0.9096.83 ± 0.97
AA(%)92.05 ± 2.1493.22 ± 0.6290.16 ± 4.8190.34 ± 2.7994.20 ± 1.3896.08 ± 0.83
K×10090.68 ± 0.9291.83 ± 0.7488.05 ± 5.8292.16 ± 1.7995.28 ± 1.0196.43 ± 1.09
Brocoli_green_weeds_199.39 ± 0.4893.91 ± 6.3782.46 ± 17.9490.25 ± 11.9595.00 ± 10.1398.47 ± 2.83
Brocoli_green_weeds_298.64 ± 2.5585.93 ± 8.4988.85 ± 2.1199.28 ± 1.4591.55 ± 4.9593.51 ± 6.67
Fallow86.65 ± 7.6993.31 ± 11.4792.04 ± 1.6171.84 ± 23.6391.97 ± 10.5993.19 ± 9.53
Fallow_rough_plow98.91 ± 1.4097.94 ± 1.5789.86 ± 0.3793.22 ± 6.4892.75 ± 4.0597.38 ± 1.91
Fallow_smooth91.71 ± 13.5494.90 ± 4.2598.32 ± 3.1895.69 ± 4.8697.39 ± 2.1798.57 ± 1.46
Stubble97.71 ± 1.6698.91 ± 1.5499.97 ± 1.5396.14 ± 3.2299.91 ± 0.1899.99 ± 0.03
Celery99.21 ± 0.5897.42 ± 1.6696.98 ± 3.36100.00 ± 0.0097.44 ± 3.3698.71 ± 1.05
Grapes_untrained89.18 ± 2.5393.71 ± 2.4075.66 ± 11.5294.24 ± 3.9896.61 ± 2.2396.85 ± 2.33
Soil_vinyard_develop99.56 ± 0.3698.41 ± 0.8698.44 ± 0.13100.00 ± 0.0099.96 ± 0.0699.68 ± 0.60
Corn_senesced_green_weeds92.34 ± 6.9395.93 ± 3.2596.76 ± 13.2997.86 ± 1.4897.64 ± 3.6798.25 ± 2.30
Lettuce_romaine_4wk82.34 ± 17.8493.25 ± 7.9395.72 ± 28.2593.15 ± 10.8595.92 ± 4.1597.69 ± 1.65
Lettuce_romaine_5wk92.30 ± 9.9592.98 ± 9.9997.41 ± 1.0198.78 ± 1.4997.06 ± 3.7497.25 ± 2.89
Lettuce_romaine_6wk93.16 ± 12.1091.48 ± 11.0196.82 ± 37.3776.05 ± 33.9493.80 ± 4.7495.21 ± 2.88
Lettuce_romaine_7wk85.10 ± 28.6798.07 ± 2.7898.23 ± 1.4095.83 ± 5.2794.53 ± 4.8496.56 ± 3.30
Vinyard_untrained78.36 ± 8.1378.57 ± 5.3684.13 ± 35.1381.07 ± 16.6891.53 ± 3.8093.13 ± 5.15
Vinyard_vertical_trellis88.30 ± 6.0886.80 ± 7.3950.94 ± 41.6462.03 ± 23.7574.11 ± 12.4182.86 ± 13.46
Table 6. Test accuracy with different preprocessing methods on the Pavia dataset.
Table 6. Test accuracy with different preprocessing methods on the Pavia dataset.
MethodEMP-RFEMP-CNNVGGT-CNNT-SSTT-SST-L
OA(%)88.29 ± 2.2491.65 ± 1.2291.72 ± 2.1292.55 ± 2.0693.37 ± 1.6093.73 ± 1.73
AA(%)81.87 ± 5.4183.13 ± 3.5784.13 ± 5.4184.84 ± 2.4385.30 ± 2.5589.45 ± 3.66
K×10084.36 ± 3.0989.42 ± 1.5689.53 ± 2.7090.54 ± 2.6461.61 ± 2.0392.10 ± 2.19
Asphalt94.22 ± 2.4591.99 ± 2.8189.85 ± 2.0592.59 ± 4.4890.88 ± 6.0986.63 ± 3.32
Meadows96.21 ± 1.7197.95 ± 1.3298.68 ± 1.2399.66 ± 0.1999.60 ± 0.5298.77 ± 0.41
Gravel70.14 ± 10.0549.02 ± 10.0768.99 ± 11.7070.42 ± 12.7498.51 ± 11.9486.16 ± 9.81
Trees85.59 ± 10.2289.25 ± 4.9686.72 ± 2.1872.63 ± 11.7682.02 ± 2.0386.70 ± 3.28
Metal sheets85.26 ± 24.1495.31 ± 5.1399.71 ± 0.2195.84 ± 3.5482.39 ± 7.64100 ± 0.00
Bare soil65.94 ± 10.9998.45 ± 1.4791.36 ± 3.2398.62 ± 2.2999.61 ± 3.5498.48 ± 5.38
Bitumen52.93 ± 22.3573.50 ± 13.3675.24 ± 22.8782.02 ± 12.9164.63 ± 11.8786.58 ± 11.03
Bricks94.29 ± 3.2593.24 ± 3.2597.24 ± 1.5995.35 ± 4.8498.85 ± 1.3596.51 ± 2.85
Shadow92.24 ± 16.3159.49 ± 21.8749.41 ± 26.6956.41 ± 27.0451.17 ± 23.3865.24 ± 28.69
Table 7. Test accuracy with different preprocessing methods on the Indian Pines dataset.
Table 7. Test accuracy with different preprocessing methods on the Indian Pines dataset.
MethodEMP-RFEMP-CNNVGGT-CNNT-SSTT-SST-L
OA(%)85.10 ± 1.5987.22 ± 1.3786.80 ± 1.0887.83 ± 0.7890.06 ± 0.6891.20 ± 1.35
AA(%)74.85 ± 2.3574.77 ± 5.5063.18 ± 1.9865.32 ± 2.7780.12 ± 4.4983.05 ± 5.94
K×10082.96 ± 1.8485.03 ± 1.6284.10 ± 1.2985.35 ± 1.0088.05 ± 0.8289.85 ± 1.17
Alfalfa60.88 ± 31.1544.81 ± 30.0935.56 ± 9.0737.78 ± 36.9584.64 ± 6.0992.36 ± 8.59
Corn-notill78.10 ± 7.9987.81 ± 7.0190.74 ± 1.9893.03 ± 3.6185.86 ± 3.8488.96 ± 6.83
Corn-min83.43 ± 11.5673.58 ± 11.1038.80 ± 20.6645.46 ± 17.2368.91 ± 22.9271.79 ± 21.13
Corn59.23 ± 16.0264.95 ± 37.0955.32 ± 27.3034.79 ± 33.5347.86 ± 40.1454.96 ± 45.01
Grass-pasture74.90 ± 10.7571.72 ± 21.0751.24 ± 11.6282.07 ± 10.0667.90 ± 21.0872.32 ± 16.20
Grass-trees94.34 ± 6.5088.79 ± 5.0994.21 ± 2.6094.02 ± 4.1991.08 ± 6.2293.55 ± 5.45
Grass-pasture-mowed32.15 ± 39.5663.95 ± 42.3013.58 ± 19.2138.52 ± 32.8477.78 ± 39.1379.26 ± 39.66
Hay-windrowed99.85 ± 0.1796.50 ± 3.0273.44 ± 18.4270.01 ± 33.4198.36 ± 2.1199.90 ± 0.20
Oats5.26 ± 15.7920.00 ± 40.0024.07 ± 34.0513.33 ± 26.6748.17 ± 39.8950.53 ± 42.96
Soybean-notill74.18 ± 10.6788.60 ± 4.9785.78 ± 4.0889.20 ± 6.5291.24 ± 3.7790.85 ± 2.44
Soybean-mintill90.84 ± 4.5592.43 ± 3.4295.70 ± 1.8294.45 ± 2.1198.23 ± 0.3297.08 ± 0.67
Soybean-clean71.13 ± 15.4871.39 ± 17.3169.35 ± 8.9177.95 ± 21.2976.98 ± 4.4576.85 ± 5.79
Wheat99.50 ± 0.9091.08 ± 7.0596.75 ± 2.7292.61 ± 8.4285.56 ± 13.4793.33 ± 10.81
Woods97.29 ± 1.3595.60 ± 4.2998.02 ± 1.9393.05 ± 4.1395.04 ± 1.4495.12 ± 1.11
Buildings-Grass-Trees82.12 ± 17.6274.28 ± 14.2058.04 ± 42.3836.49 ± 28.4180.56 ± 18.1687.02 ± 16.96
Stone-Steel-Towers94.46 ± 2.1270.88 ± 27.3730.22 ± 12.7352.31 ± 30.7483.74 ± 21.9584.95 ± 20.63
Table 8. Running time (min.) of different methods on the three HSI datasets with 200 training samples.
Table 8. Running time (min.) of different methods on the three HSI datasets with 200 training samples.
MethodsRBF-SVMEMP-SVMEMP-RFEMP-CNNCNNT-CNNSSRNVGGSSTSST-FAT-SSTT-SST-L
Datasets
Salinas0.830.060.061.601.860.223.472.5322.4222.4628.8628.89
Pavia0.710.060.070.640.790.241.620.9516.6916.7018.0118.05
Indian Pines0.470.030.030.350.550.172.270.7121.4321.4426.9126.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

He, X.; Chen, Y.; Lin, Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. https://doi.org/10.3390/rs13030498

AMA Style

He X, Chen Y, Lin Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sensing. 2021; 13(3):498. https://doi.org/10.3390/rs13030498

Chicago/Turabian Style

He, Xin, Yushi Chen, and Zhouhan Lin. 2021. "Spatial-Spectral Transformer for Hyperspectral Image Classification" Remote Sensing 13, no. 3: 498. https://doi.org/10.3390/rs13030498

APA Style

He, X., Chen, Y., & Lin, Z. (2021). Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sensing, 13(3), 498. https://doi.org/10.3390/rs13030498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop