Next Article in Journal
Spectral-Spatial Domain Attention Network for Hyperspectral Image Few-Shot Classification
Next Article in Special Issue
Global and Multiscale Aggregate Network for Saliency Object Detection in Optical Remote Sensing Images
Previous Article in Journal
Submesoscale Short-Lived Eddies in the Southwestern Taiwan Strait Observed by High-Frequency Surface-Wave Radars
Previous Article in Special Issue
A Multi-Feature Fusion-Based Method for Crater Extraction of Airport Runways in Remote-Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Level Feature Extraction Networks for Hyperspectral Image Classification

School of Information Engineering, Northwest A&F University, Xi’an 712100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(3), 590; https://doi.org/10.3390/rs16030590
Submission received: 4 November 2023 / Revised: 24 January 2024 / Accepted: 29 January 2024 / Published: 4 February 2024

Abstract

:
Hyperspectral image (HSI) classification plays a key role in the field of earth observation missions. Recently, transformer-based approaches have been widely used for HSI classification due to their ability to model long-range sequences. However, these methods face two main challenges. First, they treat HSI as linear vectors, disregarding their 3D attributes and spatial structure. Second, the repeated concatenation of encoders leads to information loss and gradient vanishing. To overcome these challenges, we propose a new solution called the multi-level feature extraction network (MLFEN). MLFEN consists of two sub-networks: the hybrid convolutional attention module (HCAM) and the enhanced dense vision transformer (EDVT). HCAM incorporates a band shift strategy to eliminate the edge effect of convolution and utilizes hybrid convolutional blocks to capture the 3D properties and spatial structure of HSI. Additionally, an attention module is introduced to identify strongly discriminative features. EDVT reconfigures the organization of original encoders by incorporating dense connections and adaptive feature fusion components, enabling faster propagation of information and mitigating the problem of gradient vanishing. Furthermore, we propose a novel sparse loss function to better fit the data distribution. Extensive experiments conducted on three public datasets demonstrate the significant advancements achieved by MLFEN.

1. Introduction

Hyperspectral image (HSI) is composed of abundant spatial and spectral data, enabling effective differentiation between various types of land cover. This technology finds extensive applications in urban planning [1,2], geological exploration [3,4], precision agriculture [5,6], and other domains. HSI classification plays a fundamental role in remote sensing as it offers a robust means of analyzing and interpreting information embedded within HSI.
Conventional machine learning approaches typically prioritize the analysis of spectral information in tasks such as k-nearest neighbor [7], support vector machine [8], random forest [9], and sparse representation [10]. While these methods are straightforward and easily scalable, they struggle when working with high-dimensional data and small sample sizes, which is known as the Hughes phenomenon [11]. To overcome the challenge, band selection is often adopted to select the most useful bands and thus reduce redundant information, such as the local feature descriptor network [12] and similarity-based ranking method [13]. In addition, dimensionality reduction techniques like principal component analysis (PCA) [14] and linear discriminant analysis [15] are commonly employed to map high-dimensional HSI data into lower-dimensional spaces. However, recent studies demonstrated that relying solely on spectral information and disregarding spatial information makes it difficult for classification algorithms to accurately capture surface object characteristics, such as spatial distribution and morphological features. This ultimately impacts classification accuracy [16]. To address this, various models that extract both spatial and spectral features were proposed, including the Markov random field [17], extended multi-attribute profile [18], Gabor filters [19], and hypergraph structure [20]. Nevertheless, these conventional machine learning methods are limited in their ability to leverage deep nonlinear features, which can result in suboptimal performance.
The advancement of deep learning techniques has greatly contributed to the progress of HSI processing [21]. Numerous models based on deep learning were proposed for HSI classification tasks. Some of the well-known baseline networks include recurrent neural networks [22], graph convolutional neural networks [23], autoencoders [24], generative adversarial networks [25], capsule networks [26], long short-term memory networks [27], and convolutional neural networks (CNNs) [28]. Among these methods, CNNs are the most widely used and can be categorized into 1D convolutional neural network (1D-CNN) [29], 2D convolutional neural network (2D-CNN) [30], and 3D convolutional neural network (3D-CNN) [31] based on their dimensions. Hu et al., introduced 1D-CNN to capture features from the spectral dimension and achieved better accuracy compared to traditional machine learning approaches. The 2D-CNN method is commonly used to extract spatial features, while the 3D-CNN is employed to capture spatial–spectral features. The hybrid spectral CNN (HybridSN) [32] combines the benefits of both 2D-CNN and 3D-CNN, leading to superior classification performance. Furthermore, Paoletti et al., utilized pyramidal bottleneck residual cells [33] to enhance performance by increasing the dimension of spectral and spatial attributes layer by layer. On the other hand, Ma et al., proposed a deconvolution network with skip connections to address issues such as the loss of important information during down-sampling and the lack of training samples [34]. While these methods are more efficient than conventional machine learning approaches, they often fail to fully leverage features at different levels. In recent years, many deep learning methods using representation learning applied to tasks such as HSI classification, segmentation and image super-resolution have also emerged, including [35,36,37,38]. For improving the performance of HSI classification tasks, researchers proposed methods based on the joint extraction of spectral–spatial features for representation learning. A central vector-oriented self-similarity network (CVSSN) [39] enhances characterization of the model by considering the association between internal pixels and their neighbors.
To emphasize regions of interest in an image while suppressing irrelevant background regions, researchers have suggested attention mechanisms, taking inspiration from the visual mechanisms of humans. Noteworthy examples of such works include the squeeze-and-excitation network (SENet) [40] and the dual attention network (DANet) [41]. SENet adjusts the scale of channel features and enables the model to focus more on important channel information, while DANet improves the model’s perception of contextual information by introducing a dual-attention mechanism. Additionally, the spectral–spatial attention network (SSAN) [42] combines the capabilities of CNN for modeling spatial interior dependence and recurrent neural network for characterizing spectral sequences. The multi-attention fusion network (MAFN) [43] incorporates attention modules that cater to spatial and spectral perspectives, mitigating issues like band redundancy and pixel interference. Furthermore, the multimodal transfer feature fusion network [44] utilizes the local attention mechanism and a multi-task learning strategy to facilitate the learning of HSIs across different domains. However, it is important to note that these attention-based methods often necessitate a substantial amount of annotated data for effective training, posing a challenge when dealing with HSIs that have limited annotated samples.
Transformer models are gaining popularity in the realm of natural language processing [45]. Researchers apply the transformer architecture, originally developed for language processing, to the image domain with the introduction of vision transformer (ViT) [46]. ViT achieves impressive results in image classification by dividing images into smaller patches and treating them as sequential data. In addition to the basic ViT model, there have been advancements in this area with the development of variants such as pyramid vision transformer [47], swin transformer [48], and data-efficient image transformers [49]. Another approach called tokens-to-token ViT [50] improves tokenization in ViT by utilizing a soft-split operation. Researchers also explored the application of ViT in HSI classification. The novel backbone network rethinking HSI classification from a sequential perspective with transformers [51] introduces ViT to the field of HSI classification, but it largely focuses on mining spectral information and lacks the utilization of spatial information. To address this limitation, various methods were proposed to combine CNNs with ViT to extract both spatial and spectral features, including the convolutional transformer network [52], spatial–spectral transformer [53], and neighborhood enhancement hybrid transformer network [54]. Other approaches such as spectral–spatial feature tokenization transformer [55] and hyperspectral image transformer (HiT) [56] were developed to extract spectral–spatial features and semantic features. However, these ViT-based methods often overlook the importance of feature reuse and information transfer.
In summary, the above-mentioned methods have faced two issues: (1) Treating HSI as linear vectors, compromising 3D attributes and spatial structure. (2) Repeatedly concatenating encoders, leading to information loss and gradient vanishing.
To overcome these challenges, on the one hand, hybrid convolution operations are utilized in order to capture the 3D characteristics and spatial structure of HSI. Specifically, to address any potential edge effects caused by the convolution operation, a novel band shift strategy is developed. Furthermore, an attention module is employed to explore the relationship between the spatial and spectral aspects of the data, thereby emphasizing the features with high discriminative capability. On the other hand, a modification is made to the arrangement of encoders in order to enhance the adaptability to the HSI classification task’s small-scale dataset. This adjustment involves eliminating the simple concatenation method to ensure that excessive information loss is avoided. To enable feature reuse, an adaptive feature fusion component is incorporated.
The major contributions in this study can be outlined as follows.
(1)
A multi-level feature extraction network (MLFEN) for HSI classification is proposed. By combining the capability of CNN’s local spatial–spectral feature capture and ViT’s global sequence modeling, MLFEN achieves effective fusion of shallow to deep features of HSIs.
(2)
A sophisticated hybrid convolutional attention module (HCAM) is suggested, which incorporates band shift, hybrid convolution, and attention mechanisms to efficiently capture and enhance multidimensional features. By seamlessly combining these essential techniques, HCAM can empower the model to gain a comprehensive understanding of the intricate details present within the image.
(3)
A novel variation of the ViT called enhanced dense vision transformer (EDVT) is introduced. EDVT is specifically designed for characterizing HSI data. To address the issue of information loss due to deep networks, EDVT incorporates a modified architecture for its encoders. Additionally, EDVT includes an adaptive feature fusion (AFF) component that enables effective information transfer and feature reuse.
(4)
A new sparse loss function for HSI classification is developed, combining the benefits of both the cross-entropy loss function and a sparsity regularization operator. This loss function contributes to sparse representation, thus effectively solving the overfitting problem and improving the robustness and accuracy of classification.
The remainder of the article is organized as follows. In Section 2, the methodology of the proposed network is presented. Section 3 details the datasets, provides specific implementation steps and parameter settings, and analyzes the experimental findings. Section 4 compares the computational cost and performance of MLFEN with other methods. Finally, Section 5 provides a summary of this work and outlines future directions.

2. Method

Figure 1 presents a visual representation of the MLFEN architecture. To begin with, the original HSIs undergo preprocessing using PCA. This process converts the HSIs into patch cubes which are then fed into two distinct modules: the hybrid convolutional attention module (HCAM) and the enhanced dense vision transformer (EDVT). For the classification task, a multi-layer perception network (MLP) is implemented as the classifier.

2.1. Hybrid Convolutional Attention Module (HCAM)

The main role of the HCAM module is to acquire the shallow joint spatial–spectral information of HSI. In order to achieve superior performance, the structure of the HCAM is both complete and compact, containing a band shift component, four convolutional blocks and an attention module (AM).

2.1.1. Band Shift

The PCA algorithm ranks the principal components according to the magnitude of the variance. The larger the variance, the more information it contains. Therefore, after PCA processing, the feature information of each band of the HSI image is arranged in descending order, which results in the most informative features being at the edge. However, the pixel points at the edges during the convolution operation are often not fully covered by the convolution kernel. This edge effect can make the mining of spectral information insufficient and further affect the convolution effect.
Based on the above analysis, a band shift strategy is proposed for mitigating the undesirable consequences of edge effects. Band shift aims to move key spectral channels to the central position across all data, while relocating less critical spectral channels towards the edges of the data. The benefits of doing so include increasing the likelihood of effective extraction of spectral features through convolutional operations, and maintaining critical spectral channels at the core position of the receptive field. With the shift strategy, the negative impact due to the edge effect of the convolution operation is mitigated, and the separability of the bands is improved thus enhancing the feature representation of the model. The mapping function g ( · ) of the band shift strategy can be expressed as:
g ( v ) = 2 ( p 2 v ) 1 , v [ 0 , p 2 ) g ( v ) = 2 ( v p 2 ) , v [ p 2 , p 1 ]
where p represents the number of principal components after PCA, g ( v ) and v denote the band index before and after shift. Applying this strategy can enhance the discrimination of target features, alleviate the band correlation problem, and help improve the accuracy and reliability of classification.

2.1.2. Multidimensional Feature Extraction

As illustrated in Figure 1, HCAM contains two 3D convolution blocks and two 2D convolution blocks with a convolution kernel size of 3 × 3 × 3 and 3 × 3 , respectively. This design first extracts the joint spatial–spectral features by 3D convolution and adaptively learns more abstract and advanced feature representations using 2D convolution. Batch normalization layers and rectified linear unit (ReLU) activation functions are added after each convolutional layer to avoid the gradient disappearance phenomenon and improve generalization ability. Compared to the single 3D-CNN method, this hybrid approach not only enables multi-level features to be obtained, but also dramatically reduces the computational cost.
To further enhance the interaction and fusion of features between different layers and save time and computational resources, the skip connection mechanism [57] is employed in HCAM. By doing so, it allows the model to more effectively learn and capture relevant features, ultimately leading to improved performance. Additionally, the implementation of skip connection in the HCAM module helps to improve computational efficiency while still maintaining robust performance.
Let X 0 R k × k × d (k and d are the patch size and number of bands of the 3D cubes) denote 3D cubes after the band shift, and X i is the output of ith block. Then X i can be expressed by:
X i = δ ( B N ( C o n v ( X i 1 ) ) ) , i = 2 m 1 , m = [ 1 , , m ] X i = δ ( B N ( C o n v ( C a t ( X i 2 , X i 1 ) ) ) ) , i = 2 m , m = [ 1 , , m ]
where Cat denotes the concatenation function, Conv means the convolutional operation and BN is the batch normalization layer. The symbol δ is the ReLU activation. When i = 4, X i is X , the input to the AM component. Through hybrid convolutional operations described above, the model comprehensively extracts shallow multidimensional spatial and spectral features.

2.1.3. Attention Module (AM)

The attention component is designed to incorporate broader contextual information into local contextual features. Inspired by the attention module proposed in the DANet [41], the local spatial–spectral features X is processed through convolution to obtain three feature mappings, namely X 1 , X 2 , and X 3 . These mappings can be represented as:
X 1 = C o n v ( X , w 1 ) + b 1 X 2 = C o n v ( X , w 2 ) + b 2 X 3 = C o n v ( X , w 3 ) + b 3
where the symbol w 1 , w 2 , and w 3 are weights of convolution layers while b 1 , b 2 , and b 3 denote biases. Then the pixel correlation is calculated by a matrix multiplication of the deformed X 1 with X 2 and the attention map S is obtained by using a softmax layer, which is defined as:
S = e x p ( X 1 X 2 ) j = 1 n e x p ( X 1 X 2 )
where n is the number of features. We perform a matrix multiplication between X 3 and S and then multiply it by a scale parameter α . Finally, the spatial attention feature is added to the input X to obtain the final result X :
X = α j = 1 n S X 3 + X
where α is initialized to 0 and continuously updated during the training process. We can deduce that the final feature X is obtained by taking a weighted sum of the features at all positions including the original features. As a result, X incorporates a global contextual view and selectively integrates contextual information based on the attention map.

2.2. Enhanced Dense Vision Transformer (EDVT)

Despite the benefits of CNNs in extracting nearby spatial contextual information, they face difficulties in capturing delicate variations in spectral data over extended distances. As a result, ViT is frequently utilized to augment the feature maps produced by CNNs, enriching the global spectral correlation and disparity information. ViT employs a conventional transformer encoder methodology, which includes MSA and MLP modules. To normalize each encoder block, layer normalization (LN) [58] is employed. The standard transformer encoder block can be represented by the following equation:
E l = M S A ( L N ( E l 1 ) ) + E l 1 E l = M L P ( L N ( E l ) ) + E l
where E l denotes the output features of lth encoder.
It is important to note that the original ViT consists of six encoders that are connected in a straightforward manner. However, there are various issues that still need to be addressed. Firstly, the connection between encoders is too simplistic for efficient feature reuse, which hinders the capturing of multi-level spectral information. Secondly, the training datasets used for HSI classification tasks are typically small and considered as limited samples in many research studies. Consequently, the network of six encoders can be seen as overly deep, resulting in a substantial loss of information.
In order to address the aforementioned issues, EDVT is proposed to minimize information loss and facilitate the reuse of multi-level features. As shown in Figure 2, EDVT comprises three encoder blocks, each consisting of an encoder and an AFF component. Taking inspiration from dense convolutional network [59], the two encoder blocks are interconnected with dense connections. These dense connections enable every encoder in the network to directly access the feature maps from all preceding encoders, thereby enhancing the extent of feature sharing and reuse. Additionally, the AFF component adaptively fuses varying numbers of features through a two-step process involving concatenation and convolution. The A F F l operation can be defined as:
E l ^ = w ¨ ( C a t ( E i n , E l 1 , E l ) )
where E l ^ is the fused feature after the lth AFF, w ¨ denotes a network weight parameter that can be learned, and C a t represents the concatenation operation.

2.3. Sparse Loss Function

Supervised learning focuses on reducing the error by regularizing the parameters. Reducing the error ensures that the model fits the training data well, while parameter regularization prevents the model from overfitting the training data. Based on these principles, a new loss function for HSI classification tasks is proposed. This loss function utilizes the cross-entropy loss function L c e [60] to minimize the error, and incorporates a sparsity regularization operator L s r o [61] as the regularization term.
The aim of the L c e function is to reduce the difference between the predicted output and the ground truth, thereby bringing the predicted value closer to the actual label. L c e is highly responsive to any inconsistency between the model output and the true label, enabling efficient gradient propagation and expediting model convergence. The calculation of L c e is as follows:
L c e = 1 C i = 1 C y i l o g ( p i )
where the symbol y i is the probability value of the ith category in the true label while p i denotes that in the predicted result. C is the number of categories.
However, when there is an imbalance in the number of samples across different categories in a dataset, the use of L c e may result in the model being biased towards predicting the more frequent categories and disregarding the less frequent ones. Additionally, L c e updates the model parameters by minimizing the prediction error rate during training, which can lead to overfitting if the training data are insufficient or the model is too complex. To address these issues, sparse regularization loss ( L s r o ) can be employed, which encourages the model to learn a more sparse feature representation. L s r o can be expressed as:
L s r o = i = 1 C y i p i
On one side, because of the scarcity of L s r o , it diminishes the preference for a larger number of categories and thus enhances the model’s performance when handling datasets with imbalances. On the other side, by minimizing the impact of noisy or irrelevant input features, L s r o as a regularization term can effectively decrease the risk of overfitting and improve the model’s ability to make generalizations. Therefore, we propose a new sparse loss function L that combines L s r o with L c e , resulting in a more robust and easier to interpret model, as demonstrated below:
L = L c e + λ L s r o
where λ represents the coefficient of L s r o . The combination of L c e and L s r o can control the parameters while characterizing the probability distribution of the sample classification.

3. Results

3.1. Dataset Description

In order to evaluate the efficacy of MLFEN, five publicly available datasets are employed: Pavia University (PU), Kennedy Space Center (KSC), Salinas (SA), Indian Pines (IP), and Houston University (HU). Illustrations in Figure 3 depict the false color maps, ground truth maps, and color bars corresponding to the PU, KSC, SA, IP and HU datasets, respectively. Table 1 shows the basic information of the five datasets, which include image size, the number of categories and bands, band range, resolution, sensor and location. The land cover categories for the five datasets, along with the number of training and test set samples, are shown in Table 2. To ensure an unbiased evaluation, we randomly choose 1% of the data from each category of the datasets as the training set, while the remaining data are kept for testing purposes.

3.2. Experimental Setup

3.2.1. Implementation Details

All experiments are conducted in the pytorch framework using a server equipped with Ubuntu 20.04, Intel(R) Xeon(R) Platinum 8375C CPU, 250 GB RAM and NVIDIA GeForce RTX 3090 GPU.
After performing PCA, 100 principal components are retained. During the training phase, 200 iterations are undertaken with a learning rate of 0.001 and a batch size of 64. The sparse loss function consists of L c e and L s r o with a coefficient of λ set to 0.01. The adam optimizer is employed with β 1 and β 2 parameters set to 0.1 and 0.99, respectively.

3.2.2. Evaluation Indicators

The performance of the proposed method is assessed using three commonly employed metrics: overall accuracy (OA), average accuracy (AA), and kappa coefficient (Kappa). Each of these evaluation indicators positively reflects the classification accuracy. To ensure reliability, the reported values for these three metrics are averaged over ten independent runs in all experiments.

3.3. Parameter Analysis

3.3.1. The Influence of the Selection of Dimensionality Reduction Method

HSI contains an abundance of redundant and noisy information in the spectral channel, which can create interference. Therefore, it is very challenging to adequately extract discriminative spectral information from the images. In this paper, comparative experiments between PCA and LDA are conducted on the PU dataset and the results are shown in Table 3. It can be concluded from the comparison that PCA is slightly better for categorization than LDA, so PCA is chosen to extract the principal spectral bands.
To explore the effect of the number of principal components on performance, parametric experiments are conducted on the PU dataset. As shown in Table 4, it can be found that both the network parameters and the running time increase with the increase in number of principal components, and the best classification results are obtained when the number is 100. Therefore, the number of principal components is set to 100.

3.3.2. The Influence of Patch Size

Patch size plays a crucial role in determining how well a classification model performs. In order to find the optimal patch size, we conducted a series of experiments within the range of 7 , 9 , 11 , 13 , 15 . Figure 4 displays the effect of parameter k on the classification performance across all three datasets.
The analysis illustrated in Figure 4 demonstrates that a patch size of 13 is the most effective. All evaluation metrics reach their peak values for three datasets when k is set to 13. When the patch size k is small, the image blocks contain less spatial neighborhood information and are vulnerable to noise interference. Conversely, if k is large, there may be an excessive amount of spatial neighborhood information, which disrupts the central pixel features. In order to achieve a balance between performance and efficiency, the value of the patch size k was established as 13.

3.3.3. The Influence of the Number of Training Samples

The performance of the model is greatly influenced by the number of training samples. Therefore, an analysis of the classification results is conducted using different sizes of training sets. To ensure the stability and robustness of the MLFEN method, we randomly selected 0.5%, 1%, 3%, 5%, and 10% labeled samples as the training data for the PU and SA datasets. For the KSC, IP and HU datasets, we selected 1%, 3%, 5%, 7%, and 10% labeled samples as the training data.
In Figure 5, we observe the impact of varying the number of training samples on the performance of the MLFEN approach. Through the experimental findings, it becomes evident that by using a mere 0.5% of training data from the PU dataset, the model can achieve an OA surpassing 92%. Remarkably, this advantage extends further to over 96% on the SA dataset. Similarly, on the KSC dataset, the model reaches an OA metric of more than 96% with the utilization of only 1% of the training samples. These results highlight the robustness of the model in handling inter-class imbalances.

3.3.4. The Influence of Coefficient λ

To examine the impact of the coefficient λ in the proposed sparse loss function on the classification performance, a series of experiments are conducted on the PU dataset. Initially, the performance of the standard L c e is confirmed. Following this, several sets of experiments are designed with varying values of λ . The outcomes are summarized in Table 5. To ensure fairness and scientific rigor in the experimental findings, all parameters except λ remained consistent with those described in Section 3.2.1.
As shown in Table 5, the lowest classification performance is achieved in the proposed MLFEN using the commonly used L c e . The addition of the L s r o regularization term to the loss function results in a higher classification accuracy than L c e alone. This can be explained by the fact that the regularization term prevents overfitting to some extent and is resistant to noise and interference. With the gradual increase of the λ hyperparameter, all three evaluation indicators show an upward trend and reach the maximum value at the value of 0.01 for the λ hyperparameter. When the value of λ is greater than 0.01, OA, AA and Kappa decrease. In view of the above experimental results, the value of λ is set to 0.01.

3.3.5. The Influence of the Depth of EDVT

To investigate the impact of the network depth of the EDVT module on the performance of the model, we conduct quantitative experiments on the PU dataset using networks of varying depths and the results are presented in Table 6.
From the table, it is evident that the performance of the short-range EDVT is relatively poor, particularly at a depth of 1 where the OA value reaches only 89.28%. This can be attributed to the difficulty in fully capturing and integrating deep global features with only one or two encoders. In contrast, the mid-range EDVT (three or four encoders) tends to yield better results, with the optimal depth parameter being three. However, as the number of network layers increases further, the classification performance experiences some regression. For instance, when there are six layers, the OA decreases to 95.69%.
In addition, we evaluate the impact of EDVT depth on computational complexity by counting the training time and testing time. The results show that the training time and testing time are the shortest when the depth is 1, which are 146.07 s and 5.87 s, respectively. The computational complexity increases gradually as the network depth increases. It can be seen that shortening the depth as much as possible can help reduce the computational complexity while ensuring high classification accuracy.
It is important to highlight that the classical ViT utilizes six encoders. However, our experiments have shown that the model can achieve good accuracy with only three encoders. This relatively shallow network design helps prevent excessive information loss and showcases the effectiveness of the EDVT module in facilitating deep feature learning.

3.4. Ablation Study

To validate the effectiveness of the band shift module, we performed ablation experiments on the PU dataset. Table 7 shows the mean values of the results of ten experiments. From these findings, we can observe that OA, AA, Kappa are improved by 1.41%, 2.76% and 1.78%, respectively, which proves that the band shift module has a positive effect on the classification results.
This paper introduces the MLFEN method and its main technical contributions. These include the HCAM module, which models and enhances the 3D spatial–spectral joint features, the EDVT module for mining deep spectral features, and the sparse loss function for reducing the fitting variance. In this section, our research goal is to analyze the impact of these structures on the PU dataset both qualitatively and quantitatively. To achieve this, we conduct several ablation experiments to test the applicability and effectiveness of these structures in MLFEN for HSI classification. The specific classification results for different components are presented in Table 8.
Table 8 demonstrate the influence of different components on the performance of MLFEN. When all three components are utilized (Case 5), the best classification results are obtained on the PU dataset. Conversely, when none of these components are employed (case 1), the worst performance is observed. Removing HCAM (Case 2) leads to a significant decrease in accuracy, with OA, AA, and Kappa decreasing by 6.08%, 11.2%, and 8.05%, respectively. This highlights the importance of extracting local joint spatial–spectral features. Similarly, removing EDVT (Case 3) results in a decrease in all three evaluation indicators, indicating that the improved EDVT model is more advantageous for HSI classification tasks compared to the original ViT model. Furthermore, it is worth noting that neither EDVT without HCAM assistance nor HCAM without EDVT support can achieve excellent performance. This suggests that HCAM and EDVT mutually enhance each other, enabling the fusion of shallow and deep features. Lastly, the inclusion of the loss function (Case 4) leads to improvements in the classification metrics OA, AA, and Kappa by 2.22%, 3.23%, and 2.85%, respectively. Despite its simplicity, the design of the sparse loss function helps mitigate overfitting and enhances the accuracy of the classification results.

3.5. Comparative Experiments

Extensive experiments were conducted on the PU, KSC, SA, IP and HU datasets to analyze and evaluate the performance of various methods. The comparison experiments include refining the evaluation of representative methods such as 2D-CNN [30], 3D-CNN [31], HybridSN [32], SSAN [42], MAFN [43], ViT [46], HiT [56], CVSSN [39], and SSFTT [53].
Regarding 2D-CNN, the network structure comprises four 2D convolutional blocks and an average pooling layer. The initial two 2D convolutional blocks consist of a 2D convolutional layer, a batch normalization layer, and a ReLU activation function layer. Meanwhile, the last two 2D convolutional blocks incorporate a dropout layer in addition to the aforementioned three layers. Each of the four convolutional kernels has a size of 3 × 3, with corresponding quantities of 30, 32, 64, and 128, respectively.
Concerning 3D-CNN, the model consists of three 3D convolutional blocks and an average pooling layer. Similar to the 2D-CNN, the first 3D convolutional block includes a 3D convolutional layer, a batch normalization layer, and a ReLU activation function layer, whereas the last two 3D convolutional blocks encompass a dropout layer in addition to the aforementioned three layers. The number of 3D convolutional kernels are 32, 64, and 128, and their size is 3 × 3 × 3.
For the other seven methods, including HybridSN, SSAN, MAFN, ViT, HiT, CVSSN and SSFTT, the networks follow the setup described in respective references.
For the proposed MLFEN method, the number of components after PCA dimensionality reduction is set to 100. The patch size after the data-processing stage is set to 13 × 13. Within the EDVT module, a 2D convolution is performed using a convolution kernel size of 3 × 3, while a 3D convolution is performed using a convolution kernel size of 3 × 3 × 3. The HCVT module utilizes three encoders.

3.5.1. Quantitative Evaluation

Table 9, Table 10, Table 11, Table 12 and Table 13 provide the comparative results of MLFEN and other state-of-the-art methods on the PU, KSC, SA, IP and HU datasets, respectively. Based on the evaluation metrics of OA, AA, and Kappa, the MLFEN approach proposed in this study displays superior performance compared to other models for all five datasets.
The PU dataset contains fewer categories and is mainly used to validate the model’s classification performance in the presence of multi-scale coarse-graining. The OA of all methods except ViT exceeds 90%, among which the OA of MAFN method reaches 96.33%, indicating that this dataset is not too difficult to classify. However, MLFEN is the most superior in all three classification metrics, with an OA improvement of 1.04% over MAFN. This could potentially be attributed to the fact that although MAFN applies the attention mechanism for spatial and spectral information, respectively, it ignores the link between them. In contrast, MLFEN considers the correlation between spatial and spectral features through its HCAM module, resulting in higher classification accuracy.
The number of samples used for training on the KSC dataset is only 50, making it suitable for evaluating the model’s classification performance with limited data. Unfortunately, the classification performance of convolution-based 2D-CNN, 3D-CNN, HybridSN, CVSSN methods, attention-mechanism-based SSAN, MAFN methods, and transformer-based ViT, HiT, SSFTT methods need to be improved. However, the MLFEN method exhibited absolute dominance, achieving the best accuracy in terms of OA, AA, Kappa, and 9 out of the 13 classes. The OA of MLFEN outperformed the other methods by margins ranging from 2.4% to 18.31%. These results indicate that MLFEN demonstrates excellent performance even with limited sample data.
The SA dataset contains 16 classes with a large disparity in the number of samples between classes. The primary purpose of using this dataset is to evaluate how well the model can address these class imbalances. However, both the transformer-based ViT and HiT methods performed poorly on the SA dataset, suggesting that they struggle to overcome the effects of class imbalances. Another transformer-based network, SSFTT, performs slightly worse for some of the classes even though it performs well on five classes. The proposed MLFEN achieves remarkably high performance on various evaluation metrics. Specifically, it achieved 99.05% OA, 99.00% AA, and 98.95% Kappa. These results clearly demonstrate that the proposed method successfully addresses the differences between classes and significantly improves classification performance.
The IP dataset contains up to 16 categories where there is a significant imbalance in the distribution of samples between categories. The main objective of utilising this dataset is to assess the ability of the models to handle these category imbalances. Strikingly, traditional transformer-based models, such as ViT and HiT, perform poorly on the IP dataset, implying that they struggle to overcome the challenges posed by category imbalances. In addition, another transformer-based network, SSFTT, performs slightly weakly for some categories, although it performs well in handling four categories. In contrast, the proposed MLFEN demonstrates significant efficacy, achieving excellent results on various evaluation metrics. Of particular note, it achieved an impressive 89.29% OA, 72.64% AA and 87.70% Kappa. These results highlight that the method successfully copes with the differences between categories, improving classification performance on the IP dataset.
After quantitative analysis of the classification result of HU, our model achieved 88.17% accuracy across the entire dataset, demonstrating classification accuracy for multiple categories. Specifically, for a specific category “Fallow_smooth”, we observe that the model has a precision of 100.00%. Similarly, MLFEN achieves 88.75% for AA and 87.21% for Kappa, the highest among the selected state-of-the-art methods.

3.5.2. Visual Analysis

The classification maps can vividly visualize the difference in effect between the proposed method MLFEN and other comparative methods. Therefore, we show in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 the classification effects of MLFEN and nine different methods on the five datasets, respectively.
In general, the classification maps obtained by the MLFEN model best fit the ground truth maps and are the clearest. Specifically, as shown in Figure 6, the category that is harder to distinguish is the blue region located in the middle of the image. The classification results of most methods for this region are adulterated with many yellow pixels, and the MAFN and the proposed MLFEN methods are the most accurate for this region. However, for the yellow region at the bottom of the image, the MAFN method incorrectly classifies many pixels as blue, whereas the MLFEN method recognizes the yellow color to a greater extent. These two regions on the image of the PU dataset confirm the superior performance of our method. Moreover, MLFEN successfully mitigated the problem of the category “Self-Blocking Bricks” being misclassified as other classes and other classes being misclassified as “Self-Blocking Bricks” that occurs in SSFTT.
Furthermore, by observing and comparing the inter-class boundary regions presented in Figure 7, it can be observed that the MLFEN classification map has the clearest boundaries. Meanwhile, we chose a zoomed-in region of interest (ROI) to further highlight the classification performance of different models more visually. The ROI can be analysed to intuitively see that the classification effect of our method is at its best.
In addition, on the SA dataset, for the category “Broccoli-green-weeds-2” in blue and the category “Corn-senesced-green-weeds” in gray, the comparison methods contain significantly more noise, indicating that these methods are unable to accurately identify feature classes. The MLFEN method, on the other hand, has less noise on category “Broccoli-green-weeds-2”, while the category in gray is clean. In the “Grapes_untrained” class (light grey area in Figure 8), some misclassification pixels appears, which indicates that the MLFEN method is slightly less effective than SSFTT in this class. However, SSFTT exhibits inadequacies in the blue area, which allows MLFEN to achieve superior results in comparison.
On the IP dataset, MLFEN achieved the best classification results for the upper middle part “Corn notill” and the right middle part “Soybean notill”, which are easily misclassified by other methods. Although MLFEN exhibits some noise, the classification outcomes for various categorical areas are generally accurate. Unlike other methods, it does not suffer from the issue where the majority of pixels within categorical regions are misclassified into incorrect categories.
As for the HU dataset, by analysing the ROI in Figure 10, it can be found that the MLFEN method achieves optimal classification in the red and blue clustered regions and achieves efficient classification. Our method demonstrates superior boundary classification capabilities, further corroborating the exceptional classification performance of the MLFEN approach.

4. Discussion

Figure 11 shows the training and testing times of the different methods on the three datasets. In order to facilitate a comprehensive and objective evaluation of the different models with regard to both computational cost and performance, the three indicators are presented in the figure in the form of broken lines. Owing to the properties of convolution such as parameter sharing and local connectivity, 2D-CNN has relatively less computation time. However, due to the increase in dimensions and parameters, the running time of 3D-CNN is substantially longer. HybridSN extracts the joint spatial–spectral features of the image by combining the above two convolutions, and the running time is shorter than pure 3D-CNN, but naturally longer than 2D-CNN. It is worth noting that the CVSSN and SSFTT methods maintain high accuracy while also having relatively short time costs. The computation times of ViT and HiT are relatively long due to the introduction of the self-attention mechanism and the fully connected layers, which require a large number of matrix multiplication and addition operations at each layer of the model. These operations require more computational resources and time. The method proposed in this paper requires significantly less time than the ViT and HiT methods, while achieving the best classification performance on all three datasets. The above analysis shows that, on the one hand, a certain extension of the computation time is within a reasonable range, considering the significant increase in performance. On the other hand, investigations toward further reducing computational costs while maintaining superior classification performance represent a direction worth exploring in the future.

5. Conclusions

In this paper, MLFEN network is proposed for mining multiple sources of information from HSIs at multiple levels. On the one hand, in order to obtain 3D attributes of HSIs and maintain the spatial structure of HSIs, the HCAM module employs hybrid convolution to mine shallow features. HCAM integrates the band shift strategy, convolution operations and attention mechanisms to achieve focusing, mining and enhancement of multidimensional features. On the other hand, in order to avoid excessive information loss and increase model adaptability to the HSI classification task, EDVT reorganizes the encoder connections and adds the AFF module to achieve adaptive feature reuse and fusion. In addition, a novel sparse loss function is proposed to enhance the sparse representation ability of the model, providing prediction results that are closer to the real data distribution. Extensive experiments conducted on five datasets fully validate the superior effectiveness of MLFEN.
In the future, we will introduce more advanced techniques to optimize the model and make it more suitable for HSI classification tasks, such as transfer learning and band selection. In addition, the fusion strategy of the two modules, EDVT and HCAM, may be an important factor affecting performance thus warranting further attention and research.

Author Contributions

S.F.; Conceptualization, Methodology, Writing—original draft. X.L.; Software, Writing—review and editing. S.T.; Supervision, Resources. W.C.; Investigation, Data curation. E.Z.; Formal analysis, Visualization, Project administration. All authors have reviewed and consented to the final version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62376225, 62006188), the QinChuangyuan High-Level Innovation and Entrepreneurship Talent Program of Shaanxi (No. 2021QCYRC4-50), and Chinese Universities Scientific Fund (No. 2452022341).

Data Availability Statement

Data is included in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HSIHyperspectral image
MLFENMulti-Level Feature Extraction Network
HCAMHybrid Convolutional Attention Module
EDVTEnhanced Dense Vision Transformer
PCAPrincipal Component Analysis
CNNConvolutional Neural Network
1D-CNN1D Convolutional Neural Network
2D-CNN2D Convolutional Neural Network
3D-CNN3D Convolutional Neural Network
HybridSNHybrid Spectral Convolutional Neural Network
CVSSNCentral Vector-oriented Self-Similarity Network
SSFTTSpectral–Spatial Feature Tokenization Transformer
SENetSqueeze-and-Excitation Network
DANetDual Attention Network
SSANSpectral-Spatial Attention Network
MAFNMulti Attention Fusion Network
ViTVision Transformer
HiTHyperspectral image Transformer
AFFAdaptive Feature Fusion
MSAMulti-head Self-Attention mechanism
MLPMulti-Layer Perception network
AMAttention Module
ReLURectified Linear Unit
LNLayer Normalization
PUPavia University dataset
KSCKennedy Space Center dataset
SASalinas dataset
IPIndian Pines dataset
HUHouston University
ROSISReflective Optics System Imaging Spectrometer
AVIRISAirborne Visible/InfraRed Imaging Spectrometer
OAOverall Accuracy
AAAverage Accuracy
KappaKappa Coefficient
ROIRegion Of Interest

References

  1. Liao, X.; Liao, G.; Xiao, L. Rapeseed Storage Quality Detection Using Hyperspectral Image Technology-An Application for Future Smart Cities. J. Test. Eval. 2023, 51, 1740–1752. [Google Scholar] [CrossRef]
  2. Ghandehari, M.; Aghamohamadnia, M.; Dobler, G.; Karpf, A.; Cavalcante, C.; Buckland, K.; Qian, J.; Koonin, S. Ground based Hyperspectral Imaging of Urban Emissions. In Proceedings of the 2016 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016; pp. 1–3. [Google Scholar] [CrossRef]
  3. Wang, J.; Zhang, L.; Tong, Q.; Sun, X. The Spectral Crust project—Research on new mineral exploration technology. In Proceedings of the 2012 4th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Shanghai, China, 4–7 June 2012; pp. 1–4. [Google Scholar] [CrossRef]
  4. Contreras, C.; Khodadadzadeh, M.; Tusa, L.; Loidolt, C.; Tolosana-Delgado, R.; Gloaguen, R. Geochemical and Hyperspectral Data Fusion for Drill-Core Mineral Mapping. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–4. [Google Scholar] [CrossRef]
  5. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of spectral–temporal response surfaces by combining multispectral satellite and hyperspectral UAV imagery for precision agriculture applications. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  6. Ang, K.L.M.; Seng, J.K.P. Big data and machine learning with hyperspectral information in agriculture. IEEE Access 2021, 9, 36699–36718. [Google Scholar] [CrossRef]
  7. Samaniego, L.; Bárdossy, A.; Schulz, K. Supervised classification of remotely sensed imagery using a modified k-NN technique. IEEE Trans. Geosci. Remote Sens. 2008, 46, 2112–2125. [Google Scholar] [CrossRef]
  8. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  9. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  10. Xue, Z.; Nie, X.; Zhang, M. Incremental Dictionary Learning-Driven Tensor Low-Rank and Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–19. [Google Scholar] [CrossRef]
  11. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory. 1968, 14, 55–63. [Google Scholar] [CrossRef]
  12. Wu, Z.; Yan, Z. Selection of optimal bands for hyperspectral local feature descriptor. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5511205. [Google Scholar] [CrossRef]
  13. Xu, B.; Li, X.; Hou, W.; Wang, Y.; Wei, Y. A similarity-based ranking method for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9585–9599. [Google Scholar] [CrossRef]
  14. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear versus nonlinear PCA for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef]
  15. Ye, Q.; Yang, J.; Liu, F.; Zhao, C.; Ye, N.; Yin, T. L1-norm distance linear discriminant analysis based on an effective iterative algorithm. IEEE Trans. Circ. Syst. Video Technol. 2016, 28, 114–129. [Google Scholar] [CrossRef]
  16. Luo, F.; Zhang, L.; Zhou, X.; Guo, T.; Cheng, Y.; Yin, T. Sparse-adaptive hypergraph discriminant analysis for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1082–1086. [Google Scholar] [CrossRef]
  17. Ghamisi, P.; Benediktsson, J.A.; Ulfarsson, M.O. The spectral-spatial classification of hyperspectral images based on Hidden Markov Random Field and its Expectation-Maximization. In Proceedings of the 2013 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Melbourne, Australia, 21–26 July 2013; pp. 1107–1110. [Google Scholar] [CrossRef]
  18. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Geosci. Remote Sens. Lett. 2010, 8, 542–546. [Google Scholar] [CrossRef]
  19. He, L.; Liu, C.; Li, J.; Li, Y.; Li, S.; Yu, Z. Hyperspectral image spectral–spatial-range Gabor filtering. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4818–4836. [Google Scholar] [CrossRef]
  20. Ji, R.; Gao, Y.; Hong, R.; Liu, Q.; Tao, D.; Li, X. Spectral-spatial constraint hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 52, 1811–1824. [Google Scholar] [CrossRef]
  21. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  22. Ma, A.; Filippi, A.M.; Wang, Z.; Yin, Z.; Huo, D.; Li, X.; Güneralp, B. Fast sequential feature extraction for recurrent neural network-based hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5920–5937. [Google Scholar] [CrossRef]
  23. Yu, Q.; Wei, W.; Pan, Z.; He, J.; Wang, S.; Hong, D. GPF-Net: Graph-Polarized Fusion Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5519622. [Google Scholar] [CrossRef]
  24. Cai, Y.; Zhang, Z.; Cai, Z.; Liu, X.; Jiang, X. Hypergraph-structured autoencoder for unsupervised and semisupervised classification of hyperspectral image. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5503505. [Google Scholar] [CrossRef]
  25. Qin, A.; Tan, Z.; Wang, R.; Sun, Y.; Yang, F.; Zhao, Y.; Gao, C. Distance Constraints-based Generative Adversarial Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5511416. [Google Scholar] [CrossRef]
  26. Paoletti, M.E.; Haut, J.M.; Fernández-Beltran, R.; Plaza, J.; Plaza, A.J.; Li, J.Y.; Pla, F. Capsule Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2145–2160. [Google Scholar] [CrossRef]
  27. Qi, W.; Zhang, X.; Wang, N.; Zhang, M.; Cen, Y. A spectral-spatial cascaded 3D convolutional neural network with a convolutional long short-term memory network for hyperspectral image classification. Remote Sens. 2019, 11, 2363. [Google Scholar] [CrossRef]
  28. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A general end-to-end 2-D CNN framework for hyperspectral image change detection. IEEE Trans. Geosci. Remote Sens. 2018, 57, 3–13. [Google Scholar] [CrossRef]
  29. Wei, H.; Yangyu, H.; Li, W.; Fan, Z.; Hengchao, L. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  30. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.; Zhang, X.; Huang, X. Hyperspectral image classification with deep learning models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  31. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  32. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  33. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep pyramidal residual networks for spectral–spatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 740–754. [Google Scholar] [CrossRef]
  34. Ma, X.; Fu, A.; Wang, J.; Wang, H.; Yin, B. Hyperspectral image classification based on deep deconvolution network with skip architecture. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4781–4791. [Google Scholar] [CrossRef]
  35. Nalepa, J.; Myller, M.; Imai, Y.; Honda, K.I.; Takeda, T.; Antoniak, M. Unsupervised Segmentation of Hyperspectral Images Using 3-D Convolutional Autoencoders. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1948–1952. [Google Scholar] [CrossRef]
  36. Zhang, S.; Zhang, X.; Li, T.; Meng, H.; Cao, X.; Wang, L. Adversarial Representation Learning for Hyperspectral Image Classification with Small-Sized Labeled Set. Remote Sens. 2022, 14, 2612. [Google Scholar] [CrossRef]
  37. Sellami, A.; Tabbone, S. Deep neural networks-based relevant latent representation learning for hyperspectral image classification. Pattern Recognition 2022, 121, 108224. [Google Scholar] [CrossRef]
  38. Zhang, K.; Zhu, D.; Min, X.; Zhai, G. Implicit Neural Representation Learning for Hyperspectral Image Super-Resolution. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar] [CrossRef]
  39. Li, M.; Liu, Y.; Xue, G.; Huang, Y.; Yang, G. Exploring the Relationship Between Center and Neighborhoods: Central Vector Oriented Self-Similarity Network for Hyperspectral Image Classification. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 1979–1993. [Google Scholar] [CrossRef]
  40. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
  41. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar] [CrossRef]
  42. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3232–3245. [Google Scholar] [CrossRef]
  43. Li, Z.; Zhao, X.; Xu, Y.; Li, W.; Zhai, L.; Fang, Z.; Shi, X. Hyperspectral Image Classification with Multiattention Fusion Network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5503305. [Google Scholar] [CrossRef]
  44. Yan, H.; Zhang, E.; Wang, J.; Leng, C.; Peng, J. MTFFN: Multimodal Transfer Feature Fusion Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6008005. [Google Scholar] [CrossRef]
  45. Lyutikova, M.N.; Korobeynikov, S.M.; Rao, U.M.; Fofana, I. Mixed Insulating Liquids With Mineral Oil for High-Voltage Transformer Applications: A Review. IEEE Trans. Dielect. Electr. Insul. 2022, 29, 454–461. [Google Scholar] [CrossRef]
  46. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2021, arXiv:2010.11929v2. [Google Scholar]
  47. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 11–17 October 2021; pp. 568–578. [Google Scholar] [CrossRef]
  48. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 11–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  49. Touvron, H.; Cord, M.; Jégou, H. Deit iii: Revenge of the vit. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 516–533. [Google Scholar] [CrossRef]
  50. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.H.; Tay, F.E.; Feng, J.; Yan, S. Tokens-to-token vit: Training vision transformers from scratch on imagenet. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 11–17 October 2021; pp. 558–567. [Google Scholar] [CrossRef]
  51. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  52. Zhao, Z.; Hu, D.; Wang, H.; Yu, X. Convolutional Transformer Network for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6009005. [Google Scholar] [CrossRef]
  53. He, X.; Chen, Y.; Lin, Z. Spatial-Spectral Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  54. Ma, C.; Jiang, J.; Li, H.; Mei, X.; Bai, C. Hyperspectral Image Classification via Spectral Pooling and Hybrid Transformer. Remote Sens. 2022, 14, 4732. [Google Scholar] [CrossRef]
  55. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  56. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar] [CrossRef]
  57. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  58. Liu, F.; Ren, X.; Zhang, Z.; Sun, X.; Zou, Y. Rethinking skip connection with layer normalization. In Proceedings of the International Conference on Computational Linguistics (COLING), Barcelona, Spain, 8–13 December 2020; pp. 3586–3598. [Google Scholar] [CrossRef]
  59. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017; pp. 4700–4708. [Google Scholar] [CrossRef]
  60. Ho, Y.; Wookey, S. The real-world-weight cross-entropy loss function: Modeling the costs of mislabeling. IEEE Access 2019, 8, 4806–4813. [Google Scholar] [CrossRef]
  61. Kumar, A.; Shaikh, A.M.; Li, Y.; Bilal, H.; Yin, B. Pruning filters with L1-norm and capped L1-norm for CNN compression. Appl. Intell. 2021, 51, 1152–1160. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed MLFEN. MLFEN is composed of two primary sub-networks, including HCAM and EDVT. HCAM consists of three key components, which are band shift, convolutional operations (e.g., 3D Block and 2D Block), and AM. While EDVT incorporates two design elements, namely dense connection and AFF.
Figure 1. The overall framework of the proposed MLFEN. MLFEN is composed of two primary sub-networks, including HCAM and EDVT. HCAM consists of three key components, which are band shift, convolutional operations (e.g., 3D Block and 2D Block), and AM. While EDVT incorporates two design elements, namely dense connection and AFF.
Remotesensing 16 00590 g001
Figure 2. The illustration of the EDVT.
Figure 2. The illustration of the EDVT.
Remotesensing 16 00590 g002
Figure 3. Visualization of five datasets. (a) PU dataset (false color image (RGB-R: 56, G: 33, B: 13)), (b) SA dataset (false color image (RGB-R: 38, G: 5, B: 20)), (c) KSC dataset(false color image (RGB-R: 59, G: 40, B: 23)), (d) IP dataset (false color image (RGB-R: 50, G: 30, B: 20)), and (e) HU dataset (false color image (RGB-R: 50, G: 30, B: 20)).
Figure 3. Visualization of five datasets. (a) PU dataset (false color image (RGB-R: 56, G: 33, B: 13)), (b) SA dataset (false color image (RGB-R: 38, G: 5, B: 20)), (c) KSC dataset(false color image (RGB-R: 59, G: 40, B: 23)), (d) IP dataset (false color image (RGB-R: 50, G: 30, B: 20)), and (e) HU dataset (false color image (RGB-R: 50, G: 30, B: 20)).
Remotesensing 16 00590 g003
Figure 4. Impact of patch size k on the performance of MLFEN.
Figure 4. Impact of patch size k on the performance of MLFEN.
Remotesensing 16 00590 g004
Figure 5. Impact of the number of training samples on the performance of MLFEN.
Figure 5. Impact of the number of training samples on the performance of MLFEN.
Remotesensing 16 00590 g005
Figure 6. Classification maps on the PU dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Figure 6. Classification maps on the PU dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Remotesensing 16 00590 g006
Figure 7. Classification maps on the KSC dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Figure 7. Classification maps on the KSC dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Remotesensing 16 00590 g007
Figure 8. Classification maps on the SA dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Figure 8. Classification maps on the SA dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Remotesensing 16 00590 g008
Figure 9. Classification maps on the IP dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Figure 9. Classification maps on the IP dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Remotesensing 16 00590 g009
Figure 10. Classification maps on the HU dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Figure 10. Classification maps on the HU dataset. (a) 2D-CNN. (b) 3D-CNN. (c) HybridSN. (d) SSAN. (e) MAFN. (f) ViT. (g) HiT. (h) CVSSN. (i) SSFTT. (j) MLFEN (Ours). (k) Ground truth. (l) Color bar.
Remotesensing 16 00590 g010
Figure 11. Running time and classification performance of different models.
Figure 11. Running time and classification performance of different models.
Remotesensing 16 00590 g011
Table 1. Descriptions of the five HSI datasets.
Table 1. Descriptions of the five HSI datasets.
DatasetImage SizeCategoriesBandsBand RangeResolutionSensorLocation
PU610 × 3409103430~860 nm1.3 mROSISPavia, Italy
KSC512 × 61413176400~2500 nm18 mAVIRISFlorida, USA
SA512 × 21716204360~2500 nm3.7 mAVIRISSalinas Valley, USA
IP145 × 14516200400~2500 nm20 mAVIRISNorth-Western Indiana, USA
HU349 × 190515144364~1046 nm2.5 mITRES CASI-1500Texas, USA
Table 2. Different categories and corresponding numbers of samples for training and testing on the PU, KSC, HU, IP and SA dataset.
Table 2. Different categories and corresponding numbers of samples for training and testing on the PU, KSC, HU, IP and SA dataset.
No.Pavia University (PU) (1%)Kennedy Space Center (KSC) (1%)Houston University (HU) (1%)
ClassTrainTestClassTrainTestClassTrainTest
1Asphalt666565Scrub7747Healthy grass131238
2Meadows18618,463Willow swamp2239Stressed grass131241
3Gravel202079Cabbage palm hammock2252Synthetic grass7690
4Trees303034Cabbage palm/oak hammock2248Tree121232
5Painted metal sheets131332Slash pine2157Soil121230
6Bare Soil504979Oak/broadleaf hammock2225Water3322
7Bitumen131317Hardwood swamp2101Residential131255
8Self-Blocking Bricks363646Graminoid marsh4423Commercial121232
9Shadows9938Spartina marsh5510Road131239
10 Cattail marsh4396Highway121215
11 Salt marsh4411Railway121223
12 Mudd flats5493Parking lot1121221
13 Water9909Parking lot25464
14 Tennis court4424
15 Running track7653
Total42342353Total505111Total15014,879
No.Indian Pines (IP) (1%)Salinas (SA) (1%)
ClassTrainTestClassTrainTest
1Alfalfa145Broccoli_green_weeds_1201889
2Corn notill141414Broccoli_green_weeds_2373689
3Corn mintill8822Fallow191957
4Corn2235Fallow_rough_plow131381
5Grass pasture48435Fallow_smooth262652
6Grass trees73657Stubble393920
7Grass pasture mowed127Celery353544
8Hay windrowed5473Grapes_untrained11211,159
9Oats119Soil_vineyard_develop626141
10Soybean notill10962Corn_senesced_green_weeds323246
11Soybean mintill252430Lettuce_remaine_4wk101058
12Soybean clean6587Lettuce_remaine_5wk191908
13Wheat2203Lettuce_remaine_6wk9907
14Woods131252Lettuce_remaine_7wk101060
15Buildings grass trees drives39347Vineyard_untrained727196
16Stone steel towers984Vineyard_vertical_trellis181789
Total2579992Total53353,496
Table 3. Classification results using PCA and LDA on the PU dataset.
Table 3. Classification results using PCA and LDA on the PU dataset.
OA (%)AA (%)Kappa (%)
PCA97.37 ± 0.3595.58 ± 0.6696.41 ± 0.41
LDA97.25 ± 0.6595.43 ± 1.3296.34 ± 0.86
Table 4. Experimental results with different PCA components on the PU dataset.
Table 4. Experimental results with different PCA components on the PU dataset.
PCAOA (%)AA (%)Kappa (%)Train(s)Params (M)
3096.5194.6195.36180.9320.12
5096.5993.7895.48183.1822.13
8097.2094.5496.28209.6825.13
10097.3795.5896.41253.9627.14
Table 5. Impact of the coefficient λ on the performance of MLFEN.
Table 5. Impact of the coefficient λ on the performance of MLFEN.
IndicatorsLceLoss ( λ Value)
0.0010.0050.010.020.050.10.2
OA (%)95.1596.4696.8597.3796.9696.8196.4495.93
AA (%)92.3594.3694.6695.5894.7094.6893.8491.97
Kappa (%)93.5695.3195.8296.4195.9795.7795.9494.59
Table 6. Impact of the number of the depth of EDVT on the performance of MLFEN.
Table 6. Impact of the number of the depth of EDVT on the performance of MLFEN.
ModuleDepthIndicators
OA (%)AA (%)Kappa (%)Train (s)Test (s)
EDVT189.2881.4085.16146.075.87
293.2387.8690.92154.196.34
397.3795.5896.41253.968.04
496.6894.6795.60261.128.52
595.9792.7794.64310.239.41
695.6992.2294.27320.749.92
Table 7. Ablation study of Band Shift on the PU dataset. Optimal results are shown in bold.
Table 7. Ablation study of Band Shift on the PU dataset. Optimal results are shown in bold.
CasesOAAAKappa
without Band Shift95.9692.8294.63
with Band Shift97.3795.5896.41
Table 8. Ablation study of the proposed MLFEN with a combination of different components on the PU dataset. Optimal results are shown in bold.
Table 8. Ablation study of the proposed MLFEN with a combination of different components on the PU dataset. Optimal results are shown in bold.
CasesComponentsIndicators
HCAMEDVTLossOA (%)AA (%)Kappa (%)
1×××90.2580.6686.93
2×91.2984.3888.36
3×95.7091.6694.28
4×95.1592.3593.56
597.3795.5896.41
Table 9. The mean and standard deviation of the ten results obtained by different methods on the PU dataset. Optimal results are shown in bold.
Table 9. The mean and standard deviation of the ten results obtained by different methods on the PU dataset. Optimal results are shown in bold.
No.2D-CNN3D-CNNHybridSNSSANMAFNViTHiTCVSSNSSFTTMLFEN
194.66 ± 2.0693.62 ± 0.7793.80 ± 1.9294.14 ± 3.0198.06 ± 3.2085.92 ± 0.9089.14 ± 3.2294.89 ± 1.8093.68 ± 1.9797.10 ± 0.80
299.11 ± 0.4998.30 ± 0.3998.80 ± 0.4599.24 ± 0.2599.07 ± 0.9398.30 ± 2.2297.43 ± 1.5098.16 ± 0.7699.86 ± 0.0899.88 ± 0.11
369.86 ± 6.0458.56 ± 2.8265.48 ± 9.8575.84 ± 8.7878.10 ± 6.9678.58 ± 0.1074.18 ± 8.4787.97 ± 7.4885.62 ± 4.1788.21 ± 3.16
486.12 ± 6.0892.67 ± 1.7091.53 ± 1.2788.79 ± 4.4295.79 ± 3.5377.84 ± 1.4986.82 ± 2.2996.86 ± 2.2689.70 ± 1.3191.96 ± 1.89
599.92 ± 0.1699.71 ± 0.2899.79 ± 0.2397.16 ± 5.0798.63 ± 0.5198.27 ± 0.4899.14 ± 0.6897.97 ± 2.6199.76 ± 0.2299.74 ± 0.32
688.86 ± 2.5574.23 ± 2.7684.94 ± 5.7993.14 ± 4.7397.39 ± 1.3497.79 ± 1.1379.78 ± 2.9995.93 ± 2.2898.56 ± 0.9199.77 ± 0.22
770.50 ± 9.3374.07 ± 4.1780.84 ± 5.4191.52 ± 4.5094.05 ± 5.8985.94 ± 0.8878.76 ± 7.2091.98 ± 6.7497.43 ± 1.7798.02 ± 2.43
888.96 ± 4.7688.52 ± 4.2089.46 ± 5.5793.44 ± 0.9191.28 ± 3.4471.48 ± 9.2185.97 ± 4.7390.11 ± 4.0688.25 ± 2.4289.04 ± 2.33
999.48 ± 0.4999.93 ± 0.0999.46 ± 0.5098.30 ± 1.9799.93 ± 0.5265.27 ± 0.7198.69 ± 1.1597.84 ± 1.9087.98 ± 3.4295.88 ± 3.30
OA (%)92.68 ± 1.3590.30 ± 0.6892.93 ± 1.1195.25 ± 0.6996.33 ± 1.5489.26 ± 0.3290.59 ± 0.4795.86 ± 0.8395.98 ± 0.5197.37 ± 0.35
AA (%)87.61 ± 3.3385.33 ± 1.8189.34 ± 2.0092.62 ± 0.7694.84 ± 0.6485.13 ± 1.4587.48 ± 1.0394.63 ± 1.2993.43 ± 0.9695.58 ± 0.66
Kappa (%)90.21 ± 1.8286.98 ± 0.9290.54 ± 1.5193.68 ± 0.9395.73 ± 0.9887.45 ± 0.9387.88 ± 0.5394.51 ± 1.0994.66 ± 0.6896.41 ± 0.41
Table 10. The mean and standard deviation of the ten results obtained by different methods on the KSC dataset. Optimal results are shown in bold.
Table 10. The mean and standard deviation of the ten results obtained by different methods on the KSC dataset. Optimal results are shown in bold.
No.2D-CNN3D-CNNHybridSNSSANMAFNViTHiTCVSSNSSFTTMLFEN
195.85 ± 2.5995.84 ± 0.7097.92 ± 0.7390.75 ± 4.8696.56 ± 2.9796.94 ± 4.6795.04 ± 3.8693.37 ± 4.0899.80 ± 0.2297.42 ± 2.51
271.31 ± 7.5756.12 ± 4.9064.52 ± 9.9259.62 ± 7.6961.44 ± 9.5390.30 ± 5.5792.25 ± 7.3262.89 ± 16.1466.27 ± 4.1966.92 ± 3.86
392.91 ± 5.1058.40 ± 9.7289.16 ± 9.7789.60 ± 6.9277.91 ± 9.2462.38 ± 5.7783.27 ± 9.9278.36 ± 13.58100.00 ± 0.0099.16 ± 1.28
419.57 ± 6.5621.95 ± 0.6025.89 ± 6.0928.26 ± 9.9139.53 ± 7.9037.92 ± 9.3741.06 ± 7.8954.51 ± 9.0186.23 ± 8.4591.35 ± 7.55
514.62 ± 4.8916.37 ± 2.5715.80 ± 8.2547.02 ± 9.4238.50 ± 5.9044.06 ± 7.4341.76 ± 2.6752.04 ± 21.4581.76 ± 0.0079.87 ± 4.62
667.62 ± 9.0252.47 ± 1.8671.97 ± 7.6450.31 ± 9.8933.19 ± 3.1338.36 ± 6.5651.86 ± 7.3179.53 ± 15.2560.00 ± 11.5195.83 ± 5.20
785.74 ± 7.9394.06 ± 8.7788.82 ± 8.7385.80 ± 6.7278.60 ± 8.0355.90 ± 3.6056.00 ± 5.7079.52 ± 19.96100.00 ± 0.00100.00 ± 0.00
884.23 ± 8.7947.27 ± 4.0380.41 ± 9.1178.47 ± 5.8781.79 ± 7.9780.79 ± 9.3384.40 ± 8.0759.45 ± 8.9897.92 ± 1.6199.66 ± 0.51
986.73 ± 2.1486.22 ± 1.3787.54 ± 1.1973.92 ± 8.9690.87 ± 8.5576.62 ± 9.2187.92 ± 5.6386.78 ± 8.8486.54 ± 3.4594.67 ± 4.65
1091.35 ± 2.8684.30 ± 2.3585.19 ± 5.2270.54 ± 9.3096.50 ± 4.1596.76 ± 4.1292.75 ± 8.2871.17 ± 17.2299.85 ± 0.3299.92 ± 0.16
1199.86 ± 0.2699.27 ± 1.18100.00 ± 0.0098.09 ± 2.5690.30 ± 6.4592.74 ± 7.0595.31 ± 4.9795.21 ± 5.83100.00 ± 0.00100.00 ± 0.00
1296.98 ± 1.0793.50 ± 1.0293.23 ± 4.3080.62 ± 9.7291.50 ± 7.5878.47 ± 8.2683.90 ± 7.9589.64 ± 7.5198.43 ± 1.4499.98 ± 0.06
13100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.49 ± 1.0299.53 ± 1.4799.98 ± 0.0799.92 ± 0.2599.07 ± 1.90100.00 ± 0.00100.00 ± 0.00
OA (%)85.29 ± 2.2478.81 ± 0.6084.31 ± 2.4277.82 ± 2.6682.65 ± 5.2279.29 ± 4.8781.63 ± 4.6681.39 ± 2.8293.73 ± 0.8096.13 ± 0.75
AA (%)77.65 ± 3.1168.90 ± 0.5575.60 ± 3.3569.88 ± 3.3872.66 ± 6.5368.17 ± 7.8972.26 ± 8.3877.04 ± 3.5490.52 ± 1.1393.98 ± 1.03
Kappa (%)83.65 ± 2.4776.36 ± 0.6782.52 ± 2.7075.30 ± 2.9880.48 ± 5.9876.58 ± 5.6779.28 ± 5.3979.26 ± 3.1593.01 ± 0.8995.69 ± 0.84
Table 11. The mean and standard deviation of the ten results obtained by different methods on the SA dataset. Optimal results are shown in bold.
Table 11. The mean and standard deviation of the ten results obtained by different methods on the SA dataset. Optimal results are shown in bold.
No.2D-CNN3D-CNNHybridSNSSANMAFNViTHiTCVSSNSSFTTMLFEN
198.59 ± 2.4597.31 ± 2.6897.41 ± 3.1399.97 ± 0.0299.56 ± 0.9199.27 ± 1.5997.53 ± 2.9499.19 ± 1.1499.82 ± 0.3099.96 ± 0.05
299.91 ± 0.1899.65 ± 0.6499.91 ± 0.0899.40 ± 1.04100.00 ± 0.0099.89 ± 0.2499.74 ± 0.5499.98 ± 0.0499.99 ± 0.0499.99 ± 0.02
399.49 ± 0.1797.82 ± 0.9997.43 ± 3.5196.17 ± 1.9899.97 ± 0.0899.24 ± 1.6697.04 ± 4.8596.86 ± 1.58100.00 ± 0.0099.98 ± 0.03
499.60 ± 0.2799.59 ± 0.1297.31 ± 4.4896.50 ± 3.4699.78 ± 0.3385.76 ± 8.9188.04 ± 8.0495.87 ± 2.6699.88 ± 0.1299.11 ± 1.44
597.58 ± 0.4590.58 ± 2.8194.22 ± 3.9692.66 ± 2.5895.94 ± 5.9397.92 ± 2.1994.52 ± 3.4398.12 ± 1.6398.65 ± 0.3199.38 ± 0.40
6100.00 ± 0.00100.00 ± 0.00100.00 ± 0.0099.96 ± 0.0799.98 ± 0.0399.86 ± 0.4098.37 ± 2.8499.97 ± 0.0799.74 ± 0.7399.90 ± 0.17
798.85 ± 0.8099.58 ± 0.1999.24 ± 0.9899.30 ± 0.4299.86 ± 0.0799.73 ± 0.3598.96 ± 1.7599.29 ± 0.8699.96 ± 0.0999.96 ± 0.04
885.21 ± 2.4279.57 ± 1.9984.01 ± 3.0484.23 ± 2.2096.38 ± 1.9692.58 ± 3.8788.93 ± 7.5592.92 ± 1.7698.93 ± 0.7499.19 ± 0.66
999.94 ± 0.0599.91 ± 0.1099.97 ± 0.0398.64 ± 2.3699.99 ± 0.0299.88 ± 0.3099.77 ± 0.4799.67 ± 0.34100.00 ± 0.00100.00 ± 0.00
1095.78 ± 0.7590.96 ± 1.8792.10 ± 1.9987.90 ± 3.0098.82 ± 0.5197.79 ± 2.1791.94 ± 7.5395.76 ± 2.1398.63 ± 0.5599.45 ± 0.35
1184.23 ± 2.3884.42 ± 0.6486.56 ± 2.5289.77 ± 4.1998.87 ± 0.8399.11 ± 1.1396.83 ± 3.3994.54 ± 5.6799.90 ± 0.0998.81 ± 1.95
1299.81 ± 0.1199.96 ± 0.0499.84 ± 0.1298.64 ± 0.9799.83 ± 0.4795.97 ± 4.0694.34 ± 7.1999.66 ± 0.5099.92 ± 0.0899.94 ± 0.10
1398.44 ± 0.6097.28 ± 0.5493.44 ± 5.9494.95 ± 3.6393.25 ± 4.6192.13 ± 8.7895.95 ± 3.7999.18 ± 1.2792.18 ± 3.2293.08 ± 4.22
1498.13 ± 0.7597.80 ± 0.5098.85 ± 0.5197.35 ± 0.5898.44 ± 0.7090.09 ± 8.7975.74 ± 8.1798.20 ± 1.4599.32 ± 0.2999.23 ± 0.65
1570.37 ± 3.8671.29 ± 3.5279.70 ± 2.7182.12 ± 3.1389.12 ± 4.1866.40 ± 6.7778.49 ± 5.5889.59 ± 2.6894.42 ± 1.5096.15 ± 2.21
1696.92 ± 0.7496.67 ± 1.3296.70 ± 1.2195.22 ± 4.0599.42 ± 0.2398.69 ± 0.5894.38 ± 2.4798.68 ± 1.4499.76 ± 0.3599.88 ± 0.16
OA (%)91.91 ± 0.2090.11 ± 0.3692.34 ± 0.4192.20 ± 0.7297.28 ± 0.6789.80 ± 2.1188.94 ± 3.1196.17 ± 0.1998.69 ± 0.2699.05 ± 0.45
AA (%)95.18 ± 0.2393.90 ± 0.3594.79 ± 0.5394.55 ± 1.0198.08 ± 0.4291.78 ± 1.8489.44 ± 3.9497.34 ± 0.4598.82 ± 0.2999.00 ± 0.39
Kappa (%)90.99 ± 0.2288.99 ± 0.4091.48 ± 0.4691.31 ± 0.8096.97 ± 0.7588.60 ± 2.3988.23 ± 3.3395.73 ± 0.2198.54 ± 0.2998.95 ± 0.50
Table 12. The mean and standard deviation of the ten results obtained by different methods on the IP dataset. Optimal results are shown in bold.
Table 12. The mean and standard deviation of the ten results obtained by different methods on the IP dataset. Optimal results are shown in bold.
No.2D-CNN3D-CNNHybridSNSSANMAFNViTHiTCVSSNSSFTTMLFEN
11.34 ± 1.170.91 ± 2.200.23 ± 0.720.46 ± 1.4415.00 ± 19.472.91 ± 1.121.70 ± 1.2647.41 ± 5.5544.67 ± 29.726.67 ± 1.93
271.80 ± 1.7340.48 ± 8.3149.56 ± 6.4341.36 ± 4.7581.21 ± 4.6651.50 ± 2.7366.00 ± 11.1759.64 ± 0.9768.80 ± 3.5086.59 ± 4.96
329.88 ± 7.3823.22 ± 2.2044.42 ± 3.3313.69 ± 7.0374.56 ± 1.8234.71 ± 4.155.67 ± 4.3453.49 ± 9.5194.96 ± 1.7390.74 ± 5.10
44.32 ± 8.776.08 ± 8.8416.87 ± 5.8626.43 ± 1.7834.74 ± 5.6520.33 ± 12.1940.18 ± 14.3462.68 ± 5.6446.00 ± 6.4185.58 ± 2.62
518.12 ± 3.7325.21 ± 3.0449.46 ± 5.8736.37 ± 3.9687.82 ± 1.6930.66 ± 6.1681.73 ± 3.0262.11 ± 7.8475.44 ± 1.2996.82 ± 3.82
697.74 ± 1.5292.59 ± 3.6592.29 ± 3.990.16 ± 8.6395.14 ± 2.4485.20 ± 1.3891.25 ± 2.7180.68 ± 6.3995.64 ± 1.4498.45 ± 1.40
70.00 ± 0.000.00 ± 0.000.77 ± 2.431.83 ± 5.371.72 ± 2.4317.95 ± 5.393.82 ± 3.2325.89 ± 2.460.00 ± 0.000.00 ± 0.00
898.38 ± 0.8195.07 ± 3.0093.97 ± 8.2887.45 ± 7.3897.53 ± 1.8280.95 ± 14.9393.16 ± 4.3797.90 ± 3.8899.49 ± 0.4297.49 ± 0.42
90.00 ± 0.001.67 ± 5.272.94 ± 1.733.05 ± 1.951.30 ± 1.847.41 ± 6.931.85 ± 0.6219.20 ± 9.800.00 ± 0.0021.52 ± 1.08
1040.28 ± 8.4622.78 ± 7.2161.99 ± 2.9036.83 ± 6.7565.30 ± 16.1761.20 ± 3.5749.94 ± 14.1964.58 ± 0.1375.17 ± 1.9387.85 ± 3.69
1169.16 ± 6.3683.38 ± 8.5372.62 ± 5.0085.34 ± 3.0681.70 ± 9.5174.16 ± 6.7971.38 ± 8.0275.06 ± 5.2792.67 ± 1.8395.21 ± 1.22
1223.34 ± 1.5325.12 ± 7.3640.51 ± 3.1316.92 ± 2.1668.35 ± 9.5121.67 ± 12.0110.18 ± 6.7041.28 ± 5.2363.05 ± 9.5375.88 ± 4.21
1373.62 ± 7.659.59 ± 2.6579.59 ± 9.8280.20 ± 8.3698.19 ± 0.6863.78 ± 5.2498.28 ± 1.2876.18 ± 3.8686.75 ± 6.3298.90 ± 1.05
1496.78 ± 4.8496.05 ± 4.1094.41 ± 6.9093.75 ± 3.8790.66 ± 7.5990.53 ± 9.5381.17 ± 7.8186.10 ± 5.5299.32 ± 0.6199.17 ± 0.68
1523.73 ± 7.9440.94 ± 3.1430.48 ± 2.3328.60 ± 4.7865.13 ± 5.9431.22 ± 9.6843.53 ± 9.4763.29 ± 9.0490.10 ± 2.5992.39 ± 4.01
1636.07 ± 4.5053.37 ± 9.2257.55 ± 4.6520.45 ± 4.8887.95 ± 7.8176.76 ± 3.4577.65 ± 7.7280.44 ± 0.6546.20 ± 6.9686.81 ± 6.80
OA (%)61.32 ± 2.5258.56 ± 2.3664.68 ± 3.3859.07 ± 2.9880.19 ± 1.6460.98 ± 2.1962.05 ± 3.7467.26 ± 2.5984.33 ± 1.2089.29 ± 2.48
AA (%)42.70 ± 4.4441.65 ± 4.3248.69 ± 6.0941.13 ± 4.0365.38 ± 2.4446.75 ± 5.3350.75 ± 2.7562.25 ± 2.7067.39 ± 2.3572.64 ± 6.31
Kappa (%)55.41 ± 2.6951.34 ± 3.0759.50 ± 3.9652.04 ± 3.7377.31 ± 1.8555.06 ± 2.5456.10 ± 4.5062.70 ± 2.9082.06 ± 1.3787.70 ± 2.88
Table 13. The mean and standard deviation of the ten results obtained by different methods on the HU dataset. Optimal results are shown in bold.
Table 13. The mean and standard deviation of the ten results obtained by different methods on the HU dataset. Optimal results are shown in bold.
No.2D-CNN3D-CNNHybridSNSSANMAFNViTHiTCVSSNSSFTTMLFEN
194.50 ± 4.7098.96 ± 1.1798.80 ± 1.5592.75 ± 5.3291.93 ± 1.1789.86 ± 5.5191.67 ± 6.3089.15 ± 6.9495.70 ± 1.4993.34 ± 1.52
283.20 ± 3.8684.66 ± 2.8986.21 ± 3.6483.79 ± 4.5890.99 ± 5.6579.72 ± 1.1593.83 ± 6.4192.33 ± 5.0691.39 ± 3.6492.52 ± 2.98
371.40 ± 21.1691.34 ± 1.8491.49 ± 5.2095.12 ± 5.3597.16 ± 1.3794.06 ± 7.6198.80 ± 1.7894.22 ± 5.8696.02 ± 0.8993.24 ± 2.36
492.50 ± 3.7291.38 ± 2.8487.36 ± 8.9991.81 ± 1.5878.56 ± 5.0973.25 ± 1.4990.70 ± 6.2386.62 ± 12.2886.57 ± 2.0684.44 ± 6.53
5100.00 ± 0.0098.76 ± 1.2299.60 ± 0.5794.63 ± 5.5199.72 ± 0.5797.76 ± 5.8998.19 ± 2.2093.12 ± 2.27100.00 ± 0.00100.00 ± 0.00
667.07 ± 5.7169.06 ± 6.0570.12 ± 3.6962.39 ± 14.0380.92 ± 3.1772.63 ± 6.9272.26 ± 12.8883.49 ± 17.1581.76 ± 2.5983.57 ± 3.44
765.33 ± 8.0358.87 ± 3.0667.84 ± 7.8054.45 ± 3.0169.08 ± 9.7873.94 ± 4.2884.37 ± 8.1382.26 ± 4.2675.28 ± 2.1078.62 ± 3.86
859.71 ± 4.8656.35 ± 3.3171.61 ± 6.0158.42 ± 9.9074.15 ± 5.4664.46 ± 2.3161.49 ± 11.8876.77 ± 7.5879.70 ± 6.6277.27 ± 1.52
973.75 ± 6.0877.38 ± 4.0458.80 ± 9.2977.86 ± 5.0967.99 ± 13.6852.99 ± 8.5776.88 ± 10.7982.33 ± 6.5679.64 ± 3.5379.80 ± 14.88
1043.30 ± 10.4668.96 ± 6.9378.29 ± 5.6978.17 ± 13.1888.37 ± 6.6271.73 ± 4.1568.60 ± 14.4573.30 ± 7.6688.72 ± 4.4293.33 ± 4.96
1170.04 ± 9.6271.38 ± 8.6677.17 ± 8.2172.25 ± 5.0775.56 ± 10.2568.79 ± 1.1875.19 ± 9.1579.90 ± 8.3876.67 ± 5.6785.65 ± 7.11
1269.34 ± 10.8665.57 ± 8.9980.37 ± 10.0760.05 ± 11.8180.91 ± 4.4073.30 ± 1.9471.46 ± 12.7577.23 ± 8.3183.38 ± 3.6389.13 ± 5.87
1367.56 ± 7.6591.26 ± 1.5563.09 ± 8.5571.83 ± 8.0674.58 ± 11.4942.66 ± 5.3358.15 ± 15.6287.66 ± 7.8890.22 ± 3.3280.72 ± 7.77
1462.93 ± 19.2991.58 ± 3.0686.07 ± 10.2481.42 ± 13.3999.63 ± 0.4397.07 ± 2.2391.34 ± 18.0174.98 ± 5.29100.00 ± 0.0099.67 ± 0.49
1598.91 ± 0.6198.77 ± 0.7398.39 ± 1.4692.57 ± 2.8299.91 ± 0.1897.93 ± 3.4598.95 ± 1.3389.05 ± 7.43100.00 ± 0.0099.98 ± 0.05
OA (%)74.89 ± 1.1879.29 ± 1.5080.81 ± 1.7377.40 ± 1.8683.25 ± 2.7975.23 ± 3.4581.93 ± 2.2483.04 ± 2.2087.09 ± 1.0388.17 ± 1.81
AA (%)74.64 ± 1.8480.95 ± 1.4081.01 ± 1.6977.83 ± 2.0684.63 ± 2.5176.01 ± 2.9982.13 ± 2.7984.16 ± 2.0988.34 ± 0.8888.75 ± 1.80
Kappa (%)72.82 ± 1.2877.62 ± 1.6279.25 ± 1.8775.58 ± 2.0181.90 ± 3.0273.30 ± 3.6580.45 ± 2.4381.67 ± 2.3986.05 ± 1.1187.21 ± 1.96
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fang, S.; Li, X.; Tian, S.; Chen, W.; Zhang, E. Multi-Level Feature Extraction Networks for Hyperspectral Image Classification. Remote Sens. 2024, 16, 590. https://doi.org/10.3390/rs16030590

AMA Style

Fang S, Li X, Tian S, Chen W, Zhang E. Multi-Level Feature Extraction Networks for Hyperspectral Image Classification. Remote Sensing. 2024; 16(3):590. https://doi.org/10.3390/rs16030590

Chicago/Turabian Style

Fang, Shaoyi, Xinyu Li, Shimao Tian, Weihao Chen, and Erlei Zhang. 2024. "Multi-Level Feature Extraction Networks for Hyperspectral Image Classification" Remote Sensing 16, no. 3: 590. https://doi.org/10.3390/rs16030590

APA Style

Fang, S., Li, X., Tian, S., Chen, W., & Zhang, E. (2024). Multi-Level Feature Extraction Networks for Hyperspectral Image Classification. Remote Sensing, 16(3), 590. https://doi.org/10.3390/rs16030590

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop