Next Article in Journal
Integrating sEMG and IMU Sensors in an e-Textile Smart Vest for Forward Posture Monitoring: First Steps
Next Article in Special Issue
A Student Facial Expression Recognition Model Based on Multi-Scale and Deep Fine-Grained Feature Attention Enhancement
Previous Article in Journal
Network Security Prediction of Industrial Control Based on Projection Equalization Optimization Algorithm
Previous Article in Special Issue
Attention Mechanism-Based Glaucoma Classification Model Using Retinal Fundus Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Joint Network of Edge-Aware and Spectral–Spatial Feature Learning for Hyperspectral Image Classification

1
College of Computer Science and Information Engineering, Harbin Normal University, Harbin 150025, China
2
Department of Municipal and Environmental Engineering, Heilongjiang Institute of Construction Technology, Harbin 150025, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(14), 4714; https://doi.org/10.3390/s24144714
Submission received: 3 June 2024 / Revised: 9 July 2024 / Accepted: 18 July 2024 / Published: 20 July 2024

Abstract

:
Hyperspectral image (HSI) classification is a vital part of the HSI application field. Since HSIs contain rich spectral information, it is a major challenge to effectively extract deep representation features. In existing methods, although edge data augmentation is used to strengthen the edge representation, a large amount of high-frequency noise is also introduced at the edges. In addition, the importance of different spectra for classification decisions has not been emphasized. Responding to the above challenges, we propose an edge-aware and spectral–spatial feature learning network (ESSN). ESSN contains an edge feature augment block and a spectral–spatial feature extraction block. Firstly, in the edge feature augment block, the edges of the image are sensed, and the edge features of different spectral bands are adaptively strengthened. Then, in the spectral–spatial feature extraction block, the weights of different spectra are adaptively adjusted, and more comprehensive depth representation features are extracted on this basis. Extensive experiments on three publicly available hyperspectral datasets have been conducted, and the experimental results indicate that the proposed method has higher accuracy and immunity to interference compared to state-of-the-art (SOTA) method.

1. Introduction

HSIs are complete spectral information for each pixel generated by hyperspectral sensors by capturing the reflection information of an object in multiple consecutive spectral bands through hyperspectral imaging technology. Compared to RGB images, HSIs also contain the information about the shape, texture, and structure of the object [1], but HSIs contain a large amount of waveband information, which allows identification and differentiation of substances with similar colors but different spectral characteristics. Thus, HSIs are widely used in scientific and industrial applications that require precise substance identification and analysis, such as medical imaging and diagnosis [2], geological and mineral exploration [3], environmental protection [4], agricultural crop monitoring [5], food safety monitoring [6], and military reconnaissance and security [7]. To fully exploit the value of HSIs, many subtasks are derived, such as classification [8,9], target detection [10,11,12], and unmixing [13,14,15]. Among these tasks, the land cover information classification task has received extensive attention.
When classifying objects in HSIs, due to the phenomenon that “the spectra of the same object may be different and the spectra of different objects may be the same” exists in HSIs [16], therefore, it is not feasible to simply apply the same methods used for the RGB image classification to the HSIs classification. To address the above challenges, researchers around the world have proposed various approaches, such as principal component analysis (PCA) [17], the Bayesian estimation method [18], SVM [19,20], and k-mean clustering [21,22].
However, with the breakthrough of deep learning, convolutional neural networks (CNNs) are gradually replacing the traditional HSI classification methods due to their stronger model generalization ability and deep feature characterization. And in recent years, in the field of HSI classification, CNNs have been rapidly developed. For example, Hu et al. used 1-D CNN [23] in order to extract the spectral information. But for the HSI classification task, using only spectral information is not enough to realize to obtain accurate classification results. Therefore, Zhao et al. proposed 2-D CNN [24] to extract spatial features. However, both 1-D CNN and 2-D CNN do not fully utilize the 3-D characteristics of HSIs. Thus, Chen et al. applied 3-D CNN [25] to the field of HSI classification in order to fuse the spatial–spectral features of HSIs, and the experimental results showed that the performance of the model was improved. Based on these experiments, many researchers have proposed hybrid convolutional methods [26,27,28,29,30,31,32]. Among them, Roy et al. proposed the HybridSN [29] with a linear structure. HybridSN contains three sequentially connected 3D convolutional layers for fusing spatial and spectral information and one 2D convolutional layer for extracting local spatial features, respectively. In addition, Zhong et al. proposed the SSRN [32], where they introduced residual connectivity between convolutional blocks to promote backpropagation of the gradient. However, these convolution-based methods are limited by the convolutional kernel, which can only learn the information within the coverage of the convolutional kernel, thus restricting the representation of global features.
During this period, Dosovitskiy et al. proposed the vision transformer (ViT) [33], and the multi-head attention mechanism proposed in the paper greatly alleviated the problem of sensory field limitation of convolution-based methods. Since then, a large number of methods combining ViT and CNN appeared [33,34,35,36,37,38,39,40,41,42,43,44,45], such as SSTN [45], SSFTT [37], CTMixer [43], and 3DCT [39]. These methods have both global feature perception capability based on ViT methods and local feature fusion capability based on CNN methods. Thus, compared to methods [46,47,48,49] that only use CNN to obtain spectral–spatial features, these mixed-use CNN and ViT methods are able to extract more comprehensive spectral–spatial features due to the use of a global attention mechanism.
However, the characteristics of HSIs, especially the importance of edge features between different classes and spectral bands in the classification process, are not fully considered in these methods that use a mixture of CNN and ViT. To enhance edge features, edge data augmentation methods are often employed. Traditional image edge data augmentation methods usually apply edge detection operators (e.g., Laplacian, Canny, Sobel operators) [50] directly on the original image to obtain the edge information, which is then used for subsequent model training by directly superimposing with the original image. However, in the field of HSI classification, due to the existence of the characteristic that the boundaries of the same object may be different in different spectral bands, processing the original data in this way will cause a large amount of noise, which will affect the subsequent classification performance. In order to minimize the effects of superimposed noise, Tu et al. applied edge-preserving filtering to the edge portion in their proposed MSFE [51] with pyramidal structure, but the MSFE does not take into account the fact that different spectral bands play different roles in the classification process.
Therefore, inspired by the above work, and in order to enhance the image features and weaken the noise interference of the initial HSI, we adopt a dynamic learning approach to obtain the edge information and the decision weights of different spectral bands. Then, we use a mixture of attention mechanisms and CNN on this basis with the aim of obtaining global spectral–spatial features. Figure 1 shows the edge-aware and spectral–spatial feature extraction network which we propose. The network contains two parts: an edge feature augment block and a spectral–spatial feature extraction block. Different from traditional data augmentation that is not dynamically learnable, our edge feature augment block adaptively learns the degree of edge feature enhancement in different spectral bands, which reduces the high-frequency noise. In addition, in the spectral attention block, we adaptively adjust the weights of different spectral bands for classification and then perform feature extraction on its basis. To sum up, there are three main contributions:
  • We propose a novel feature extraction network (ESSN) with richer and more efficient representation of edge features and spectral–spatial features compared to existing networks;
  • We designed a novel edge feature augment block. The block consists of an edge-aware part and a dynamic adjustment part. Compared with edge data augmentation methods that are not dynamically learnable, this block greatly reduces edge distortion and noise amplification;
  • We propose a spectral–spatial features extraction block. It contains a spectral attention block, a spatial attention block, and a 3D–2D hybrid convolution block. Among them, the spectral attention block and the spatial attention block gain an effective feature by enhancing the information favorable for classification and suppressing noise and other interfering information. The convolution block fuses the above features.
The subsequent sections are composed as follows. Our proposed method is described in Section 2. In Section 3, we describe our experimental environment and make a detailed comparison with other SOTA methods in the same environment. We perform sensitivity analysis experiments and ablation experiments aimed at verifying the importance of each part of the model in Section 4. In Section 5, we distill the paper and suggest directions for model improvement.

2. Methodology

Figure 1 shows the whole process of HSI classification. It consists of a data preprocessing block, the backbone of the proposed network, and a linear classifier.
Real objects are given a hyperspectral image (HSI) after passing through a hyperspectral sensor. Assuming that the HSI is I r a w R H × W × C . H ,   W ,   C are the height, width, and number of spectral bands of the raw HSI image, respectively. In HSI, each pixel can be represented by the vector X p i x e l = ( V 1 , V 2 , , V c ) , where V c represents the pixel value on the Cth spectrum. Obviously, the greater the number of spectra, the richer the information, but this greatly slows down computational efficiency. Therefore, we adopt the PCA technique to preprocess the HSI data to improve the efficiency, maintain the height and width unchanged, and reduce the spectral number from C to P . We denote the HSI after PCA dimensionality reduction as I p c a R H × W × P , where P denotes the number of the spectra after PCA dimensionality reduction. In order to obtain a suitable input format for the network, we crop the image into patches I p a t c h R h × w × P with pixel-centered dimensions as h ,   w ,   P , where h ,   w ,   P represent the height, width, and spectral number of the patch, respectively. The data preprocessing block is shown in Figure 1. Note that the same symbols appearing in this section represent the same meaning.
The backbone of ESSN contains both an edge feature augment block and a global spectral–spatial feature extraction block, and we will describe the content of ESSN in as much detail as possible. Finally, we use a linear classifier to determine the class of each pixel.

2.1. Edge Feature Augment Block

As shown in Figure 2, the points where the model fails to predict are mostly at the intersection of different categories, which is due to the high feature similarity between certain categories on the one hand, and the possible boundary blurring between certain categories [52].
Previously, edge data augmentations were usually used to strengthen the edge for the above problems. However, the direct superposition of edge information may produce strong edge noise, leading to the confusion of similar categories. Therefore, we propose a novel edge feature augment block, as shown in Figure 1, which can adaptively adjust the model’s emphasis on the edges of a region by learning the importance of the edge information in the input data, and personalize the edge information.

Laplacian of Gaussian Operator

The Laplacian of Gaussian operator is generated by the convolution of the Laplace operator and Gaussian filtering operator. The Laplace operator is particularly sensitive to regions of the image that change abruptly, and therefore has a better performance in edge-awareness tasks. Due to the prevalence of Gaussian noise in images captured by electronic devices, which seriously affects the accuracy of edge perception, hyperspectral images need to be processed with Gaussian filtering before perceiving the edges. The Gaussian filtering operator and Laplace operator can be expressed by Equations (1) and (2), respectively:
G σ x , y = 1 2 π σ 2 e ( x 2 + y 2 2 σ 2 )
L x , y = 2 I x 2 + 2 I y 2
where x , y denote the spatial coordinate positions of the HSIs, σ is the Gaussian standard deviation, and I represents the value of the pixel on the image.
Convolutional operations have the law of union, so we use the result of the convolution of the Gaussian filter operator with the Laplace operator as a new edge-aware operator ( L o G ) and then use the L o G to convolve the image to obtain the image edges. The L o G expression is shown in Equation (3).
L o G x , y = 1 π σ 4 [ 1 x 2 + y 2 2 σ 2 ] e ( x 2 + y 2 2 σ 2 )
Due to the discrete representation of hyperspectral images, we discretize Equation (3) to obtain an approximate L o G operator for practical use. As shown in Figure 3, we list the L o G operators for the two cases σ < 0.5 and σ = 1.4 , respectively. Then, let the result after edge-aware operator be I L o G with the following expression:
I L o G = D W C o n v L o G ( I p a t c h ) R h × w × P
where D W C o n v L o G ( ) indicates depthwise separable convolution with the kernel of L o G .
In the edge feature augment block, because of the characteristic that “the spectra of the same object may be different and the spectra of different objects may be the same”, strengthening the edge features at the same rate in different spectral bands will generate interference noise, so we design a learnable parameter γ R 1 × P for adjusting the degree of feature augment in different spectra. We explore the importance of γ in Section 4.2. And in order to make the network more flexible and the optimization process smoother and more efficient, we use residual connectivity. The output ( I L o G o u t ) of the module is shown below:
I L o G o u t = I p a t c h + γ ( φ ( I L o G ) I p a t c h ) R h × w × P
where φ ( ) indicates activation function of sigmoid. denotes the dot product of the corresponding position.

2.2. Spectral–Spatial Feature Extraction Block

2.2.1. Spectral Attention Block

HSIs are rich in spectral information; to make it easier to see, we show the specific image of each spectral band by means of a grayscale map, as shown in Figure 4. Obviously, the importance of different spectra in the decision-making process is different [53]. The spectral attention helps the model in adaptive adjustment of weights for different spectra and in enhancing the representation of these spectra during the learning process. This helps the model to suppress the influence of task-irrelevant spectra.
In the spectral attention block, to strengthen the correlation between encoded and decoded data, we use residual concatenation. Let the input features of the block be I i n p u t R h × w × P , then the results after global maximal pooling and global mean pooling are v m a x and v m e a n , respectively, where global maximum pooling is complementary to global average pooling. The results are shown below.
v m a x ( p ) = M a x ( I i n p u t ( p ) ) R 1 × P
v m e a n ( p ) = 1 h × w x = 1 h y = 1 w I i n p u t ( x , y , p ) R 1 × P
To reduce the size of parameters, the pooled features are entered into a shared multi-layer perception (MLP), and the results h m a x and h m e a n are obtained. Let the rate of dimensionality reduction be r and the weights of the two MLP layers be, in order, W 1 R p × ( p / r ) and W 2 R ( p / r ) × p . h m a x and h m e a n are as follows:
h m a x = W 2 ( W 1 ( v m a x ) ) R 1 × P
h m e a n = W 2 ( W 1 ( v m e a n ) ) R 1 × P
Adaptive spectral weights W h of the input feature map are obtained by adding h m a x and h m e a n and passing through the sigmoid activation function. W h is shown below:
W h = σ ( h m a x + h m e a n ) R 1 × P
where σ ( · ) is the activation function of sigmoid.
Finally, let the output of the spectral attention block be I o u t p u t R h × w × P , the expression is as follows:
I o u t p u t = I i n p u t + I i n p u t W h R h × w × P

2.2.2. Spatial Attention Block

In contrast to traditional convolutional operations, which focus on only a portion of the input data, the spatial attention mechanism [54] can adaptively adjust the area of attention over the global spatial range of the input data and give more importance and weight to these locations during preprocessing, thus improving the recognition accuracy and efficiency of the model.
In Figure 5, the structure of spatial attention in the spatial attention block is illustrated. Considering that the spatial information of the same location may behave differently on different spectral bands, we first fused the local spatial features by 2D convolution and then projected the convolved feature maps to obtain Q , K , and V , respectively:
Q = C o n v 2 D ( I ) W Q R h w × p
K = C o n v 2 D ( I ) W K R h w × p
V = C o n v 2 D ( I ) W V R h w × P
Then, the attention map A t t n can be calculated as follows:
A t t n = S o f t m a x ( Q · K T / d k ) R h w × h w
where d k is the dimension of K . Let O u t p u t s a be the output in the network of Figure 4, the expression is as follows:
O u t p u t s a = A t t n · V = S o f t m a x ( Q · K T / d k ) V R h w × P
Finally, we reshape O u t p u t s a to the size of h × w × P for subsequent processing.
In Figure 1, the spatial attention block contains two spatial attention parts; we use two convolutional kernels of different sizes on two spatial attention parts to enhance the perceptual region. In addition, in order to strengthen spatial expression, we use a residual structure.

2.2.3. 2D–3D Convolution

2D convolution layer can extract spatial features, and 3D convolution layer can extract spectral features. Therefore, as shown in Figure 1, 2D convolution and 3D convolution are used in the spectral–spatial feature extraction block. In the 2D–3D convolution block, three consecutive 3D convolutional layers with a different kernel and one 2D convolutional layer are included in the 2D–3D convolutional block. A detailed description is shown below.
In the 3D convolution layer, a single 3D convolution can be regarded as a 3D convolution kernel sliding along the three dimensions (H, W, C). During the convolution process, the spatial and spectral information of the neighboring spectra are fused. And the values of the nth feature map of the mth layer at the spatial location of ( x , y , z ) are as follows:
v m , n i , j , k = φ ( b m , n + τ = 1 d l 1 λ = h m c m ρ = w m h m σ = c m w m ω m , n , τ σ , ρ , λ × v m 1 , τ i + σ , j + ρ , k + λ )
where φ ( · ) is the activation function, and b m , n and ω m , n are the bias parameters and weight values of kernel corresponding to the nth feature map of the mth layer, respectively. d l 1 indicates the number of feature maps in the (l − 1)th layer and the depth of ω m , n . The height, width, and spectral dimension of the kernel are ( 2 h m + 1 ) , ( 2 w m + 1 ) , and ( 2 c m + 1 ) , respectively.
In the 2D convolution layer, the convolution kernel slides over the entire space, and the output of the convolution is the sum of the dot products between the kernel and the input data. During the convolution process, the information of different spectra in the same space is fully integrated. In 2D convolution, the values of the nth feature map of the mth layer at the spatial location of ( x , y , z ) are as follows:
v m , n i , j = φ ( b m , n + τ = 1 d l 1 ρ = w m h m σ = c m w m ω m , n , τ σ , ρ × v m 1 , τ i + σ , j + ρ )
the parameters appearing in Equation (18) represent the same meaning as in Equation (17).

3. Comparison Experiments

3.1. HSI Datasets

We apply the model to three public datasets: the Indian Pines (IP), the Kennedy Space Center (KSC), and the Pavia University (PU) to evaluate the validity of model.
The IP size is 145 × 145 pixels, and the spatial resolution is 20 m. In the experiment, noisy bands and water absorption have been removed, and 200 spectral bands have been filtered out. The false-color map, ground truth map, training ratios, and 16 vegetation classes of the IP dataset are shown in Figure 6.
The KSC size is 512 × 614 pixels, and the spatial resolution is 18 m. In the experiment, noisy bands and water absorption have been removed, and 176 spectral bands have been filtered out. The false-color map, ground truth map, training ratios, and 13 wetland categories of the KSC dataset are shown in Figure 7.
The PU size is 610 × 340 pixels, and the spatial resolution is 1.3 m. In the experiment, noisy bands have been removed, and 103 spectral bands have been filtered out. The false-color map, ground truth map, training ratio, and 9 urban land cover categories of the PU dataset are shown in Figure 8.
In order to avoid the effect of dataset randomness as much as possible, all experiments under this section are trained and tested on ten randomly generated identical training sets and corresponding test sets.

3.2. Experimental Setting

3.2.1. Measurement Indicators

The overall accuracy (OA), average accuracy (AA), and kappa coefficient (κ) metrics are being used for quantitative evaluation to fairly consider ESSN against other comparative methods. The larger the value of the model corresponding to the above three metrics, the better the model performs.

3.2.2. Environment Configuration

The software environment for all experiments is PyTorch1.12.0, cuDNN8.0, CUDA11.7, and Python3.8. The hardware environment is an Intel i5-12490F CPU, an NVIDIA GeForce RTX 3060 GPU, RAM 32 GB, and 1 TB of memory. The Stochastic Gradient Descent (SGD) optimizer was chosen as the initial optimizer for all experiments, the cross-entropy loss function is used to calculate the loss, the learning rate is set to 0.01, the batch size is 100, and the patch size is 15 × 15. One hundred epochs are applied to each dataset.

3.3. Comparison Experiment Result

We have experimented fully in the same environment and compared it with other SOTA models. In total, eight comparison models were selected, including 2-D CNN [24], 3-D CNN [25], HybridSN [29], SSRN [32], SSTN [45], SSFTT [37], CTMixer [43], and 3DCT [39]. In order to attain a relatively fair comparison result, only the model was changed during the comparison process, and the other parts were kept unchanged, so that the influence of other external factors could be minimized.
Table 1, Table 2 and Table 3 show the mean classification results and standard deviation for each category on the IP, KSC, and PU datasets, respectively, as well as the evaluation metrics for each model. Results are in percentage terms. Figure 9, Figure 10 and Figure 11 show the performance of each model on the same training set on the IP, KSC, and PU datasets, respectively. And in Figure 9, Figure 10 and Figure 11, the background is labeled as black pixels, the category to be predicted is labeled as colored pixels.
By looking at the classification result plots, it can be found that the points of classification errors of each model occurred mostly in the regions where multiple categories appeared densely, so we zoomed in on the localized multi-category regions.

3.3.1. Comparative Results on the IP

By observing Table 1, it shows that ESSN outperforms other methods, and the overall results are OA (98.93%), AA (96.49%), and Kappa (98.78%). And compared with the second-best model (HyBridSN [29]), all performance metrics are improved, 0.21% higher in OA, 0.68% higher in AA, and 0.24% higher in Kappa, respectively. This indicates that the inclusion of edge feature augment and the spectral–spatial attention block is beneficial to the feature representation of the model. In addition, it is observed that all models perform poorly on classes ‘1’ and ‘9’, and by analyzing the IP dataset, it can be found that the overall number of classes ‘1’ and ‘9’ is too small. Consequently, when using the approach of proportional random selection for the training dataset, an inadequate number of samples is obtained for this particular class, resulting in suboptimal classification performance by all models for that class.
In addition, from the overall plot in Figure 9, ESSN has the best overall performance, and from the local plots with high category density, ESSN also outperforms the other models in places with high category density. This indirectly shows that the edge feature enhancement is effective. Also, Figure 9 verifies that most of the places where the prediction is wrong are common areas with different categories.

3.3.2. Comparative Results on the KSC

By observing Table 2, it shows that ESSN is higher than the others. The overall results are OA (99.18%), AA (98.79%), and Kappa (99.08%). Compared with the second-best model (HyBridSN [29]), all performance metrics are improved, 0.19% higher in OA, 0.32% higher in AA, and 0.21% higher in Kappa, respectively. In addition, it can be found that ESSN performance is optimal in most categories, and on a small number of categories, the classification results of our model do not differ much from the best classification results.
Based on Figure 10, it can be seen that ESSN outperforms the other models on the KSC dataset in general. As can be seen in the local zoomed-in image, at the edges of the categories in the category-dense region, ESSN’s performance is also better than the other models. The reason for obtaining the above results is that in KSC, in general, each category is more independent, and there are not many sample points that are in close proximity to other categories. The performance of ESSN on the edges is much sharper than the method without edge feature augment.

3.3.3. Comparative Results on the PU

By observing Table 3, the overall performance of ESSN is OA (99.45%), AA (98.82%), and Kappa (99.27%). Compared with the second-best model (3DCT [39]), all performance metrics are improved, where OA is improved by 0.09%, AA by 0.14%, and Kappa by 0.12%, respectively. Combining Table 3 and Figure 7 and Figure 11, compared with the HybridSN [29], which has the highest number of optimal classification accuracies in a single category, it can be found that ESSN’s classification accuracies are competitive, if not optimal, in categories ‘1’, ‘2’, ‘7’, and ‘8’, but ESSN outperforms it significantly in categories ‘3’, ‘4’, and ‘9’. The same result can be observed in the localized zoomed-in plots in Figure 11. This explains why the average classification accuracy of ESSN is higher than that of other methods, even though ESSN’s single-class classification accuracy is not optimal on most categories.

3.4. Depletion of Resources

We take the parameter size, training time, and testing time as resource consumption training metrics, and smaller values are better for all indicators. All results are stored in Table 4.
Based on Table 4, compared with 3DCT, although ESSN contains the largest number of parameters, ESSN has a shorter training time. And by comparing SSTN and 2-D CNN, it can be found that the parameter size of the model does not necessarily have a linear relationship with the training time.

4. Discussion and Analysis

4.1. Parametric Analysis

In this section, we perform a sensitivity analysis on three parameters: patch size, training ratio, and operator of L o G , respectively, and explore their impact on model performance.
In Table 5, ESSN performs better on IP and PU when patch size is selected as 15 × 15 but does not perform as well on KSC as when patch size is 19 × 19. Considering the amount of computation, patch size 15 × 15 is selected as the optimal size. In addition, as the patch size increases, the OA on the KSC becomes larger and larger, and combined with the full ground truth map of KSC, there are two reasons for this result: one is that as the patches increase, each patch contains more spatial information, and thus the model can learn more key elements from it. The other is that as the patches increase, the longer distance edges are gradually incorporated into the model’s observation range. In addition, by looking at the ground truth plots of IP and PU, it can be seen that there is more category intermingling in these two datasets. Increasing the patch can obtain a larger perceptual field which is beneficial to the model but, at the same time, will introduce more noise and confusing information which is detrimental to the model. Therefore, ESSN shows a performance on both IP and KSC that first increases with increasing patch and then decreases with increasing patch. From Table 5, the point of patch size of 15 × 15 is the cutoff point where the model performance goes from up to down.
As seen in Figure 12, apparently, as expected, the performance of all models improves with increasing training samples, with OA gradually approaching 100%. In addition, ESSN has a large advantage when the number of training samples is insufficient, and the OA of ESSN gradually decreases with the increase in samples used for training, and eventually the performance of all models gradually converges to the same. All in all, ESSN outperforms other models on different training ratios.
Figure 13 shows the performance on different cases. Comparing ‘a’ with ‘c’, it is clear that if traditional data enhancement methods are used without processing the raw data with learnable parameters, the performance is not as good as when using edge feature augment blocks. In addition, comparing ‘a’ with ‘e’, it can be found that traditional data augment does not have a positive effect. Especially on the KSC dataset, it greatly reduces its classification performance. When using the edge feature augment block, it can be seen that the performance of the different operators of L o G is very close. And comparing ‘e’ with ‘b’, ‘c’, ‘d’, the classification capabilities of the model are improved when the edge feature augment block is added. Comparing ‘b’ with ‘c’, the performance gap between the two on all the datasets used in the experiment is super small. The reason is that the L o G operators used for both ‘b’ and ‘c’ are discrete approximations at σ < 0.5 , and the difference between the two is the difference in the angle at which the rotational invariance is satisfied, with the L o G operator corresponding to ’b’ having invariant results for rotations in the 90° direction, and the L o G operator corresponding to ‘c’ having invariant results for rotations in the 45° direction. In this study, raw data are not rotationally transformed, so the difference between the two is not significant, and both are better than the case corresponding to ’e’. Obviously, ‘b’ performs optimally on the KSC, and ‘c’ performs optimally on the IP and PU. After comprehensive consideration, the L o G operator corresponding to ‘c’ is chosen.

4.2. Ablation Experiment

In this section, for the influence of external factors, the training samples in the ablation experiments are kept the same as those in the experiments in Section 3. Thus, effects arising from hyperparameters and randomized training samples are excluded.
PCA operation is used in the data preprocessing part, but it also causes loss of spectral information when extracting the principal components of hyperspectral images. Therefore, we explore the effect of PCA operation on the comprehensive performance of the model by conducting ablation experiments of PCA operation on three datasets: IP, KSC, and PU. The experimental results are shown in Figure 14. When PCA operation is used to downscale to 50 dimensions, it can be clearly observed from subplot (a) in Figure 14 that the classification accuracy of the model under PCA operations is not very different from the classification performance of the model without PCA operations, but from subplot (b) in Figure 14, it can be found that the time cost of the model can be greatly reduced by using PCA operations. With the comprehensive consideration of model classification accuracy and time cost, we adopt PCA to preprocess the original data.
In addition, in order to analyze the effect of each component in the model on the model performance, a total of seven combinations were considered in this experiment. All the results are shown in Table 6, which depicts the edge feature augment block with “Edge Block” instead, the spectral attention block with “Spectral Block” instead, and the spatial attention block with “Spatial Block” instead, respectively.
Comparing ‘7’ with ‘1‘ to ‘6’, when all blocks are employed, the model has the best performance.
Comparing ‘7’ to ‘6’, ‘5’ to ‘3’, and ‘4‘to ‘2’, respectively, it can be seen that adding the edge feature augment block improves the model’s performance in IP, KSC, and PU. Combined with the experimental conclusions about edges drawn in Section 4.1, it can be found that the edge feature augment block is different from the traditional edge data augmentation and plays a positive role in improving the network performance. Comparing ‘7’ to ‘5’, ‘6’ to 3, and ‘4’ to ‘1’, respectively, it can be found that the ability to model will be strengthened on IP, KSC, and PU when adding the spectral attention block. Comparing ‘7’ to ‘4’, ‘6’ to ‘2’, and ‘5’ to ‘1’, adding the spectral attention block is also beneficial in improving the model’s performance on IP, KSC, and PU.
Comparing ‘7’ with ‘1’ to ‘6’, it is found that ‘7’ performs optimally on IP, KSC, and PU, which indicates that the combination of edge feature augment block, spectral attention block, and spatial attention block is effective. Then, comparing ‘4’ with ‘1’ and ‘2’, ‘5’ with ‘1’ and ‘3’, and ‘6’ with ‘2’ and ‘3’, it can be found that the performance of the combination of two blocks is better than a single block.
Comparing ‘1’, ‘2’, and ‘3’, it can be found that the model using only the edge feature augment block and the model using only the spectral attention block perform similarly on IP, KSC, and PU, while the model using only the spatial attention block does not perform as well. This is because for the models using either the edge feature augment block or the spectral attention block, this has the effect of removing noise, whereas using the spatial attention block directly introduces this invalid noise into the network, which adversely affects the model’s capabilities.
Comparing ‘4’, ‘5’, and ‘6’, the model performs similarly on all datasets at this point, and better than the model in ‘1’, ‘2’, and ‘3’. Comparing ‘3’ with ‘5’ and ‘6’, it can be found that the performance of the model gains a significant improvement, which greatly illustrates the need for edge feature augmentation and noise reduction processing of HSI using edge feature augment block and spectral attention block.
Overall, the edge feature augment block and the spectral attention block play a great role in suppressing the noise in HSIs, and combining them with the spatial attention block will result in better performance than other combinations.

5. Conclusions

In this paper, a novel feature extraction network (ESSN) is proposed for efficiently extracting local edge features and global spectral–spatial features from HSIs. In the ESSN, firstly, the edge feature augment block performs edge-aware and selective feature enhancement efficiently compared to the traditional edge data augmentation using the L o G operator with no learnable parameters. Secondly, due to the presence of a large amount of noise in some of the spectra in the HSI, different spectra do not have the same importance for the classification decision, so we introduce the spectral attention block to enhance the effective spectra and suppress the noise. Also, due to the geometric constraints of the convolutional operation, we introduce spatial attention to model the pixel–pixel interactions at all locations. Finally, we fuse representations of the feature maps reconstructed by the above methods through the 2D–3D convolution block to obtain the final feature representations. The experimental results show that ESSN performs competitively on the IP, KSC, and PU datasets.
Although ESSN has better performance in HSI classification, further improvements are needed. Afterwards, we will continue the following studies:
  • Exploring better edge-aware algorithms so as to reduce noise interference from isolated nodes;
  • Reduce the parameter size to speed up training and increase efficiency.

Author Contributions

Conceptualization, Y.H. and C.Y.; Methodology, X.Y.; Validation, J.Z. and S.Q.; Data curation, Y.S.; Writing—original draft, J.Z.; Writing—review & editing, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, C.; He, Z.; Lou, A.; Plaza, A. RGB-to-HSV: A frequency-spectrum unfolding network for spectral super-resolution of RGB videos. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5609318. [Google Scholar] [CrossRef]
  2. Fei, B. Hyperspectral imaging in medical applications. In Data Handling in Science and Technology; Elsevier: Amsterdam, The Netherlands, 2019; Volume 32, pp. 523–565. [Google Scholar]
  3. Liu, L.; Feng, J.; Han, L.; Zhou, J.; Xu, X.; Liu, R. Mineral mapping using spaceborne Tiangong-1 hyperspectral imagery and ASTER data: A case study of alteration detection in support of regional geological survey at Jintanzi-Malianquan area, Beishan, Gansu Province, China. Geol. J. 2018, 53, 372–383. [Google Scholar] [CrossRef]
  4. Stuart, M.B.; McGonigle, A.J.; Willmott, J.R. Hyperspectral imaging in environmental monitoring: A review of recent developments and technological advances in compact field deployable systems. Sensors 2019, 19, 3071. [Google Scholar] [CrossRef] [PubMed]
  5. Thomas, S.; Kuska, M.T.; Bohnenkamp, D.; Brugger, A.; Alisaac, E.; Wahabzada, M.; Behmann, J.; Mahlein, A.-K. Benefits of hyperspectral imaging for plant disease detection and plant protection: A technical perspective. J. Plant Dis. Prot. 2018, 125, 5–20. [Google Scholar] [CrossRef]
  6. Qin, J.; Kim, M.S.; Chao, K.; Chan, D.E.; Delwiche, S.R.; Cho, B.-K. Line-scan hyperspectral imaging techniques for food safety and quality applications. Appl. Sci. 2017, 7, 125. [Google Scholar] [CrossRef]
  7. Shimoni, M.; Haelterman, R.; Perneel, C. Hypersectral imaging for military and security applications: Combining myriad processing and sensing techniques. IEEE Geosci. Remote Sens. Mag. 2019, 7, 101–117. [Google Scholar] [CrossRef]
  8. Deng, F.; Pu, S.; Chen, X.; Shi, Y.; Yuan, T.; Pu, S. Hyperspectral image classification with capsule network using limited training samples. Sensors 2018, 18, 3153. [Google Scholar] [CrossRef] [PubMed]
  9. Ma, A.; Filippi, A.M.; Wang, Z.; Yin, Z. Hyperspectral image classification using similarity measurements-based deep recurrent neural networks. Remote Sens. 2019, 11, 194. [Google Scholar] [CrossRef]
  10. Bakhshi, G.; Shahtalebi, K.; Momeni, M. A new adaptive algorithm for target detection in hyperspectral images. Infrared Phys. Technol. 2019, 99, 222–230. [Google Scholar] [CrossRef]
  11. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Sparse representation for target detection in hyperspectral imagery. IEEE J. Sel. Top. Signal Process. 2011, 5, 629–640. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Du, B.; Zhang, L. Regularization framework for target detection in hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2013, 11, 313–317. [Google Scholar] [CrossRef]
  13. Shao, Y.; Lan, J.; Zhang, Y.; Zou, J. Spectral unmixing of hyperspectral remote sensing imagery via preserving the intrinsic structure invariant. Sensors 2018, 18, 3528. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, S.; Li, J.; Li, H.-C.; Deng, C.; Plaza, A. Spectral–spatial weighted sparse regression for hyperspectral image unmixing. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3265–3276. [Google Scholar] [CrossRef]
  15. Ma, Y.; Jin, Q.; Mei, X.; Dai, X.; Fan, F.; Li, H.; Huang, J. Hyperspectral unmixing with Gaussian mixture model and low-rank representation. Remote Sens. 2019, 11, 911. [Google Scholar] [CrossRef]
  16. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  17. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear versus nonlinear PCA for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef]
  18. SahIn, Y.E.; Arisoy, S.; Kayabol, K. Anomaly detection with Bayesian Gauss background model in hyperspectral images. In Proceedings of the 2018 26th Signal Processing and Communications Applications Conference (SIU), Izmir, Turkey, 2–5 May 2018; pp. 1–4. [Google Scholar]
  19. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  20. Ye, Q.; Huang, P.; Zhang, Z.; Zheng, Y.; Fu, L.; Yang, W. Multiview learning with robust double-sided twin SVM. IEEE Trans. Cybern. 2021, 52, 12745–12758. [Google Scholar] [CrossRef] [PubMed]
  21. Chen, C.; Chen, H. Unsupervised Segmentation of Hyperspectral Images Using Gauss-Markov Random Fields and Pca. In Frontiers Of Remote Sensing Information Processing; World Scientific: Singapore, 2003; pp. 329–345. [Google Scholar]
  22. Ye, Q.; Gao, W.; Huang, Q. Automatic text segmentation from complex background. In Proceedings of the 2004 International Conference on Image Processing, 2004. ICIP’04., Singapore, 24–27 October 2004; pp. 2905–2908. [Google Scholar]
  23. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  24. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  25. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  26. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Cai, W.; Yu, C.; Yang, N.; Cai, W. Multi-feature fusion: Graph neural network and CNN combining for hyperspectral image classification. Neurocomputing 2022, 501, 246–257. [Google Scholar] [CrossRef]
  27. Gao, H.; Yang, Y.; Li, C.; Zhou, H.; Qu, X. Joint alternate small convolution and feature reuse for hyperspectral image classification. ISPRS Int. J. Geo-Inf. 2018, 7, 349. [Google Scholar] [CrossRef]
  28. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  29. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  30. Wang, L.; Peng, J.; Sun, W. Spatial–spectral squeeze-and-excitation residual network for hyperspectral image classification. Remote Sens. 2019, 11, 884. [Google Scholar] [CrossRef]
  31. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A fast dense spectral–spatial convolution network framework for hyperspectral images classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  32. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  33. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  34. Chen, X.; Kamata, S.-I.; Zhou, W. Hyperspectral image classification based on multi-stage vision transformer with stacked samples. In Proceedings of the TENCON 2021—2021 IEEE Region 10 Conference (TENCON), Auckland, New Zealand, 7–10 December 2021; pp. 441–446. [Google Scholar]
  35. Ding, Y.; Zhao, X.; Zhang, Z.; Cai, W.; Yang, N. Graph sample and aggregate-attention network for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5504205. [Google Scholar] [CrossRef]
  36. Roy, S.K.; Deria, A.; Shah, C.; Haut, J.M.; Du, Q.; Plaza, A. Spectral–spatial morphological attention transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503615. [Google Scholar] [CrossRef]
  37. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  38. Tang, P.; Zhang, M.; Liu, Z.; Song, R. Double attention transformer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5502105. [Google Scholar] [CrossRef]
  39. Wang, Y.; Yu, X.; Wen, X.; Li, X.; Dong, H.; Zang, S. Learning a 3D-CNN and Convolution Transformers for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5504505. [Google Scholar]
  40. Wu, Y.; Weng, Q.; Lin, J.; Jian, C. RA-ViT: Patch-wise radially-accumulate module for ViT in hyperspectral image classification. J. Phys. Conf. Ser. 2022, 2278, 012009. [Google Scholar] [CrossRef]
  41. Yan, H.; Zhang, E.; Wang, J.; Leng, C.; Basu, A.; Peng, J. Hybrid Conv-ViT network for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5506105. [Google Scholar] [CrossRef]
  42. Yang, X.; Cao, W.; Lu, Y.; Zhou, Y. Hyperspectral image transformer classification networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5528715. [Google Scholar] [CrossRef]
  43. Zhang, J.; Meng, Z.; Zhao, F.; Liu, H.; Chang, Z. Convolution transformer mixer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6014205. [Google Scholar] [CrossRef]
  44. Zhao, F.; Zhang, J.; Meng, Z.; Liu, H.; Chang, Z.; Fan, J. Multiple vision architectures-based hybrid network for hyperspectral image classification. Expert Syst. Appl. 2023, 234, 121032. [Google Scholar] [CrossRef]
  45. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.-S. Spectral–spatial transformer network for hyperspectral image classification: A factorized architecture search framework. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5514715. [Google Scholar] [CrossRef]
  46. Liang, H.; Bao, W.; Shen, X.; Zhang, X. HSI-mixer: Hyperspectral image classification using the spectral–spatial mixer representation from convolutions. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6013005. [Google Scholar] [CrossRef]
  47. Shen, X.; Bao, W.; Liang, H.; Zhang, X.; Ma, X. Grouped collaborative representation for hyperspectral image classification using a two-phase strategy. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5505305. [Google Scholar] [CrossRef]
  48. Xu, H.; He, W.; Zhang, L.; Zhang, H. Unsupervised spectral–spatial semantic feature learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5526714. [Google Scholar] [CrossRef]
  49. Zhang, S.; Xu, M.; Zhou, J.; Jia, S. Unsupervised spatial-spectral cnn-based feature learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5524617. [Google Scholar] [CrossRef]
  50. Ziou, D.; Tabbone, S. Edge detection techniques—An overview. Pattern Recognit. Image Anal. Adv. Math. Theory Appl. 1998, 8, 537–559. [Google Scholar]
  51. Tu, B.; Li, N.; Fang, L.; He, D.; Ghamisi, P. Hyperspectral image classification with multi-scale feature extraction. Remote Sens. 2019, 11, 534. [Google Scholar] [CrossRef]
  52. Al-Khafaji, S.L.; Zhou, J.; Bai, X.; Qian, Y.; Liew, A.W.-C. Spectral-spatial boundary detection in hyperspectral images. IEEE Trans. Image Process. 2021, 31, 499–512. [Google Scholar] [CrossRef] [PubMed]
  53. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  54. Rizzolatti, G.; Craighero, L. Spatial attention: Mechanisms and theories. Adv. Psychol. Sci. 1998, 2, 171–198. [Google Scholar]
Figure 1. Framework of the classification process using ESSN. Note that BN and ReLU after each convolution operation have been omitted.
Figure 1. Framework of the classification process using ESSN. Note that BN and ReLU after each convolution operation have been omitted.
Sensors 24 04714 g001
Figure 2. Comparison of prediction and truth plots. (a) Partial ground truth of IP, (bd) Predicted classification maps.
Figure 2. Comparison of prediction and truth plots. (a) Partial ground truth of IP, (bd) Predicted classification maps.
Sensors 24 04714 g002
Figure 3. L o G operators for the two cases σ < 0.5 and σ = 1.4 . (a,b) corresponding to σ < 0.5 , (c) corresponding to σ = 1 .
Figure 3. L o G operators for the two cases σ < 0.5 and σ = 1.4 . (a,b) corresponding to σ < 0.5 , (c) corresponding to σ = 1 .
Sensors 24 04714 g003
Figure 4. Binarized images of different spectral bands of PU dataset. (ac) derived from raw PU dataset, (df) derived from PU dataset after PCA.
Figure 4. Binarized images of different spectral bands of PU dataset. (ac) derived from raw PU dataset, (df) derived from PU dataset after PCA.
Sensors 24 04714 g004
Figure 5. The framework of spatial attention in Figure 1.
Figure 5. The framework of spatial attention in Figure 1.
Sensors 24 04714 g005
Figure 6. Specific information on the IP dataset.
Figure 6. Specific information on the IP dataset.
Sensors 24 04714 g006
Figure 7. Specific information on the KSC dataset.
Figure 7. Specific information on the KSC dataset.
Sensors 24 04714 g007
Figure 8. Specific information on the PU dataset.
Figure 8. Specific information on the PU dataset.
Sensors 24 04714 g008
Figure 9. Performance on IP. (a) Full ground truth; (b) 2-D CNN (97.11%); (c) 3-D CNN (97.19%); (d) HybridSN (98.72%); (e) SSRN (98.61%); (f) SSTN (97.66%); (g) SSFTT (98.57%); (h) CTMixer (98.42%); (i) 3DCT (98.54%); (j) ESSN (98.93%).
Figure 9. Performance on IP. (a) Full ground truth; (b) 2-D CNN (97.11%); (c) 3-D CNN (97.19%); (d) HybridSN (98.72%); (e) SSRN (98.61%); (f) SSTN (97.66%); (g) SSFTT (98.57%); (h) CTMixer (98.42%); (i) 3DCT (98.54%); (j) ESSN (98.93%).
Sensors 24 04714 g009
Figure 10. Performance on KSC. (a) Full ground truth; (b) 2-D CNN (91.35%); (c) 3-D CNN (93.04%); (d) HybridSN (98.99%); (e) SSRN (98.60%); (f) SSTN (63.16%); (g) SSFTT (92.35%); (h) CTMixer (95.14%); (i) 3DCT (97.87%); (j) ESSN (99.18%).
Figure 10. Performance on KSC. (a) Full ground truth; (b) 2-D CNN (91.35%); (c) 3-D CNN (93.04%); (d) HybridSN (98.99%); (e) SSRN (98.60%); (f) SSTN (63.16%); (g) SSFTT (92.35%); (h) CTMixer (95.14%); (i) 3DCT (97.87%); (j) ESSN (99.18%).
Sensors 24 04714 g010
Figure 11. Performance on PU. (a) Full ground truth; (b) 2-D CNN (97.15%); (c) 3-D CNN (96.97%); (d) HybridSN (99.36%); (e) SSRN (99.12%); (f) SSTN (96.01%); (g) SSFTT (99.13%); (h) CTMixer (99.02%); (i) 3DCT (99.36%); (j) ESSN (99.45%).
Figure 11. Performance on PU. (a) Full ground truth; (b) 2-D CNN (97.15%); (c) 3-D CNN (96.97%); (d) HybridSN (99.36%); (e) SSRN (99.12%); (f) SSTN (96.01%); (g) SSFTT (99.13%); (h) CTMixer (99.02%); (i) 3DCT (99.36%); (j) ESSN (99.45%).
Sensors 24 04714 g011
Figure 12. OA (%) at different training ratios on IP (left), KSC (middle), and PU (right), respectively.
Figure 12. OA (%) at different training ratios on IP (left), KSC (middle), and PU (right), respectively.
Sensors 24 04714 g012
Figure 13. Classification accuracy (%) on IP (left), KSC (middle), and PU (right) in different cases. The ‘a’ denotes traditional edge data augmentation using the L o G operator with no learnable parameters. In turn, ‘b’, ‘c’, and ‘d’ correspond to the use of the (a), (b), and (c) operators in Figure 3, respectively. ‘e’ indicates no processing on edges, corresponding to case ‘6’ in Table 6.
Figure 13. Classification accuracy (%) on IP (left), KSC (middle), and PU (right) in different cases. The ‘a’ denotes traditional edge data augmentation using the L o G operator with no learnable parameters. In turn, ‘b’, ‘c’, and ‘d’ correspond to the use of the (a), (b), and (c) operators in Figure 3, respectively. ‘e’ indicates no processing on edges, corresponding to case ‘6’ in Table 6.
Sensors 24 04714 g013
Figure 14. Model classification accuracy (subplot (a)) and train time cost (subplot (b)) after downscaling to different dimensions using PCA operations. ‘raw’ indicates that PCA is not used.
Figure 14. Model classification accuracy (subplot (a)) and train time cost (subplot (b)) after downscaling to different dimensions using PCA operations. ‘raw’ indicates that PCA is not used.
Sensors 24 04714 g014
Table 1. Classification accuracy (%) of different models on IP.
Table 1. Classification accuracy (%) of different models on IP.
Class2-D CNN3-D CNNHybridSNSSRNSSTNSSFTTCTMixer3DCTESSN
168.33 ± 22.1488.60 ± 15.8289.55 ± 13.29 82.06 ± 15.17 83.19 ± 20.67 90.92 ± 13.1592.86 ± 12.4284.95 ± 18.20 89.30 ± 12.81
296.29 ± 1.7996.24 ± 1.3098.32 ± 0.62 98.25 ± 0.52 97.75 ± 0.74 98.53 ± 0.6998.03 ± 0.80 97.83 ± 0.76 98.39 ± 0.58
397.40 ± 0.9495.81 ± 2.5299.19 ± 0.87 99.02 ± 0.72 98.28 ± 1.35 98.80 ± 0.8898.41 ± 1.00 98.56 ± 0.83 99.34 ± 0.56
492.37 ± 7.6898.26 ± 2.5598.77 ± 2.20 99.19 ± 1.53 97.86 ± 3.25 98.63 ± 2.2299.23 ± 1.6597.92 ± 2.56 98.86 ± 2.04
596.16 ± 1.7697.86 ± 1.3798.34 ± 1.21 98.16 ± 1.37 96.60 ± 1.85 98.27 ± 1.1598.16 ± 1.22 98.29 ± 1.53 98.80 ± 1.14
699.27 ± 0.7199.16 ± 0.6899.39 ± 0.54 99.39 ± 0.43 98.52 ± 0.93 99.50 ± 0.4599.25 ± 0.49 99.66 ± 0.24 99.48 ± 0.46
751.12 ± 28.9592.69 ± 14.0292.66 ± 12.10 92.66 ± 13.15 32.47 ± 37.73 92.20 ± 11.9793.42 ± 7.14 92.68 ± 12.09 97.29 ± 4.57
8100.00 ± 0.0099.81 ± 0.39100.00 ± 0.0099.98 ± 0.0799.95 ± 0.14 99.93 ± 0.2199.98 ± 0.07 99.98 ± 0.07 99.98 ± 0.07
944.75 ± 35.7063.97 ± 31.0972.91 ± 37.31 71.08 ± 35.9536.14 ± 40.49 65.35 ± 39.8166.79 ± 37.46 72.89 ± 34.52 73.31 ± 36.21
1096.60 ± 2.0696.38 ± 1.3398.96 ± 0.55 98.56 ± 0.6597.52 ± 0.94 99.15 ± 0.7598.09 ± 0.92 98.50 ± 0.63 99.10 ± 0.55
1198.5 ± 0.6998.16 ± 0.7198.99 ± 0.50 99.00 ± 0.6098.33 ± 0.59 98.84 ± 0.4798.58 ± 0.70 98.75 ± 0.58 99.14 ± 0.47
1293.41 ± 3.0092.84 ± 3.3696.11 ± 1.98 97.31 ± 1.47 94.38 ± 3.35 95.10 ± 2.6095.98 ± 2.75 97.24 ± 1.73 97.43 ± 1.62
1399.31 ± 0.6697.89 ± 2.2999.18 ± 1.96 99.38 ± 1.51 98.50 ± 3.17 99.64 ± 0.6598.34 ± 3.17 99.22 ± 1.50 99.01 ± 1.60
1499.72 ± 0.1798.97 ± 0.3599.86 ± 0.1499.67 ± 0.30 99.71 ± 0.29 99.51 ± 0.5099.78 ± 0.18 99.78 ± 0.29 99.78 ± 0.24
1596.02 ± 3.5096.13 ± 3.4299.46 ± 0.99 97.49 ± 2.30 99.00 ± 1.10 98.51 ± 1.9299.54 ± 0.7898.56 ± 1.79 99.17 ± 0.92
1689.27 ± 9.0493.84 ± 3.2591.22 ± 9.55 91.85 ± 4.78 87.85 ± 10.57 92.19 ± 4.6593.17 ± 5.34 95.50 ± 2.5495.49 ± 2.45
OA97.11 ± 0.3497.19 ± 0.6098.72 ± 0.26 98.61 ± 0.24 97.66 ± 0.44 98.57 ± 0.3298.42 ± 0.26 98.54 ± 0.30 98.93 ± 0.17
AA88.91 ± 4.3094.16 ± 3.7195.81 ± 3.40 95.19 ± 3.33 88.50 ± 4.46 95.32 ± 3.3195.60 ± 2.99 95.64 ± 3.52 96.49 ± 3.14
Kappa96.71 ± 0.3996.80 ± 0.6898.54 ± 0.30 98.41 ± 0.28 97.33 ± 0.51 98.37 ± 0.3698.20 ± 0.30 98.33 ± 0.36 98.78 ± 0.19
Table 2. Classification accuracy (%) of different models on KSC.
Table 2. Classification accuracy (%) of different models on KSC.
Class2-D CNN3-D CNNHybridSNSSRNSSTNSSFTTCTMixer3DCTESSN
197.76 ± 1.4598.87 ± 0.9299.88 ± 0.21 100.00 ± 0.0081.15 ± 28.39 97.73 ± 1.7689.94 ± 29.92 99.82 ± 0.29 99.90 ± 0.21
268.97 ± 8.8181.87 ± 7.9097.22 ± 2.75 96.16 ± 2.32 6.46 ± 10.6886.63 ± 7.1290.07 ± 15.79 94.28 ± 4.99 97.99 ± 2.17
390.74 ± 8.5686.23 ± 6.4295.43 ± 2.7093.43 ± 2.00 0.15 ± 0.3382.29 ± 8.7590.27 ± 20.86 93.36 ± 3.78 95.30 ± 3.06
469.68 ± 5.9874.79 ± 8.1294.14 ± 3.50 90.06 ± 3.50 17.25 ± 16.6062.24 ± 13.2587.87 ± 17.11 91.24 ± 3.83 94.89 ± 3.43
582.52 ± 5.1490.06 ± 2.6597.86 ± 2.05 97.25 ± 2.42 30.86 ± 21.5190.96 ± 4.6193.22 ± 8.70 97.28 ± 2.45 98.76 ± 0.88
678.21 ± 13.1785.85 ± 4.3598.79 ± 1.05 96.46 ± 3.01 10.88 ± 8.6885.79 ± 10.3982.66 ± 24.62 95.40 ± 5.79 99.13 ± 1.04
786.47 ± 10.4494.57 ± 5.5599.28 ± 1.10 100.00 ± 0.000.00 ± 0.0095.18 ± 11.8588.92 ± 29.72 96.29 ± 6.00 100.00 ± 0.00
894.26 ± 3.6792.27 ± 3.2499.65 ± 0.45 99.53 ± 0.60 73.43 ± 18.1291.23 ± 9.5195.73 ± 5.44 98.53 ± 1.82 99.62 ± 0.41
996.36 ± 4.6388.58 ± 3.5199.43 ± 1.06 99.20 ± 1.57 54.91 ± 29.1188.89 ± 5.7898.56 ± 2.53 97.67 ± 2.36 99.55 ± 0.89
1088.23 ± 3.6493.14 ± 3.5598.72 ± 1.87 99.88 ± 0.28 73.91 ± 19.82 95.33 ± 3.0598.50 ± 2.01 97.25 ± 1.56 99.31 ± 1.22
1198.13 ± 1.4297.44 ± 1.71100.00 ± 0.00100.00 ± 0.0089.62 ± 9.5498.44 ± 1.2799.85 ± 0.36 100.00 ± 0.00100.00 ± 0.00
1287.56 ± 2.8495.48 ± 1.5499.73 ± 0.44 99.71 ± 0.24 72.68 ± 10.91 93.71 ± 5.2599.73 ± 0.36 98.67 ± 1.18 99.83 ± 0.31
1399.93 ± 0.1299.88 ± 0.16100.00 ± 0.00100.00 ± 0.0097.82 ± 5.43 100.00 ± 0.0099.99 ± 0.04 100.00 ± 0.00100.00 ± 0.00
OA91.35 ± 1.2293.04 ± 0.9898.99 ± 0.30 98.60 ± 0.34 63.16 ± 6.48 92.35 ± 92.3595.14 ± 9.21 97.87 ± 0.56 99.18 ± 0.32
AA87.60 ± 1.8390.69 ± 1.4698.47 ± 0.49 97.82 ± 0.48 46.85 ± 6.65 89.88 ± 1.8993.49 ± 11.31 96.91 ± 0.99 98.79 ± 0.49
Kappa90.36 ± 1.3692.25 ± 1.1098.87 ± 0.34 98.44 ± 0.38 58.36 ± 7.49 91.48 ± 1.2294.61 ± 10.20 97.61 ± 0.62 99.08 ± 0.36
Table 3. Classification accuracy (%) of different models on PU.
Table 3. Classification accuracy (%) of different models on PU.
Class2-D CNN 3-D CNNHybridSNSSRNSSTNSSFTTCTMixer3DCTESSN
196.44 ± 1.0896.12 ± 0.8899.85 ± 0.2299.62 ± 0.26 95.23 ± 4.78 99.70 ± 0.18 99.59 ± 0.35 99.63 ± 0.27 99.76 ± 0.35
299.99 ± 0.0199.79 ± 0.1299.99 ± 0.0199.99 ± 0.0199.82 ± 0.20 99.99 ± 0.0199.98 ± 0.04 99.95 ± 0.04 99.99 ± 0.03
388.773 ± 3.3088.99 ± 3.7395.79 ± 2.50 92.99 ± 2.91 79.55 ± 11.40 96.49 ± 2.24 97.87 ± 1.6394.83 ± 2.46 97.38 ± 1.69
493.36 ± 1.5292.46 ± 2.2096.95 ± 1.09 97.32 ± 0.44 91.64 ± 3.41 95.34 ± 1.40 94.02 ± 1.99 98.32 ± 0.6797.39 ± 0.86
599.95 ± 0.0999.97 ± 0.7099.99 ± 0.05 99.82 ± 0.25 96.35 ± 3.8699.86 ± 0.17 99.91 ± 0.11 100.00 ± 0.0099.98 ± 0.05
698.37 ± 0.9399.24 ± 0.7899.99 ± 0.04 99.97 ± 0.10 96.85 ± 8.47 100.00 ± 0.0099.84 ± 0.47 99.86 ± 0.22 100.00 ± 0.00
798.17 ± 2.0299.50 ± 0.6699.99 ± 0.0299.53 ± 0.62 88.32 ± 15.45 99.73 ± 0.55 99.92 ± 0.21 99.86 ± 0.21 99.97 ± 0.09
889.91 ± 2.9890.51 ± 1.7899.11 ± 0.7397.82 ± 0.76 95.70 ± 3.84 97.64 ± 1.15 98.15 ± 1.15 98.62 ± 0.73 98.78 ± 0.67
993.44 ± 3.8185.33 ± 3.1295.07 ± 2.57 96.62 ± 2.01 83.98 ± 9.21 95.63 ± 1.63 91.68 ± 2.23 97.93 ± 1.3596.16 ± 1.51
OA97.15 ± 0.3196.97 ± 0.4099.36 ± 0.12 99.12 ± 0.15 96.01 ± 1.93 99.13 ± 0.15 99.02 ± 0.24 99.36 ± 0.18 99.45 ± 0.10
AA95.38 ± 0.5494.66 ± 0.7398.52 ± 0.33 98.19 ± 0.27 91.94 ± 3.53 98.27 ± 0.35 97.88 ± 0.43 98.78 ± 0.36 98.82 ± 0.24
Kappa96.22 ± 0.4195.99 ± 0.5399.15 ± 0.16 98.83 ± 0.20 94.68 ± 2.60 98.85 ± 0.19 98.71 ± 0.32 99.15 ± 0.24 99.27 ± 0.13
Table 4. Resource consumption for each model.
Table 4. Resource consumption for each model.
MethodParams (M)IPKSCPU
Train Time (s)Test Time (s)Train Time (s)Test Time (s)Train Time (s)Test Time (s)
2-D CNN0.04200.852.54197.802.12199.255.64
3-D CNN0.62253.214.03255.092.64268.2410.52
HybridSN4.34319.424.93319.783.16337.1615.85
SSRN0.47291.304.56293.643.07308.8514.61
SSTN0.02228.263.08227.892.36236.128-17
SSFTT0.23207.702.54205.242.19208.705.86
CTMixer0.60262.063.96267.032.70280.8112.20
3DCT3.84346.956.19349.573.77384.9022.05
ESSN4.37332.875.24330.773.36354.4117.32
Table 5. OA (%) on different patch sizes.
Table 5. OA (%) on different patch sizes.
Patch Size11 × 1113 × 1315 × 1517 × 1719 × 19
IP98.62 ± 0.3198.72 ± 0.2498.93 ± 0.1798.76 ± 0.2898.73 ± 0.25
KSC98.60 ± 0.3998.93 ± 0.4499.18 ± 0.3299.37 ± 0.2099.58 ± 0.20
PU99.18 ± 0.1099.37 ± 0.1699.45 ± 0.1099.42 ± 0.1699.37 ± 0.16
Table 6. Classification accuracy (%) of different combinations of block.
Table 6. Classification accuracy (%) of different combinations of block.
CaseEdge BlockSpectral BlockSpatial BlockMetricIPKSCPU
1××OA98.74 ± 0.2499.00 ± 0.2999.39 ± 0.11
AA95.72 ± 3.0198.52 ± 0.4898.79 ± 0.32
Kappa98.68 ± 0.2798.88 ± 0.3299.21 ± 0.14
2××OA98.75 ± 0.2599.06 ± 0.3299.38 ± 0.08
AA96.02 ± 3.2898.77 ± 0.4598.59 ± 0.24
Kappa98.58 ± 0.2999.06 ± 0.3699.11 ± 0.10
3××OA76.25 ± 8.9755.88 ± 13.0289.70 ± 4.60
AA70.23 ± 11.2246.93 ± 12.7283.67 ± 8.98
Kappa73.08 ± 9.9850.30 ± 14.5486.33 ± 6.09
4×OA98.82 ± 0.3199.12 ± 0.1899.40 ± 0.14
AA95.86 ± 3.4198.72 ± 0.2898.65 ± 0.44
Kappa98.66 ± 0.3599.03 ± 0.2099.21 ± 0.18
5×OA98.77 ± 0.2099.16 ± 0.2399.40 ± 0.15
AA95.95 ± 3.5598.75 ± 0.3598.61 ± 0.43
Kappa98.59 ± 0.2399.04 ± 0.2699.17 ± 0.20
6×OA98.77 ± 0.2799.07 ± 0.2499.39 ± 0.14
AA96.43 ± 3.2598.75 ± 0.3298.57 ± 0.29
Kappa98.57 ± 0.3099.07 ± 0.2799.12 ± 0.19
7OA98.93 ± 0.1799.18 ± 0.3299.45 ± 0.10
AA96.49 ± 3.1498.79 ± 0.4998.82 ± 0.24
Kappa98.78 ± 0.1999.08 ± 0.3699.27 ± 0.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, J.; Sun, Y.; Hao, Y.; Qin, S.; Yang, C.; Li, J.; Yu, X. A Joint Network of Edge-Aware and Spectral–Spatial Feature Learning for Hyperspectral Image Classification. Sensors 2024, 24, 4714. https://doi.org/10.3390/s24144714

AMA Style

Zheng J, Sun Y, Hao Y, Qin S, Yang C, Li J, Yu X. A Joint Network of Edge-Aware and Spectral–Spatial Feature Learning for Hyperspectral Image Classification. Sensors. 2024; 24(14):4714. https://doi.org/10.3390/s24144714

Chicago/Turabian Style

Zheng, Jianfeng, Yu Sun, Yuqi Hao, Senlong Qin, Cuiping Yang, Jing Li, and Xiaodong Yu. 2024. "A Joint Network of Edge-Aware and Spectral–Spatial Feature Learning for Hyperspectral Image Classification" Sensors 24, no. 14: 4714. https://doi.org/10.3390/s24144714

APA Style

Zheng, J., Sun, Y., Hao, Y., Qin, S., Yang, C., Li, J., & Yu, X. (2024). A Joint Network of Edge-Aware and Spectral–Spatial Feature Learning for Hyperspectral Image Classification. Sensors, 24(14), 4714. https://doi.org/10.3390/s24144714

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop