Next Article in Journal
Practical Verification of Diagnostic Capabilities of Methods for the Recreation of Voltage Fluctuations
Previous Article in Journal
Design and Simulation of Low-Orbit Satellite Broadcast Signal Receiving and Processing Terminal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration

1
College of Automation, Jiangsu University of Science and Technology, No. 666 Changhui Road, Zhenjiang 212100, China
2
Systems Science Laboratory, Jiangsu University of Science and Technology, No. 666 Changhui Road, Zhenjiang 212100, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(16), 3271; https://doi.org/10.3390/electronics13163271
Submission received: 11 July 2024 / Revised: 11 August 2024 / Accepted: 16 August 2024 / Published: 18 August 2024
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)

Abstract

:
In recent years, graph convolutional neural networks (GCNs) and convolutional neural networks (CNNs) have made significant strides in hyperspectral image (HSI) classification. However, existing models often encounter information redundancy and feature mismatch during feature fusion, and they struggle with small-scale refined features. To address these issues, we propose DCG-Net, an innovative classification network integrating CNN and GCN architectures. Our approach includes the development of a double-branch expanding network (E-Net) to enhance spectral features and efficiently extract high-level features. Additionally, we incorporate a GCN with an attention mechanism to facilitate the integration of multi-space scale superpixel-level and pixel-level features. To further improve feature fusion, we introduce a feature aggregation module (FAM) that adaptively learns channel features, enhancing classification robustness and accuracy. Comprehensive experiments on three widely used datasets show that DCG-Net achieves superior classification results compared to other state-of-the-art methods.

1. Introduction

Hyperspectral images (HSIs) collected from satellites or aircrafts comprise hundreds of contiguous bands and contain abundant spectral–spatial information [1]. HSIs provide detailed spectral and spatial data, essential for precise classification. Consequently, HSIs are crucial in fields such as target detection, ocean exploration, crop monitoring, and environmental investigation [2,3,4,5,6,7,8]. HSI classification, a key tool in HSI processing, aims to assign category labels to each pixel based on spectral and spatial information, facilitating detailed analysis of target areas for precise management and research.
Initially, HSI classification relied on machine learning methods such as random forests [9], support vector machines (SVMs) [10], principal component analysis (PCA) [11], linear discriminant analysis (LDA) [12], and sparse representation classification (SRC) [13]. These traditional supervised methods focused solely on spectral data, ignoring spatial information, which limited their classification effectiveness. Additionally, these methods lacked robustness and accuracy.
Compared to traditional machine learning, deep learning enables end-to-end, automatic, advanced feature extraction. Consequently, numerous deep learning-based models have been proposed [14,15,16]. Convolutional neural networks (CNNs) have significant advantages in HSI processing and have proven to be particularly effective [17,18]. Specifically, 1D-CNNs [19], 2D-CNNs [20,21,22], and 3D-CNNs [23,24] extract spectral, spatial, and spectral–spatial features, respectively. Zhong et al. [25] introduced residual blocks in their SSRN model, while Wang et al. [26] (FDSSC) incorporated dense connections, enhancing network training and representation. Additionally, Yang et al. [27] developed a two-channel CNN (TCCNN) for separate spectral and spatial feature extraction. The attention mechanism [28,29,30] can extract more discriminative features and reduce the interference of irrelevant features in classification, making it widely used in HSI classification. For instance, Li et al. [31] (DBDA) added channel attention and spatial attention blocks to their network.
Recently, graph neural networks (GNNs) [32] have become crucial in HSI classification. GNNs handle graph structures, overcoming CNN limitations with non-Euclidean data and better extracting internal HSI relationships. For instance, Mou et al. [33] proposed a nonlocal graph convolutional network to classify HSIs using both labeled and unlabeled data. However, this method is computationally intensive because it directly uses each pixel of the entire image as a graph node. To reduce computational demand, Wan et al. [34] proposed multi-scale dynamic graph convolution (MDGCN), using superpixels instead of individual pixels for graph construction. Liu et al. [35] introduced a classification method based on multi-level superpixel structured graphs (MSSGU) to learn spectral features at different scales. Ding et al. [36] proposed a multi-feature fusion network (MFGCN) incorporating multi-scale GCNs and multi-scale CNNs to combine hyperpixel-level and pixel-level features. Additionally, Liu et al. [37] developed a CNN-enhanced GCN (CEGCN), and Dong et al. [38] proposed a convolutional neural and graph attention network weighted feature fusion model (WFCG).
Although these models have achieved good classification results, challenges remain. First, feature extraction for small-scale refined samples remains a prominent issue. Second, CNN and GCN feature extraction and fusion occur only at a single spatial scale, ignoring features at different scales. Finally, multiple features cannot be accurately fused, often resulting in information redundancy. To address these issues, we propose an innovative classification network (DCG-Net) that integrates CNN and GCN architectures, merging superpixel-level graph convolutional networks with pixel-level convolutional networks [39]. Our contributions include designing an expanding network (E-Net) for feature extraction, integrating a GCN with an attention mechanism, and developing a feature aggregation module (FAM) for adaptive channel feature learning. We demonstrate our model’s effectiveness through comprehensive experiments on three widely used datasets, achieving superior classification results compared to state-of-the-art methods.
The main contributions of this study are as follows:
  • Innovative classification network (DCG-Net): We introduced DCG-Net, a novel classification network that integrates convolutional neural networks (CNNs) and graph convolutional networks (GCNs). This hybrid architecture is designed to effectively capture both large-scale regular features and small-scale fine features in HSIs, addressing issues of information redundancy and feature mismatch commonly encountered in existing models.
  • Expanding network (E-Net) for enhanced feature extraction: We developed a double-branch expanding network (E-Net) based on a CNN architecture. E-Net enhances the spectral features of HSIs and efficiently extracts high-level features by projecting the image information into higher spatial dimensions. This approach balances the extraction of both high-level and fine-grained features.
  • Feature aggregation module (FAM): We designed a feature aggregation module (FAM) that adaptively learns channel features. The FAM dynamically calibrates channel responses within the network, enhancing the model’s feature representation and extraction capabilities.

2. Related Work

2.1. HSI Classification Based on Convolutional Neural Network

Convolutional Neural Networks (CNNs) can automatically extract both spatial and spectral features from HSIs. They support an end-to-end learning framework, enabling direct learning and classification of features from the original HSIs.
Lee et al. [40] (CDCNN) achieved feature mapping of HSIs by integrating local spatial–spectral information of neighboring pixels through multi-scale dynamic convolution. To alleviate the training difficulties of deep networks, Zhong et al. [25] incorporated residual blocks into the spectral and spatial network using the 3D cube approach as training samples to reduce network parameters. Wang et al. [26] employed end-to-end fast and dense spectral–spatial convolution to facilitate the rapid classification of HSIs. Additionally, Yang et al. [27] designed a dual-channel CNN for separate spectral and spatial feature extraction. To overcome the limitations of fixed kernel shapes in conventional convolution, Liu et al. [41] used dynamic convolution for classification, allowing the kernel shape to adapt according to the spatial distribution of the HSI. This deformable convolution helps suppress irregularities and accidental features at category boundaries, enhancing the ability to learn features in cross-category regions. Furthermore, the attention mechanism has been applied to CNN networks. For instance, Li et al. [31] designed two branches to capture a large number of spectral and spatial features in the HSI, applying channel attention blocks and spatial attention blocks to improve feature extraction accuracy. Given that CNN networks often require a large number of computational parameters and are computationally expensive during the training phase, Wang et al. [42] (LMAFN) designed a lightweight spectral–spatial attention feature fusion network based on the network search framework (NAS), which reduces computational parameters by adjusting the weights of different channels through adaptive passbands.
Although various structural CNN networks have achieved good results in addressing HSI classification problems, significant challenges remain in handling the internal connections within HSIs. This difficulty arises from the complex nature of hyperspectral data, which often involves non-Euclidean relationships.

2.2. HSI Classification Based on Hyperpixel-Based Graph Convolutional Network

Using a superpixel segmentation [43] strategy to create the graphical structure accurately captures the core information of the image and significantly reduces the parameter requirements of the graph convolutional network when processing HSIs. This approach reduces model complexity while ensuring a sensitive and accurate interpretation of image details.
Recently, various superpixel-based graph convolutional models for HSI classification have been proposed. For example, Wan et al. [34] (MDGCN) used a multi-scale simple linear iterative clustering (SLIC) approach to obtain superpixels, capturing multi-scale feature information of the HSI as input nodes for the GCN. To address the problem of superpixel segmentation accuracy, GAT-AGSM [44] designed the SFS mechanism, which is able to assign different channel weights to the superpixel features. This can make the superpixel segmentation generate finer classification results. To fuse pixel-level and superpixel features, Liu et al. [37] first used SLIC to segment and encode the HSI into a graph structure, then they constructed an association matrix between pixels and superpixels to capture superpixel-level features with the GCN and combined them with the pixel-level features extracted by the CNN. To learn spectral features at different scales, Liu et al. [35] (MSSGU) constructed a multilevel graph structure by obtaining superpixels using a region merging-based superpixel segmentation method, then they built graph structures at different scales. This method effectively fuses multi-scale feature information, thereby improving the classification accuracy of HSIs. In superpixel segmentation, pixels with similar attributes in an image are organized into separate units, enhancing model processing speed, the accuracy of feature extraction, and overall model expressiveness. This demonstrates the significant value of superpixels in image processing and analysis. Superpixel segmentation organizes pixels with similar attributes into distinct units. This approach improves processing speed, enhances feature extraction accuracy, and increases the model’s expressiveness. This demonstrates the significant value of superpixels in image processing and analysis.
Although these methods have achieved good classification results, they still face challenges in processing fine edge features. To address these issues, we constructed superpixel graph structures at various spatial scales using superpixel-based graph convolution. Simultaneously, we fused GCNs and CNNs at multiple spatial scales to effectively enhance texture feature extraction.

3. Proposed Method

In this paper, we propose DCG-Net, a novel HSI classification method. As shown in Figure 1, HSIs exhibit complex edge features and unbalanced sample distributions. To enhance the extraction of detailed and small-scale sample features, we propose a expanding network (E-Net) based on a double-branch encoder–decoder architecture. The E-Net effectively extracts regular spatial–spectral features and amplifies spatial features to refine details and edge features through high-dimensional mapping. To enhance the extraction of large-scale sample features, we incorporate a superpixel-based GCN. This integration enables the fusion of pixel-level local spatial features with superpixel-level large-scale spatial features. The graph convolution module is embedded in the encoder and decoder of E-Net, facilitating a more effective fusion of local and global features. Additionally, to enhance dual-channel feature aggregation, we designed a feature aggregation module based on the attention mechanism. This module further optimizes feature extraction, enhancing the accuracy and robustness of the classification process.
We first use the superpixel segmentation module to construct the spatial topology of HSIs and preserve their large-scale structural information. Then, we incorporate the attention mechanism into GCN to enhance the correlation between graph nodes. Next, to fully extract small-scale sample features, we integrate E-Net and GCN. This enables the extraction of both pixel-level and superpixel spectral features, along with multi-spatial-scale feature extraction. Then, we use the dynamically integrated feature aggregation module at different scales. Finally, we predict the obtained features using a Softmax classifier.

3.1. Expanding Network

HSIs exhibit complex edge features and unbalanced sample distributions. To enhance the extraction of advanced features and small-scale samples, we designed a double-branch encoding and decoding expanding network (E-Net), as shown in Figure 2. E-Net is capable of projecting HSI to spatially higher dimensions and effectively combining GCN, enhancing the extraction of pixel-level and superpixel features for multi-spatial-scale feature extraction. We will now introduce E-Net in detail.
In E-Net, the first branch network consists of an encoder–decoder structure operating at the original image size. The encoder of the second branch network upscales the feature output from the first branch encoding to a higher spatial scale and performs feature extraction. The decoder of the second branch network downsamples these features and performs convolution operations, enabling efficient extraction of small-scale sample features at this level.
Specifically, the encoder and decoder use a customized convolution module built upon the CNN architecture. In this module, input features undergo batch normalization first, followed by downsampling with a ( 1 × 1 ) pointwise convolution. Subsequently, nonlinear transformation is performed using the Leaky ReLU activation function [45]. Then, the features are extracted by a ( 3 × 3 ) depth-separable convolution, followed by another nonlinear transformation using the Leaky ReLUs [45] activation function. After the first branch of the encoder, the input features are up-sampled, and the resulting output features are then up-sampled as follows:
Y u = T C o n v Y ,
where Y represents the feature output from the first branch encoder, T C o n v denotes the transpose convolution that doubles the spatial size of the input features, and Y u is the resulting feature with doubled spatial dimensions. The feature output Y from the first branch encoder is also fed into the first branch GCN network. The specific operation of the GCN network will be described below. After extracting the superpixel-level features using the GCN network, we obtain the new feature Y g . At this point, Y g is input into the customized convolution module for decoding, and the superpixel-level features are fused with the pixel-level features to produce the feature Y c .
Simultaneously, the upsampled feature Y u passes through the second branch encoder for additional pixel-level feature extraction. The extracted feature is also input into the GCN network to obtain the upsampled superpixel-level feature output Y u g , which is subsequently decoded to yield Y u c . This decoded feature is then downsampled, as follows:
Y p = M a x P o o l i n g Y u c ,
where M a x P o o l i n g refers to the MaxPool2d function and Y p represents the feature output after downsampling. At this point, the features Y p and Y c have the same spatial size. They are then concatenated to generate the output features of E-Net.

3.2. Superpixel Structured Graph

Graph convolutional networks accept only graph-structured data as input, necessitating the construction of a graph structure. We use a superpixel segmentation method to construct the graph structure. This method groups pixels with similar features into units, simplifying the graph representation while preserving crucial image information.
As shown in Figure 3, to reduce information redundancy and noise pollution of the HSI, we first apply a principal component analysis (PCA) to reduce the dimensionality of the original HSI. Subsequently, we apply superpixel segmentation to obtain a specified number of superpixels [43]. Each superpixel is treated as a node, and we construct the graph structure using the K-nearest neighbor (KNN) method [46]. Specifically, assuming the number of superpixels is Z, we define G = V , E as an unweighted graph, where V and E represent the nodes and edges of the graph, respectively. Here, the edges E are represented by the adjacency matrix A R Z × Z , where A i , j denotes the edge of the node in the j-th column of the i-th row. The adjacency matrix A is constructed using the K-nearest neighbor approach and is computed as follows:
A i , j = 1 , if p i is one of the K - Nearest - Neighbors of p j 0 , otherwise ,
where p i and p j denote the i-th and j-th superpixel in the superpixel, respectively.
Meanwhile, to establish the relationship between pixel features and superpixel features, influenced by MSSGU [35], we construct the association matrix M R ( H W ) × Z . As shown in Figure 3, we define G k = V k , E k as the original HSI 4-connected graph. We take the first three principal components extracted by the PCA as the node features in G k and define x i as the i-th pixel of the original HSI after dimensionality reduction via PCA. The final correlation matrix M i , j is computed using the following formula:
M i , j = 1 , if x i p i 0 , otherwise
Finally, node V can be represented by the node matrix H. The computation can be expressed as follows:
H = M ^ i , j T V k ,
where M ^ i , j denotes the association matrix normalized by columns, M ^ i , j = M i , j / m M m , j , and V k is the node matrix of the 4-connectivity graph of the original HSI. Specifically, we construct the 4-connectivity map of the original HSI to obtain its node matrix V k , which is directly constructed from the original pixels. At this stage, V k represents the pixel-level node information. Consequently, the constructed H-matrix allows the node features to be transferred from the pixel level to the superpixel level. At this point, in order to transfer from superpixel-level features to pixel-level features, we construct the adjacency matrix H ^ , which can be calculated as follows:
H ^ = M ^ i , j V k
For the 2nd step of DCG-Net, we perform bilinear interpolation on the original HSI to construct a new HSI graph. We then repeat the process to construct a graph structure that aligns with the spatial scale.

3.3. Graph Convolutional Network

We used superpixel segmentation to divide the image into multiple superpixels, which forms the basis for constructing the graph structure. We then employed a graph convolutional network (GCN) [47] to focus on extracting features at the superpixel level, improving the model’s feature extraction capabilities. Additionally, we incorporated an attention mechanism into the GCN to enhance the correlation between graph nodes.
We will outline the specific implementation method, beginning with the calculation of the attention coefficient between a node and its neighbors:
A i , j = S i g m o i d ( H M · ( H M ) T ) ,
A i , j denotes the attention coefficient between the i-th node and the j-th neighbor, where M is the learnable parameter and H is the adjacency matrix. Next, the attention coefficient matrix is obtained using normalization A ^ i , j :
A ^ i , j = S o f t m a x ( A i , j ) ,
Finally, the attention coefficient matrix A ^ i , j is weighted and averaged over the node features, and Leaky ReLU [45] is used as the activation function to perform the graph convolution:
f g H , A = L e a k y R e L U ( j = 1 W A ^ i , j · o u t i , j ) ,
where H represents the node matrix, A is the adjacency matrix, W is the number of neighbors, and o u t i , j denotes the feature representation between the j-th neighbor and the i-th node.

3.4. Feature Aggregation Module

After obtaining the dual-branch features, we need to perform feature fusion to dynamically calibrate the channel response within the network and enhance the model’s feature perception. As shown in Figure 4, we designed a feature aggregation module (FAM) that includes a channel attention module and a customized convolution module. Next, these two modules will be introduced in detail.
Channel attention module: We use the channel attention module to dynamically reweight the fused features. As shown in Figure 4, we apply global adaptive pooling to the input features, compressing their spatial scale from B C × H × W to B × C × 1 × 1 to extract global information, expressed as follows:
Y = A v g P o o l ( X ) ,
where X is the input feature and Y is the feature output. Then, the features pass through two MLP layers. The first MLP layer conducts a fully connected operation followed by ReLU activation, and the second MLP layer also performs a fully connected operation followed by sigmoid activation to constrain the attention weights within the range of (0, 1). This process is specifically illustrated as follows:
Y s 1 = Linear ( Y , c h i n r e d u c t i o n ) Y s 2 = ReLU ( Y s 1 ) Y s 3 = Linear ( Y s 2 , c h i n ) Y o u t = Sigmoid ( Y s 3 ) ,
where Y s 1 , Y s 2 , Y s 3 are the intermediate features, Y o u t is the normalized attention feature, c h i n is the number of input feature channels, and reduction is the feature reduction factor. Finally, the attention features are applied to each channel.
Customized convolution module: After extracting features using the channel attention module, we input these features into the customized convolution module for further extraction of pixel-level features. In this module, features are first batch normalized, then downsampled using ( 1 × 1 ) point-wise convolution and subjected to a nonlinear transformation using the Leaky ReLU activation function [45]. Next, features are processed with ( 3 × 3 ) depthwise separable convolution for extraction, followed by another nonlinear transformation using Leaky ReLU [45], resulting in the final output features.
Through these processes, we globally and adaptively learn the weights of each channel, assigning varying degrees of importance. We then use the customized convolution module to enhance the model’s representation capability and feature extraction, thus enhancing the robustness and accuracy of classification.

4. Experimental Results and Analysis

4.1. Experiment Design

To comprehensively evaluate DCG-Net, we used three well-known datasets: Indian Pines, Salinas, and Kennedy Space Center. We compared DCG-Net with six state-of-the-art methods: contextual CNN-based deep networks (CDCNNs) [40], dual-branch dual attention networks (DBDAs) [31], dual-branch multi-attention networks (DBMAs) [48], fast dense spectral–spatial convolutional networks (FDSSCs) [26], spectral–spatial residual networks (SSRNs) [25], and CNN-enhanced graph convolutional networks (CEGCNs) [37]. The overall accuracy (OA), average accuracy (AA), and Kappa statistics were used as the performance evaluation criteria.

4.1.1. HSI Datasets

To validate the robustness and effectiveness of DCG-Net, we used three datasets: Indian Pines (IN) [49], Salinas (SA) [49], and Kennedy Space Center (KSC) [49], as shown in Figure 5, Figure 6 and Figure 7. In Figure 5, Figure 6 and Figure 7, ‘Train’ indicates the number of training samples, ‘Val’ represents the number of validation samples, and ‘Test’ represents the number of test samples.
Indian Pines dataset: This dataset was captured by the US AVIRIS sensor in the Indian remote sensing experiment (IRSE) area, with an image size of 145 × 145 and a spatial resolution of 20 m. The 20 bands covering the water vapor absorption region are removed, and the remaining 200 spectral bands are taken. In addition, there are a total of 16 unevenly distributed classes containing 10,366 labeled samples. Figure 5 shows a pseudo-color visualization of this dataset, where each color denotes a distinct label category.
Kennedy Space Center dataset: This dataset comprises images captured by the U.S. at the Kennedy Space Center, Florida using the AVIRIS sensor with a size of 512 × 614 pixels and a resolution of 18 m. The water vapor absorption and noise bands were removed, and the remaining 176 spectral bands were taken. This HSI has 13 categories, with a total of 5211 labeled samples. Figure 6 shows a pseudo-color visualization of this dataset, where each color denotes a distinct label category.
Salinas dataset: This dataset was captured by the United States in Salinas Valley in California using the AVIRIS sensor. Its image size is 512 × 217, and it has a spatial resolution of 3.7 m. Twenty water vapor absorption bands were removed, and the remaining 204 spectral bands were taken. This HSI has 16 categories, totaling 54,129 labeled samples. Figure 7 shows a pseudo-color visualization of this dataset, where each color denotes a distinct label category.

4.1.2. Evaluation Indices

Four performance metrics were used in this experiment to evaluate the classification ability of the model: per-class accuracy (PA), overall accuracy (OA), average accuracy (AA), and statistics (Kappa). The formulas for each metric are as follows:
P A = T i F i ,
O A = i T i i F i ,
A A = 1 I i P A ,
K a p p a = O A i T i F i i F i 2 1 i T i F i i F i 2 ,
In Equations (12)–(15), I indicates that the dataset has I categories of samples, i represents the i category of samples, T i indicates the number of samples in the i-th category, and F i indicates the number of samples correctly categorized in the i category.

4.1.3. Environment Configuration

All experiments were conducted in PyCharm using Python 3.11, PyTorch 2.0.1, and an NVIDIA GeForce RTX 3090 Ti GPU server manufactured by Kuankes Co., Shanghai, China. The loss function used in this experiment is category-weighted cross entropy. All activation functions are Leaky ReLU. The network parameters were updated using the Adam optimizer with a learning rate of 0.0005, and the total number of iterations was set to 600. The number of superpixels was set to 512, and the number of neighbors for KNN was set to 15.

4.2. Experiment Results

4.2.1. Comparative Analysis of Classification Performance

In this section, we compare the proposed DCG-Net with six recent methods. For a fair comparison, we used the hyperparameter settings reported in the respective references. Each experiment was repeated ten times to calculate the mean and standard deviation for each metric. The statistical classification results of all the methods are shown in Table 1, Table 2 and Table 3. Additionally, to further analyze the classification performance of our proposed network, the classification results for all the methods are visualized in Figure 8, Figure 9 and Figure 10.
(a) Results of comparative experiments on the IN dataset. For the IN dataset, as shown in Figure 5, the training, validation, and test sets comprised 5%, 5%, and 90%, respectively. The proposed DCG-Net achieved an overall accuracy (OA) of 98.37%, an average accuracy (AA) of 97.64%, and a Kappa coefficient of 98.15; the highest among all the compared models, as shown in Table 1. Compared to CDCNN, DBMA, FDSSC, SSRN, DBDA, and CEGCN, DCG-Net improved the OA by 22.59%, 4.45%, 2.11%, 3.65%, 1.57%, and 1.16%, respectively. The IN dataset has an uneven sample distribution across its classes, especially with very few training samples in classes 1, 7, 9, and 16. In CDCNN, DBMA, SSRN, DBDA, and FDSSC, these classes exhibit poor classification performance due to limited sample sizes. For instance, despite CEGCN’s combined double-branch structure of convolutional and graph networks, it still lacks the ability to effectively extract features from small-scale samples. Conversely, DCG-Net achieved classification accuracies of over 93% in these four challenging classes, demonstrating its strong ability to extract features from small-scale samples.
Figure 8 shows the predicted labeled images for the seven methods for the IN dataset. CDCNN displays significant noise in its classification results, while DBMA, SSRN, DBDA, FDSSC, and CEGCN exhibit varying levels of noise and misclassification. In contrast, the proposed DCG-Net generates smoother and more accurate classification results with fewer noise artifacts and errors. This visualization result clearly demonstrates the superiority of our proposed method.
(b) Results of comparative experiments on the KSC dataset. The KSC dataset, known for its high spatial resolution and low noise levels, is particularly suitable for classification tasks. As shown in Figure 6, due to the dataset’s limited sample size, only 5% of the samples were used as training data, resulting in fewer than 30 training samples per category. As shown in Table 2, DCG-Net achieved exceptional performance, with an overall accuracy (OA) of 99.49%, an average accuracy (AA) of 99.43%, and a Kappa coefficient of 99.44; the highest among all the compared models. Specifically, DCG-Net improved the OA by 11.30%, 5.07%, 1.82%, 4.19%, 2.67%, and 0.31% compared to CDCNN, DBMA, FDSSC, SSRN, DBDA, and CEGCN, respectively. Our model achieved 100% classification accuracy in six categories: 2, 4, 6, 7, 8, and 13, highlighting its robust classification abilities.
Figure 9 shows the predicted labeled images of these seven methods on the KSC dataset. CDCNN’s predicted images exhibit more errors, while DBMA, SSRN, DBDA, FDSSC, and CEGCN show fewer errors but still some noise. In contrast, DCG-Net demonstrates a superior classification performance with minimal noise, closely aligning with the true labels.
(c) Results of comparative experiments on the SA dataset. The SA dataset, characterized by simpler scenarios, explicit feature information, and a more continuous distribution, is relatively easier to train for classification purposes. Consequently, as shown in Figure 7, the dataset was divided into a training set of 0.5%, a validation set of 0.5%, and a test set of 99%. As indicated in Table 3, DCG-Net achieved 99.53% OA, 99.22% AA, and a Kappa coefficient of 99.49; the highest among all the methods. Compared to CDCNN, DBMA, FDSSC, SSRN, DBDA, and CEGCN, DCG-Net improved OA by 10.69%, 3.01%, 3.51%, 6.36%, 2.78%, and 0.79%, respectively. In individual categories 1, 2, 3, and 9, DCG-Net achieved a 100% accuracy, while CEGCN achieved a 100% accuracy only in categories 2 and 9. DBMA, FDSSC, and SSRN achieved a 100% accuracy only in category 1, while CDCNN and DBDA did not achieve a 100% accuracy in any category. This highlights DCG-Net’s superior classification performance across all the categories.
Figure 10 shows the predicted labeled images of these seven methods on the SA dataset. Compared with the other six algorithms, DCG-Net demonstrates superior classification results with minimal noise, closely aligning with the ground truth and showcasing excellent visual classification performance.
DCG-Net achieved the highest OA, AA, and Kappa scores across all three datasets, demonstrating superior classification effectiveness. DCG-Net’s double-branch structure enhances the extraction of features from small sample ranges, while the graph network excels at capturing global features from larger ranges. This dual-branch network structure enhances the complementarity between features at different scales, enabling the model to effectively adapt to diverse target characteristics and achieve superior classification performance. In regions with large-scale samples, DCG-Net produces smoother classification maps, effectively suppressing noise. Conversely, in regions with small-scale samples, DCG-Net delivers finer-grained classification results with a higher accuracy, highlighting its robustness.

4.2.2. Comparison of Running Time between Different Methods

To compare the direct running efficiency for each model, this section records the running time of the different methods. The experimental setup is the same as in Section 4.2.1. Table 4 shows the time cost of each model for the three datasets.
Table 4 shows that the running efficiency of DBMA, FDSSC, SSRN, and DBDA across the three datasets is generally low. This is because these models are based on 3D cubes, which contain a large number of parameters and require longer running and testing times. CDCNN is a 2D-CNN-based model with a comparatively shorter training time. However, all five models utilize local blocks of HSIs as inputs, resulting in long computation times. CEGCN and DCG-Net employ the whole HSIs as input, facilitating parallel computation, and the GCN-based approach enhances the processing speed. Given that the 2nd branch of DCG-Net processes super-resolution hyperspectral data, its structure is more complex, leading to longer processing times than CEGCN. Nonetheless, among all the compared methods, DCG-Net achieved the best classification results, underscoring the superiority of our model.

4.2.3. Comparison of Classification Performance with Different Proportions of Training Samples

The proportion of training samples significantly impacts model classification effectiveness. In this study, we investigate how varying proportions of training samples influence classification performance across different models. We assess training sample proportions of 1%, 3%, 5%, and 7% for the IN and KSC datasets and 0.1%, 0.3%, 0.5%, and 0.7% for the SA dataset. As show in Figure 11, it is evident that for the SA dataset, characterized by distinct data boundaries, our model achieves a 98.78% classification accuracy using just 0.3% of the samples for training. This demonstrates our model’s reduced dependency on training sample size compared to the other models. For both the KSC and IN datasets, significant classification results were achieved using just 3% of the samples for training.
In contrast, the classification performance of the other networks varied significantly across different training sample sizes, whereas DCG-Net maintained a stable performance. This highlights DCG-Net’s robust generalization ability, consistently delivering excellent classification performance, even with limited training samples.

5. Discussion

The classification performance of the proposed DCG-Net is heavily influenced by the characteristics of the constructed graph, particularly the number of nodes and their connectivity properties. The number of superpixels dictates the number of nodes in the graph, while the value of K in the KNN algorithm determines its connectivity characteristics. We conducted two extended experiments to investigate the impact of these factors on DCG-Net’s classification performance.
Furthermore, to assess the effectiveness of each module in our proposed DCG-Net, we performed ablation experiments focusing on the two branches of DCG-Net and the feature aggregation module.

5.1. Effects of the Number of Superpixels

The number of superpixels is proportional to the number of nodes in the graph, affecting the classification results of DCG-Net. In this section, the same number of superpixels is set for both the 1st and 2nd branches, with values of 64, 128, 256, 512, and 768. Figure 12 illustrates the impact of varying numbers of superpixels on the classification results across the three datasets. Specifically, for the IN dataset, Figure 12 shows that increasing the number of superpixels leads to a higher number of graph nodes and improves classification accuracy. When the number of superpixels reaches or exceeds 512, the segmentation results are satisfactory, with little difference in the classification accuracy compared to using 768 superpixels. For the SA dataset, changing the number of superpixels has a minimal effect on the classification accuracy. For the KSC dataset, the segmentation effect is poor, with 64 superpixels, but it improves as the number of superpixels increases.
Therefore, the number of superpixels determines the graph’s node count. More superpixels result in more nodes. Consequently, as each node represents less sample information, more details are retained, leading to improved classification performance.

5.2. Effects of the Value of K in the KNN Algorithm

In our study, the structure of a graph, particularly its edges, is crucial for defining its connectivity characteristics. Utilizing the K-nearest neighbors (KNN) algorithm, we designate a predetermined number of neighboring edges for each node within the graph. To examine how altering the number of neighboring edges influences the classification outcomes, we experimentally set the parameter ‘K’ in the KNN algorithm to various values—specifically, 5, 10, 15, 20, and 25. It is important to note that the quantity of superpixels directly correlates with the total number of nodes in the graph; for consistency across experiments, we maintain this number at 512.
Figure 13 shows the classification performance across three distinct datasets under different configurations of ‘K’. It can be observed that setting the number of neighboring edges too low results in poor node correlations and decreases the network’s adaptive learning ability. Conversely, having too many neighboring edges causes the nodes to cover different categories, which adversely affects training effectiveness. In the analysis of Figure 13, when ‘K’ is set to 15, the classification results for all the datasets are the best.

5.3. Ablation Study

In order to fully evaluate the algorithms we propose in this paper, we conducted a comprehensive test of the double-branch DCG-Net and feature aggregation module using ablation experiments. First, we evaluated the 1st branch and the 2nd branch of DCG-Net on three datasets separately, then we tested the impact of the feature aggregation module on the network. The experimental results are shown in Table 5.
As shown in Table 5, the absence of any branch or module in DCG-Net impacts the classification results. For the Indian Pines and Salinas datasets, FAM improves all three metrics, particularly the average accuracy (AA). For the Kennedy Space Center dataset, FAM significantly boosts all three metrics. For the 2nd branch, on the Kennedy Space Center and Salinas datasets, the improvement in classification accuracy was smaller due to the distinct class boundaries. However, for the Indian Pines dataset, the classification accuracy showed a significant improvement. The 2nd branch refines the classification, enhancing the extraction of fine and edge features. It addresses the 1st branch’s limitations in classifying small samples. The feature aggregation module enhances the overall classification accuracy of the network by dynamically learning channel features and combining them with the CNN, thus improving the model’s robustness and adaptability.

6. Conclusions

In this study, we proposed DCG-Net, an innovative classification network that combines convolutional neural networks (CNNs) and graph convolutional networks (GCNs) to enhance hyperspectral image (HSI) classification. Our approach overcame the limitations of traditional methods by combining pixel-level and superpixel-level features, effectively handling both large-scale regular features and small-scale fine features. The key contributions of our work include the development of the expanding network (E-Net) for enhanced feature extraction, the integration of GCNs with an attention mechanism, and the creation of a feature aggregation module (FAM) for adaptive channel feature learning.
The comprehensive experiments conducted on three widely used hyperspectral datasets, Indian Pines, Salinas, and Kennedy Space Center, demonstrate the superiority of DCG-Net. Our model consistently outperformed state-of-the-art methods in terms of overall accuracy (OA), average accuracy (AA), and the kappa coefficient. Specifically, DCG-Net achieved an OA of 98.37% on the Indian Pines dataset, 99.53% on the Salinas dataset, and 99.49% on the Kennedy Space Center dataset. These results highlight the robustness and effectiveness of our proposed network in various HSI classification scenarios. Our findings indicate that the combination of CNN and GCN architectures, along with the use of attention mechanisms and feature aggregation techniques, significantly improves the classification performance of HSI.
However, our model has shortcomings in memory usage and running speed, and it requires further optimization to improve its performance and efficiency. Future research can explore further enhancements to this hybrid approach, including the incorporation of additional data augmentation techniques and the investigation of alternative graph structures to further optimize feature extraction and classification accuracy.

Author Contributions

Conceptualization, W.Z. and X.S.; methodology, W.Z.; software, W.Z. and Q.Z.; writing—original draft preparation, W.Z. and Q.Z.; writing—review and editing, W.Z. and X.S.; visualization, W.Z. and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Data Availability Statement

All the datasets can be obtained at: http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 20 May 2011). The code corresponding to the proposed method https://github.com/229Kevin/DCG-Net (accessed on 10 July 2024).

Acknowledgments

The authors would like to thank the editor and reviewers for their insights and comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, C.; He, Z.; Lou, A.; Plaza, A. RGB-to-HSV: A Frequency-Spectrum Unfolding Network for Spectral Super-Resolution of RGB Videos. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–18. [Google Scholar] [CrossRef]
  2. Transon, J.; D’Andrimont, R.; Maugnard, A.; Defourny, P. Survey of Hyperspectral Earth Observation Applications from Space in the Sentinel-2 Context. Remote Sens. 2018, 10, 157. [Google Scholar] [CrossRef]
  3. Guo, T.; Luo, F.; Zhang, L.; Tan, X.; Liu, J.; Zhou, X. Target Detection in Hyperspectral Imagery via Sparse and Dense Hybrid Representation. IEEE Geosci. Remote Sens. Lett. 2020, 17, 716–720. [Google Scholar] [CrossRef]
  4. Su, Y.; Gao, L.; Jiang, M.; Plaza, A.; Sun, X.; Zhang, B. NSCKL: Normalized Spectral Clustering with Kernel-Based Learning for Semisupervised Hyperspectral Image Classification. IEEE Trans. Cybern. 2023, 53, 6649–6662. [Google Scholar] [CrossRef] [PubMed]
  5. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  6. Gu, Y.; Chanussot, J.; Jia, X.; Benediktsson, J.A. Multiple Kernel Learning for Hyperspectral Image Classification: A Review. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6547–6565. [Google Scholar] [CrossRef]
  7. Gerhards, M.; Schlerf, M.; Mallick, K.; Udelhoven, T. Challenges and Future Perspectives of Multi-/Hyperspectral Thermal Infrared Remote Sensing for Crop Water-Stress Detection: A Review. Remote Sens. 2019, 11, 1240. [Google Scholar] [CrossRef]
  8. Lv, Z.; Wang, F.; Sun, W.; You, Z.; Falco, N.; Benediktsson, J.A. Landslide Inventory Mapping on VHR Images via Adaptive Region Shape Similarity. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5630211. [Google Scholar] [CrossRef]
  9. Ham, J.; Chen, Y.; Crawford, M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  10. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  11. Prasad, S.; Bruce, L.M. Limitations of Principal Components Analysis for Hyperspectral Target Recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  12. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of Hyperspectral Images with Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  13. Tang, Y.Y.; Yuan, H.; Li, L. Manifold-Based Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7606–7618. [Google Scholar] [CrossRef]
  14. Bai, J.; Yuan, A.; Xiao, Z.; Zhou, H.; Wang, D.; Jiang, H.; Jiao, L. Class Incremental Learning With Few-Shots Based on Linear Programming for Hyperspectral Image Classification. IEEE Trans. Cybern. 2022, 52, 5474–5485. [Google Scholar] [CrossRef] [PubMed]
  15. Wu, Z.; Sun, J.; Zhang, Y.; Zhu, Y.; Li, J.; Plaza, A.; Benediktsson, J.A.; Wei, Z. Scheduling-Guided Automatic Processing of Massive Hyperspectral Image Classification on Cloud Computing Architectures. IEEE Trans. Cybern. 2021, 51, 3588–3601. [Google Scholar] [CrossRef]
  16. Liu, Q.; Zhou, F.; Hang, R.; Yuan, X. Bidirectional-Convolutional LSTM Based Spectral-Spatial Feature Learning for Hyperspectral Image Classification. Remote Sens. 2017, 9, 1330. [Google Scholar] [CrossRef]
  17. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  18. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 6999–7019. [Google Scholar] [CrossRef]
  19. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  20. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar] [CrossRef]
  21. Zhang, S.; Xu, M.; Zhou, J.; Jia, S. Unsupervised Spatial-Spectral CNN-Based Feature Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  22. Huang, L.; Chen, Y. Dual-Path Siamese CNN for Hyperspectral Image Classification with Limited Training Samples. IEEE Geosci. Remote Sens. Lett. 2021, 18, 518–522. [Google Scholar] [CrossRef]
  23. Li, Y.; Zhang, H.; Shen, Q. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  24. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral Image Classification Using a Hybrid 3D-2D Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  25. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
  26. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A Fast Dense Spectral–Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef]
  27. Yang, J.; Zhao, Y.Q.; Chan, J.C.W. Learning and Transferring Deep Joint Spectral–Spatial Features for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4729–4742. [Google Scholar] [CrossRef]
  28. Zhao, C.; Qin, B.; Feng, S.; Zhu, W.; Sun, W.; Li, W.; Jia, X. Hyperspectral Image Classification with Multi-Attention Transformer and Adaptive Superpixel Segmentation-Based Active Learning. IEEE Trans. Image Process. 2023, 32, 3606–3621. [Google Scholar] [CrossRef] [PubMed]
  29. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Visual Attention-Driven Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  30. Pu, C.; Huang, H.; Yang, L. An attention-driven convolutional neural network-based multi-level spectral–spatial feature learning for hyperspectral image classification. Expert Syst. Appl. 2021, 185, 115663. [Google Scholar] [CrossRef]
  31. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of Hyperspectral Image Based on Double-Branch Dual-Attention Mechanism Network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  32. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2017. [Google Scholar] [CrossRef]
  33. Mou, L.; Lu, X.; Li, X.; Zhu, X.X. Nonlocal Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8246–8257. [Google Scholar] [CrossRef]
  34. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale Dynamic Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3162–3177. [Google Scholar] [CrossRef]
  35. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. Multilevel Superpixel Structured Graph U-Nets for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  36. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Cai, W.; Yu, C.; Yang, N.; Cai, W. Multi-feature fusion: Graph neural network and CNN combining for hyperspectral image classification. Neurocomputing 2022, 501, 246–257. [Google Scholar] [CrossRef]
  37. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. CNN-Enhanced Graph Convolutional Network with Pixel- and Superpixel-Level Feature Fusion for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8657–8671. [Google Scholar] [CrossRef]
  38. Dong, Y.; Liu, Q.; Du, B.; Zhang, L. Weighted Feature Fusion of Convolutional Neural Network and Graph Attention Network for Hyperspectral Image Classification. IEEE Trans. Image Process. 2022, 31, 1559–1572. [Google Scholar] [CrossRef] [PubMed]
  39. Lu, T.; Li, S.; Fang, L.; Jia, X.; Benediktsson, J.A. From Subpixel to Superpixel: A Novel Fusion Framework for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4398–4411. [Google Scholar] [CrossRef]
  40. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep Feature Extraction and Classification of Hyperspectral Images Based on Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  41. Liu, Q.; Xiao, L.; Yang, J.; Chan, J.C.W. Content-Guided Convolutional Neural Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6124–6137. [Google Scholar] [CrossRef]
  42. Wang, J.; Huang, R.; Guo, S.; Li, L.; Zhu, M.; Yang, S.; Jiao, L. NAS-Guided Lightweight Multiscale Attention Fusion Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8754–8767. [Google Scholar] [CrossRef]
  43. Wei, X.; Yang, Q.; Gong, Y.; Ahuja, N.; Yang, M.H. Superpixel Hierarchy. IEEE Trans. Image Process. 2018, 27, 4838–4849. [Google Scholar] [CrossRef]
  44. Bai, J.; Shi, W.; Xiao, Z.; Regan, A.C.; Ali, T.A.A.; Zhu, Y.; Zhang, R.; Jiao, L. Hyperspectral Image Classification Based on Superpixel Feature Subdivision and Adaptive Graph Structure. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  45. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 17–19 June 2013; Volume 30, p. 3. [Google Scholar]
  46. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5966–5978. [Google Scholar] [CrossRef]
  47. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the 30th International Conference on Neural Information Processing Systems NIPS’16, Barcelona, Spain, 5–10 December 2016; pp. 3844–3852. [Google Scholar]
  48. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-Branch Multi-Attention Mechanism Network for Hyperspectral Image Classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef]
  49. Maveganzones. Hyperspectral Remote Sensing Scenes. Available online: https://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 20 May 2011).
Figure 1. DCG-Net architecture diagram. Our model is structured into four primary stages: superpixel segmentation, feature extraction, feature aggregation, and feature classification. We employ a dual-branch architecture, where each branch processes both pixel-level and superpixel-level features. The 1st branch retains the original image resolution for feature processing, whereas the 2nd branch processes upscaled feature images to facilitate multi-scale feature analysis.
Figure 1. DCG-Net architecture diagram. Our model is structured into four primary stages: superpixel segmentation, feature extraction, feature aggregation, and feature classification. We employ a dual-branch architecture, where each branch processes both pixel-level and superpixel-level features. The 1st branch retains the original image resolution for feature processing, whereas the 2nd branch processes upscaled feature images to facilitate multi-scale feature analysis.
Electronics 13 03271 g001
Figure 2. E-Net structure diagram. Both the encoder and decoder use a customized convolution module. E-Net can effectively combine GCN in the dual-branch encoding and decoding processes to achieve the effective combination of different spatial features.
Figure 2. E-Net structure diagram. Both the encoder and decoder use a customized convolution module. E-Net can effectively combine GCN in the dual-branch encoding and decoding processes to achieve the effective combination of different spatial features.
Electronics 13 03271 g002
Figure 3. Flowchart of superpixel segmentation and graph structure construction. (a) Original hyperspectral image. (b) Hyperspectral image after PCA downscaling, with the 4-connected graph as a feature graph constructed using the pixels of (b) as nodes. (c) Hyperspectral image after superpixel segmentation, with the superpixel graph structure as a feature graph constructed using the superpixels of (c) as nodes. The orange dots represent the graph’s nodes, and the orange dotted lines represent the graph’s edges.
Figure 3. Flowchart of superpixel segmentation and graph structure construction. (a) Original hyperspectral image. (b) Hyperspectral image after PCA downscaling, with the 4-connected graph as a feature graph constructed using the pixels of (b) as nodes. (c) Hyperspectral image after superpixel segmentation, with the superpixel graph structure as a feature graph constructed using the superpixels of (c) as nodes. The orange dots represent the graph’s nodes, and the orange dotted lines represent the graph’s edges.
Electronics 13 03271 g003
Figure 4. Feature aggregation module structure diagram. The module begins with channel feature learning through the channel attention module, then it continues with feature extraction using the customized convolution module.
Figure 4. Feature aggregation module structure diagram. The module begins with channel feature learning through the channel attention module, then it continues with feature extraction using the customized convolution module.
Electronics 13 03271 g004
Figure 5. Description of Indian Pines dataset.
Figure 5. Description of Indian Pines dataset.
Electronics 13 03271 g005
Figure 6. Description of Kennedy Space Center dataset.
Figure 6. Description of Kennedy Space Center dataset.
Electronics 13 03271 g006
Figure 7. Description of the Salinas dataset.
Figure 7. Description of the Salinas dataset.
Electronics 13 03271 g007
Figure 8. Classification maps of different methods for the Indian Pines dataset.
Figure 8. Classification maps of different methods for the Indian Pines dataset.
Electronics 13 03271 g008
Figure 9. Classification maps of different methods for the Kennedy Space Center dataset.
Figure 9. Classification maps of different methods for the Kennedy Space Center dataset.
Electronics 13 03271 g009
Figure 10. Classification maps of different methods for the Salinas dataset.
Figure 10. Classification maps of different methods for the Salinas dataset.
Electronics 13 03271 g010
Figure 11. Comparison of classification performance of seven methods with different training set ratios for three datasets.
Figure 11. Comparison of classification performance of seven methods with different training set ratios for three datasets.
Electronics 13 03271 g011
Figure 12. Comparison results of different superpixel numbers on three datasets.
Figure 12. Comparison results of different superpixel numbers on three datasets.
Electronics 13 03271 g012
Figure 13. Comparison results of different K values on three datasets.
Figure 13. Comparison results of different K values on three datasets.
Electronics 13 03271 g013
Table 1. Classification results of different methods in terms per-class accuracy, OA (%), AA (%), and Kappa for the Indian Pines dataset.
Table 1. Classification results of different methods in terms per-class accuracy, OA (%), AA (%), and Kappa for the Indian Pines dataset.
ClassCDCNNDBMAFDSSCSSRNDBDACEGCNDCG-Net
131.8690.6976.0293.6890.0869.7998.95
268.6590.6876.0293.3696.1496.7198.03
362.5795.5393.4492.6497.4598.0598.42
458.5294.1696.1888.9597.1191.6198.09
596.3797.0298.7799.4297.9393.0294.94
690.1698.3099.0898.3178.7399.4099.35
739.5470.1976.1588.5678.3990.9094.09
888.1798.1496.8698.0499.8699.9799.97
974.3275.4377.1284.6379.9623.3393.12
1066.5289.3293.9791.8194.2794.7196.88
1175.8195.0297.3094.2897.6598.8198.61
1249.9192.6495.7894.2195.7896.9997.60
1393.3998.3999.8499.3396.0499.3699.26
1488.6997.2996.8098.2598.8099.6799.99
1578.8287.8696.8093.6197.0394.1898.73
1695.3294.1997.4291.8386.6293.5296.23
OA (%)75.78 ± 5.0293.92 ± 1.5496.26 ± 2.4994.72 ± 0.7296.80 ± 0.5997.21 ± 0.5198.37 ± 0.58
AA (%)72.41 ± 7.9791.55 ± 1.5692.93 ± 6.7393.81 ± 1.7493.78 ± 2.2590.00 ± 2.3397.64 ± 0.86
Kappa72.22 ± 5.9793.07 ± 1.7695.73 ± 2.8693.98 ± 0.8296.35 ± 0.6896.82 ± 0.5898.15 ± 0.66
Table 2. Classification results of different methods in terms of per-class accuracy, OA (%), AA (%), and Kappa for the Kennedy Space Center dataset.
Table 2. Classification results of different methods in terms of per-class accuracy, OA (%), AA (%), and Kappa for the Kennedy Space Center dataset.
ClassCDCNNDBMAFDSSCSSRNDBDACEGCNDCG-Net
194.71100.0098.6496.5499.9499.5699.61
285.4193.3897.3890.8796.4698.41100.00
379.2879.5890.8287.5783.2799.5899.43
453.5775.1190.5675.6283.3295.08100.00
547.4065.9689.8569.6290.8590.5397.76
671.2291.4899.1493.0798.4499.48100.00
772.1687.7493.6971.6788.5898.76100.00
884.1095.3898.7198.1399.6499.77100.00
991.0996.0799.7698.4999.91100.0099.89
1092.8597.0299.8699.24100.00100.0098.67
1198.2499.8998.7798.9699.26100.0098.43
1296.2598.1798.1299.5299.4399.2998.86
1399.91100.00100.0099.8999.91100.00100.00
OA (%)88.14 ± 2.5794.42 ± 1.9197.67 ± 1.1295.30 ± 2.6596.82 ± 1.3199.18 ± 0.5199.49 ± 0.34
AA (%)82.02 ± 3.6190.75 ± 2.9396.65 ± 1.1390.71 ± 8.0095.31 ± 1.5198.50 ± 0.8999.43 ± 0.55
Kappa86.78 ± 2.8893.78 ± 2.1397.41 ± 1.2594.76 ± 2.9696.46 ± 1.4599.08 ± 0.5799.44 ± 0.38
Table 3. Classification results of different methods in terms of per-class accuracy, OA (%), AA (%), and Kappa for the Salinas dataset.
Table 3. Classification results of different methods in terms of per-class accuracy, OA (%), AA (%), and Kappa for the Salinas dataset.
ClassCDCNNDBMAFDSSCSSRNDBDACEGCNDCG-Net
197.34100.00100.00100.00100.0099.95100.00
297.2599.9299.9299.7599.96100.00100.00
390.3797.7997.7594.2198.1699.86100.00
498.0893.5597.3297.8794.6199.5295.99
594.6098.5499.3898.8998.9898.1098.28
696.7999.5699.9599.8599.8399.8399.82
797.0999.4199.5199.3298.9099.9899.77
878.5694.1890.5188.4593.0497.6499.70
999.0099.6099.5099.3699.35100.00100.00
1088.2197.1397.1996.9398.2996.9798.56
1184.9895.3594.8495.6896.1099.7597.48
1295.9499.1898.2198.6999.16100.0099.84
1396.9299.2999.4497.9199.6499.8099.86
1493.7693.4996.9996.7494.9698.6397.91
1573.7590.9090.7581.1693.1797.1199.92
1694.2799.0499.9399.2199.9198.9799.93
OA (%)88.80 ± 1.0396.52 ± 0.8996.02 ± 1.1593.17 ± 3.5096.75 ± 0.6498.74 ± 0.5099.53 ± 0.29
AA (%)92.31 ± 1.2697.31 ± 0.6697.58 ± 0.5296.50 ± 0.7497.75 ± 0.4199.13 ± 0.1899.22 ± 0.48
Kappa87.52 ± 1.1496.12 ± 0.9995.56 ± 1.2992.42 ± 3.8296.38 ± 0.7299.13 ± 0.5599.49 ± 0.29
Table 4. Comparison of running times between different methods.
Table 4. Comparison of running times between different methods.
DatesetTime (s)CDCNNDBMAFDSSCSSRNDBDACEGCNDCG-Net
Indian PinesTrain47.10451.78584.42407.85436.54.7417.52
Test0.493.742.181.683.411.520.60
SalinasTrain25.40237.50312.99213.95229.7117.9279.95
Test2.8221.8712.679.6919.691.742.92
Kennedy Space CenterTrain23.30175.87230.31164.04180.0060.46278.95
Test0.241.670.990.761.532.438.39
Table 5. Comparison of different modules (× indicates that the module was not used in these experiments, ✓ indicates that the module was used).
Table 5. Comparison of different modules (× indicates that the module was not used in these experiments, ✓ indicates that the module was used).
Indian Pines
No.1st Branch2nd BranchFAMOA (%)AA (%)Kappa
1××97.78 ± 0.3497.52 ± 0.8997.47 ± 0.39
2××97.29 ± 0.4497.12 ± 1.2096.19 ± 0.50
3×98.33 ± 0.5796.60 ± 0.1998.10 ± 0.66
498.37 ± 0.4797.64 ± 0.8698.15 ± 0.66
Salinas
No.1st Branch2nd BranchFAMOA (%)AA (%)Kappa
1××99.40 ± 0.2098.96 ± 0.5099.38 ± 0.25
2××94.37 ± 0.4693.83 ± 1.2293.74 ± 0.52
3×99.40 ± 0.2098.96 ± 0.5099.38 ± 0.25
499.53 ± 0.2999.22 ± 0.4899.49 ± 0.29
Kennedy Space Center
No.1st Branch2nd BranchFAMOA (%)AA (%)Kappa
1××98.65 ± 0.5898.50 ± 0.8698.59 ± 0.64
2××94.37 ± 0.4693.83 ± 1.2293.74 ± 0.52
3×98.89 ± 0.6798.47 ± 1.0398.76 ± 0.75
499.49 ± 0.3499.43 ± 0.5599.44 ± 0.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, W.; Sun, X.; Zhang, Q. DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration. Electronics 2024, 13, 3271. https://doi.org/10.3390/electronics13163271

AMA Style

Zhu W, Sun X, Zhang Q. DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration. Electronics. 2024; 13(16):3271. https://doi.org/10.3390/electronics13163271

Chicago/Turabian Style

Zhu, Wenkai, Xueying Sun, and Qiang Zhang. 2024. "DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration" Electronics 13, no. 16: 3271. https://doi.org/10.3390/electronics13163271

APA Style

Zhu, W., Sun, X., & Zhang, Q. (2024). DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration. Electronics, 13(16), 3271. https://doi.org/10.3390/electronics13163271

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop