Next Article in Journal
Utö Observatory for Analysing Atmospheric Ducting Events over Baltic Coastal and Marine Waters
Previous Article in Journal
ECOSTRESS Reveals the Importance of Topography and Forest Structure for Evapotranspiration from a Tropical Forest Region of the Andes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Classification Based on Dense Pyramidal Convolution and Multi-Feature Fusion

1
College of Computer Science and Technology, China University of Petroleum (East China), Qingdao 266580, China
2
State Key Laboratory of Integrated Services Networks, Xidian University, Xi’an 710071, China
3
The Sixth Research Institute of China Electronics Corporation, Beijing 100083, China
4
School of Information Engineering, Huzhou University, Huzhou 313000, China
5
College of Science, China University of Petroleum (East China), Qingdao 266580, China
6
Shandong Provincial Key Laboratory of Computer Networks, Shandong Computer Science Center (National Supercomputer Center in Jinan), Qilu University of Technology (Shandong Academy of Sciences), Jinan 250013, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(12), 2990; https://doi.org/10.3390/rs15122990
Submission received: 4 April 2023 / Revised: 29 May 2023 / Accepted: 6 June 2023 / Published: 8 June 2023

Abstract

:
In recent years, hyperspectral image classification techniques have attracted a lot of attention from many scholars because they can be used to model the development of different cities and provide a reference for urban planning and construction. However, due to the difficulty in obtaining hyperspectral images, only a limited number of pixels can be used as training samples. Therefore, how to adequately extract and utilize the spatial and spectral information of hyperspectral images with limited training samples has become a difficult problem. To address this issue, we propose a hyperspectral image classification method based on dense pyramidal convolution and multi-feature fusion (DPCMF). In this approach, two branches are designed to extract spatial and spectral features, respectively. In the spatial branch, dense pyramid convolutions and non-local blocks are used to extract multi-scale local and global spatial features in image samples, which are then fused to obtain spatial features. In the spectral branch, dense pyramidal convolution layers are used to extract spectral features in image samples. Finally, the spatial and spectral features are fused and fed into fully connected layers to obtain classification results. The experimental results show that the overall accuracy (OA) of the method proposed in this paper is 96.74%, 98.10%, 98.92% and 96.67% on the four hyperspectral datasets, respectively. Significant improvements are achieved compared to the five methods of SVM, SSRN, FDSSC, DBMA and DBDA for hyperspectral classification. Therefore, the proposed method can better extract and exploit the spatial and spectral information in image samples when the number of training samples is limited. Provide more realistic and intuitive terrain and environmental conditions for urban planning, design, construction and management.

Graphical Abstract

1. Introduction

Hyperspectral images, also known as hyperspectral remote sensing images, are stereoscopic images captured by aerospace vehicles equipped with hyperspectral imagers. They consist of two spatial dimensions and one spectral dimension. The spectral dimension contains 10 s or even 100 s of spectral bands, which provide it with broad prospects for applications such as military target detection [1]; atmospheric and environmental research [2]; forest vegetation cover detection [3]; and change area detection [4]. Hyperspectral image classification is a commonly used technique in the applications listed above. However, the excessive redundancy of spectral information and the limited number of training samples pose a great challenge for hyperspectral image classification.
In the early research on hyperspectral image classification, methods such as support vector machine (SVM) [5], multinomial logistic regression (MLR) [6] and sparse representation classification (SRC) [7] were proposed, which directly take the original input as the training sample and use it to train the classifier through the spectral information of the hyperspectral image. However, such methods ignore two problems: (1) the large amount of redundant information in spectral bands makes it difficult to train the model; (2) hyperspectral images have high spatial correlation and contain abundant spectral information. To solve problem (1), dimensionality reduction strategies [8,9] (feature selection [10] and feature extraction [11]) are applied to hyperspectral image classification tasks. To solve problem (2), morphological contours [12] and Gabor features [13] are used to ex-tract spatial information, and the morphological kernel [14] and composite kernel [15] methods are used to extract spectral–spatial information. Although the aforementioned methods improve the accuracy of the classifier, it is difficult to achieve better classification results in complex scenes because these methods use shallow models and rely heavily on labeled samples, which cannot extract the deep features of the samples.
Deep learning (DL) has shown strong capabilities in automatically extracting nonlinear and hierarchical features, and thus has been widely used in information extraction [16], image classification [17], semantic segmentation [18] and object detection [19]. Therefore, some hyperspectral image classification methods based on deep learning are proposed. In [20], Zhou et al. used a stacked auto-encoder (SAE) to extract spectral and spatial features and used logistic regression to obtain classification results. In [21], Szegedy C, et al. used a restricted Boltzmann machine (RBM) and deep belief network (DBN) for classification. In [22], Ma et al. used a spatially updated deep auto-encoder (DAE) to extract spectral–spatial features and designed a different co-representation mechanism to handle narrow-scale training sets. In [23], Zhang et al. utilized a recursive auto-encoder to learn the spatial and spectral information and adopted a weighting scheme to fuse the spatial information. Although these methods can extract the spectral–spatial features of hyperspectral images to a certain extent, they destroy the spatial structure. Since convolutional neural networks (CNNs) can exploit spatial features while preserving the original spatial structure, some methods based on CNNs have been proposed. In [24], Zhao et al. employed a CNN as the feature extractor. In [25], Zhang et al. proposed a method based on differentiated region convolutional neural network (DRCNN), which uses different image patches within the neighborhood of the target pixel as the input of the CNN, and the input data is effective reinforcement. In [26], Lee et al. proposed a contextual deep CNN (CDCNN) with deeper and wider network layers.
In general, deep-level features in the image can be extracted by increasing the depth of the network, but this also causes problems such as difficulty in model training and gradient vanishing. The residual network (ResNet) [27] and dense convolutional network (DenseNet) [28] solved this problem quite efficiently, and such networks can also extract deep-level features without increasing the depth of the network structure. Inspired by ResNet, the literature [29] proposed a spectral-spatial residual network (SSRN) that contains a spectral residual block and a spatial residual block for sequentially extracting spectral features and spatial features. Inspired by DenseNet, some literature [30] has proposed a fast dense spectral-spatial convolutional network (FDSSC), which achieves better performance while reducing the training time. Although the aforementioned methods solve the feature extraction problem using CNN, but in the process of model training, the attention of the convolutional layer to features is not the same. In order to optimize the extracted features, the attention mechanism is used to process different features differently. It is also a research hotspot in recent years. One study [31] proposed a feedback attention-based dense CNN network, while another [32] proposed a dual-branch multi-attention mechanism network (DBMA) based on the convolutional block attention module (CBAM) [33]. Moreover, [34] proposed a dual-branch dual-attention mechanism network (DBDA) based on the dual-attention network (DANet) [35]. Although these methods are very effective, the extraction and utilization of spatial information and spectral information of hyperspectral images are not sufficient, resulting in the inability to obtain better classification results in the case of limited training samples.
Inspired by pyramidal convolution (PyConv) [36] and DenseNet, for the two problems of missing features and insufficient feature extraction, this paper proposes a hyperspectral image classification method based on dense pyramidal convolution and multi-feature fusion (DPCMF). The proposed method consists of two branches: the spatial branch and the spectral branch, both of which are designed to capture spectral and spatial features, respectively. In the spatial branch, principal component analysis (PCA) is performed to achieve dimensionality reduction for image samples, whereas noise and redundant spectral information are removed while retaining significant spectral information. Then, the dense pyramid convolution module and non-local block [37] are used to extract multi-scale local spatial information and global spatial information from image samples. Finally, the multi-scale local spatial information and global spatial information are fused to obtain a spatial feature map. In the spectral branch, convolutional neural networks are first used to perform convolutions on image samples, and then the spectral information in the images is extracted through dense pyramid convolution to obtain spectral feature maps. Lastly, the spatial and spectral feature maps are fused and fed into the classification module to obtain the classification results. The three main contributions of this paper are described below.
  • To address the problem of missing features, a hyperspectral image classification method (DPCMF) based on dense pyramidal convolution and multi-feature fusion is proposed. This method uses spatial and spectral branches to extract local spatial features and spectral features, respectively. Multi-scale local spatial information is extracted using non-local block segmentation and fused to obtain hyperspectral feature maps.
  • To address the problem of insufficient feature extraction, in the feature extraction part of the spatial and spectral branches, a combination of pyramidal convolution and DenseNet is used. Without increasing the depth of the network, the convolutional kernels are arranged in descending order to extract deep-level features at different scales.
  • DPCMF achieves state-of-the-art classification accuracies on four datasets with limited training samples.
The rest of this paper is arranged as follows: Section 2 presents the specific implementation of the proposed method; Section 3 and Section 4 present and analyze the experimental results; and Section 5 summarizes the conclusions of the article and proposes future directions for research.

2. Materials and Methods

In this section, we give a brief introduction to the structure of dense pyramidal convolution layers, non-local blocks and DPCMF networks.

2.1. Dense Pyramidal Convolution

In a traditional CNN, deeper features are generally extracted by increasing the network depth, but the problem of vanishing or exploding gradients occurs simultaneously, making it difficult to train deep models. DenseNet improves feature availability without increasing the network depth through dense connections, and it also solves the problem of gradient vanishing and gradient exploding. As shown in Figure 1, the purpose of feature reuse is achieved in DenseNet by building dense connections between all previous layers and subsequent layers to fuse shallow and deep features. The output X i of the ith layer can be expressed as:
  X i = H i X 0 ,   X 1 , , X i 1
where H i represents a nonlinear transformation including convolution, activation function, and batch normalization (BN), and [] means the output X 0 , X 1 , . . , X i 1 of the 0th layer to the (i−1)th layer are spliced along the channel dimension.
During the training process of DenseNet, convolution kernels of uniform size are normally used to extract image features, which will lead to insufficient actual receptive field of the convolutional neural network, inability to extract multi-scale information in image samples, and insufficient feature extraction. PyConv uses convolution kernels of different sizes to extract image features, which solves this problem effectively. Based on DenseNet and PyConv, we propose a dense pyramid structure as a feature extraction module for hyperspectral images. The three-dimensional convolutional neural network (3D-CNN) is obtained by adding a dimension of convolution calculation on basis of the original convolutional neural network, and the dimensionality of the convolution kernel is increased from the original two dimensions to three dimensions. Unlike 2D convolutions, which focus on exploring single feature information, 3D convolutions can slide through three dimensions, mining rich information from all feature maps. A hyperspectral image is a cubic structure that contains both spectral and spatial information, and 3D convolutions not only affect the spatial dimension, but they also involve multiple continuous spectral bands simultaneously. Therefore, spatial and spectral information can be obtained by applying 3D convolutional neural networks to extract features from hyperspectral images. As shown in Figure 2, 3D convolutions with convolution kernels arranged in descending order according to their sizes are used as the basic structure to extract images that are not used in the image samples. Scale information, using dense connections, concatenates shallow large-scale features and deep small-scale features to obtain a feature map that combines multi-scale information in image samples.
The convolution process of the 3D convolutional layer is shown in Figure 3. The input is n k feature maps with a size of p k × p k × b k , and after n k + 1 3D-CNN layers with a certain convolution kernel size, n k + 1 feature maps with a size of p k + 1 × p k + 1 × b k + 1 are generated. The ith output of the (k + 1)th 3D-CNN layer is expressed as:
  X i k + 1 = R j = 1 n k X ^ j k × H i k + 1 + c i k + 1
where   X ^ j k R p × p × k represents the jth input feature map of the (k+1)th layer,   X i k + 1 represents the final output of the kth layer, H i k + 1 and c i k + 1 represent the weights and biases of the (k+1)th layer and R(∙) represents the batch normalization (BN) and nonlinear activation function.

2.2. Non-Local Block

In a CNN, the convolution operation only focuses on the local receptive field. If the receptive field of a neuron is to be increased, this can only be achieved by stacking convolutional and pooling layers. However, this approach would considerably increase the computational load and the number of parameters. Therefore, in this paper, we use non-local blocks to capture the connections between distant pixels and obtain the global information in the image sample by computing the weights of all position features in the image sample. Figure 4 shows the structure of a non-local block, where X is the input feature map and Y is the output feature map with the same shape as X. θ, Φ, g and h represent convolution kernels with a size of 1 × 1 × 1, where the number of θ, Φ and g is 1/2 of the number of channels of X and the number of h is the number of channels of X. The convolutional layer first reduces the number of channels of the input X to 1/2 of the original one through θ, Φ and g, thus reducing the computational load. This process can be expressed as:
θ x i = W θ x i + b θ
Φ x i = W Φ x i + b Φ
g x i = W g x i + b g
where W and b denote the weight matrix and bias in the convolution process, respectively. After three convolutions, the outputs θ x i , Φ x i and g x i with the number of channels being 1/2 of that in X are obtained, and the transposes of θ x i and Φ x i are point multiplication operations. This process can be expressed as:
  f x i , x j = θ x i T Φ x j
where f x i , x j represents the influence of position j on position i. The greater the value of f, the greater the influence of position j on position i. The main idea is to perform point multiplication between the obtained output and the eigenvalue g x j of position j, restore the output through the convolution layer h until it has the same shape as the original input X and finally add the obtained weight matrix to the original input. This process can be expressed as:
  y i = f x i , x j g x j
  z i = h y i + x i
where   y i is the weight of position i in the non-local block and z i denotes the final output of position i in the non-local block. The final output Z of the entire non-local block can be expressed as:
  Z = h 1 C x j f x i , x j g x j + X
where C(x) is a normalization parameter, and its value is the number of positions in X.

2.3. Framework of the DPCMF Network

The structure of the DPCMF network is shown in Figure 5. In this section, the Indian Pines (IP) dataset is taken as an example to describe the architecture of the DPCMF model. The Indian Pines dataset contains 145 × 145 pixels, and each pixel has 200 spectral bands. The size of this dataset is 145 × 145 × 200, the number of pixels with labels is 20,249, the other pixels are the background, the number of label categories is 16, and the number of convolution kernels is 12.
In the spatial branch, the original image samples are processed to a size of 145 × 145 × 100 by PCA, which reduces the amount of calculation and the number of model parameters while retaining the main information. Then, the image is split into training samples with a size of 9 × 9 × 100. In the local information extraction module, each 9 × 9 × 100 image sample is first convolved into a feature map with 24 channels and a size of 9 × 9 × 1 through the convolution layer and then input into the spatial block that consists of three layers. The convolution kernels are arranged in descending order according to their sizes, which are 7 × 7 × 1, 5 × 5 × 1, and 3 × 3 × 1. Each convolutional layer is followed by a batch normalization layer and a ReLU [38] activation function, and finally, a feature map with a channel number of 60 and a size of 9 × 9 × 1 is obtained. In the global information extraction module, the feature map with the same shape as the input is obtained through the non-local module and then input into the convolutional layer to obtain a feature map with a channel number of 60 and a size of 9 × 9 × 1. After the global and local information in the image samples is extracted by the two modules, the feature maps of the two parts are concatenated to obtain the spatial feature maps. The specific implementation details of the spatial branch are given in Table 1.
In the spectral branch, the image is split into training samples with a size of 9 × 9 × 200, each sample is first convolved into a 9 × 9 × 97 feature map through the convolutional layer and then input into the spectral block. The spectral block consists of three convolutional layers. The convolutional kernels are arranged in descending order according to their sizes, which are 1 × 1 × 7, 1 × 1 × 5 and 1 × 1 × 3, respectively. After each convolutional layer, a batch normalization layer and a ReLU activation function are added, and the image sample is input to the convolution kernel in the convolutional layer with a size of 1 × 1 × 97 to obtain the spectral feature map. The specific implementation details of the spectral branch are listed in Table 2.
In the classification module, the spatial feature map obtained by the spatial branch and the spectral feature map obtained by the spectral branch are subjected to feature fusion to obtain a feature map with a channel number of 180 and a size of 9 × 9 × 1; then, a feature map with a size of 1 × 180 is obtained through the global average pooling (GAP) layer; finally, the classification result is obtained through the fully connected layer and the Softmax activation function. The implementation details of the classification module are given in Table 3.

3. Experimental Results

In order to verify the effectiveness of the method proposed in this paper, four public hyperspectral datasets, namely, Indian Pines (IP), Pavia University (UP), Salinas Valley (SV) and Botswana (BS), are used to conduct experiments. The accuracy of each method is measured by three evaluation indicators: overall accuracy (OA), average accuracy (AA) and Kappa coefficient. OA represents the proportion of correctly classified samples to the total test samples, and AA represents the average accuracy across all categories. The Kappa coefficient represents the level of consistency between the true value and the classification result. The greater the values of these three evaluation metrics, the better the classification results.

3.1. Introduction and Division of the Dataset

Indian Pines (IP): Indian Pines was imaged by the Airborne Visible Infrared Imaging Spectrometer (AVIRIS) in 1992 on an Indian pine tree in Indiana, USA and marked with a size of 145 × 145. The imaging wavelength range of AVIRIS is 0.4–2.5 μm, and it continuously images ground objects in 220 continuous bands. However, because the 104th–108th, 150th–163rd and 220th bands cannot be reflected by water, 200 bands covering 16 types of ground objects are truly used for training.
Pavia University (UP): Pavia University is part of the hyperspectral data of the city of Pavia in Italy in 2003 collected by the German Airborne Reflective Optics Spectrographic Imaging System (ROSIS-03). The size of this dataset is 610 × 340. The spectral imager continuously captures images with 115 bands within the wavelength range of 0.43–0.86 μm, and the spatial resolution of the resulting images is 1.3 m. Among these bands, 12 bands were eliminated due to noise effects, and 103 bands covering 9 types of ground objects were used for real-world training.
Salinas Valley (SV): The Salinas Valley dataset was acquired by the AVIRIS sensor in Salinas Valley, California. The size of this dataset is 512 × 217, the spatial resolution is 3.7 m, and it contains 224 continuous bands. In total, 20 water-absorbing bands (108–112, 154–167, 224) were removed, and 204 bands covering 16 types of ground objects were actually used for training.
Botswana (BS): The Botswana dataset was acquired by the NASA EO-1 satellite in the Okavango Delta of Botswana in May 2001. The size of this dataset is 1476 × 256. The sensor on EO-1 has a wavelength range of 400–2500 nm and a spatial resolution of about 20 m. Among the 242 bands, the noise bands (1–9, 56–81, 98–101, 120–133, 165–186) were removed, and 145 bands covering 14 types of ground objects were actually used for training.
Before conducting the experiments, we split each dataset into three parts, namely, training set, validation set, and test set. The training set is used to update model parameters, the validation set is used to monitor the temporary models generated during the training phase, and the test set is used to evaluate the optimal model. For different datasets, the proportions of the three parts are different. The division of Indian Pines (IP) is shown in Table 4, the division of Pavia University (UP) is shown in Table 5, the division of Salinas Valley (SV) is shown in Table 6 and the division of Botswana (BS) is shown in Table 7.

3.2. Experimental Setting

To validate the classification performance of the, we conducted experiments to compare the DPCMF network with the SVM, SSRN, FDSSC, DBMA and DBDA classification networks. All experiments were performed on Intel (R) Xeon (R) 4208 CPU @ 2.10 GHz processor with Nvidia GeForce RTX Running on the 2060Ti graphics card system. The programming language used is Python. All classification networks were implemented using PyTorch, PyCharm was used as the compiler, the batch size was set to 16, RMSprop was used as the optimizer, the initial learning rate was set to 0.00005, and the cross-entropy loss function was used for experiments.

3.3. Classification Maps and Categorized Results

3.3.1. Classification Maps and Categorized Results for the IP Dataset

In this experiment, 3% of the samples were used as training samples, 3% as validation samples, and 94% as test samples. The categorized results of different methods on the IP dataset are listed in Table 8, and the classification maps are shown in Figure 6.

3.3.2. Classification Maps and Categorized Results for the UP Dataset

In this experiment, 0.5% of the samples were used as training samples, 0.5% as validation samples, and 99% as test samples. The categorized results of different methods on the UP dataset are listed in Table 9, and the classification maps are shown in Figure 7.

3.3.3. Classification Maps and Categorized Results for the SV Dataset

In this experiment, 0.5% of the samples were used as training samples, 0.5% as validation samples, and 99% as test samples. The categorized results of different methods on the SV dataset are listed in Table 10, and the classification maps are shown in Figure 8.

3.3.4. Classification Maps and Categorized Results for the BS Dataset

In this experiment, 1.2% of the samples were used as training samples, 1.2% as validation samples, and 97.6% as test samples. The categorized results of the different methods on the BS dataset are listed in Table 11, and the classification maps are shown in Figure 9.

3.4. Impact of Convolution Kernel Size

In the process of feature extraction, the size of the convolution kernel affects the degree to which information is extracted. In this chapter experiment, convolution kernels of different sizes were used to extract features from hyperspectral images. To evaluate the impact of the convolution kernel size on the experimental results, experiments were conducted using convolution kernels of the same size instead of Pyconv. The experimental results are shown in Table 12. In the table, DPCMF_3 represents the use of a 3 × 3 × 1 convolution kernel in the spatial block and a 1 × 1 × 3 convolution kernel in the spectral block; DPCMF_5 represents the use of a 5 × 5 × 1 convolution kernel in the spatial block and a 1 × 1 × 5 convolution kernel in the spectral block; DPCMF_7 represents the use of a 7 × 7 × 1 convolution kernel in the spatial block and a 1 × 1 × 7 convolution kernel in the spectral block; and DPCMF_9 represents the use of a 9 × 9 × 1 convolution kernel in the spatial block and a 1 × 1 × 9 convolution kernel in the spectral block.

3.5. Impact of Training Sample Size

To verify the impact of the number of training samples on the experimental results, experiments were conducted using a varying number of samples as training samples. For the IP dataset, 0.5%, 1%, 3%, 5% and 10% of the samples were used as training sets, and the experimental results are shown in Table 13. For the UP and SV datasets, 0.1%, 0.5%, 1%, 3% and 5% of the samples were used as training sets, and the experimental results are shown in Table 14 and Table 15, respectively. For the BS dataset, 0.5%, 1.2%, 3%, 5% and 10% of the samples were used as training sets, and the experimental results are shown in Table 16.

3.6. Ablation Experiment

To verify the impact of the spatial block, spectral block, and non-local block on OA, experiments were conducted on these three modules using four datasets, as shown in Table 17. Table 17 displays the classification results of DPCMF, DPCMF-AE, DPCMF-AN, DPCM-EN and DPCMF-D on the four datasets. DPCMF-AE represents the absence of the non-local block, DPCMF-AN represents the absence of the spectral block, DPCMF-EN represents the absence of the spatial block, and DPCMF-D represents the absence of the Dense structure. The dataset partitioning process is consistent with that described in the previous section.

4. Discussion

The experimental results are shown in Figure 10a, DPCMF method achieves significant improvements in all of the following three metrics: OA, AA and Kappa. In terms of time, due to the large input volume of convolutional neural networks and the need for more training parameters, the time cost of the DPCMF method is higher than that of the SVM method, but in terms of classification accuracy, the accuracy level of the SVM method is much lower than those of other deep learning methods. In most cases, the DPCMF method takes less time than other deep learning-based methods.
For the IP dataset, the OA of the DPCMF method is 96.74%, which is 27.39%, 4.79%, 1.29%, 3.89% and 0.61% higher than the OA levels of the other five methods, its AA is 96.12%, which is 30.26%, 7.18%, 1.44%, 9.94% and 0.43% higher than the AA levels of the other five methods, and its Kappa coefficient is 96.32%, which is 31.67%, 5.51%, 1.5%, 4.47% and 0.89% higher than the Kappa of the other five methods. The classification accuracy for each category has reached more than 90%. Compared to the other ground objects, the classification accuracy is lower for the Grass-pasture-mowed because there are fewer training samples for the ground objects, and it is difficult to extract a large amount of feature-related information from a small number of samples during model training.
For the UP dataset, the OA of the DPCMF method is 98.10%, which is 15.03%, 9.78%, 1.18%, 2.82% and 0.98% higher than the OA levels of the other five methods, its AA is 98.33%, which is 16.09%, 5.19%, 1.55%, 2.64% and 0.99% higher than the AA levels of the other five methods, and its Kappa coefficient is 97.77%. The classification accuracy for all categories has reached more than 94%. Compared to the alternative ground-truth features, Self-blocking bricks have lower classification accuracy because the category features of the ground-truth features are not obvious, and it is difficult to extract features with elevated discriminative degree.
For the SV dataset, the OA of the DPCMF method is 98.92%, which is 10.83%, 4.5%, 2.4%, 1.97.05% and 1.61% higher than the OA levels of the other five methods, its AA is 98.76%, which is 7.31%, 2.14%, 0.95%, 1.79% and 1.42% higher than the AA levels of the other five methods, and its Kappa coefficient is 98.73%, which is 12.03%, 4.96%, 3.35%, 2.13% and 0.92% higher than the Kappa coefficient of the other five methods. The classification accuracy for each category has reached more than 91%.
For the BS dataset, the OA of the DCFE method is 96.67%, which is 18.04%, 6.41%, 3.50%, 1.48% and 0.28% higher than the OA levels of the other five methods, its AA is 97.08%, which is 17.51%, 6.26%, 2.63%, 1.36% and 0.58% higher than the AA levels of the other five methods, and its Kappa coefficient is 96.57%, which is 19.70%, 7.10%, 3.98%, 1.78% and 0.48% than the Kappa coefficient of the other five methods. Compared to the other ground features, Floodplain Meadows 2 has lower classification accuracy because there are fewer training samples for this ground feature, the features of this ground feature are more complex, and feature extraction is more difficult during the training of the model.
From the results of the Figure 10b and Table 12, as the size of the convolutional kernel increased, the OA gradually increased, indicating that larger kernels can better capture a wider range of features. However, when the size of the kernel increased to a certain extent, the OA started to decrease because excessively large kernels may capture noise or irrelevant information, thereby reducing the accuracy of the model. In this group of experiments, the best OA can be achieved through pyramidal convolutions. The reason is that the pyramidal convolution module uses multi-scale kernels to capture features of different scales, thereby improving the model generalization ability. Choosing an appropriate combination of convolutional kernels can also improve the accuracy of the model. For example, for certain tasks, the use of smaller kernels may be more suitable because they can better capture local features. Therefore, when choosing the size of the convolutional kernel, it is necessary to ensure a good balance according to the specific task requirements and data characteristics in order to obtain the best experimental results.
In the experiments on the impact of training sample size, the classification accuracy of the SVM, SSRN, FDSSC, DBMA, DBDA and DPCMF methods all improved as the number of training samples increased. Moreover, the performance gap between different models also narrowed with the increase in the number of training samples. The results of these experiments show that, in the case of limited number of training samples, the DPCMF method can better extract multi-class features in the samples by using densely connected pyramidal convolution layers to capture spectral features and multi-scale spatial features and using non-local modules to capture global spatial information. Therefore, it achieved good classification results.
From the results of Figure 10c and Table 17, it can be seen that the absence of any module will reduce the model accuracy. DPCMF-AE performed poorly because of inadequate perception of global spatial features. When processing images, it is not only necessary to understand the characteristics of each pixel in the image, but it is also necessary to understand the global information such as the structure, background, layout and composition of the image. Such global information can help the model better understand the image and improve its performance in image processing. DPCMF-AN performed poorly because spectral images contain multiple continuous spectral bands, and each spectral band corresponds to different spectral information of different wavelengths. Therefore, they have high dimensionality and rich information that can be used to accurately describe the spectral characteristics of objects. The inability to extract spectral features results in the lack of important spectral information, making the model unable to distinguish and classify different objects. DPCMF-EN performed poorly because considering only the spectral information of pixels is often insufficient to provide sufficient information. For example, when classifying vegetation and non-vegetation, the spectral information of vegetation may vary at different positions, making it difficult to distinguish vegetation from non-vegetation using only spectral information. In this case, it is necessary to consider the spatial information of pixels, the positional relationship between pixels in the image, to improve classification accuracy. DPCMF-D performed poorly because the Dense structure plays an important role in feature extraction. The Dense structure can share features between different layers through feature reuse, which can effectively improve the network expressive ability and enable the network to better learn the complex features of input data. This mechanism can avoid the problem of vanishing gradients that is often encountered in traditional deep networks, thus improving the network feature extraction ability. The parameter sharing between different layers in the Dense structure can greatly reduce the number of parameters that need to be trained in the network, thereby reducing the network complexity. This makes the network more compact and lightweight, helping to avoid overfitting and improve the network generalization ability. Since each layer in the Dense structure can accept inputs from all previous layers, the network can learn the features of input data more quickly. In addition, the Dense structure can use a shallower network structure to achieve the same performance as traditional networks, which can reduce training time and computational cost. Therefore, the Dense structure is essential in the DPCMF network.

5. Conclusions

In this paper, we propose a hyperspectral image classification method based on dense pyramidal convolution and multi-feature fusion to address the difficulty in adequately extracting and exploiting the spatial and spectral information of hyperspectral images when the sample size is limited. In this approach, two branches—i.e., spatial and spectral branches—are designed, and in each branch, dense pyramidal convolution layers are used as feature extractors. In the spatial branch, multiple local and global spatial features in image samples are extracted using dense pyramidal convolution and non-local blocks. In the spectral branch, the spectral features in the image samples are extracted by the dense pyramidal convolution module. Finally, the spatial and spectral features are fused through fully connected layers to obtain classification results.
The results of experiments conducted to compare the proposed method with the SVM, SSRN, FDSSC, DBMA and DBDA methods on four public hyperspectral datasets (Indian Pines, Pavia University, Salinas Valley, and Botswana) show that the DPCMF method achieves the best experimental results in terms of OA, AA and Kappa coefficients. In the follow-up study, we will continue to build more efficient classification models to resolve the problem of limited sample size and further improve the current model classification accuracy for hyperspectral images.

Author Contributions

Conceptualization, J.Z., H.J., S.S., J.W., P.Z., W.Z. and L.W.; Writing—original draft, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the Shandong Provincial Natural Science Foundation, China under Grant ZR2020MF006 and ZR2022LZH015, partially supported by the Industry-university Research Innovation Foundation of Ministry of Education of China under Grant 2021FNA01001, partially supported by the Major Scientific and Technological Projects of CNPC under Grant ZD2019-183-006, partially supported by the Open Foundation of State Key Laboratory of Integrated Services Networks (Xidian University) under Grant ISN23-09, partially supported by Zhejiang Provincial Natural Science Foundation of China under Grant LZ22F020002.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chang, C.I. (Ed.) Hyperspectral Data Exploitation: Theory and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  2. Nolin, A.W.; Dozier, J. A hyperspectral method for remotely sensing the grain size of snow. Remote Sens. Environ. 2000, 74, 207–216. [Google Scholar] [CrossRef]
  3. Awad, M.; Jomaa, I.; Arab, F. Improved Capability in Stone Pine Forest Mapping and Management in Lebanon Using Hyperspectral CHRIS-Proba Data Relative to Landsat ETM+. Photogramm. Eng. Remote Sens. 2014, 80, 725–731. [Google Scholar] [CrossRef]
  4. Marinelli, D.; Bovolo, F.; Bruzzone, L. A novel change detection method for multitemporal hyperspectral images based on binary hyperspectral change vectors. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4913–4928. [Google Scholar] [CrossRef]
  5. Pal, M. Ensemble of support vector machines for land cover classification. Int. J. Remote Sens. 2008, 29, 3043–3049. [Google Scholar] [CrossRef]
  6. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised hyperspectral image segmentation using multinomial logistic regression with active learning. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4085–4098. [Google Scholar] [CrossRef] [Green Version]
  7. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–Spatial Hyperspectral Image Segmentation Using Subspace Multinomial Logistic Regression and Markov Random Fields. IEEE Trans. Geosci. Remote Sens. 2012, 50, 809–823. [Google Scholar] [CrossRef]
  8. Xue, G.; Chang, Q.; Wang, J.; Zhang, K.; Pal, N.R. An adaptive neuro-fuzzy system with integrated feature selection and rule extraction for high-dimensional classification problems. IEEE Trans. Fuzzy Syst. 2022. [Google Scholar] [CrossRef]
  9. Gong, X.; Yu, L.; Wang, J.; Zhang, K.; Bai, X.; Pal, N.R. Unsupervised feature selection via adaptive autoencoder with redundancy control. Neural Netw. 2022, 150, 87–101. [Google Scholar] [CrossRef]
  10. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Classification for hyperspectral imagery based on sparse representation. In Proceedings of the 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, Reykjavik, Iceland, 14–16 June 2010; pp. 1–4. [Google Scholar]
  11. Khan, Z.; Shafait, F.; Mian, A. Joint group sparse PCA for compressed hyperspectral imaging. IEEE Trans. Image Process. 2015, 24, 4934–4942. [Google Scholar] [CrossRef]
  12. Li, J.; Marpu, P.R.; Plaza, A.; Bioucas-Dias, J.M.; Benediktsson, J.A. Generalized composite kernel framework for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4816–4829. [Google Scholar] [CrossRef]
  13. Li, W.; Du, Q. Gabor-filtering-based nearest regularized subspace for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1012–1022. [Google Scholar] [CrossRef]
  14. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral–spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  15. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite kernels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
  16. Zeng, D.; Liu, K.; Chen, Y.; Zhao, J. Distant supervision for relation extraction via piecewise convolutional neural networks. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1753–1762. [Google Scholar]
  17. Li, G.; Li, L.; Zhu, H.; Liu, X.; Jiao, L. Adaptive multiscale deep fusion residual network for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8506–8521. [Google Scholar] [CrossRef]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015, Proceedings, Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  19. Wang, R.J.; Li, X.; Ling, C.X. Pelee: A real-time object detection system on mobile devices. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  20. Zhou, P.; Han, J.; Cheng, G.; Zhang, B. Learning compact and discriminative stacked autoencoder for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4823–4833. [Google Scholar] [CrossRef]
  21. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  22. Ma, X.; Wang, H.; Jie, G. Spectral–Spatial Classification of Hyperspectral Image Based on Deep Auto-Encoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 4073–4085. [Google Scholar] [CrossRef]
  23. Zhang, X.; Liang, Y.; Li, C.; Huyan, N.; Jiao, L.; Zhou, H. Recursive autoencoders-based unsupervised feature learning for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1928–1932. [Google Scholar] [CrossRef] [Green Version]
  24. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  25. Zhang, M.; Li, W.; Du, Q. Diverse region-based CNN for hyperspectral image classification. IEEE Trans. Image Process. 2018, 27, 2623–2634. [Google Scholar] [CrossRef]
  26. Lee, H.; Kwon, H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [Green Version]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  29. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  30. Wang, W.; Dou, S.; Jiang, Z.; Sun, L. A Fast Dense Spectral–Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef] [Green Version]
  31. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.I. Feedback attention-based dense CNN for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5501916. [Google Scholar] [CrossRef]
  32. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-branch multi-attention mechanism network for hyperspectral image classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef] [Green Version]
  33. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  34. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef] [Green Version]
  35. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  36. Duta, I.C.; Liu, L.; Zhu, F.; Shao, L. Pyramidal convolution: Rethinking convolutional neural networks for visual recognition. arXiv 2020, arXiv:2006.11538. [Google Scholar]
  37. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  38. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Architecture of DenseNet.
Figure 1. Architecture of DenseNet.
Remotesensing 15 02990 g001
Figure 2. Architecture of the dense pyramidal convolution module.
Figure 2. Architecture of the dense pyramidal convolution module.
Remotesensing 15 02990 g002
Figure 3. Architecture of a 3D-convolutional neural network (3D-CNN).
Figure 3. Architecture of a 3D-convolutional neural network (3D-CNN).
Remotesensing 15 02990 g003
Figure 4. Architecture of the non-local block.
Figure 4. Architecture of the non-local block.
Remotesensing 15 02990 g004
Figure 5. Structure of the DPCMF network.
Figure 5. Structure of the DPCMF network.
Remotesensing 15 02990 g005
Figure 6. Classification maps for the IP dataset with 3% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Figure 6. Classification maps for the IP dataset with 3% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Remotesensing 15 02990 g006
Figure 7. Classification maps for the UP dataset with 0.5% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Figure 7. Classification maps for the UP dataset with 0.5% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Remotesensing 15 02990 g007
Figure 8. Classification maps for the SV dataset with 0.5% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Figure 8. Classification maps for the SV dataset with 0.5% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Remotesensing 15 02990 g008
Figure 9. Classification maps for the BS dataset with 1.2% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Figure 9. Classification maps for the BS dataset with 1.2% training samples. (a) Ground-truth (GT); (bg) classification maps from disparate algorithms.
Remotesensing 15 02990 g009
Figure 10. Experiment of DPCMF network on different datasets.
Figure 10. Experiment of DPCMF network on different datasets.
Remotesensing 15 02990 g010
Table 1. Implementation details of the spatial branch.
Table 1. Implementation details of the spatial branch.
Layer NameKernel SizeOutput Size
PCA(145 × 145 × 100)
Conv(3 × 3 × 1)(9 × 9 × 1, 24)
Spatial Block-1(7 × 7 × 1)(9 × 9 × 1, 12)
Spatial Block-2(5 × 5 × 1)(9 × 9 × 1, 12)
Spatial Block-3(3 × 3 × 1)(9 × 9 × 1, 12)
Non-local block(9 × 9 × 100)
Conv(3 × 3 × 1)(9 × 9 × 1, 60)
Concatenate(9 × 9 × 1, 120)
Table 2. Implementation details of the spectral branch.
Table 2. Implementation details of the spectral branch.
Layer NameKernel SizeOutput Size
Conv(1 × 1 × 7)(9 × 9 × 97)
Spectral Block-1(1 × 1 × 7)(9 × 9 × 97, 12)
Spectral Block-2(1 × 1 × 5)(9 × 9 × 97, 12)
Spectral Block-3(1 × 1 × 3)(9 × 9 × 97, 12)
Conv(1 × 1 × 97)(9 × 9 × 1, 60)
Table 3. Implementation details of the classification module.
Table 3. Implementation details of the classification module.
Layer NameKernel SizeOutput Size
Concatenation layer(9 × 9 × 1, 180)
GAP layer(1 × 180)
Fully Connected layer(1 × 16)
Table 4. Samples of the IP dataset for training, validation and testing.
Table 4. Samples of the IP dataset for training, validation and testing.
OrderClassTotalTrainVal.Test
1Alfalfa463340
2Corn-notill142842421344
3Corn-mintill8302424782
4Corn23777223
5Grass-pasture4831414455
6Grass-trees7302121688
7Grass-pasture-mowed283322
8Hay-windrowed4781414450
9Oats203314
10Soybean-notill9722929914
11Soybean-mintill245573732309
12Soybean-clean5931717559
13Wheat20566193
14Woods126537371191
15Buildings-Grass-Tree-Drives3861111364
16Stone-Steel-Towers933387
Total102493073079635
Table 5. Samples of the UP dataset for training, validation and testing.
Table 5. Samples of the UP dataset for training, validation and testing.
OrderClassTotalTrainVal.Test
1Asphalt663133336465
2Meadows18,649939318,463
3Gravel209910102079
4Corn306415153034
5Trees1345661333
6Bare Soil502925254979
7Bitumen1330661318
8Self-Blocking Bricks368218183646
9Shadows94744939
Total42,77621021042,356
Table 6. Samples of the SV dataset for training, validation and testing.
Table 6. Samples of the SV dataset for training, validation and testing.
OrderClassTotal TrainVal.Test
1Brocoli-green-weeds-1200910101989
2Brocoli-green-weeds-2372618183690
3Fallow1976991958
4Fallow-rough-plow1394661382
5Fallow-smooth267813132652
6Stubble395919193921
7Celery357917173545
8Grapes-untrained11,271565611,159
9Soil-vineyard-develop620331316141
10Corn-senesced-green-weeds327816163246
11Lettuce-romaine-4wk1068551058
12Lettuce-romaine-5wk1927991909
13Lettuce-romaine-6wk91644908
14Lettuce-romaine-7wk1070551060
15Vineyard-untrained726836367196
16Vineyard-vertical-trellis1807991789
Total54,12926326353,603
Table 7. Samples of the BS dataset for training, validation and testing.
Table 7. Samples of the BS dataset for training, validation and testing.
OrderClassTotalTrainVal.Test
1Water27033264
2Hippo grass1012297
3Floodplain grasses125133245
4Floodplain grasses221533209
5Reeds126933263
6Riparian26933263
7Fierscar225933253
8Island interior20333197
9Acacia woodlands31444306
10Acacia shrublands24833242
11Acacia grasslands30544297
12Short mopane18122177
13Mixed mopane26933263
14Exposed soils951193
Total324840403168
Table 8. Categorized results for the IP dataset with 3% training samples.
Table 8. Categorized results for the IP dataset with 3% training samples.
ClassSVMSSRNFDSSCDBMADBDADPCMF
1/%24.1960.0097.6783.3397.72100
2/%56.7191.4799.1292.2796.4397.62
3/%65.0993.5195.8592.3797.8195.83
4/%39.6388.9510010097.56100
5/%87.3310098.3598.2498.3092.86
6/%83.8795.9588.4498.4096.7595.24
7/%57.5086.2082.7539.5988.0090.26
8/%89.2894.5010099.10100100
9/%22.5869.5693.3326.31100100
10/%66.7084.3587.3183.9891.1293.10
11/%62.5091.8699.0995.6598.6397.26
12/%51.8686.7489.0185.0593.5594.12
13/%94.7998.9798.9210097.97100
14/%90.4294.7496.2493.7394.4097.30
15/%62.8295.0994.4194.3792.9890.91
16/%98.4691.1194.3896.5195.4593.44
OA/%69.3591.9595.4592.8596.1396.74
AA/%65.8688.9494.6886.1895.6996.12
Kappa/%64.6590.8194.8291.8595.4396.32
Training Time/s11.8552.30128.74113.2577.2572.41
Test Time/s1.623.126.328.967.027.25
Table 9. Categorized results for the UP dataset with 0.5% training samples.
Table 9. Categorized results for the UP dataset with 0.5% training samples.
ClassSVMSSRNFDSSCDBMADBDADPCMF
1/%80.2668.4599.5593.4995.2796.97
2/%86.9492.9598.5796.9896.8198.92
3/%71.7399.5210096.7299.26100
4/%96.4499.0898.0796.9398.2096.67
5/%90.8597.7799.7099.6298.61100
6/%77.0298.1599.8199.5298.8398.00
7/%69.7010094.7497.11100100
8/%67.3083.1280.7082.2589.0994.44
9/%99.8999.2599.8898.59100100
OA/%83.0788.3296.9295.2897.1298.10
AA/%82.2493.1496.7895.6997.3498.33
Kappa/%77.0784.1495.9193.7096.2397.77
Training Time/s5.6911.8831.3810.6521.2121.08
Test Time/s2.095.2113.5613.1211.0512.98
Table 10. Categorized results for the SV dataset with 0.5% training samples.
Table 10. Categorized results for the SV dataset with 0.5% training samples.
ClassSVMSSRNFDSSCDBMADBDADPCMF
1/%99.8499.34100100100100
2/%98.9510097.2010099.62100
3/%89.8787.7199.2898.38100100
4/%97.3095.2296.2392.0289.3591.67
5/%93.5599.5499.9699.1199.63100
6/%99.7999.7999.9299.8499.26100
7/%91.3399.5210098.7496.69100
8/%74.7384.5199.8093.1597.9598.21
9/%97.6999.6799.0999.2399.07100
10/%90.0199.2899.1797.4293.76100
11/%75.9294.0494.3780.7192.8996.00
12/%95.1996.9699.8499.44100100
13/%94.8610010099.44100100
14/%89.2698.2299.0696.3395.8197.00
15/%75.8592.1083.8197.1893.4897.22
16/%99.03100100100100100
OA/%88.0994.4296.5296.9597.3198.92
AA/%91.4596.6297.8196.9797.3498.76
Kappa/%86.7093.7795.3896.6097.8198.73
Training Time/s10.2784.99123.28149.1281.2078.22
Test Time/s4.1316.3231.0541.3323.6023.66
Table 11. Categorized results for the BS dataset with 1.2% training samples.
Table 11. Categorized results for the BS dataset with 1.2% training samples.
ClassSVMSSRNFDSSCDBMADBDADPCMF
1/%10098.4794.0297.7695.9797.76
2/%70.7094.6210098.9898.00100
3/%84.1087.89100100100100
4/%65.9586.8096.8989.4085.7788.28
5/%82.6274.5087.5092.2793.9694.36
6/%65.7180.1969.7680.1387.0488.65
7/%78.7790.3510096.9310099.21
8/%65.8787.1195.6010099.32100
9/%75.1893.7610094.4291.04100
10/%69.8281.5691.0492.7710089.7
11/%95.49100100100100100
12/%93.1010088.88100100100
13/%76.2596.25100100100100
14/%90.4110010097.43100100
OA/%78.6390.2693.1795.1996.3996.67
AA/%79.5790.8294.4595.7296.5097.08
Kappa/%76.8789.4792.5994.7996.0996.57
Training Time/s1.6211.2323.0221.0917.9620.39
Test Time/s0.382.012.653.112.102.32
Table 12. Impact of kernel size on OA.
Table 12. Impact of kernel size on OA.
Kernel SizeIPUPSVBS
DPCMF_3/%89.1393.1592.1290.61
DPCMF-5/%93.2896.2295.4294.22
DPCMF-7/%95.1297.0297.3894.67
DPCMF-9/%94.3796.5896.8895.13
Table 13. OA for different proportions of training samples in the IP dataset.
Table 13. OA for different proportions of training samples in the IP dataset.
Algorithms0.5%1%3%5%10%
SVM/%48.5355.9569.3574.7480.55
SSRN/%64.9981.4090.520.95597.84
FDSSC/%70.7584.7196.1497.2198.02
DBMA/%59.3377.6493.1493.7596.91
DBDA/%56.9778.8196.1996.5897.55
DPCMF/%73.2585.1496.7497.9598.56
Table 14. OA for different proportions of training samples in the UP dataset.
Table 14. OA for different proportions of training samples in the UP dataset.
Algorithms0.1%0.5%1%3%5%
SVM/%70.5983.0788.4590.3593.29
SSRN/%78.3294.8597.1199.4399.69
FDSSC/%88.9797.0297.7499.5099.58
DBMA/%89.8795.0696.3799.1099.49
DBDA/%88.0197.1198.4099.0799.33
DPCMF/%91.3598.1098.8999.9999.99
Table 15. OA for different proportions of training samples in the SV dataset.
Table 15. OA for different proportions of training samples in the SV dataset.
Algorithms0.1%0.5%1%3%5%
SVM/%78.6588.0989.8991.2492.47
SSRN/%67.2295.3596.3297.2398.14
FDSSC/%88.8395.8596.4897.5298.85
DBMA/%92.1595.9096.6697.6298.21
DBDA/%94.2397.7098.3198.9599.36
DPCMF/%96.0498.9299.2599.9099.99
Table 16. OA for different proportions of training samples in the BS dataset.
Table 16. OA for different proportions of training samples in the BS dataset.
Algorithms0.5%1.2%3%5%10%
SVM/%73.5378.6387.8289.0692.76
SSRN/%84.0794.2795.5298.1999.15
FDSSC/%87.9890.8096.3397.2499.46
DBMA/%93.3694.8795.8898.0199.04
DBDA/%96.2796.3997.3898.6499.33
DPCMF/%96.3796.6799.1099.5899.91
Table 17. Ablation experiments on four datasets.
Table 17. Ablation experiments on four datasets.
AlgorithmsIPUPSVBS
DPCMF/%96.7498.1098.9296.67
DPCMF-AE/%94.2595.1695.4292.19
DPCMF-AN/%90.3890.3091.3789.24
DPCMF-EN/%82.1382.9785.2181.98
DPCMF-D/%95.1194.2693.2890.56
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Zhao, L.; Jiang, H.; Shen, S.; Wang, J.; Zhang, P.; Zhang, W.; Wang, L. Hyperspectral Image Classification Based on Dense Pyramidal Convolution and Multi-Feature Fusion. Remote Sens. 2023, 15, 2990. https://doi.org/10.3390/rs15122990

AMA Style

Zhang J, Zhao L, Jiang H, Shen S, Wang J, Zhang P, Zhang W, Wang L. Hyperspectral Image Classification Based on Dense Pyramidal Convolution and Multi-Feature Fusion. Remote Sensing. 2023; 15(12):2990. https://doi.org/10.3390/rs15122990

Chicago/Turabian Style

Zhang, Junsan, Li Zhao, Hongzhao Jiang, Shigen Shen, Jian Wang, Peiying Zhang, Wei Zhang, and Leiquan Wang. 2023. "Hyperspectral Image Classification Based on Dense Pyramidal Convolution and Multi-Feature Fusion" Remote Sensing 15, no. 12: 2990. https://doi.org/10.3390/rs15122990

APA Style

Zhang, J., Zhao, L., Jiang, H., Shen, S., Wang, J., Zhang, P., Zhang, W., & Wang, L. (2023). Hyperspectral Image Classification Based on Dense Pyramidal Convolution and Multi-Feature Fusion. Remote Sensing, 15(12), 2990. https://doi.org/10.3390/rs15122990

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop