Next Article in Journal
Industrial Fault Detection Based on Discriminant Enhanced Stacking Auto-Encoder Model
Previous Article in Journal
Semi-Supervised Group Emotion Recognition Based on Contrastive Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification

School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(23), 3992; https://doi.org/10.3390/electronics11233992
Submission received: 9 November 2022 / Revised: 29 November 2022 / Accepted: 29 November 2022 / Published: 1 December 2022

Abstract

:
Deep-learning-based methods have been widely used in hyperspectral image classification. In order to solve the problems of the excessive parameters and computational cost of 3D convolution, and loss of detailed information due to the excessive increase in the receptive field in pursuit of multi-scale features, this paper proposes a lightweight hybrid convolutional network called the 3D lightweight receptive control network (LRCNet). The proposed network consists of a 3D depthwise separable convolutional network and a receptive field control network. The 3D depthwise separable convolutional network uses the depthwise separable technique to capture the joint features of spatial and spectral dimensions while reducing the number of computational parameters. The receptive field control network ensures the extraction of hyperspectral image (HSI) details by controlling the convolution kernel. In order to verify the validity of the proposed method, we test the classification accuracy of the LRCNet based on three public datasets, which exceeds 99.50% The results show that compare with state-of-the-art methods, the proposed network has competitive classification performance.

1. Introduction

Hyperspectral images (HSIs) contain a large amount of spectral and spatial data, which provide abundant information based on the spectral characteristics of the objects and retain the overall shape of an object and its association with the surrounding objects [1]. Considering the characteristics of HSI data, it is important to analyze and extract the spectral and spatial features. HSI processing technology possesses the capability to satisfy military and civilian needs, such as medical image processing, agriculture and geological exploration, and sea resource investigation [2,3,4,5,6,7,8]. Consequently, hyperspectral image classification (HSIC) has become a research hotspot in image processing and remote sensing [9].
In the early works of HSIC, convolutional neural networks (CNNs) were usually used to extract the features [10,11,12,13,14]. Cheng et al. proposed a simple, effective method in order to extract hierarchical deep spatial features for HSI classification by exploring the power of off-the-shelf CNN models [15]. Makantasis et al. exploited a convolutional neural network to encode pixels’ spectral and spatial information, and a multi-layer perceptron to conduct the classification task [16]. Many deep neural networks have been developed to handle HSIC tasks. Jiao et al. applied fully convolutional networks (FCNs) to the HSIC task for the first time by combining the weighted extracted features and spectral information [17]. Sun et al. introduced a supervised network for better performance and proposed a fully convolutional segmentation network (FCSN) [18]. Kang et al. believed that the CNN-based methods are unable to effectively extract the discriminant features and proposed a dual-path network (DPN) by combining a residual network and dense convolutional network to perform HSIC [19]. In order to obtain additional neighborhood information, Soucy et al. proposed the clustering ensemble U-Net (CEU-Net) by combining clustering methods with U-Net [20]. Si et al. used DeepLab v3+ technology and a support vector machine (SVM) classifier for HSI feature extraction and classification recognition [21].
The aforementioned methods perform 2D convolution-based operations. However, some researchers [22] believe that 2D CNNs cannot effectively extract the features because they do not consider the correlation information between channels [22]. On the contrary, 3D convolution effectively combines the spatial and spectral features to improve the accuracy. Based on a 3D CNN for feature extraction [13], Hamida et al. presented an efficient method that enables a joint spectral and spatial information process [23]. He et al. proposed a multi-scale 3D deep convolutional neural network (M3D-DCNN) in order to jointly learn both 2D multi-scales and 1D spectral features from HSI data [24]. Zhong et al. introduced additional residual blocks by using 3D convolutional layers and proposed the spectral–spatial residual network (SSRN) [25]. Roy et al. proposed HybridSN by combining 2D and 3D convolutions to obtain a higher classification accuracy [22]. Zhu et al. proposed the 3D deep capsule network based on the abundant feature representation capability [26]. Sun et al. proposed the cubic capsule network (EMAP-Cubic-Caps) in order to overcome the shortcomings, including the inability to capture fine spatial features, and loss of important information of PCA dimensionality reduction [27].
Despite the fact that 3D convolution effectively obtains the joint features from spatial and spectral dimensions, it has its own limitations [22,28,29]. A network that incorporates 3D convolutions has a large number of parameters, leading to a higher computational cost [22]. In addition, most networks using 3D convolution do not consider the impact of controlling the receptive field size on the classification accuracy, and they emphasize expanding the receptive fields to obtain a better performance [30,31]. Actually, due to the low spatial resolution of HSIs, there is a considerable loss of detail in large receptive fields [9,32]. Although multiple pooling operations are useful for acquiring multi-scale features, they also have an adverse effect on the classification accuracy due to the loss of detailed features, creating confusion among similar category features [31]. However, if the receptive field is too small, it is not possible to consider the multi-scale features. This may lead to underfitting in the network. Therefore, a suitable receptive field is essential for enhancing the network’s classification accuracy.
Recently, researchers have tried to address the above problems. Considering the low spatial resolution of HSIs, Pan et al. proposed the dilated semantic segmentation network (DSSNet) [31]. The authors presented a concept for controlling the receptive field of convolution kernels at 13   ×   13 [31]. However, it is difficult to focus on the joint features of the space and spectrum because DSSNet still extracts features using 2D convolutions. Li et al. argued that existing networks do not effectively combine 2D and 3D convolutions, so they alternately used 2D and 3D units to solve the redundancy of the model structure [33]. Although the method proposed by them can reduce the size of the model, it does not specifically reduce the consumption of 3D convolution. Howard et al. proposed depthwise separable convolution in order to reduce the number of parameters and computations in the 2D convolution process. The authors used the proposed convolution in the lightweight network Mobilenetv1 [34]. Fırat et al. introduced 2D depthwise separable convolution in HSIC tasks to decrease the computational cost [35]. Sandler et al. upgraded the depthwise separable convolution and proposed the inverted residual structure [36]. These methods lessen the number of parameters and computational costs that convolutions introduce. Additionally, these methods are aimed at 2D convolution instead of 3D convolution. We also found that in other research fields, researchers have modified and introduced 3D depthwise separable convolution to reduce the computational cost [37,38,39]. In order to effectively obtain the multi-scale information from HSIs, Gong et al. proposed the multi-scale squeeze-and-excitation pyramid pooling network (MSPN) [28]. The classification accuracy of this network is affected due to the introduction of a pooling layer without controlling the size of the receptive field. Although there are solutions available for addressing the aforementioned issues, it is difficult to resolve the defects.
From the previous work, two problems can be summarized as follows. First, although 3D convolutions effectively capture the spectral and spatial features, the number of parameters and computations introduced during the training process is large [22]. Second, the spatial resolution of HSIs is usually low and some details are presented only based on a few pixels [31,32]. It is noteworthy that the details may disappear after multiple pooling operations, and the lost details cannot be retrieved by up-sampling [40]. If the network is too deep, it may lose some details during the feature extraction process due to the large receptive field of the convolution kernel. As a result, the classification accuracy is affected.
For the purpose of solving the above problems, the 3D lightweight receptive control network (LRCNet) is proposed in this paper. We combine 2D and 3D convolution to effectively integrate the features from the spatial and spectral dimensions. Next, in order to lower the computational cost and reduce the number of parameters, we employ depthwise separable convolution and convert it from 2D to 3D format. In order to reduce the negative impact of a low spatial resolution, we control the size of the receptive field based on dilated convolutions. Below is a summary of this work’s contributions:
  • The application of 3D depthwise separable convolution decreases the computational costs of 3D convolution. Additionally, 3D depthwise convolution can effectively capture spatial and spectral features, while pointwise convolution can extract information from adjacent spectral bands, improving the learning ability of the spectral domain.
  • The receptive field control strategy is adopted. To prevent the loss of detailed information when learning multi-scale features, the receptive filed is gradually increased through dilated convolution. Moreover, the receptive field is left unchanged during 3D convolution to enhance the robustness of the model and lower the risk of overfitting.
  • The experimental results show that the proposed method has a better classification accuracy in three public datasets, indicating that the model is competitive.
The rest of this paper is organized as follows: The LRCNet architecture and the functional block are presented in Section 2. The experimental results and analysis are discussed in Section 3, and the conclusion is presented in Section 4.

2. Methods

The proposed LRCNet’s architecture is depicted in Figure 1. For the input HSI, we use principal component analysis (PCA) to reduce the dimensions of the data. Next, a 3D depthwise separable convolutional network comprising three 3D-DW modules is used. Afterwards, a reshape operation is applied, and the resulting data are used as the input of the receptive field control network. This network is followed by a fully connected (FC) module, which consists of three FC layers. Finally, the classification results are obtained.

2.1. Initial Data Input and Processing

As an HSI contains mixed land categories, there is a similarity between different categories. Additionally, a significant percentage of spectral bands exhibit redundancy, which makes it difficult to train models [16]. As shown in Figure 1, in order to reduce the impact of redundant HSI data during the training process, we use PCA before further processing [41]. Assume that the initial hyperspectral image is represented by I H × W × D , where I represents the input HSI data, H and W represent the height and width of the input data, respectively, and D represents the number of bands in the input images. The data cube after dimension reduction based on PCA is X H × W × B , where X represents the data cube and B represents the number of bands after dimension reduction. Next, we divide X into equal sizes and obtain P S × S × B , where P represents the data cube after partition, B is the number of bands, and S × S represents the height and width of P .

2.2. 3D Depthwise Separable Convolutional Network

Depthwise separable convolution was first proposed by Howard et al. and used in Mobilenetv1 [34]. The standard convolution is split into two parts through the depthwise separable convolution. The first part is the depthwise convolution, which is utilized to extract the features from each input channel separately. The second part is the pointwise convolution, which uses 1   ×   1 convolution to combine the output of the depthwise convolution.
Compared with the standard convolution, the depthwise separable convolution significantly reduces the number of parameters and the computational complexity of the convolution layer. We assume that the size of the input feature map is H × W × C i n and the parameters of a standard convolution layer are K 2 D × K 2 D × C i n × C o u t , where H and W represent the height and width of the input data, respectively, C i n denotes the number of channels in the input feature map, K 2 D represents the size of the convolution kernel for performing 2D convolutions, and C o u t represents the number of output channels. If the feature map size is still H × W , we set C o s t S as the computational complexity of the standard 2D convolution. Next, C o s t S is calculated as follows [34]:
C o s t S = K 2 D · K 2 D · C i n · C o u t · H · W
If 2D depthwise separable convolution is adopted, we assume its computational cost is C o s t D W . C o s t D W consists of two parts. The first part denotes the computational cost of the 2D depthwise convolution, and the second part denotes the computational cost of the 2D pointwise convolution. The costs are represented by C o s t D and C o s t P , respectively. In order to compare the computational costs of the 2D depthwise separable convolution and standard 2D convolution, we assume that the size of the convolution kernel is K 2 D , the numbers of input and output channels are C i n and C o u t , respectively, and the height and width of the input data are H and W , respectively. Next, C o s t D W is calculated as follows:
C o s t D W = C o s t D + C o s t P = K 2 D · K 2 D · C i n · H · W + C i n · C o u t · H · W
By comparing the computational costs of the two convolutions, the ratio of the computation is obtained as follows:
C o s t D W C o s t S = K 2 D · K 2 D · C i n · H · W + C i n · C o u t · H · W K 2 D · K 2 D · C i n · C o u t · H · W = 1 C o u t + 1 K 2 D 2
For convenience, we define the computational cost factor ξ 2 D as the ratio of the computational cost of the current 2D convolution to that of the standard 2D convolution, as shown in Equation (4):
ξ 2 D = C o s t D W C o s t S
Generally, the values of C o u t and K 2 D are greater than 2; thus, ξ 2 D < 1 can be obtained from Equations (3) and (4), which shows that 2D depthwise separable convolution can effectively decrease the computational costs. If a convolution kernel of size 3   ×   3 is used, the computational cost of 2D depthwise separable convolution can be reduced by about 9 times as compared with the standard 2D convolution. Therefore, a lightweight network can be created using depthwise separable convolution, which can also increase the network’s training effectiveness.
In the 2D depthwise convolution part, the features are extracted separately from each input channel. If 2D depthwise convolution is adopted, the connection between different bands of the same pixel is ignored, and the spectral features cannot be learned completely. Moreover, it is easy to ignore the relationship between spatial and spectral features in channel-by-channel convolutions. Although pointwise convolution addresses this defect, there are still many features that cannot be obtained.
Considering the limitations of 2D depthwise separable convolution, we propose the 3D depthwise separable convolution technique, which can fully extract the spatial–spectral features and learn joint features from multiple bands to enhance the classification performance. As each 3D convolution convolves a data block, it is possible to capture the features of adjacent groups of bands. Figure 2 depicts the structure of the proposed 3D depthwise separable convolution (3D-DW) module. The proposed technique also splits the standard 3D convolution into halves, including 3D depthwise convolution and 3D pointwise convolution.
In addition, the proposed 3D depthwise separable convolution retains the advantages of 2D depthwise separable convolutions. Note that the computational complexity of 3D depthwise separable convolution is lower as compared to the standard 3D convolution.
Assume that the size of the input data cube is C i n × H × W × B , where C i n is the number of input channels, B is the number of bands, and H and W are the height and width of the data cube, respectively. The number of parameters in a standard 3D convolution is K 3 D × K 3 D × K 3 D × C i n × C o u t , where K 3 D is the size of the 3D convolution kernel and C o u t is the number of output channels. If the space size of the output data cube remains unchanged, we consider C o s t 3 D - S as the computational cost of the standard 3D convolution. C o s t 3 D - S is computed as follows:
C o s t 3 D - S = K 3 D · K 3 D · K 3 D · C i n · C o u t · B · H · W
If 3D depthwise separable convolution is adopted, we assume its computational cost is C o s t 3 D D W . C o s t 3 D D W consists of two parts, i.e., computational cost of 3D depthwise convolution, and computational cost of 3D pointwise convolution, which are denoted as C o s t 3 D D and C o s t 3 D P , respectively. To compare the computational costs of 3D depthwise separable convolution with those of the standard 3D convolution, we assume that the size of the convolution kernel is K 3 D , the numbers of input channels and output channels are C i n and C o u t , respectively, and the height and width of the input data are H and W , respectively. Next, C o s t 3 D D W is calculated as follows:
C o s t 3 D D W = C o s t 3 D D + C o s t 3 D P = K 3 D · K 3 D · K 3 D · C i n · H · W · B + C i n · C o u t · H · W · B
To compare the computational costs of the convolutions, we define the computational cost factor ξ 3 D as follows:
ξ 3 D = C o s t 3 D - S C o s t 3 D - D W = K 3 D · K 3 D · K 3 D · C i n · H · W · B + C i n · C o u t · B K 3 D · K 3 D · K 3 D · C i n · C o u t · B · H · W = 1 C o u t + 1 K 3 D 2
Since C o u t 2 and K 3 D 2 , ξ 3 D < 1 is obtained using Equation (11). Therefore, it is evident that 3D depthwise separable convolution greatly reduces the computational cost.
Figure 3 shows the difference between the filters of the 3D depthwise separable convolution and the filters of the standard 3D convolution. Since each input layer channel is convolved separately in depthwise convolution, it is difficult to efficiently utilize the feature information from many channels in the same spatial position. The convolution kernels of the 3D depthwise convolution have three dimensions, so each convolution kernel extracts features from a group of adjacent bands, effectively avoiding the defects of depthwise convolution. Additionally, the number of channels is adjusted, and features are captured again using 3D pointwise convolution. Note that the size of the convolution kernel is only 1   ×   1   ×   1 . Therefore, as compared with the standard convolution, the 3D depthwise separable convolution has significantly fewer parameters and a lower computational cost.
The 3D depthwise separable convolutional network contains three 3D-DW modules. After each depthwise convolution and pointwise convolution, batch normalization (BN) is applied, along with the ReLU activation function. The parameters of the three 3D-DW modules are different. Since all the bands corresponding to each pixel in the HSI image collectively reflect the features of a pixel, it is necessary to aggregate the information from multiple bands as much as possible when extracting the features. Therefore, we set the size of the convolution kernels for the 3D depthwise convolution to ( 7 , 3 , 3 ) , ( 5 , 3 , 3 ) , and ( 3 , 3 , 3 ) .
In addition, the stride and padding parameters of the depthwise convolution and pointwise convolution are set to 1 and 0, respectively. As a result, the number of channels can be increased without changing the height and width of the input images. Due to the low spatial resolution of HSIs, it is easy to lose small features if the data size is compressed too early. This operation ensures that the receptive field of the convolution kernels does not increase during the 3D convolution and that spectral and spatial dimension information can be aggregated.
The essence of 3D depthwise convolution is still 3D convolution. For pixels with spatial position ( x , y , z ) in the j th feature map of the i th layer, we assume that the activation value v i , j x , y , z is expressed as follows [22]:
v i , j x , y , z = f ( b i , j + τ = 1 d l 1 λ = η η ρ = γ γ σ = δ δ ω i , j , τ σ , ρ , λ × v i 1 , j x + σ , y + ρ , z + λ )
where f represents the ReLU activation function, b i , j represents the bias parameter for the j th feature map of the i th layer, d l 1 denotes the number of feature maps in the ( l 1 ) th layer and the depth of kernel ω i , j for the j th feature map of the i th layer, 2 γ + 1 is the width of the convolution kernel, 2 δ + 1 is the height of the convolution kernel, and 2 η + 1 is the depth of the convolution kernel along the spectral dimension.

2.3. Receptive Field Control Network

This network includes a standard 2D convolution layer and two dilated convolution layers. The BN operation and ReLU activation function are also applied after each convolution operation. Since the data format output by the 3D convolution includes four dimensions, we multiply the number of bands and channels to reshape the data into three dimensions. However, this operation results in too many channels of data input. To avoid the impact of data redundancy on the training results, we compress the number of channels using a standard 2D convolution layer. Two dilated convolution layers are added to increase the receptive field for obtaining the multi-scale features. The dilation convolution can also obtain the features between neighbors, which can help to improve the classification accuracy. The stride parameter of the two dilated convolutions is 1, the padding parameter is 0, and the dilation rate is 2. The lower side of each 2D convolution layer in Figure 1 also shows the size of their receptive fields. It can be seen that the size of the final receptive field is 11 × 11 .
Assume that the receptive field after convolution is r o u t , and the receptive field for introducing the dilated convolution operation is [31]
r o u t = ( r i n 1 ) · s t r i d e + ( t × 2 + 1 )
where r i n represents the size of the receptive field of the upper layer, s t r i d e is the stride parameter of the convolution layer, and t represents the dilation parameter of the dilated convolution.

2.4. Fully Connected Module

The proposed LRCNet consists of three fully connected (FC) layers. The first FC layer converts the feature map output by the last dilation convolution layer into a 1D vector with 256 nodes. Due to the similarity between the classes of HSI data, we further compress 256 nodes into 128 nodes by using an FC layer. As a result, the influence of the feature location on the classification results is reduced in order to improve the network’s robustness. To reduce the risk of overfitting, a dropout layer is added after each FC layer. Finally, we use an FC layer with the number of nodes equal to the number of classes in the dataset.
Suppose that the 1D vector output by the FC layer is A = ( a 1 , a 2 , a 3 a i 1 , a i ) , where a i represents the i th element of A . Next, a i is calculated as follows:
a i = b i + κ = 1 q ( G κ · W i , κ )
where G κ represents the κ th feature map, W i , κ denotes the weight matrix of the κ th feature map of the i th element, and q denotes the total number of feature maps output by the receptive field control network.

2.5. Classification Result Output

After the third FC layer, we map the output to ( , 0 ] by using the logsoftmax function. Since the softmax activation function performs exponential operations, overflow or underflow may occur during the calculation; therefore, by using the logsoftmax activation function, problems can be avoided, data stability can be improved, and the operation can be sped up [31]. Assuming that x h 1 × C is the output vector of pixel h after passing through the FC layer, where C is the number of object categories in the dataset, the output is expressed as:
y c = log e x h ( c ) n = 1 C e x h ( n )
where y c represents the possibility that x h belongs to category c , and x h ( c ) is the c th element in x h .
The cross-entropy loss is chosen as the loss function. Assuming the loss of classifying pixel h is l o s s C E ( h , c t ) , l o s s C E ( h , c t ) can be calculated as:
l o s s C E ( h , c t ) = y c t = log e x h ( c t ) n = 1 C e x h ( n )
where c t represents the correct class of pixel h , and x h ( c t ) is the element in x h that belongs to class c t .

3. Results

3.1. Dataset Introduction

In this work, we used three public datasets to verify the performance of LRCNet in HSIC tasks [42], including Indian Pines (IndianP), Pavia University (PaviaU), and Salinas Valley (SalinasV).

3.2. Experimental Setup

We used a GTX 1080 Ti with 10GB of memory for training the network. The hyperparameters of LRCNet were set as follows. We set the learning rate to 0.00008, epochs to 100, and batch size to 128. We divided the input image into small windows of 25 × 25 pixels and reduced the band number to 30. The Adam algorithm was adopted to optimize the learning rate. The cross-entropy function was selected as the loss function. We reserved 30% of the data for testing and 70% of the data for training the network. In this paper, we used the OA, kappa, and AA metrics to evaluate the classification performance. OA represents the number of correctly classified test samples, AA represents the average accuracy of each class, and kappa combines the diagonal and non-diagonal terms of the confusion matrix and is a robust measure of consistency.

3.3. Classification Results and Analysis

We compared the proposed LRCNet with other methods in detail, including two classical methods, i.e., SVM and 2D-CNN [16,43], five 3D convolutional networks, i.e., 3D-CNN, HybridSN, M3D-CNN, SSRN, and 3D-Caps [22,23,24,25,26], a method for controlling the receptive field named DSSNet [31], and some state-of-the-art methods such as EMAP-Cubic-Caps (EMAP-C-C), MSPN, and SST-M [27,28,44].
Table 1 shows the OA, AA, and kappa values of the different methods based on the three public datasets. The proposed LRCNet clearly performs well on the three datasets, and its classification accuracy has a certain competitiveness. Based on the PaviaU and SalinasV datasets, the classification accuracy of LRCNet is close to 100. However, the classification performance obtained using the IndianP dataset reaches the ideal result, and the AA index score is only 98.40%. Based on the confusion matrix shown in Figure 4, we find that the proposed LRCNet easily misjudges the Soybean-mintill class as the Corn-notill class and Grass-pasture class. The additional observations of the ground truth map of IndianP show that the three classes are very close in the image. We infer that the details of the three categories are wrongly fused together during feature learning.
Figure 5, Figure 6 and Figure 7 show the classification results obtained using LRCNet, HybridSN, and DSSNet on three public datasets. The results show that the classification results obtained with the proposed LRCNet are close to the ground truth.
In order to further verify the LRCNet’s high classification performance, we performed a one-sample t-test experiment to study if the mean values of OA, kappa, and AA were substantially different from those in Table 1. We repeated 10 tests on three datasets, and the results are shown in Table 2. The results in Table 1 and Table 2 do not significantly differ from one another. It can be inferred that the outstanding classification performance of LRCNet is not an accidental result.
Table 3 displays the proposed LRCNet’s classification performance under different training set partition ratios. IndianP was used in the experiment. It can be found that the decline in the OA and kappa indicators is small. When the training set only accounts for 10%, the score of the OA indicator also reaches 97.90%. It is evident that the proposed LRCNet learns most of the spatial and spectral features by using a small number of training samples.
Table 4 shows the number of parameters and the computational cost of training the proposed LRCNet and HybridSN by using the IndianP dataset. It is evident from Table 4 that the proposed method effectively reduces the number of parameters and the computational cost of 3D convolution. The LRCNet has only 3,857,330 parameters, and the floating point operations per second (Flops) is only 95.71MB, which is 152.91MB less than that of HybridSN. Therefore, it can be verified that the 3D depthwise separable technology ensures classification accuracy and reduces the calculation cost.

3.4. Ablation Experiments

In order to confirm the impact of the 3D-DW module on the classification accuracy, we performed ablation experiments. We contrasted the proposed LRCNet and the two modified networks in terms of classification performance. In Net1, we replaced the third 3D-DW module with a 2D-DW module. Similarly, in Net2, we replaced the second and third 3D-DW modules with two 2D-DW modules.
We tested the classification performance of the three networks based on IndianP. Table 5 displays the results. It is evident from the results that the classification performance reduces significantly after the 3D-DW module is replaced with the 2D-DW module, and the AA index decreases the most. When only one 3D-DW module is used, the score of the AA index is only 84.29%. When the number of 3D-DW modules is increased to 2, the score of the AA index increases to 93.75%. For some classes with a small number, it is difficult to learn the corresponding features using the 2D-DW module, and they are often classified as other classes during classification. Although the scores of the OA and kappa indicators do not drop considerably, the scores of the AA indicator are much lower. This shows the validity of the 3D-DW module, which learns the joint features of the spectral and spatial dimensions.
In order to verify the effectiveness of the receptive field control strategy, we also conducted ablation experiments. We changed the dilation rate of the dilated convolution in the receptive field control network by using three different values, i.e., 1, 3, and 4. The corresponding networks are Net3, Net4, and Net5.
We tested the classification accuracies of the three networks and compared them with LRCNet. The results are presented in Table 6. It is evident from the results that no matter whether the size of the receptive field continues to increase or decrease, the classification accuracy decreases by varying degrees. When the receptive field is 7 × 7 , it is too small to capture the multi-scale information. Therefore, for smaller categories, the network is unable to learn all the features and is prone to incorrect classification results. When the receptive field is increased to 15 × 15 and 19 × 19 , OA decreases to 98.68% and 98.45%, respectively. It can be inferred that a large receptive field loses some details. More details are lost when the receptive field is larger. Therefore, the strategy for controlling the receptive field of LRCNet is successful.

4. Conclusions

In this work, we proposed the LRCNet for performing HSIC tasks, which is an end-to-end framework comprising two functional modules, including a 3D depthwise separable convolutional network, which is used to reduce the computational cost of the convolution and the number of parameters, and the receptive field control network, which is used to control the receptive field for capturing multi-scale features. In the 3D convolutional network, we propose the 3D depthwise separable convolution technique to decrease the number of parameters and the computational costs. In the receptive field control network, we use dilated convolutions to control the receptive field for capturing the multi-scale features and to avoid the loss of detailed features. In the ablation study, we found that the strategy of mixing 3D and 2D convolutions and controlling the receptive field can enhance the classification accuracy. In order to verify the classification performance of the proposed LRCNet, we tested it on three public datasets and obtained competitive results. The proposed method can be applied to accurately identify and classify ground objects in hyperspectral images.

Author Contributions

All authors contributed substantially to this study. Individual contributions were as follows: conceptualization, Q.Z.; methodology, C.L. and F.G.; software, C.L. and T.W.; validation, T.W. and S.D.; formal analysis, C.L.; investigation, Q.Z. and Z.Y.; resources, F.G.; writing—original draft preparation, C.L. and Q.Z.; writing—review and editing, F.G. and C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Zhejiang Provincial Innovative Incubation Projects for University Students (Emerging Artists Talent Program), grant number 2021R407016, and the Open Foundation of Key Laboratory of Submarine Geosciences, MNR, grant number KLSG2002.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank those who provided help in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, H.; Zou, J.; Zhang, L. EMS-GCN: An End-to-End Mixhop Superpixel-Based Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5526116. [Google Scholar] [CrossRef]
  2. Feng, J.; Zhao, N.; Shang, R.; Zhang, X.; Jiao, L. Self-Supervised Divide-and-Conquer Generative Adversarial Network for Classification of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5536517. [Google Scholar] [CrossRef]
  3. Bayramoglu, N.; Kaakinen, M.; Eklund, L.; Heikkila, J. Towards virtual H&E staining of hyperspectral lung histology images using conditional generative adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 64–71. [Google Scholar]
  4. Han, Y.; Shi, X.; Yang, S.; Zhang, Y.; Hong, Z.; Zhou, R. Hyperspectral Sea Ice Image Classification Based on the Spectral-Spatial-Joint Feature with the PCA Network. Remote Sens. 2021, 13, 2253. [Google Scholar] [CrossRef]
  5. Hu, K.; Weng, C.; Zhang, Y.; Jin, J.; Xia, Q. An Overview of Underwater Vision Enhancement: From Traditional Methods to Recent Deep Learning. J. Mar. Sci. Eng. 2022, 10, 241. [Google Scholar] [CrossRef]
  6. Zhou, J.; Yang, T.; Zhang, W. Underwater vision enhancement technologies: A comprehensive review, challenges, and recent trends. Appl. Intell. 2022, 1–28. [Google Scholar] [CrossRef]
  7. Ye, P.; Han, C.; Zhang, Q.; Gao, F.; Yang, Z.; Wu, G. An Application of Hyperspectral Image Clustering Based on Texture-Aware Superpixel Technique in Deep Sea. Remote Sens. 2022, 14, 5047. [Google Scholar] [CrossRef]
  8. Zhang, Q.; Zheng, E.; Wang, Y.; Gao, F. Recognition of ocean floor manganese nodules by deep kernel fuzzy C-means clustering of hyperspectral images. J. Image Graph. 2021, 26, 1886–1895. [Google Scholar]
  9. Li, S.; Song, W.; Fang, L.; Chen, Y.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, Q.; Meng, Z.; Li, X. Locality adaptive discriminant analysis for spectral–spatial classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2077–2081. [Google Scholar] [CrossRef]
  11. Yuan, Y.; Lin, J.; Wang, Q. Hyperspectral image classification via multitask joint sparse representation and stepwise MRF optimization. IEEE Trans. Cybern. 2015, 46, 2966–2977. [Google Scholar] [CrossRef]
  12. Pan, B.; Shi, Z.; Xu, X. Hierarchical guidance filtering-based ensemble classification for hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4177–4189. [Google Scholar] [CrossRef]
  13. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  14. Pan, B.; Shi, Z.; Xu, X. MugNet: Deep learning for hyperspectral image classification using limited samples. ISPRS J. Photogramm. Remote Sens. 2018, 145, 108–119. [Google Scholar] [CrossRef]
  15. Cheng, G.; Li, Z.; Han, J.; Yao, X.; Guo, L. Exploring hierarchical convolutional features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6712–6722. [Google Scholar] [CrossRef]
  16. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  17. Jiao, L.; Liang, M.; Chen, H.; Yang, S.; Liu, H.; Cao, X. Deep fully convolutional network-based spatial distribution prediction for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5585–5599. [Google Scholar] [CrossRef]
  18. Sun, H.; Zheng, X.; Lu, X. A supervised segmentation network for hyperspectral image classification. IEEE Trans. Image Process. 2021, 30, 2810–2825. [Google Scholar] [CrossRef]
  19. Kang, X.; Zhuo, B.; Duan, P. Dual-path network-based hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2018, 16, 447–451. [Google Scholar] [CrossRef]
  20. Soucy, N.; Sekeh, S.Y. CEU-Net: Ensemble Semantic Segmentation of Hyperspectral Images Using Clustering. arXiv 2022, arXiv:2203.04873. [Google Scholar]
  21. Si, Y.; Gong, D.; Guo, Y.; Zhu, X.; Huang, Q.; Evans, J.; He, S.; Sun, Y. An Advanced Spectral–Spatial Classification Framework for Hyperspectral Imagery Based on DeepLab v3+. Appl. Sci. 2021, 11, 5703. [Google Scholar] [CrossRef]
  22. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  23. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  24. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  25. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  26. Zhu, K.; Chen, Y.; Ghamisi, P.; Jia, X.; Benediktsson, J.A. Deep convolutional capsule network for hyperspectral image spectral and spectral-spatial classification. Remote Sens. 2019, 11, 223. [Google Scholar] [CrossRef] [Green Version]
  27. Sun, L.; Song, X.; Guo, H.; Zhao, G.; Wang, J. Patch-wise semantic segmentation for hyperspectral images via a cubic capsule network with EMAP features. Remote Sens. 2021, 13, 3497. [Google Scholar] [CrossRef]
  28. Gong, H.; Li, Q.; Li, C.; Dai, H.; He, Z.; Wang, W.; Li, H.; Han, F.; Tuniyazi, A.; Mu, T. Multiscale information fusion for hyperspectral image classification based on hybrid 2D-3D CNN. Remote Sens. 2021, 13, 2268. [Google Scholar] [CrossRef]
  29. Ghaderizadeh, S.; Abbasi-Moghadam, D.; Sharifi, A.; Zhao, N.; Tariq, A. Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7570–7588. [Google Scholar] [CrossRef]
  30. Xu, H.; Yao, W.; Cheng, L.; Li, B. Multiple spectral resolution 3D convolutional neural network for hyperspectral image classification. Remote Sens. 2021, 13, 1248. [Google Scholar] [CrossRef]
  31. Pan, B.; Xu, X.; Shi, Z.; Zhang, N.; Luo, H.; Lan, X. DSSNet: A simple dilated semantic segmentation network for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1968–1972. [Google Scholar] [CrossRef]
  32. Yokoya, N.; Chan, J.C.-W.; Segl, K. Potential of resolution-enhanced hyperspectral data for mineral mapping using simulated EnMAP and Sentinel-2 images. Remote Sens. 2016, 8, 172. [Google Scholar] [CrossRef] [Green Version]
  33. Li, Q.; Wang, Q.; Li, X. Exploring the relationship between 2D/3D convolution for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8693–8703. [Google Scholar] [CrossRef]
  34. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  35. Firat, H.; Asker, M.E.; Hanbay, D. Hybrid 3D Convolution and 2D Depthwise Separable Convolution Neural Network for Hyperspectral Image Classification. Balk. J. Electr. Comput. Eng. 2022, 10, 35–46. [Google Scholar] [CrossRef]
  36. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition, 2018; pp. 4510–4520.
  37. Jiang, Y.; Han, W.; Ye, L.; Lu, Y.; Liu, B. Two-Stream 3D MobileNetV3 for Pedestrians Intent Prediction Based on Monocular Camera. In Proceedings of the International Conference on Neural Computing for Advanced Applications, Jinan, China, 8–10 July 2022; pp. 247–259. [Google Scholar]
  38. Hou, B.; Liu, Y.; Ling, N.; Liu, L.; Ren, Y. A Fast Lightweight 3D Separable Convolutional Neural Network With Multi-Input Multi-Output for Moving Object Detection. IEEE Access 2021, 9, 148433–148448. [Google Scholar] [CrossRef]
  39. Alalwan, N.; Abozeid, A.; ElHabshy, A.A.; Alzahrani, A. Efficient 3D deep learning model for medical image semantic segmentation. Alex. Eng. J. 2021, 60, 1231–1239. [Google Scholar] [CrossRef]
  40. Stergiou, A.; Poppe, R. Adapool: Exponential adaptive pooling for information-retaining downsampling. arXiv 2021, arXiv:2111.00772. [Google Scholar]
  41. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–Spatial Feature Tokenization Transformer for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  42. Graña, M.; Veganzons, M.A.; Ayerdi, B. Hyperspectral Remote Sensing Scenes. Available online: https://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 5 August 2022).
  43. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  44. Bai, J.; Wen, Z.; Xiao, Z.; Ye, F.; Zhu, Y.; Alazab, M.; Jiao, L. Hyperspectral Image Classification Based on Multibranch Attention Transformer Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5535317. [Google Scholar] [CrossRef]
Figure 1. The architecture of the proposed LRCNet.
Figure 1. The architecture of the proposed LRCNet.
Electronics 11 03992 g001
Figure 2. The proposed 3D-DW module.
Figure 2. The proposed 3D-DW module.
Electronics 11 03992 g002
Figure 3. A comparison of a 3D standard convolution kernel and a 3D depthwise separable convolution kernel.
Figure 3. A comparison of a 3D standard convolution kernel and a 3D depthwise separable convolution kernel.
Electronics 11 03992 g003
Figure 4. The confusion matrix obtained using the proposed method by using the IndianP, PaviaU, and SalinasV datasets in the first, second, and third matrices, respectively.
Figure 4. The confusion matrix obtained using the proposed method by using the IndianP, PaviaU, and SalinasV datasets in the first, second, and third matrices, respectively.
Electronics 11 03992 g004
Figure 5. The classification map for the IndianP dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Figure 5. The classification map for the IndianP dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Electronics 11 03992 g005
Figure 6. The classification map for the PaviaU dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Figure 6. The classification map for the PaviaU dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Electronics 11 03992 g006
Figure 7. The classification map for the SalinasV dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Figure 7. The classification map for the SalinasV dataset: (a) original image, (b) ground truth, and (ce) classification maps obtained using LRCNet, HybridSN, and DSSNet, respectively.
Electronics 11 03992 g007aElectronics 11 03992 g007b
Table 1. The classification accuracy comparison for the IndianP, PaviaU, and SalinasV datasets (%).
Table 1. The classification accuracy comparison for the IndianP, PaviaU, and SalinasV datasets (%).
MethodIndianPPaviaUSalinasV
OAKappaAAOAKappaAAOAKappaAA
SVM [43]85.3083.1079.0394.3692.5092.9892.9592.1194.60
2D-CNN [16]89.4887.9686.1497.8697.1696.5597.3897.0898.84
3D-CNN [23]91.1089.9891.5896.5395.5197.5793.9693.3297.01
M3D-CNN [24]95.3294.7096.4195.7694.5095.0894.7994.2096.25
SSRN [25]99.1999.0798.9399.9099.8799.9199.9899.9799.97
HybridSN [22]99.5699.5198.5099.8599.8099.88100100100
3D-Caps [26]90.2090.1593.0088.3484.9390.1488.9587.7494.35
DSSNet [31]97.6197.2796.3199.6299.5099.2298.5198.3497.56
EMAP-C-C [27]98.2096.7297.9598.8198.4298.4998.5598.3899.08
MSPN [28]96.0995.5391.5396.5695.4294.5597.0096.6697.33
SST-M [44]99.0898.9599.0199.6199.4899.23///
LRCNet99.6099.5498.4099.9799.9699.95100100100
Table 2. The classification accuracy of LRCNet repeated experiments on three datasets (%).
Table 2. The classification accuracy of LRCNet repeated experiments on three datasets (%).
DatasetsOAKappaAA
IndianP99.31 ± 0.3499.22 ± 0.3898.40 ± 0.72
PaviaU99.95 ± 0.0499.93 ± 0.0699.92 ± 0.04
SalinasV99.99 ± 0.0199.99 ± 0.0199.98 ± 0.02
Table 3. The classification accuracies obtained using a small number of training samples (%).
Table 3. The classification accuracies obtained using a small number of training samples (%).
Proportion of Training SamplesOAKappaAA
30%99.6099.5498.40
20%98.6898.5095.10
10%97.9097.6088.16
Table 4. The comparison of the number of parameters and computational cost.
Table 4. The comparison of the number of parameters and computational cost.
MethodParamsFlops (MB)
LRCNet3,857,33095.71
HybridSN5,122,176248.62
Table 5. The classification accuracy obtained using a small number of training samples (%).
Table 5. The classification accuracy obtained using a small number of training samples (%).
NetworksArchitecture of 3D-DW PartOAKappaAA
LRCNetthree 3D-DW modules99.6099.5498.40
Net1two 3D-DW modules and one 2D-DW module97.1696.7593.75
Net2one 3D-DW module and two 2D-DW modules97.0996.6884.29
Table 6. The classification accuracy of different receptive fields (%).
Table 6. The classification accuracy of different receptive fields (%).
NetworksDilation RateReceptive FieldOAKappaAA
LRCNet211 × 1199.6099.5498.40
Net317 × 799.3799.2898.06
Net4315 × 1598.6898.4997.25
Net5419 × 1998.4598.2492.94
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, C.; Wang, T.; Dong, S.; Zhang, Q.; Yang, Z.; Gao, F. Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification. Electronics 2022, 11, 3992. https://doi.org/10.3390/electronics11233992

AMA Style

Lin C, Wang T, Dong S, Zhang Q, Yang Z, Gao F. Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification. Electronics. 2022; 11(23):3992. https://doi.org/10.3390/electronics11233992

Chicago/Turabian Style

Lin, Chengle, Tingyu Wang, Shuyan Dong, Qizhong Zhang, Zhangyi Yang, and Farong Gao. 2022. "Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification" Electronics 11, no. 23: 3992. https://doi.org/10.3390/electronics11233992

APA Style

Lin, C., Wang, T., Dong, S., Zhang, Q., Yang, Z., & Gao, F. (2022). Hybrid Convolutional Network Combining 3D Depthwise Separable Convolution and Receptive Field Control for Hyperspectral Image Classification. Electronics, 11(23), 3992. https://doi.org/10.3390/electronics11233992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop