Next Article in Journal
Consolidated Convolutional Neural Network for Hyperspectral Image Classification
Next Article in Special Issue
SAR Ship–Iceberg Discrimination in Arctic Conditions Using Deep Learning
Previous Article in Journal
A Strategy of Parallel SLIC Superpixels for Handling Large-Scale Images over Apache Spark
Previous Article in Special Issue
ClassHyPer: ClassMix-Based Hybrid Perturbations for Deep Semi-Supervised Semantic Segmentation of Remote Sensing Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Two-Branch Convolutional Neural Network Based on Multi-Spectral Entropy Rate Superpixel Segmentation for Hyperspectral Image Classification

1
Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, Collaborative Innovation Center of Quantum Information of Shaanxi Province, International Research Center for Intelligent Perception and Computation, Joint International Research Laboratory of Intelligent Perception and Computation, School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
School of Electronic Engineering, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(7), 1569; https://doi.org/10.3390/rs14071569
Submission received: 23 February 2022 / Revised: 16 March 2022 / Accepted: 21 March 2022 / Published: 24 March 2022
(This article belongs to the Special Issue Deep Learning for Remote Sensing Image Classification)

Abstract

:
Convolutional neural networks (CNNs) can extract advanced features of joint spectral–spatial information, which are useful for hyperspectral image (HSI) classification. However, the patch-based neighborhoods of samples with fixed sizes are usually used as the input of the CNNs, which cannot dig out the homogeneousness between the pixels within and outside of the patch. In addition, the spatial features are quite different in different spectral bands, which are not fully utilized by the existing methods. In this paper, a two-branch convolutional neural network based on multi-spectral entropy rate superpixel segmentation (TBN-MERS) is designed for HSI classification. Firstly, entropy rate superpixel (ERS) segmentation is performed on the image of each spectral band in an HSI, respectively. The segmented images obtained are stacked band by band, called multi-spectral entropy rate superpixel segmentation image (MERSI), and then preprocessed to serve as the input of one branch in TBN-MERS. The preprocessed HSI is used as the input of the other branch in TBN-MERS. TBN-MERS extracts features from both the HSI and the MERSI and then utilizes the fused spectral–spatial features for the classification of HSIs. TBN-MERS makes full use of the joint spectral–spatial information of HSIs at the scale of superpixels and the scale of neighborhood. Therefore, it achieves excellent performance in the classification of HSIs. Experimental results on four datasets demonstrate that the proposed TBN-MERS can effectively extract features from HSIs and significantly outperforms some state-of-the-art methods with a few training samples.

1. Introduction

Hyperspectral images (HSIs) are obtained by simultaneously imaging ground objects in a certain area on continuous multiple spectral bands. The classification of HSIs is an important technology in remote sensing, as HSIs contain rich spatial and spectral information [1,2,3], which is widely applied in the fields of military target recognition, geological resource detection, agricultural crop monitoring, and archaeological relic restoration [4,5,6,7]. HSIs have a very high spectral dimension and rich spectral information, but the spectral information is mixed with noise information brought by atmospheric environment and imaging instruments. The classification accuracy of HSIs is often not good enough if the original information is directly used. Effective feature extraction of HSIs is one of the important means to improve classification accuracy [8,9,10]. In addition, due to the high cost of manual labeling, the labels of HSIs are usually limited. How to achieve higher classification accuracy with fewer labeled samples is an important direction in the research of HSI classification [11].
In recent decades, the classification of HSIs has received more and more attention. Researchers have made considerable progress [12,13,14]. Only the spectral information of HSI was used for classification in early research, such as K-nearest neighbors (KNN) [15], support vector machine (SVM) [16,17], sparse representation [18], and so on. These methods classified samples according to the spectral feature and were easy to implement, but the classification accuracy was relatively limited. Subsequently, some methods [19,20] combining spatial and spectral information were proposed, which led to a better classification performance. However, due to the high dimension of HSIs and the intra-class differences of ground objects, how to obtain better classification accuracy is still a difficult problem that needs to be investigated further.
With the vigorous development of deep learning, researchers began to apply deep learning to HSI classification. A lot of methods were proposed such as the stacked auto-encoder (SAE) [19], deep belief networks (DBNs) [20], and convolutional neural networks (CNNs) [21,22,23]. Among all these methods, CNN-based methods are the most widely used. CNN-based methods take the patch-based neighborhoods of samples as input and perform feature extraction through convolution. The spatial features and the spectral features, or the joint features of spatial and spectral information can be extracted by convolution from the patch-based neighborhoods, and the classification accuracy of HSIs is greatly improved. Hamida et al. [24] studied the effect of different three-dimensional (3D) convolutional networks on HSI classification and designed a 3DCNN network consisting of 4 layers of 3D convolution and a fully connected layer to extract the spectral–spatial features from HSIs. Hamida’s 3DCNN model made good use of joint features and achieved a higher classification accuracy. Zhong et al. [25] proposed a spectral–spatial residual network (SSRN), in which they designed a spatial feature extraction module and a spectral feature extraction module using multiple 3D convolutional layers according to the residual structure and investigated the effects of number and order of modules in feature extraction. This end-to-end feature extraction and classification network of HSIs achieved high classification accuracy. Roy et al. [26] proposed a hybrid spectral convolutional neural network (HybridSN), which firstly used multiple 3D convolutional layers for spectral–spatial feature extraction, and then used a two-dimensional (2D) convolutional layer to extract advanced spatial feature from the obtained feature maps. The 2D convolutional layer reduced the complexity of the model and improved the classification accuracy.
The CNN-based methods mostly take the patch-based neighborhoods of samples as input, which means that the spatial information only comes from the neighborhood, and the spatial information outside the neighborhood is ignored. Some researchers tried to propose some methods that utilized a wider area of spatial information. Superpixel segmentation is a commonly used means to obtain spatial distribution information of ground objects in a wide area. Leng et al. [27] used entropy rate superpixel (ERS) segmentation to perform multi-scale superpixel segmentation on the first principal component of HSIs and further extracted the spectral–spatial information within each superpixel for classification. Jiang et al. [28] proposed a superpixel-wise principal component analysis (PCA) approach for unsupervised feature extraction of HSIs (SuperPCA). SuperPCA firstly performed multi-scale ERS on the first principal component of HSIs, and then applied the proposed superpixel-wise principal component analysis approach to reduce the dimension of HSIs within the obtained superpixels. Finally, it trained classifiers over the data after dimension reduction on each scale, respectively, and obtained the results through decision fusion. This method achieved higher classification accuracy even when the training samples were limited.
A two-branch network is a commonly used structure to process multi-source data or extract a variety of different features [29,30,31,32]. The two branches may have different structures or same structures with different parameters. According to the different inputs of the two branches, each branch processes a certain type of information. Feature fusion will usually be performed at the end of network. The spectral–spatial attention network proposed by Mei et al. [29] is a two-branch network. One branch uses the structure of recurrent neural network to extract spectral information, the other branch uses CNN structures to extract spatial information, and an attention mechanism is applied to make the network able to extract key fusion features. Mu et al. [30] proposed a low-rank based method for HSI classification called a two-branch network combined with robust PCA, where both the low rank and the sparse components were preserved and used for feature extraction in two independent convolutional branches. This method constructed a convenient model for HSI classification by discarding the low-rank subspace estimation and combining denoising, feature extraction, feature fusion, and classification into an end-to-end network, which maintained better classification performance even for the cases of small samples and class imbalance.
The above methods have made great progress in improving the classification results of HSIs, but some issues still need to be addressed. These methods mainly use the rich spectral features of HSIs and extract joint spectral–spatial features from the patch-based neighborhoods of the samples. In fact, the information of each band of HSIs is quite different, and each band has rich spatial information that has not been fully utilized. If the various bands of the HSI are processed separately, more representative spatial information can be extracted. On the other hand, the pixels corresponding to the same ground object, which may distribute in a wide area, should belong to the same category. Mining and making use of such priori knowledge can improve the classification results of HSIs further. In high-resolution images, the same kind of ground objects are distributed in a continuous area with an irregular shape, and the size is much larger than that of the patch-based neighborhoods. When the classifier just takes the patch-based neighborhoods as input, it can only extract features in the scale of patch-based neighborhoods, and the spatial distribution information of the ground objects outside the patches is ignored.
To solve the above problems, this paper designs a two-branch convolutional neural network based on multi-spectral entropy rate superpixel segmentation (TBN-MERS) for HSI classification. By performing ERS segmentation on each band of HSIs, TBN-MERS can divide a certain area of the same kind of ground objects into a same superpixel, so as to obtain the superpixel-scale spatial information of each band. The preprocessed HSI and the preprocessed multi-spectral ERS image (MERSI) are then used as the two inputs of the two-branch convolutional neural network, which can extract the joint information of the spectral–spatial features at the scale of both the patch-based neighborhoods and the superpixels.
The main contributions of this paper are as follows. (1) ERS segmentation is performed on all bands of HSIs, and the obtained MERSI contains rich superpixel-scale spatial information, which greatly improves the classification accuracy of HSIs. (2) A two-branch convolutional neural network is designed, which can effectively extract and fuse the preprocessed HSI and preprocessed MERSI, and the fusion feature further improves the classification accuracy. (3) TBN-MERS can obtain much better classification results, even with very limited samples.
The rest of this paper is as follows. Section 2 introduces the related backgrounds and the details of TBN-MERS. Section 3 first analyzes the effect of the structure and the parameters in TBN-MERS, then describes the comparison experiments between TBN-MERS and its variants, and finally, compares TBN-MERS with other state-of-the-art methods on four datasets. The last section summarizes the work of this paper and provides the direction of the further work in the future.

2. Materials and Methods

2.1. Multi-Spectral ERS

Superpixel segmentation is a kind of commonly used preprocessing method in computer vision fields such as target detection and image segmentation [33,34,35]. Superpixel segmentation is to divide the pixels in the image into different groups with certain properties. The pixels belonging to the same superpixel are similar in texture, brightness, color, or other characteristics [36]. The purpose of superpixel segmentation is to achieve the following effects. (a) Each superpixel contains only one class of ground object. (b) The set of superpixel boundaries is a super set of the boundaries of ground objects [37].
The methods of generating superpixels can be categorized into clustering-based methods and graph-based methods. Clustering-based methods start from a rough initial clustering of pixels, and iteratively refine the clustering to form superpixels until certain convergence criteria are met, such as Mean Shift [38], Watershed [39], and Simple Linear Iterative Clustering (SLIC) [36]. The graph-based methods regard each pixel in the image as a node of the graph. The weight of edges between two nodes is calculated based on the similarity of them. Graph-based methods include Normalized Cuts [40], Felzenszwalb [41], Entropy Rate Superpixel Segmentation (ERS) [37], and so on.
Mean Shift [38] is an iterative method to find the pattern among pixels in the feature space by maximizing the local density function. The pixels that converge to the same pattern form a superpixel. Watershed [39] regards the image as a geographical topological structure, and the value of a pixel in the image represents the altitude. The connection of the pixels with large values is regarded as a ridge, the area formed by the pixels with small values is regarded as a valley, and the closed valley is regarded as a superpixel. SLIC [36] maps the image to a new feature space, clusters it according to the distance metric, and regards different clusters as superpixels. Normalized Cuts [40] uses contour and texture to recursively segment the image and minimizes the global cost function defined on the segmentation boundary. Felzenszwalb [41] regards the image as graph and clusters pixels on the graph, so that each superpixel is the smallest spanning tree of pixels. ERS [37] finds multiple disconnected sub-graphs from the graph of the image by iteratively optimizing the entropy rate loss function, and each sub-graph represents a superpixel.
Mean Shift is robust to local differences but tends to straddle multiple objects. Watershed runs very quickly, but the research of Levinshtein et al. [42] and Veksler et al. [43] show that the superpixels produced by Watershed often contain multiple object categories. SLIC has excellent performance in efficiency and segmentation effect, and is one of the commonly used segmentation algorithms. Normalized Cuts generates superpixels with the uniform size and compact shape, but it requires a lot of calculations and runs slowly. Felzenszwalb is very efficient and generates a good embedding representation. However, Ren et al. [34] proved that it tended to sacrifice details so that it generated smooth boundaries, and the boundary recall rate was not high enough. ERS has high efficiency and high boundary recall rate, and the obtained superpixels are compact and homogeneous. In this paper, ERS is applied for superpixel segmentation.
ERS regards superpixel segmentation as an undirected graph clustering problem [37]. The image is represented by a graph G = ( V , E ) , in which V is a set of all pixels, and E is the set of edges that are between every pixel and its adjacent pixels. The purpose of ERS is to find a sub-graph G = ( V , A ) , A E . G = ( V , A ) represents the image after superpixel segmentation. The objective function of ERS includes two parts: (1) the entropy rate H ( A ) calculated by random walk on the graph; and (2) the balance term B ( A ) of the clustering distribution. In the process of superpixel segmentation, a larger entropy rate H ( A ) will make the obtained superpixels more compact and homogeneous, which is conducive to containing only a single ground object in each superpixel, and the balance item B ( A ) controls the size of the cluster, avoiding it becoming too smooth while retaining the boundary of ground object. The objective function of ERS is as follows:
A * = a r g m a x ( H ( A ) + α B ( A ) ) ,       s . t . A E
where α is a trade-off parameter of H ( A ) and B ( A ) .
ERS has been applied to a variety of HSI classification methods to extract spatial features, but the existing methods generally only perform ERS on the first principal component of HSIs and do not consider the spatial differences between different bands. To fully utilize the spatial information of HSIs in different bands, and to correlate the spatial information with spectral information better, we perform ERS on images of all bands in HSIs.
Taking the Pavia University (PU) dataset as the input, the process of multi-spectral ERS of HSIs proposed in this paper is shown in Figure 1. First, the HSI is separated according to bands, and then the values in the image of each band are scaled to the interval [0, 255]. Suppose X represents the set of all pixels of the image in a certain band, x i represents the value of certain pixel in X , Z ( x i ) is the corresponding value after scaling, m i n ( X ) is the minimum value of X , and m a x ( X ) is the maximum value of X , then the scaling formula is:
Z ( x i ) = x i m i n ( X ) m a x ( X ) m i n ( X ) × 255
We perform ERS on the scaled image of each band to obtain the superpixel segmentation image of each band. ERS will divide the pixels into n c superpixels. The superpixels are numbered 1 , 2 , , n c respectively. In the obtained segmentation image, the values of all pixels in the same superpixel are all equal to the serial number of the superpixel. Finally, the segmentation images are stacked one by one according to order of the bands to form the MERSI.

2.2. Two-Branch Convolutional Neural Network (TBN-MERS)

CNNs are mainly composed of convolutional layers. The convolutional layer uses multiple convolution kernels to perform convolution on the multi-spectral feature map input to obtain a more advanced feature map. The convolutional layer has the characteristics of local connection and weight sharing; that is, each node of the output is only connected to some nodes of the input, and the weights of connections between the nodes in different positions in output and corresponding nodes in input are the same. The convolution operation can keep roughly the same spatial structure of the output data with that of the input data, so it is widely used in the feature extraction of two-dimensional (2D) and three-dimensional (3D) images. Commonly used convolutional layers include 2D convolutional layers and 3D convolutional layers. The 2D convolutional layer mainly performs feature extraction from the spatial dimension, and the 3D convolutional layer can perform feature extraction from the spatial dimension and the spectral dimension at the same time. We use 2D convolutional layers, 3D convolutional layers, and fully connected layers to build neural networks. The HybridSN [26] proposed by Roy et al. is an end-to-end single-branch network, which is very effective in the feature extraction of HSIs. It has achieved high classification accuracy on multiple datasets. We design a two-branch convolutional neural network (TBN-MERS) based on the structure of HybridSN.
Taking the PU dataset as the input, Figure 2 shows the network structure of TBN-MERS. TBN-MERS has two feature extraction branches, with the same structure, where 3 layers of 3D convolutional layer and 1 layer of 2D convolutional layer are cascaded, and after each 3D convolutional layer and 2D convolutional layer, there is a BatchNorm layer and a ReLU activation layer. The outputs of the two network branches are then flattened as vectors and are added by the element addition operation to implement feature fusion. Subsequently, 3 cascaded fully connected layers are used for classification, where one dropout layer follows each of the first two fully connected layers. Finally, the Softmax function is applied to convert the network output into a probability vector. In Figure 2, we use Conv3D to represent the cascade of a 3D convolutional layer, a BatchNorm layer, and a ReLU layer, use Conv2D to represent the cascade of a 2D convolutional layer, a BatchNorm layer, and a ReLU layer, use FC to represent the cascade of a fully connected layer and a Dropout layer, and use fc represents the cascade of a fully connected layer and a Softmax layer. The hyper-parameters in the two-branch neural network are shown in Table 1. In Table 1, Conv3D_1, Conv3D_2, and Conv3D_3 represent different Conv3Ds that are cascaded in two branches of the same structure, and FC_1, FC_2 represent different FCs that are cascaded in the network.
The BatchNorm layer standardizes the input batch data and utilizes scaling variables γ and translation variables β to adjust the mean and variance of the standardized batch data in order to get a better distribution of values. Let B = { x 1 ,   x 2 ,   ,   x m } represent a batch of data and BN ( x i ) represent the corresponding output of sample x i after BatchNorm layer; then, the BatchNorm operation is as follows:
BN ( x i ) = γ x ^ i + β
In Formula (3), x ^ i = x i μ B σ B 2 + ε , where σ B 2 = 1 m i = 1 m ( x i μ B ) 2 , μ B = 1 m i = 1 m x i , and ε is a very small number to stabilize the value.
The ReLU activation function can introduce nonlinearity to the neural network and make the gradient propagate more efficiently. For the input data, the ReLU activation function can be expressed as:
ReLU ( x i ) = x i , x i > 0 0 , x i 0

2.3. The Process of TBN-MERS

The overall process of proposed TBN-MERS is shown in Figure 3, and the details are described as follows.
The data cube of one HSI is represented as X w × h × b , where w is the width and h is the height of the image, and b is the number of spectral bands. Let Y w × h represent the corresponding label image, and the value y i j at the location (i, j) of the label map Y is chosen from the set { 0 , 1 , 2 , , c } , where c represents the total number of classes.
X k w × h , k [ 1 , b ] represents the image in the kth band. The value of all pixels on X k is scaled to the interval [0, 255], and ERS algorithm is applied to obtain the corresponding segmentation image S k . After obtaining the segmentation images of all spectral bands, we stack them according to the order of the bands to obtain the data cube of superpixel segmentation image S w × h × b corresponding to X .
Firstly, the HSI data X are preprocessed, and the obtained data cube is denoted by X . The preprocessing operation is to standardize the image of each band. Let x i represent the value of a certain pixel in X k , and let S t a n d a r d i z e ( x i ) represent the value of this pixel after standardization; then, the standardization formula is:
S t a n d a r d i z e ( x i ) = x i μ k σ k
where μ k is the mean, and σ k is the standard deviation of all pixel values in X k .
Secondly, the MERSI data S is preprocessed, and the obtained data cube is denoted by S . The preprocessing operation is to normalize the image of each band. Let s i represent the value of a certain pixel in S k , N o r m a l i z e ( s i ) represent the value of this pixel after normalization, m i n ( S k ) represent the minimum value of S k , and m a x ( S k ) represent the maximum value of S k ; then, the normalization formula is:
N o r m a l i z e ( s i ) = s i m i n ( S k ) m a x ( S k ) m i n ( S k )
Thirdly, we generate the overall sample set. Each sample is corresponding to a pixel in the HSI and consists of two parts: the patch-based neighborhood p i x p × p × b taken from the preprocessed data cube X , where p is the height or width of the patch and b is the number of the bands in the HSI; the patch-based neighborhood p i s p × p × b taken from the preprocessed data cube S . We randomly select a certain number of samples from each class to form the training set and draw an equal number of samples from the remaining samples of each class to form the validation set, and the remaining samples are used as the test set.
Finally, we feed the two parts p i x and p i s of each sample x i into the two branches of TBN-MERS and train the network. We test the model on the verification set every epoch. When the classification accuracy of the model on the validation set no longer rises, the training is completed. We use the best model on the validation set to predict the samples in testing set.
The cross-entropy loss function is used to train the network. The cross-entropy loss function is as follows:
L = 1 N i N j c y i j log ( p i j )
where N is the number of samples in a batch, x i is the ith sample in the batch, c is the number of classes, and p i j is the probability value that the sample x i belongs to the class j predicted by the model. When the sample x i belongs to the class j, the value of y i j is 1; otherwise, it is 0.
The overall process of TBN-MERS is shown in Figure 3.

3. Results

3.1. Datasets

We conducted experiments on the Indian Pines (IP), Pavia University (PU), Salinas (SA), and Houston (HU) datasets.
IP was captured by the American AVIRIS sensor in the Indian remote sensing experimental area with a spatial resolution of 20 m. The AVIRIS sensor divided the spectrum from 375 μm to 2200 μm into 220 bands. After removing the 20 bands contaminated by noise, the remaining 200 spectral bands were retained. IP included 16 classes of ground objects, including grass pasture, woods, and wheat, etc. The image size of IP was 145 × 145. After removing the background pixels in the image, a total of 10,249 labeled samples could be used for classification. Figure 4a is the true label image of IP. Table 2 lists the specific classes of IP and the number of labeled samples of each class.
PU was captured by the ROSIS sensor at the University of Pavia in northeastern Italy with a spatial resolution of 1.3 m. The ROSIS sensor detected 115 bands in the wavelength range of 0.43 μm to 0.86 μm. After 12 noisy bands are removed, data from the remaining 103 bands are used for experiments. PU includes nine classes of ground objects, such as asphalt, meadows, gravel, etc. The image size of PU is 610 × 340. After removing background pixels in the image, a total of 42,776 labeled samples could be used for classification. Figure 4b is the true label image of PU. Table 3 lists the specific classes of PU and the number of labeled samples of each class.
SA was captured by the American AVIRIS sensor in the Salinas Valley in California. It had a spatial resolution of 3.7 m. This image had 224 bands originally. Removing the bands that could not be reflected by water, the remaining 204 bands were generally used for classification. SA included 16 classes of ground objects, such as fallow, celery, etc. It contained 111,104 pixels in total, of which 56,975 pixels were background pixels, and 54,129 pixels could be used for classification. The image size of SA was 512 × 217. Figure 4c is the true label image of SA. Table 4 lists the specific classes of SA and the number of labeled samples of each class.
HU was captured by the ITRES CASI-1500 sensor with a spatial resolution of 2.5 m, provided by the 2013 IEEE GRSS Data Fusion Contest. HU contained 144 bands ranging from 364 nm to 1046 nm. HU included 15 classes of ground objects such as trees, soil, residential, etc. After removing the background pixels in the image, a total of 15,268 labeled samples could be used for classification. Figure 4d is the true label image of HU. Table 5 lists the specific classes of HU and the number of labeled samples of each class.

3.2. Experimental Settings

In order to analyze the effect of different factors on the classification performance, we conducted a series of experiments, including: (1) the difference between two-branch network and single-branch network; (2) the influence of the number of superpixels in ERS and the patch size of input samples; (3) the difference of multi-spectral methods based on two kinds of superpixel segmentation methods, ERS and SLIC; (4) the different effect of segmentation images obtained by applying ERS to images on multiple bands and the first principal component; and (5) a comparison between TBN-MERS and other methods.
In multi-spectral ERS, the number of superpixels in each band nc is set to 50 for IP and SA and 200 for PU and HU. In ERS, the trade-off parameter α is set to 0.5. For IP, PU, and HU datasets, 50 samples are randomly selected from each class (10 samples are selected from the class with less than 50 samples) to form the training set. For SA, five samples are randomly selected from each class to form the training set. For each dataset, the number of randomly selected samples in the validation set is the same as that of the training set. The remaining samples are used to form the test set. In TBN-MERS, the training batch size is set to 32, and SGD optimizer is adopted. The learning rate is set to 0.0005. In SVM, only the spectral vector corresponding to the local center pixel is taken as the input. In TBN-MERS, Net-X, and Net-S (in Section 3.3.1), the patch size p of samples from IP, PU, and SA is set to 5, and the patch size p of samples from HU is set to 7. To achieve the best effect, the patch size of samples in HybridSN and 3DCNN is set to 25. The hyper-parameters of SSRN and SuperPCA are set as described in the original papers. We carry out all the methods for comparison, and all the experimental results in this paper are the averages of five result values obtained by independent runs for five times on RTX TITAN.
Three commonly used metrics are adopted in this paper to evaluate the classification results of HSIs obtained by different methods: Overall Accuracy (OA), Average Accuracy (AA), and Kappa coefficient (Kappa). OA represents the proportion of correctly classified samples to all samples. AA represents the average value of classification accuracy of each class. Kappa represents the degree of consistency between the classification result and the true labels. The larger the three metrics are, the better the classification performance is.
Suppose n is the total number of samples, m is the total number of classes, N i is the total number of samples of the actual class i, N i is the total number of samples predicted as class i, and C i i represents the number of samples whose true class is i and is predicted to be class i; then, OA can be expressed by Formula (8):
O A = 1 n i = 1 m C i i
AA can be expressed by Formula (9):
AA = 1 m i = 1 m C i i N i
Kappa can be expressed by Formula (10):
K a p p a = P o P e 1 P e
where P o = 1 n i = 1 m C i i , P e = 1 n 2 i = 1 m N i × N i .

3.3. The Results and Analyses of Experiments

3.3.1. The Difference between Two-Branch Network and Single-Branch Network

In order to prove that the two-branch neural network plays an important role in feature extraction and feature fusion, we designed two single-branch networks (Net-X and Net-S) for comparison. Net-X means that only preprocessed HSI is used for feature extraction and classification, and Net-S means that only preprocessed MERSI is used for feature extraction and classification, whereas TBN-MERS performs feature extraction, feature fusion, and classification using both the preprocessed HSI and MERSI. The hyper-parameter settings of Net-X and Net-S are the same as those of TBN-MERS. The experimental results are shown in Table 6.
From Table 6, we can see that the contribution of HSI and MERSI to the final classification results is quite different on different datasets. On IP and SA datasets, Net-S performs much better than Net-X, indicating that MERSI contains more useful information of these two datasets for classification. As for PU and HU datasets, Net-X is a little inferior to Net-S on PU dataset but is much superior to Net-S on HU dataset, indicating that the features extracted from HSIs are also important for the final classification accuracy. Compared to Net-X and Net-S, TBN-MERS obtains the best classification accuracy on all the four datasets, showing the necessity of the feature fusion. TBN-MERS obtains more advanced features and complementary information by feature fusion, which thus improves the final classification results greatly. In a word, the proposed two-branch network TBN-MERS is effective in fusing the features of both the preprocessed HSI and MERSI, which is thus more efficient and robust in classification.

3.3.2. The Influence of the Number of Superpixels and the Patch Size

There are two important parameters in the proposed TBN-MERS: the number of superpixels nc and patch size p. nc is a hyper-parameter in ERS, which determines how many superpixels ERS divides the image into. If nc is large, it means that TBN-MERS performs a fine segmentation on the image and the size of each superpixel is small, obtaining detailed information of the image. On the contrary, if nc is small, it means that the size of each superpixel is big, and TBN-MERS pays more attention to the large-scale features, such as the consistency in a wide area. The patch size p is the size of patch-based neighborhoods of the input samples. In brief, nc determines the richness of superpixel-scale spatial information, and p determines the richness of neighborhood-scale spatial information. We conduct experiments on different combinations of nc and p and plot the results in Figure 5, where different colors indicate the results corresponding to different p.
From Figure 5, it can be seen that TBN-MERS with a small nc and a small p achieves better OA results on IP. Such results are reasonable. It can be observed from Figure 4a that IP has good consistency in large areas. A smaller nc means that the features are extracted from the superpixels with a larger size, and the pixels in the same superpixel have better consistency, whereas a larger nc makes TBN-MERS extract the features from the smaller superpixels, which will destroy the internal homogeneousness of the area. On the other hand, superpixel-scale information has a more important effect on the classification of IP, as shown in Section 3.3.1, where Net-S performs much better than Net-X on IP. Therefore, if a larger p is used, it may introduce redundant information on the neighborhood scale and will interfere in the feature extraction and keep the model from extracting the useful features well. In short, a small nc and a small p are good choices for IP. For PU, the reverse is true. TBN-MERS with a larger nc can achieve better classification results. When nc is large, a larger p can further achieve better classification results. It can be observed from Figure 4b that the scene of PU is relatively complex, and the distribution of the ground objects is irregular, so a larger nc (corresponding to a smaller size of superpixel) enables the model to extract fine-grained superpixel-scale information, and a larger p enables the model to extract richer neighborhood-scale information. For SA, which is homogeneous in large areas, a smaller nc and a suitable p achieve better classification results, which shows that the superpixel-scale information is more important for the classification of SA. For HU, the classification effect is not sensitive to the parameter changes, and many suitable nc and p can lead to good classification results.

3.3.3. The Difference of Multi-Spectral Methods Based on Two Kinds of Superpixel Segmentation Methods, ERS and SLIC

TBN-MERS introduces the superpixel-scale spatial information in each band using the results of ERS segmentation as the priori information, which significantly improves the final classification results. How much does the final classification results depend on the effect of the superpixel segmentation? How will the proposed model of TBN-MERS perform if ERS is replaced by another superpixel segmentation method? We will investigate these issues in this section. As introduced in Section 2.1, there are many superpixel segmentation methods, among which both SLIC and ERS have excellent segmentation effect. We design a variant of TBN-MERS by replacing ERS with SLIC in the model of TBN-MERS, which is named as TBN-MSLIC. The comparison results between TBN-MSLIC and TBN-MERS are shown in Table 7. From Table 7, it can be seen that both TBN-MSLIC and TBN-MERS can achieve good performance with higher classification accuracy than that of Net-X in Section 3.3.1, indicating that the model of TBN-MERS is effective and can obtain higher classification accuracy than the single-branch network, which only uses the spectral information and the spatial information at the scale of patch-based neighborhood, even if ERS in TBN-MERS is replaced by other superpixel segmentation method. In addition, TBN-MERS performs better than TBN-MSLIC, especially on IP and SA datasets with large homogeneous regions. According to the analyses shown in Section 3.3.1, the superpixel segmentation image contains more useful information of IP and SA datasets for classification. Therefore, the model using the better superpixel segmentation method will achieve better classification performance on IP and SA datasets. Figure 6 shows the results of superpixel segmentation obtained by SLIC and ERS on the 103rd band of PU set, respectively. The number of superpixels in Figure 6a,b is both 200. From Figure 6, it can be seen that both SLIC and ERS can extract the superpixel-scale spatial information. For example, both SLIC and ERS can segment the contour of the parking area at bottom left. However, there are also some differences. The boundary obtained by SLIC is smoother, without too many local details, whereas the boundary obtained by ERS is finer, and both the overall outlines and local details are well-described. In summary, the model of TBN-MERS is effective, where ERS can obtain better superpixel segmentation results than SLIC, enabling TBN-MERS to achieve good classification results on different datasets.

3.3.4. The Different Effects of Segmentation Images Obtained by Applying ERS to Images on Multiple Bands and the First Principal Component

Some existing methods [27,28] used to extract the spectral–spatial information within each superpixel for classification by performing ERS on the first principal component of HSIs. However, different bands contain not only different spectral information but also different spatial information; the spectral–spatial information can hardly be sufficiently utilized by only choosing the first principal component. Figure 7 shows the results segmented by ERS on the first principal component and some bands of PU. Since each region in the image has different radar reflectivity on different bands, it can be seen from Figure 7 that the superpixel segmentation results obtained by ERS on different bands have significant differences. Comparing with the true label image of PU, the result of the first principal component seems to have good effect at first glance, but a different conclusion could be drawn after a careful comparison. For instance, for the parking area at the bottom left of the image (marked by a big rectangle in Figure 7a,c), the segmentation result of the 51st band is better. For the circular building areas (marked by a small rectangle in Figure 7a,d,e), the 77th and 103rd bands have a better segmentation effect. It can also be noticed that the contours of each area on most bands are roughly similar, but there are differences in local details. This exactly reflects the difference between different bands. The segmentation result of the first principal component can obtain the segmentation result of the overall contour, which contains the mainly superpixel-scale spatial information. The obtained superpixels of different bands are overlapped and interlaced, which not only contain a wealth of superpixel-scale spatial information, but also include very fine local spatial information of different regions in different bands.
In order to prove that the segmentation image obtained by multi-spectral ERS contributes to better classification results, we designed another variant of TBN-MERS, which is denoted by PC-TBN-MERS. The details of PC-TBN-MERS are as follows. The dimension of preprocessed HSI is reduced by principal component analysis, and the first principal component is retained. ERS is applied on the first principal component to obtain a superpixel segmentation result. The superpixel segmentation result is copied and stacked to form a data cube until it has the same size as that of the HSI. We take the obtained data cube and the preprocessed HSI as the input of two-branch neural network. Therefore, the only difference between PC-TBN-MERS and TBN-MERS lies in that the former only performs ERS on the first principal component (noted as FPC-ERS), and the latter performs ERS on all the bands (i.e., multi-spectral ERS). The comparison results are shown in Table 8.
From the metrics in Table 8, it can be seen that TBN-MERS has achieved much better performance than PC-TBN-MERS on IP and SA datasets, demonstrating that multi-spectral ERS is necessary. In addition, as shown in Figure 4, IP and SA datasets have several large homogeneous regions. This result also indicates that FPC-ERS is able to extract details of the image, but it is not good at finding the homogeneousness in a large area. However, multi-spectral ERS is good at both aspects, making TBN-MERS obtain better classification results on all kinds of tested datasets.

3.3.5. The Effect of Using Different Fusion Methods in TBN-MERS

From the experimental results in Section 3.3.1, it can be seen that the proposed two-branch network TBN-MERS is effective in fusing the features of both the preprocessed HSI and MERSI, where the data fusion is performed after the flatten layer. If the data fusion is performed at earlier stages, could the final classification accuracy be higher or not? We will investigate this issue in this section.
We designed two variants of TBN-MERS. In the feature extraction process, the features of the same depth layer from different branches are fused through skip connections, including the introduction of MERSI features into the HIS feature extraction process, marked as TBN(M2H)-MERS, whose network structure is shown in Figure 8a, and the introduction of HIS features into MERSI feature extraction process, marked as TBN(H2M)-MERS, whose network structure is shown in Figure 8b. Table 9 lists the corresponding experimental results. From Table 9, it can be found that neither TBN(M2H)-MERS nor TBN(H2M)-MERS achieves better classification results than TBN-MERS.
Why does information fusion at earlier stages not lead to better classification results? In our opinion, the reason is as follows. The data of HSIs contain the values of the electromagnetic reflection intensity of ground objects, and the data of MERSI contain the label values obtained by superpixel segmentation. The data distribution of HSI and MERSI is quite different. Therefore, if the feature fusion is carried out at earlier stages of the network, it will cause interference, which is not conducive to learning key features. In contrast, after the flatten layer, the network has extracted the advanced features of HSI and MERSI, and the features at this time represent the essential information, where the feature fusion will achieve better classification results. Therefore, we choose the structure of TBN-MERS to implement the information fusion, which is more concise and more effective.

3.3.6. Comparison between TBN-MERS and Other Methods

We compare TBN-MERS with five existing methods, including SVM [16], 3DCNN [24], SSRN [25], HybridSN [26], and SuperPCA [28]. SVM only takes the spectral features as input and can achieve satisfactory classification results even with limited samples. It has the advantages of good robustness and good generalization, so it is often used as a benchmark for classification of HSIs. Archibald discussed the application of SVM in hyperspectral image classification in 2007 [16]. We refer to the settings in that paper [16] and train the SVM classifier based on radial basis function to classify HSIs according to the spectral features of HSIs. SSRN was proposed by Zhong et al. in 2017 [25]. SSRN combined multiple 3D convolutional layers into spatial feature extraction modules and spectral feature extraction modules according to the residual structure to extract spectral–spatial features. SSRN connected these modules in series to perform end-to-end feature extraction and classification. In 2018, Hamida et al. [24] studied the classification effect of 3D convolutional networks with different numbers and different structures on HSIs. We choose the 3DCNN network structure consisting of four layers of 3D convolution and a fully connected layer proposed by Hamida for comparison. HybridSN was proposed by Roy et al. in 2019 [26]. It firstly used multiple 3D convolutional layers to extract the joint spectral–spatial features of HSIs, then used 2D convolution to extract the spatial information of the feature map, and finally, adopted a multi-layer fully connected layer to perform classification. It not only ensured the feature extraction ability of the model, but also reduced the model parameters. Jiang et al. proposed SuperPCA in 2018 [28]. In SuperPCA, the first principal component of HSI was segmented by multi-scale ERS, and then the principal component analysis (PCA) was applied to reduce the dimension of HSIs within the obtained superpixels. Subsequently, SuperPCA trained classifiers on the reduced data at each scale and obtained the final classification results through decision fusion.
The results and analyses of six different methods on four datasets are as follows. In the following tables, the best values are marked in bold, and the second best values are marked with underlines for the convenience of the readers. The numbers in parentheses are the standard deviation.
(1) Results and analyses of IP: Table 10 shows the OA, AA, and Kappa of the six different methods on IP. From Table 10, it can be seen that TBN-MERS achieves an overall classification accuracy of 98.13%, which is increased by 26.27%, 15.97%, 9.75%, 4.36%, and 3.07% over SVM, 3DCNN, SSRN, HybridSN, and SuperPCA, respectively. Except that the classification accuracy of TBN-MERS in class 3 is slightly lower than that of HybridSN, TBN-MERS has achieved the highest classification accuracy in other classes. Both TBN-MERS and SuperPCA use ERS for feature extraction, and TBN-MERS and HybridSN use a similar network structure for feature extraction. Yet, TBN-MERS has achieved higher classification accuracy, showing that the use of multi-spectral ERS and the structure of the two-branch neural network both contribute to the final results.
The samples in different classes in IP are seriously unbalanced. For instance, the samples of class 1, class 7, and class 9 are very few, the number of which is 46, 28, and 20, respectively. When the number of training set samples is limited, the imbalance of the samples in different categories may make it difficult for the model to learn the features of the classes with limited samples. As we can see in Table 10 for SSRN, the classification accuracy values of these three classes are zero. In class 1, 3DCNN, HybridSN, SuperPCA and TBN-MERS achieve good classification results. In class 7 and class 9, only HybridSN and TBN-MERS achieve good classification results with the value of 100%. These results demonstrate that TBN-MERS is able to learn the features of classes with limited samples effectively, indicating it has the characteristic of good generalization and robustness.
Figure 9 shows the predicted label images of the six different methods on IP. From Figure 9, we can see that there is a certain amount of salt and pepper noise in the results of SVM and 3DCNN, whereas the results of HybridSN, SSRN, SuperPCA, and TBN-MERS are relatively close to the true label image. On the whole, the visual effect of the classification obtained by TBN-MERS is the best among all the compared methods.
(2) Results and analyses of PU: Table 11 shows the OA, AA, and Kappa of six different methods on PU. From Table 11, it can be seen that TBN-MERS achieves an overall classification accuracy of 99.86%, which is increased by 17.13%, 11.48%, 4.78%, 4.32%, and 6.62% over SVM, 3DCNN, SSRN, HybridSN, and SuperPCA, respectively. In each class, TBN-MERS has achieved the highest classification accuracy. Both HybirdSN and SSRN directly use hyperspectral data as input. SuperPCA uses the superpixel segmentation results of the first principal component for feature extraction. Compared with these methods, TBN-MERS has achieved a significant improvement in classification accuracy. For PU, the features extracted by the multi-spectral ERS and the combined spectral–spatial information extracted by TBN-MERS play an important role in improving the classification accuracy. Figure 10 shows the predicted label images of the six different methods. From Figure 10, we can see that SVM, 3DCNN, and SuperPCA have more misclassifications, whereas HybridSN, SSRN, and TBN-MERS results are very close to the true label image. On the whole, the visual effect of the classification obtained by TBN-MERS is the most satisfactory among all the compared methods.
(3) Results and analyses of SA: Table 12 shows the OA, AA, and Kappa of six different methods on SA. In particular, due to the relatively simple scenarios of SA and the continuous distribution of ground objects, the classification of SA is easier to a degree compared with other datasets. To fully explore the classification performance of different methods, in the phase of training, only five training samples are randomly selected for each class. From Table 12, it can be seen that TBN-MERS achieves an overall classification accuracy of 99.31%, which is increased by 28.79%, 16.84%, 18.46%, 8.49%, and 25.67% over SVM, 3DCNN, SSRN, HybridSN, and SuperPCA, respectively. As considering a certain class, TBN-MERS has not achieved the highest classification accuracy in many classes. The accuracy of TBN-MERS is lower than 3DCNN in class 5, class 6, and class 16. It is lower than SSRN in class 9. It is lower than HybridSN in class 14 and class 16. It is lower than SuperPCA in class 15. However, the classification accuracy gap is very small in these classes, whereas the accuracy of TBN-MERS far exceeds other methods in other classes, such as class 8 and class 10. This shows that TBN-MERS has stronger generalization ability and better robustness for different types of ground objects on SA. Compared with other datasets, SA has the characteristics of concentrated ground object distribution, relatively regular shape of ground objects, high ground object consistency, and high consistency within the sample class, so it is especially suitable for multi-spectral ERS. The obtained superpixels well mark the characteristics of the distribution of ground objects. Even if the samples are limited and the neighborhood-scale spatial information is limited, TBN-MERS can also classify the test samples well by fully extracting the superpixel-scale spatial information of SA and filtering out the most representative features for classification. Figure 11 shows the predicted label images of the six different methods. From Figure 11, we can see that the predicted label images of SVM, 3DCNN, SSRN, HybridSN, and SuperPCA all have many pixels that are wrongly classified, whereas the result of TBN-MERS is almost the same as the true label image. On the whole, the visual effect of the classification obtained by TBN-MERS is the best among all the compared methods.
(4) Results and analyses of HU: Table 13 shows the OA, AA, and Kappa of the six different methods on HU. HU scene is relatively complex, the ground objects are small, and the distribution is not continuous, which certainly introduce some difficulty for the classification. From Table 13, it can be seen that TBN-MERS achieves an overall classification accuracy of 97.52%, which is increased by 39.65%, 14.16%, 4.75%, 3.06%, and 5.68% over SVM, 3DCNN, SSRN, HybridSN, and SuperPCA, respectively. As considering a certain class, the accuracy of TBN-MERS is lower than HybridSN in class 5 and class 6 and lower than SSRN in class 8. TBN-MERS has achieved the highest classification accuracy in other classes. For HU, SSRN and HybridSN have achieved relatively high classification accuracy, indicating that the neighborhood-scale spatial information has an important effect on the classification results. SuperPCA also has achieved good classification accuracy, indicating that the superpixel-scale spatial information has an important effect on the classification accuracy. TBN-MERS has achieved the highest classification accuracy, indicating that the two-branch network fully combines the superpixel-scale and neighborhood-scale spatial information through feature fusion. Figure 12 shows the predicted label images of the six different methods. From Figure 12, we can see that some misclassified samples can be seen in the predicted label images of SVM and 3DCNN, whereas SSRN, HybridSN, SuperPCA, and TBN-MERS are relatively similar with the true label image. On the whole, the visual effect of the classification obtained by TBN-MERS is the most satisfactory among all the compared methods.
(5) Comparison about the training loss of deep learning methods on different datasets: For the deep learning methods including 3DCNN, SSRN, HybridSN, and our TBN-MERS, we compare their convergence curves of training loss on different datasets, as shown in Figure 13. Figure 13a–d show the training loss curves of four deep learning methods on IP, PU, SA, and HU. In Figure 13, we can see that 3DCNN, HybridSN, and TBN-MERS converge faster, and their final loss is smaller than SSRN. The convergence of TBN-MERS in the training process is not the fastest, and there are certain fluctuations in the training process, but in the end, the highest classification accuracy can be achieved on the testing set, which illustrates that the two-branch structure and the feature fusion of superpixel-scale and neighborhood-scale make TBN-MERS more robust and have better generalization performance.
(6) Comparison about the training time of several deep learning methods: To compare the training time of several deep learning methods, further experiments are conducted, and the results are shown in Table 14. TBN-MERS needs to implement the superpixel segmentation first and then carry out the training stage, so we record the segmentation time, training time, and total time. From Table 14, in terms of training time, the training time of TBN-MERS is the smallest, because it uses a smaller patch size, making it require less computation. However, its segmentation time is relatively long, so the total time is not the minimal. At the same time, comparing the segmentation time of different datasets, we can find that the segmentation time is longer for datasets with larger size and more channels. This cost is acceptable, as the final classification results are improved greatly, and the increasing of the total time is limited.

4. Discussion

It can be seen from the above results that the performance of different methods on different datasets is different, but TBN-MERS always achieves best classification accuracy with limited samples. The main reason why TBN-MERS can achieve highest classification accuracy is as follows. On the one hand, we apply ERS to each band of the preprocessed HSI. Each superpixel segmentation image obtained contains the spatial distribution of ground objects in the superpixel scale. The segmentation results of different bands are different, and the obtained superpixels on different bands are crossing and overlapping, which reflects the difference of the spectral information contained in different bands indirectly. That is to say, the superpixel segmentation images not only directly contain rich spatial information but also indirectly contain spectral information. In addition, in the superpixel segmentation images, the values of all pixels within the same superpixel are all equal to the serial number of the superpixel (i.e., the number in the range of 1 to nc), which is essentially a filtering and denoising process for HSIs. Therefore, a single-branch network that uses only preprocessed segmentation images (i.e., Net-S in Section 3.3.1) can also achieve good classification accuracy. On the other hand, the proposed two-branch network not only utilizes superpixel-scale information in the HSI, but also extracts lots of detailed spectral–spatial information at the neighborhood-scale from the preprocessed HSI. The feature fusion designed in TBN-MERS further improves the classification accuracy and improves the robustness and generalization of TBN-MERS.
From the results of above experiments, we also find some limitations of TBN-MERS. Compared with several other deep learning methods, TBN-MERS mainly relies on the extra features provided by multi-spectral ERS to improve the classification accuracy. The time cost of multi-spectral ERS increases as the size of dataset becomes larger, which will affect the efficiency of TBN-MERS to a certain extent. Fortunately, such extra time cost is limited, as shown in Table 14, which is worthwhile, considering the significant improvement in classification results.

5. Conclusions

This paper proposed a two-branch convolutional neural network based on multi-spectral entropy rate superpixel segmentation (TBN-MERS) for hyperspectral image (HSI) classification. The multi-spectral entropy rate superpixel (ERS) segmentation, i.e., performing ERS segmentation on each band of the HSI, enables TBN-MERS to extract rich spectral–spatial information at the superpixel scale with the effect of filtering and denoising, which significantly improves the final classification accuracy. The feature fusion implemented by a two-branch network in TBN-MERS, which utilize both the superpixel-scale information and the spectral–spatial information at the neighborhood-scale in the HSI, improves the classification accuracy further, and improves the robustness and generalization of TBN-MERS.
On the whole, TBN-MERS can achieve good classification results on different kinds of datasets with limited training samples. For datasets with large homogeneous areas, TBN-MERS is able to extract such homogeneousness within the same regions effectively by using the multi-spectral ERS segmentation to extract superpixel-scale information and obtains high classification accuracy finally. For datasets with fragmentary and dispersed ground objects, where the superpixel-scale information plays a less important role, TBN-MERS can fully mine the neighborhood-scale information and fuse it with the superpixel-scale information to ensure the high classification accuracy. In summary, both the multi-spectral ERS segmentation and the feature fusion contribute to the excellent classification results of TBN-MERS. TBN-MERS is designed with a concise and flexible frame, and its classification performance could be improved further if a better spectral–spatial feature extraction scheme is adopted in the first branch or a segmentation algorithm with performance better than ERS is applied in the second branch. In the future, we will try to design deep neural networks to implement unsupervised or semi-supervised image segmentation, which may help to improve the classification results further.

Author Contributions

Conceptualization, Z.D. and C.M.; methodology, Z.D. and C.M.; software, Z.D.; investigation, Z.D.; data curation, Y.L.; writing—original draft preparation, Z.D.; writing—review and editing, C.M. and Y.L.; visualization, Z.D. and Y.L.; supervision, C.M.; funding acquisition, C.M. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62077038, 61672405, 61871306, 62176196, 61836009 and 61871310.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available at http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes#Indian_Pines (accessed on 16 February 2022) (Indian Pines), http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes#Pavia_University_scene (accessed on 16 February 2022) (Pavia University), http://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes#Salinas_scene (accessed on 16 February 2022) (Salinas), and https://hyperspectral.ee.uh.edu/?page_id=459 (accessed on 16 February 2022) (Houston).

Acknowledgments

The authors would like to thank the Hyperspectral Image Analysis group and the NSF Funded Center for Airborne Laser Mapping (NCALM) at the University of Houston for providing the data sets used in this study, and the IEEE GRSS Data Fusion Technical Committee for organizing the 2013 Data Fusion Contest.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Plaza, A.; Benediktsson, J.A.; Boardman, J.W.; Brazile, J.; Bruzzone, L.; Camps-Valls, G.; Chanussot, J.; Fauvel, M.; Gamba, P.; Gualtieri, A. Recent advances in techniques for hyperspectral image processing. Remote Sens. Environ. 2009, 113, S110–S122. [Google Scholar] [CrossRef]
  2. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. An augmented linear mixing model to address spectral variability for hyperspectral unmixing. IEEE Trans. Image Processing 2018, 28, 1923–1938. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Yao, J.; Meng, D.; Zhao, Q.; Cao, W.; Xu, Z. Nonconvex-sparsity and nonlocal-smoothness-based blind hyperspectral unmixing. IEEE Trans. Image Processing 2019, 28, 2991–3006. [Google Scholar] [CrossRef] [PubMed]
  4. Zhong, P.; Wang, R. Multiple-spectral-band CRFs for denoising junk bands of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2260–2275. [Google Scholar] [CrossRef]
  5. Li, W.; Prasad, S.; Fowler, J.E.; Bruce, L.M. Locality-preserving dimensionality reduction and classification for hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2011, 50, 1185–1198. [Google Scholar] [CrossRef] [Green Version]
  6. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  7. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Processing Mag. 2013, 31, 45–54. [Google Scholar] [CrossRef] [Green Version]
  8. Li, J.; Zhang, H.; Huang, Y.; Zhang, L. Hyperspectral image classification by nonlocal joint collaborative representation with a locally adaptive dictionary. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3707–3719. [Google Scholar] [CrossRef]
  9. Plaza, A.; Martínez, P.; Plaza, J.; Pérez, R. Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations. IEEE Trans. Geosci. Remote Sens. 2005, 43, 466–479. [Google Scholar] [CrossRef] [Green Version]
  10. Zhang, H.; Li, J.; Huang, Y.; Zhang, L. A nonlocal weighted joint sparse representation classification method for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 2056–2065. [Google Scholar] [CrossRef]
  11. Cao, X.; Yao, J.; Xu, Z.; Meng, D. Hyperspectral image classification with convolutional neural network and active learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4604–4616. [Google Scholar] [CrossRef]
  12. Chakrabarti, A.; Zickler, T. Statistics of real-world hyperspectral images. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR) 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 193–200. [Google Scholar]
  13. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum. IEEE Trans. Image Processing 2010, 19, 2241–2253. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Green, R.O.; Eastwood, M.L.; Sarture, C.M.; Chrien, T.G.; Aronsson, M.; Chippendale, B.J.; Faust, J.A.; Pavri, B.E.; Chovit, C.J.; Solis, M. Imaging spectroscopy and the airborne visible/infrared imaging spectrometer (AVIRIS). Remote Sens. Environ. 1998, 65, 227–248. [Google Scholar] [CrossRef]
  15. Bruzzone, L.; Cossu, R. A multiple-cascade-classifier system for a robust and partially unsupervised updating of land-cover maps. IEEE Trans. Geosci. Remote Sens. 2002, 40, 1984–1996. [Google Scholar] [CrossRef]
  16. Archibald, R.; Fann, G. Feature selection and classification of hyperspectral images with support vector machines. IEEE Geosci. Remote Sens. Lett. 2007, 4, 674–677. [Google Scholar] [CrossRef]
  17. Shao, Z.; Zhang, L.; Zhou, X.; Ding, L. A novel hierarchical semisupervised SVM for classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1609–1613. [Google Scholar] [CrossRef]
  18. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification via kernel sparse representation. IEEE Trans. Geosci. Remote Sens. 2012, 51, 217–231. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  20. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  21. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef] [Green Version]
  22. Yue, J.; Mao, S.; Li, M. A deep learning framework for hyperspectral image classification using spatial pyramid pooling. Remote Sens. Lett. 2016, 7, 875–884. [Google Scholar] [CrossRef]
  23. Cao, X.; Zhou, F.; Xu, L.; Meng, D.; Xu, Z.; Paisley, J. Hyperspectral image classification with Markov random fields and a convolutional neural network. IEEE Trans. Image Processing 2018, 27, 2354–2367. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef] [Green Version]
  25. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  26. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  27. Leng, Q.; Yang, H.; Jiang, J.; Tian, Q. Adaptive multiscale segmentations for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5847–5860. [Google Scholar] [CrossRef]
  28. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A superpixelwise PCA approach for unsupervised feature extraction of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef] [Green Version]
  29. Mei, X.; Pan, E.; Ma, Y.; Dai, X.; Huang, J.; Fan, F.; Du, Q.; Zheng, H.; Ma, J. Spectral-spatial attention networks for hyperspectral image classification. Remote Sens. 2019, 11, 963. [Google Scholar] [CrossRef] [Green Version]
  30. Mu, C.; Zeng, Q.; Liu, Y.; Qu, Y. A two-branch network combined with robust principal component analysis for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2020, 18, 2147–2151. [Google Scholar] [CrossRef]
  31. Varga, D. Composition-preserving deep approach to full-reference image quality assessment. Signal Image Video Processing 2020, 14, 1265–1272. [Google Scholar] [CrossRef]
  32. Saenz, H.; Sun, H.; Wu, L.; Zhou, X.; Yu, H. Detecting phone-related pedestrian distracted behaviours via a two-branch convolutional neural network. IET Intell. Transp. Syst. 2021, 15, 147–158. [Google Scholar] [CrossRef]
  33. Mori, G.; Ren, X.; Efros, A.A.; Malik, J. Recovering human body configurations: Combining segmentation and recognition. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 27 June–2 July 2004; p. II. [Google Scholar]
  34. Ren, X.; Malik, J. Learning a classification model for segmentation. In Proceedings of the Proceedings Ninth IEEE International Conference on Computer VisionComputer Vision, Nice, France, 13–16 October 2003; p. 10. [Google Scholar]
  35. Kohli, P.; Torr, P.H. Robust higher order potentials for enforcing label consistency. Int. J. Comput. Vis. 2009, 82, 302–324. [Google Scholar] [CrossRef] [Green Version]
  36. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Liu, M.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR) 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  38. Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef] [Green Version]
  39. Vincent, L.; Soille, P. Watersheds in digital spaces: An efficient algorithm based on immersion simulations. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 583–598. [Google Scholar] [CrossRef] [Green Version]
  40. Shi, J.; Malik, J. Normalized cuts and image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar]
  41. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient graph-based image segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
  42. Levinshtein, A.; Stere, A.; Kutulakos, K.N.; Fleet, D.J.; Dickinson, S.J.; Siddiqi, K. Turbopixels: Fast superpixels using geometric flows. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 2290–2297. [Google Scholar] [CrossRef] [Green Version]
  43. Veksler, O.; Boykov, Y.; Mehrani, P. Superpixels and supervoxels in an energy optimization framework. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; pp. 211–224. [Google Scholar]
Figure 1. The process of the multi-spectral entropy rate superpixel segmentation.
Figure 1. The process of the multi-spectral entropy rate superpixel segmentation.
Remotesensing 14 01569 g001
Figure 2. The network structure of TBN-MERS (Conv3D represents the cascade of 3D convolutional layer, BatchNorm layer and ReLU layer; Conv2D represents the cascade of 2D convolutional layer, BatchNorm layer and ReLU layer; FC represents the cascade of fully connected layer and dropout layer; fc represents the cascade of the fully connected layer and the Softmax layer).
Figure 2. The network structure of TBN-MERS (Conv3D represents the cascade of 3D convolutional layer, BatchNorm layer and ReLU layer; Conv2D represents the cascade of 2D convolutional layer, BatchNorm layer and ReLU layer; FC represents the cascade of fully connected layer and dropout layer; fc represents the cascade of the fully connected layer and the Softmax layer).
Remotesensing 14 01569 g002
Figure 3. The overall process of TBN-MERS.
Figure 3. The overall process of TBN-MERS.
Remotesensing 14 01569 g003
Figure 4. True label images of the four datasets (a) IP, (b) PU, (c) SA, and (d) HU.
Figure 4. True label images of the four datasets (a) IP, (b) PU, (c) SA, and (d) HU.
Remotesensing 14 01569 g004
Figure 5. OA obtained on four datasets with different number of superpixels and patch size (a) IP, (b) PU, (c) SA, and (d) HU.
Figure 5. OA obtained on four datasets with different number of superpixels and patch size (a) IP, (b) PU, (c) SA, and (d) HU.
Remotesensing 14 01569 g005
Figure 6. The segmentation image obtained by SLIC and ERS on the 103rd band of PU (a) SLIC (b) ERS.
Figure 6. The segmentation image obtained by SLIC and ERS on the 103rd band of PU (a) SLIC (b) ERS.
Remotesensing 14 01569 g006
Figure 7. The segmentation results of ERS on different bands of PU: (a) the first principal component, (b) the 1st band, (c) the 51st band, (d) the 77th band, and (e) the 103rd band.
Figure 7. The segmentation results of ERS on different bands of PU: (a) the first principal component, (b) the 1st band, (c) the 51st band, (d) the 77th band, and (e) the 103rd band.
Remotesensing 14 01569 g007
Figure 8. Network structures of two variants of TBN-MERS: (a) TBN(M2H)-MERS and (b) TBN(H2M)-MERS.
Figure 8. Network structures of two variants of TBN-MERS: (a) TBN(M2H)-MERS and (b) TBN(H2M)-MERS.
Remotesensing 14 01569 g008
Figure 9. The predicted label images of different methods on IP: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Figure 9. The predicted label images of different methods on IP: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Remotesensing 14 01569 g009
Figure 10. The predicted label images of different methods on PU: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Figure 10. The predicted label images of different methods on PU: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Remotesensing 14 01569 g010
Figure 11. The predicted label images of different methods on SA: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) Super-PCA, and (f) TBN-MERS.
Figure 11. The predicted label images of different methods on SA: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) Super-PCA, and (f) TBN-MERS.
Remotesensing 14 01569 g011
Figure 12. The predicted label images of different methods on HU: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Figure 12. The predicted label images of different methods on HU: (a) SVM, (b) 3DCNN, (c) SSRN, (d) HybridSN, (e) SuperPCA, and (f) TBN-MERS.
Remotesensing 14 01569 g012aRemotesensing 14 01569 g012b
Figure 13. Training loss curves of four deep learning methods on different datasets: (a) IP, (b) PU, (c) SA, and (d) HU.
Figure 13. Training loss curves of four deep learning methods on different datasets: (a) IP, (b) PU, (c) SA, and (d) HU.
Remotesensing 14 01569 g013
Table 1. The hyper-parameters of the TBN-MERS.
Table 1. The hyper-parameters of the TBN-MERS.
LayersKernel SizeNumber of KernelsPadding
Cov3D_1(3,3,7)81
Cov3D_2(3,3,5)161
Cov3D_3(3,3,3)321
Cov2D(3,3)641
FC_1--256--
FC_2--128--
fc--Number of classes--
Table 2. The information of each class in IP.
Table 2. The information of each class in IP.
NumberClassSamplesColor
1Alfalfa46
2Corn-notill1428
3Corn-mintill830
4Corn237
5Grass-pasture483
6Grass-trees730
7Grass-pasture-mowed28
8Hay-windrowed478
9Oats20
10Soybean-notill972
11Soybean-mintill2455
12Soybean-clean593
13Wheat205
14Woods1265
15Buildings-Grass-Trees-Drives386
16Stone-Steel-Towers93
Total 10,249
Table 3. The information of each class in PU.
Table 3. The information of each class in PU.
NumberClassSamplesColor
1Asphalt6631
2Meadows18,649
3Gravel2099
4Trees3064
5Painted metal sheets1345
6Bare Soil5029
7Bitumen1330
8Self-Blocking Bricks3682
9Shadows947
Total 42,776
Table 4. The information of each class in SA.
Table 4. The information of each class in SA.
NumberClassSamplesColor
1Brocoli_green_weeds_12009
2Brocoli_green_weeds_23726
3Fallow1976
4Fallow_rough_plow1394
5Fallow_smooth2678
6Stubble3959
7Celery3579
8Grapes_untrained11,271
9Soil_vinyard_develop6203
10Corn_senesced_green_weeds3278
11Lettuce_romaine_4wk1068
12Lettuce_romaine_5wk1927
13Lettuce_romaine_6wk916
14Lettuce_romaine_7wk1070
15Vinyard_untrained7268
16Vinyard_vertical_trellis1807
Total 54,129
Table 5. The information of each class in HU.
Table 5. The information of each class in HU.
NumberClassSamplesColor
1Healthy grass1251
2Stressed grass1254
3Synthetic grass732
4Trees1244
5Soil1242
6Water339
7Residential1268
8Commercial1244
9Road1252
10Highway1227
11Railway1288
12Parking Lot 11233
13Parking Lot 2531
14Tennis Court463
15Running Track700
Total 15,268
Table 6. Classification results of Net-X, Net-S, and TBN-MERS.
Table 6. Classification results of Net-X, Net-S, and TBN-MERS.
NetworksMetricsIPPUSAHU
Net-XOA (%)81.0294.6385.3194.68
AA (%)87.7094.3391.7495.39
Kappa (×100)78.4492.8883.5994.25
Net-SOA (%)97.5995.9097.8990.58
AA (%)98.9796.3798.6591.90
Kappa (×100)97.2394.5897.6689.83
TBN-MERSOA (%)98.1399.7499.3597.51
AA (%)99.0199.7099.3197.88
Kappa (×100)97.8599.6699.2897.31
Table 7. The comparison results of TBN-MSLIC and TBN-MERS.
Table 7. The comparison results of TBN-MSLIC and TBN-MERS.
MethodsMetricsIPPUSAHU
TBN-MSLICOA (%)95.0199.2095.1696.59
AA (%)97.4899.5596.5597.13
Kappa (×100)94.2898.9494.6296.31
TBN-MERSOA (%)98.1399.7499.3597.51
AA (%)99.0199.7099.3197.88
Kappa (×100)97.8599.6699.2897.31
Table 8. Classification results of PC-TBN-MERS and TBN-MERS.
Table 8. Classification results of PC-TBN-MERS and TBN-MERS.
MethodsMetricsIPPUSAHU
PC-TBN-MERSOA (%)86.5897.8093.7996.28
AA (%)91.9197.7895.4096.68
Kappa (×100)84.7297.0893.1095.98
TBN-MERSOA (%)98.1399.7499.3597.51
AA (%)99.0199.7099.3197.88
Kappa (×100)97.8599.6699.2897.31
Table 9. Classification results of TBN(M2H)-MERS, TBN(H2M)-MERS, and TBN-MERS.
Table 9. Classification results of TBN(M2H)-MERS, TBN(H2M)-MERS, and TBN-MERS.
MethodsMetricsIPPUSAHU
TBN(M2H)-MERSOA (%)97.7497.7499.1294.81
AA (%)98.9998.2998.9495.72
Kappa (×100)97.4197.0199.0294.39
TBN(H2M)-MERSOA (%)93.6298.9497.3496.34
AA (%)96.4199.3097.1496.87
Kappa (×100)92.7098.6097.0496.05
TBN-MERSOA (%)98.1399.7499.3597.51
AA (%)99.0199.7099.3197.88
Kappa (×100)97.8599.6699.2897.31
Table 10. Classification results of different methods on IP with 50 samples per class.
Table 10. Classification results of different methods on IP with 50 samples per class.
Class NumberSVM [16]3DCNN [24]SSRN [25]HybridSN [26]SuperPCA [28]TBN-MERS
166.1199.370.00100.00100.00100.00
263.9178.0376.7187.8093.7698.21
364.1770.6486.2796.7689.1094.64
484.3861.16100.0099.3595.19100.00
590.7188.9594.6797.5097.0098.33
692.7993.5297.4196.9195.2999.94
785.5598.880.00100.0085.71100.00
895.2397.60100.00100.0099.53100.00
988.0094.840.00100.0090.00100.00
1072.1464.1388.2193.5581.1397.22
1157.0490.1187.3189.4785.0796.92
1270.0173.2374.1194.9193.7498.96
1398.5891.9599.31100.0099.35100.00
1483.0994.7198.0197.9997.20100.00
1570.4777.0499.9497.9198.81100.00
1697.6777.4099.05100.0097.83100.00
OA (%)71.86
(0.34)
82.16
(1.47)
88.38
(1.58)
93.77
(1.22)
95.06
(1.24)
98.13
(0.21)
AA (%)79.99
(0.93)
84.47
(1.58)
75.06
(1.07)
97.01
(0.55)
96.70
(1.00)
99.01
(0.09)
Kappa
(×100)
68.22
(0.41)
79.72
(1.65)
86.73
(1.79)
92.88
(1.38)
94.32
(1.42)
97.85
(0.24)
Table 11. Classification results of different methods on PU with 50 samples per class.
Table 11. Classification results of different methods on PU with 50 samples per class.
Class NumberSVM [16]3DCNN [24]SSRN [25]HybridSN [26]SuperPCA [28]TBN-MERS
177.3495.2792.5587.3370.6699.74
280.6294.7898.4098.3880.9399.89
379.5966.6797.6294.7493.7099.50
495.0893.3070.3894.1080.2398.95
599.3299.4299.6799.8496.60100.00
679.9775.0099.7599.8687.73100.00
793.0663.1897.5199.9392.97100.00
884.8684.2199.3889.0091.8899.53
999.8696.0868.3493.73100.00100.00
OA (%)82.73
(1.78)
88.38
(0.99)
95.08
(1.27)
95.54
(0.71)
93.24
(0.67)
99.86
(0.07)
AA (%)87.75
(0.42)
85.32
(0.94)
91.51
(1.00)
95.21
(0.70)
94.42
(0.37)
99.77
(0.08)
Kappa
(×100)
77.78
(2.08)
84.69
(1.25)
93.47
(1.66)
94.09
(0.94)
91.10
(0.85)
99.64
(0.09)
Table 12. Classification results of different methods on SA with 5 samples per class.
Table 12. Classification results of different methods on SA with 5 samples per class.
Class NumberSVM [16]3DCNN [24]SSRN [25]HybridSN [26]SuperPCA [28]TBN-MERS
198.3573.4499.6798.65100.00100.00
281.5097.1395.1797.3273.53100.00
337.4499.4896.1497.5487.27100.00
497.9998.5293.8083.9470.2799.98
593.9299.4880.0791.7351.6397.23
697.2099.9699.0997.9285.3199.39
799.0291.0392.9099.2572.97100.00
837.3166.5471.3285.4533.3798.95
997.9786.99100.0094.2053.3499.93
1024.3286.1682.3891.0366.2798.75
1183.5595.9973.9394.9396.3399.34
1296.1393.5562.5997.0078.1599.76
1398.7789.2749.9896.1195.2899.62
1488.4585.8994.7099.2671.7498.74
1557.5159.0052.2875.95100.0099.59
1671.6798.7877.8897.6774.9797.03
OA (%)70.52
(1.19)
82.47
(0.83)
80.85
(4.37)
90.82
(1.89)
73.64
(3.53)
99.31
(0.13)
AA (%)78.82
(0.81)
88.83
(1.01)
82.62
(5.78)
93.62
(0.89)
81.35
(2.46)
99.27
(0.13)
Kappa
(×100)
67.37
(1.29)
80.39
(0.92)
78.71
(4.88)
89.78
(2.12)
70.92
(3.65)
99.23
(0.15)
Table 13. Classification results of different methods on HU with 50 samples per class.
Table 13. Classification results of different methods on HU with 50 samples per class.
Class NumberSVM [16]3DCNN [24]SSRN [25]HybridSN [26]SuperPCA [28]TBN-MERS
184.8286.9186.7189.3491.5195.25
284.3079.0887.3993.3283.3199.13
398.9190.6097.7499.9799.71100.00
490.4682.4284.5894.6792.2998.29
586.6688.8699.7599.8997.1599.83
684.2966.3894.26100.0087.8997.92
721.6787.2188.6888.1182.9296.33
819.8877.2191.2285.3777.5590.78
982.0785.3988.5887.6581.9592.74
103.2182.9999.1799.0891.59100.00
1156.0783.7698.5097.4990.9599.75
122.1675.7493.8898.0378.3698.41
1310.2791.3894.9099.1272.9799.95
1496.6176.71100.00100.0096.85100.00
1599.0494.3798.05100.0098.92100.00
OA (%)57.87
(0.56)
83.36
(2.27)
92.77
(0.77)
94.46
(0.52)
91.84
(1.69)
97.52
(0.25)
AA (%)61.36
(0.63)
83.27
(1.97)
93.56
(0.73)
95.47
(0.43)
92.24
(1.39)
97.89
(0.20)
Kappa
(×100)
54.74
(0.61)
82.04
(2.45)
92.18
(0.83)
94.00
(0.56)
91.18
(1.83)
97.32
(0.27)
Table 14. Time for deep learning methods on different datasets.
Table 14. Time for deep learning methods on different datasets.
MethodsOperationIPPUSAHU
3DCNNTraining530 s121 s421 s299 s
SSRNTraining794 s257 s114 s566 s
HybridSNTraining 335 s184 s50 s536 s
TBN-MERSSegmentation121 s152 s230 s646 s
Training70 s60 s38 s114 s
Total time191 s212 s268 s760 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mu, C.; Dong, Z.; Liu, Y. A Two-Branch Convolutional Neural Network Based on Multi-Spectral Entropy Rate Superpixel Segmentation for Hyperspectral Image Classification. Remote Sens. 2022, 14, 1569. https://doi.org/10.3390/rs14071569

AMA Style

Mu C, Dong Z, Liu Y. A Two-Branch Convolutional Neural Network Based on Multi-Spectral Entropy Rate Superpixel Segmentation for Hyperspectral Image Classification. Remote Sensing. 2022; 14(7):1569. https://doi.org/10.3390/rs14071569

Chicago/Turabian Style

Mu, Caihong, Zhidong Dong, and Yi Liu. 2022. "A Two-Branch Convolutional Neural Network Based on Multi-Spectral Entropy Rate Superpixel Segmentation for Hyperspectral Image Classification" Remote Sensing 14, no. 7: 1569. https://doi.org/10.3390/rs14071569

APA Style

Mu, C., Dong, Z., & Liu, Y. (2022). A Two-Branch Convolutional Neural Network Based on Multi-Spectral Entropy Rate Superpixel Segmentation for Hyperspectral Image Classification. Remote Sensing, 14(7), 1569. https://doi.org/10.3390/rs14071569

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop