Next Article in Journal
Hardware-Aware Design of Speed-Up Algorithms for Synthetic Aperture Radar Ship Target Detection Networks
Next Article in Special Issue
MADANet: A Lightweight Hyperspectral Image Classification Network with Multiscale Feature Aggregation and a Dual Attention Mechanism
Previous Article in Journal
Hybrid Cross-Feature Interaction Attention Module for Object Detection in Intelligent Mobile Scenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Images Weakly Supervised Classification with Noisy Labels

Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150080, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(20), 4994; https://doi.org/10.3390/rs15204994
Submission received: 16 August 2023 / Revised: 9 October 2023 / Accepted: 14 October 2023 / Published: 17 October 2023
(This article belongs to the Special Issue Recent Advances in the Processing of Hyperspectral Images)

Abstract

:
The deep network model relies on sufficient training samples to achieve superior processing performance, which limits its application in hyperspectral image (HSI) classification. In order to perform HSI classification with noisy labels, a robust weakly supervised feature learning (WSFL) architecture combined with multi-model attention is proposed. Specifically, the input noisy labeled data are first subjected to multiple groups of residual spectral attention models and multi-granularity residual spatial attention models, enabling WSFL to refine and optimize the extracted spectral and spatial features, with a focus on extracting clean samples information and reducing the model’s dependence on labels. Finally, the fused and optimized spectral-spatial features are mapped to the multilayer perceptron (MLP) classifier to increase the constraint of the model on the noisy samples. The experimental results on public datasets, including Pavia Center, WHU-Hi LongKou, and HangZhou, show that WSFL is better at classifying noise labels than excellent models such as spectral-spatial residual network (SSRN) and dual channel residual network (DCRN). On Hangzhou dataset, the classification accuracy of WSFL is superior to DCRN by 6.02% and SSRN by 7.85%, respectively.

1. Introduction

Hyperspectral remote sensing uses a large number of narrow electromagnetic wave channels to obtain spatial, radiation, and spectral triple information of ground objects, which obtains information about ground objects through the band range of visible light and infrared light in the electromagnetic spectrum [1,2]. Due to the characteristic of “combination of image and spectrum” in hyperspectral images, they contain a much higher degree of ground information. By fully utilizing this feature, accurate classification of ground objects can be achieved [3]. Therefore, hyperspectral remote sensing has been widely applied in urban planning [4], environmental monitoring [5], and precision agriculture [6,7].
Early hyperspectral image classification mostly used supervised classification methods, whose performance relied on high-quality labels [8]. Pal et al. mapped hyperspectral data to a high-dimensional space, and found an optimal segmentation hyperplane in the space to maximize the distance between different types and achieve the best classification effect [9]. Cui et al. considered the relationship between hyperspectral data classes, effectively combining a sparse representation classifier and K-nearest neighbor to increase classification accuracy [10].
Due to intra-class complexity and the scarcity of labeled samples, it is challenging to achieve high-precision ground object classification only through spectral features [11]. Gu et al. proposed a multi-kernel learning (MKL) architecture to learn spectral and spatial information, combined with a support vector machine (SVM) [12]. Liu et al. proposed a multi-morphic superpixel method to extract spectral and spatial features and complete the classification task [13]. The extracted features of these methods require manual adjustment, which is a complex process.
Recently, owing to the powerful automatic feature extraction ability of deep learning, great progress has been made in supervised learning algorithms. With the local feature extraction capability of convolutional neural networks (CNN) [14,15], 2DCNN [16,17] and 3DCNN [18,19] have achieved impressive results in spectral, spatial, and spectral-spatial feature extraction. The use of supervised classification methods to process hyperspectral images requires manual annotation of data. In the process of obtaining labeled data, due to the high complexity of terrain, the lack of on-site surveys by legal persons, and the subjective influence of human factors on the quality of annotated data labels, the existence of noise in labels cannot be avoided; that is, the labels do not match their actual categories [20]. When the labels in the training data are incomplete, inaccurate, or only partially labeled, this is a weakly supervised learning method [21].
At present, there are roughly two methods for dealing with noise labeling problems, one of which is noise removal. For example, Kang et al. utilized domain transformation recursive filtering to enhance the distinguishability of spectral-spatial features and used a constrained energy minimization strategy to correct noisy samples [22]. Tu et al. proposed a spatial density peak clustering algorithm that utilizes a local density strategy to obtain inter-sample density values, where noisy samples are removed through decision functions [23,24]. Fang et al. used a confidence learning framework to accurately detect label errors [25].
The noise samples also have rich spatial-spectral information, which is also beneficial and can be effectively utilized. Another method is to design a more robust network model that directly utilizes a training set with incorrect labels [26,27]. Sukhbaatar et al. introduced a constrained linear “noise” layer on the softmax layer, so that the network output can simulate the noise label distribution and significantly improve the performance of the deep network [28]. Jiang et al. learned noisy labels through a small loss strategy, which indicates that deep neural networks will use small loss samples as “clean” samples and only use such samples for feedback propagation to update network parameters [29]. These methods have excellent noise processing capabilities, but their ability to handle complex noisy samples is limited, which limits the further improvement of classification accuracy. Xu et al. proposed a simple and efficient dual channel residual network (DCRN), which, respectively, extracted more refined features from the spectral and spatial dimensions of hyperspectral data to reduce the impact of noisy labels on the model to achieve an excellent training effect [30].
The presence of noisy samples brings many uncertainties to hyperspectral image classification, which limits the performance of many network models. However, due to the scarcity of hyperspectral image data, even noise samples cannot be easily discarded, otherwise a lot of information will be lost. In addition, deep learning models have the problem of easily memorizing clean samples in the early stages and gradually memorizing noisy samples in the later stages. As the number of iterations increases, the model gradually begins to fit noisy samples, which leads to the model continuously learning the features of noisy samples in later training and ignoring more important clean samples. Finally, there is a supervised learning model that is highly dependent on correctly labeled data, which will lead to the influence of noise samples being equal to clean samples. How to reduce the dependence of the model on noisy samples and further improve the performance of weakly supervised classification models has become a research hotspot in hyperspectral image classification.
Therefore, this paper designs a novel weakly supervised feature learning (WSFL) classification model for processing HSI with noisy samples. In order to better utilize noisy samples and preserve the diversity of features, this method adaptively learns features through multi-model attention feature learning. By comparing the similarity between samples, it obtains clean samples’ features with higher information content, reducing the weight of noisy samples and the impact of noisy samples on the model. Secondly, in order to reduce the ability of the model to fit noisy samples in the later stage, multiple sets of residual spectral attention models were designed in the spectral dimension. The spectral features in the later stage were differentiated in the space of multiple sets of spectral features to avoid excessive concentration of single layer spectral features and memory of noisy samples. In addition, in order to obtain more high-quality features and reduce the dependence of the model on samples, a multi-granularity residual spatial attention model was designed in the spatial dimension. In the multi-granularity space, the spatial features were further refined to obtain finer spatial features. Finally, in order to eliminate the adverse effects of local connectivity in the model and focus on the spatial structure information of more data, a MLP model was introduced, with a focus on learning spectral-spatial features to enhance the overall model’s feature capture ability. The main contributions are summarized as follows:
  • This paper proposes a weakly supervised feature learning architecture combined with multi-model attention, which can build a more robust network that can classify noisy samples more stably and accurately;
  • In order to enhance the constraint of spectral dimension on noisy samples, multiple sets of residual spectral attention models were designed to enhance the ability to learn clean samples and weaken the model’s fitting ability for noisy samples;
  • In order to improve the utilization of clean samples in weakly supervised models, a multi-granularity residual spatial attention model was designed to gradually extract clean sample information from spatial dimensions and obtain more significant features;
  • We introduced a MLP model to further extract spectral-spatial features, eliminate the adverse effects of local connectivity of the model, pay more attention to the spatial structure information of the data, and improve the overall model’s anti-interference ability against noise.

2. Methodology

In this section, we will introduce the main architecture of WSFL in detail as shown in Figure 1, including multi-group residual spectral attention model (MGRSAM), multi-granularity residual spatial attention model (MRSAM), MLP model, noise loss function and Lion optimizer. In addition, samples labeled incorrectly are called noisy samples, and samples labeled correctly are called clean samples. First, the 3D data cube is input to MGRSAM and MRSAM to extract spectral and spatial features. In MGRSAM, the first two convolutional layers are used to perform coarse feature extraction on the spectral dimension. Subsequently, the extracted features are mapped to multiple sets of spectral feature spaces through the Group Convolution (GConv) layer to reduce the model’s ability to fit noise samples. In addition, the features of the first layer are mapped into this space by means of skip connections, which solves the problem of gradient descent due to the increase in network depth. Secondly, the output features are mapped to the spectral feature attention space, focusing on extracting clean samples’ features and suppressing the influence of noisy samples.

2.1. Spectral and Spatial Feature Extraction

In order to improve the robustness and generalization ability of image classification with noisy labels, this paper addresses two aspects separately. On the one hand, in response to the fact of neural networks easily remembering clean samples in the early stage and gradually remembering noisy samples in the later stage, this paper designs multiple sets of residual spectral attention models in the spectral dimension of hyperspectral data. In the early training of the spectral dimension, rough extraction of spectral dimension features is performed to obtain higher quality feature maps and enhance noise resistance. Secondly, in the later training, in order to avoid the model overfitting the features of noisy samples, the input features are mapped to multiple sets of spectral feature spaces, and the later spectral features are processed in a grouping manner. Each set of spectral features is finely extracted to avoid a single layer of spectral features being too concentrated, thus memorizing the noisy samples. Secondly, while reducing the fitting of noise samples, the ability to fit clean samples is strengthened, and more clean spectral features are learned through the spectral attention model.
On the other hand, a multi-granularity residual spatial attention model is designed in the spatial dimension of HSI to solve the problem that supervised learning is too dependent on labeled samples and easy to overfit noisy samples. In early training of spatial dimensions, learning more discriminative spatial features through the spatial attention space weakens the spatial feature weights of noisy samples. Secondly, we map the input features to a multi-granularity space, extract important features in the spatial domain, obtain the similarity features of a large number of positive and negative sample pairs, mine the feature representation information of the dataset, obtain predictive tags, and reduce the dependence of supervised learning on labels. These two parts will be introduced in detail as follows.

2.1.1. Spectral Feature Extraction

This paper carefully designed the network architecture in the spectral dimension of hyperspectral data. In the early training of spectral dimension, the focus is on obtaining higher quality feature maps to improve the anti-noise ability of spectral dimension in the early stage. Secondly, in order to prevent the model from overfitting noisy samples in the later training, different channels are grouped to avoid excessive concentration of noise features in one layer, reducing the ability of spectral dimension to overfitting noisy samples in the later stage. However, this approach also reduces the ability to fit clean samples. To achieve this, spectral feature attention space is used to focus on more discriminative clean sample features among numerous features, while suppressing unnecessary noise information and enhancing the spectral dimension’s ability to fit clean samples in the later stage. Multiple residual spectral attention models are composed of convolutional layers, spectral feature attention spaces, multiple spectral feature spaces, and residual blocks, as shown in Figure 2.
In the early training process of spectral dimension, we use a 1 × 1 × 7 convolutional layer for coarse feature extraction. The step of the first convolutional kernel is (1, 1, 2) to remove information redundancy in adjacent bands, allowing the model to focus on more important spectral features and maintain the original spatial correlation, improving the early noise resistance of the spectral dimension model.
In the later stage of spectral dimension training, when the concatenated features of the spectral dimension are mapped to multiple sets of spectral feature spaces, different channels are grouped to prevent noise features from being too concentrated in one layer, and to avoid fitting noisy samples in that layer and affecting the overall training of the spectral dimension in the later stage. Multiple spectral feature spaces can reduce the ability of later models to fit noisy samples. In addition, feature extraction for each group of spectral features can also better explore spectral information and enhance the noise resistance of multiple spectral feature spaces.
As shown in Figure 3, there are multiple groups of spectral feature spaces, where Group = 3 and the size of each group of feature maps is H × W × C 1 / 3 , which corresponds to the height, width, and number of channels. The size of each group of convolution kernels is h 1 × w 1 × C 1 / 3 , which corresponds to the height, width, and number of channels of the convolution kernels. Convolution is performed in the corresponding group, and the output features are obtained by stacking the output features.
Although the fitting ability of noisy samples can be effectively reduced in multiple sets of feature spectral spaces, it also reduces the fitting ability of clean samples. In order to enhance the feature capture ability of clean samples in the later stage of spectral dimension, the features Z c output from multiple sets of spectral feature spaces are mapped to the spectral attention space. In this space, different features have different weights, and the allocation of weights is based on similarity. Features with high similarity are considered clean sample features, while features with low similarity are considered noisy sample features.
Firstly, we transform Z c from top to bottom to obtain K c , Q c , and V c representing the key vector, query vector, and numerical vector, respectively. The subscript c represents the channel attention module, P × P is the spatial dimension with a channel count of 48. We calculate and compare the similarity between K c and Q c , as shown in Formula (1).
f c = K c Q c T
Then, the Softmax classifier is used to obtain the pixel weight matrix W c , where W c ( i , j ) represents the similarity of pixel i to pixel j.
W c ( i , j ) = e f ( i , j ) j = 1 p × p e f ( i , j )
The spectral attention feature is obtained by multiplying V c and W c T , and the spectral attention feature obtained by weighting V c with W c T is a discriminative feature.
A c = W c T V c
Finally, we transform the channel attention feature into A c * , so that its dimension is the same as the input feature, and add the spectral attention feature to the later training, as shown in Formula (4).
Z c * = A c * + Z c

2.1.2. Spatial Feature Extraction

For spatial feature extraction, residual networks can effectively extract spatial features and prevent overfitting. However, the presence of noise samples can easily lead to a portion of the noise sample features being transmitted to the lower layer after each jump connection. In order to obtain more significant semantic features, noise information around the target pixel is suppressed at different spatial positions, highlighting clean sample features, improving the model’s efficiency in feature utilization, and reducing the model’s dependence on annotated data.
This paper designs a network architecture in the hyperspectral spatial dimension. In the early training of the spatial dimension, the attention space of spatial features is used to generate a weight value for each pixel in the input patch. This is done to suppress the negative impact of noisy samples on feature extraction and thereby strengthen spatial texture features. The size of the patch is 7 × 7. Compared to smaller domains, larger domains mean that the input contains more spatial information, which will also increase the number of noisy samples. Hence, relying solely on spatial feature attention space cannot completely reduce the interference of noise samples in spatial dimensions. Therefore, in the later training of the spatial dimension, each layer of feature maps is separated from one another through multi-granularity space, and each feature map is further subdivided into 3 × 3 regions, where multi-granularity refers to the processing and extraction of features in feature maps at different levels. In a multi-granularity space, emphasis is placed on the ground feature information within the granularity to obtain more discriminative spatial features and further enhance the feature capture ability for clean samples. The multi-granularity residual spatial attention model consists of convolutional layers, spatial feature attention spaces, multi-granularity spaces, and residual blocks, with an architecture shown in Figure 4.
In early training of the spatial dimensions, 3D convolution is first used to map hyperspectral data to a multi-granular residual spatial attention model with only one band, and the convolution kernel size is 1 × 1 × 102. Then the coarse feature extraction is performed using the first layer of convolution, and the domain patch size of 7 × 7 is mapped to the attention space of spatial features. In order to better process spatial information, both the second and third convolutional layers use a 2D convolution size of 3 × 3 extracts of spatial features, and after each convolution, they are cascaded with a BatchNorm (BN) layer and a ReLU layer.
Firstly, a convolution operation is performed on input Z s to obtain three feature maps, namely K s , Q s , and V s from top to bottom, where the subscript s represents the weight of the convolutional layer for spatial attention modules W K , W Q , and W V , respectively. B K , B Q , and B V represent bias terms. The three feature maps are obtained as shown in Formula (5).
K s = C o n v ( Z s , W K ) + B K Q s = C o n v ( Z s , W Q ) + B Q V s = C o n v ( Z s , W V ) + B V
We transform three feature maps to obtain K s T , Q s , and V s , and multiply K s T by Q s to calculate the correlation of pixels in the spatial feature map, as shown in Formula (6).
f s = K s T Q s
Then, Softmax is used to obtain the pixel weight matrix W s , where W s ( i , j ) represents the impact of pixel i on pixel j. Similarly, a larger weight value indicates a stronger correlation between spatial pixels, as shown in Equation (7).
W S ( i , j ) = e f ( i , j ) j = 1 p × p e f ( i , j )
Subsequently, by multiplying V s and W s T to obtain spatial attention features, the spatial features with significant weight are more helpful in improving classification results, as shown in Formula (8).
A s = W s T V s
Finally, we transform the spatial attention feature into A S * , and add the spatial attention feature to the input until convergence, as shown in Formula (9).
Z S * = A S * + Z S
During the space dimension post training, we use residual blocks to concatenate the output of the first layer network with the output Z s * of the spatial attention feature space in a spatial dimension of 7 × 7, which is divided into 3 × 3 regions in a multi-granularity space. The region forms different particles, with varying levels of information contained within each particle. The purpose of multi-granularity is to reduce the concentration of noisy sample features in a certain part of the feature map, thereby affecting the feature information of adjacent clean samples. The multi-granularity space is shown in Figure 5. After multi-granularity, the features exist in the form of particles, weakening the interference of other particles and re-extracting more significant feature information from each particle. The particle size is the size of the convolution kernel in deep convolution. Assuming that the feature map of the multi-granularity residual spectral attention model is I R w × h × c , deep convolution divides the feature map into several semantic markers of different granularity. w is the width of the feature map and h is the height of the feature map; c is the number of bands. Ti can be obtained by Formula (10).
Ti = DWConv2Di(I′), i = 1, 2, …
i represents the ith granularity branch, and DWConv2D represents a two-dimensional deep convolution operation. By setting the size of deep convolution, the granularity can be adjusted.

2.2. Spectral-Spatial Feature Extraction

At present, most algorithms for noise label classification only use a single network model for classification, which has lower binding force on noise labels and higher misjudgment rate compared to multiple network models. Based on the above reasons, after redesigning the architecture of spectral and spatial dimensions, this article also uses the MLP model to further obtain spectral-spatial features for the fused spectral-spatial features. The MLP model is cascaded with multiple sets of residual spectral attention models and multi-granularity residual spatial attention models to form a weakly supervised feature learning architecture, and noise labels are constrained by different models in different dimensions.
As a neural network with fewer constraints, MLP can eliminate the adverse effects of local connectivity, thus enabling the model to have strong discrimination ability for small differences in the local field of view, effectively extract deep features, achieve accurate acquisition of spectral-spatial structure information, and further reduce the interference of noisy samples in the model [31].Therefore, this paper introduces the MLP neural network as the final model for processing spectral-spatial dimensions.
In this section, first, the concat function is used to simultaneously integrate spectral and spatial information of the data, integrating dimension 128 × 7 × 7 spectral information and dimension 24 × 7 × 7 spatial information, resulting in dimension 152 × 7 × 7 feature maps that combine spectral and spatial information, followed by the use of the average pooling layer size of 7 × 7 to reduce the size of feature maps while maintaining spatial information, thereby reducing the number of parameters that need to be optimized in the network, resulting in a vector size of 152 × 1. Finally, the vector is input to an MLP composed of the full connection layer, the GELU activation function, and the Dropout layer, and propagates forward to complete the final classification. Through the multi-layer perceptron classifier, the spectral-spatial dimension feature information can be further obtained, and the feature information of the noise label can be constrained to the greatest extent.
Next, MLP will be introduced as shown in Figure 6, which consists of three parts: full connection layer (FC), the GELU activation function, and the Dropout layer, in which the layers are fully connected. By introducing the GELU activation function to process data, when the input is negative, the input will be mapped to a non-zero value, so as to avoid the problem that some neurons of the ReLU activation function are invalid, and retain the characteristic information of the model in the negative signal, increasing the learning ability of MLP models for small differences within local features. In addition, by randomly discarding the values of 0.1% of neurons through the Dropout layer, overfitting of the model is avoided.

2.3. Lion Optimization

Hyperspectral data contain noise samples, which makes each batch of training have many confounding data points. If a small training batch is used, when the number of noise samples extracted from the batch is greater than the number of clean samples, the model will not be able to fully learn the features of clean samples. Therefore, we increase the number of batch sizes and the number of learnable samples per batch during each training phase. However, the currently popular AdamW optimizers often apply small batch sizes.
A d a m W : = m t = β 1 × m t 1 + ( 1 β 1 ) × g t v t = β 2 × v t 1 + ( 1 β 2 ) × g t 2 m ^ t = m t / ( 1 β 1 t ) v ^ t = v t / ( 1 β 2 t ) u t = m ^ t / ( v ^ t +   ε ) + λ t θ t 1 θ t = θ t 1 η t μ t
Compared to AdamW and various adaptive optimizers that require both first-order and second-order moments to be saved simultaneously, Lion only requires momentum, reducing the additional memory footprint by half, which will be beneficial for training large models and batch sizes [32]. Therefore, this paper introduces the Lion optimizer to simplify the process of parameter updating. Taking the tth iteration of gradient descent as an example, the Lion optimizer process is shown in Formula (15).
  L i o n : = u t = s i g n ( β 1 × m t 1 + ( 1 β 1 ) × g t ) + λ t θ t 1 θ t = θ t 1 η t μ t m t = β 2 × m t 1 + ( 1 β 2 ) × g t
When the input value is positive, sign is 1, and when the input value is negative, sign is −1. m t and v t are the first order momentum term and the second order momentum term, respectively, β 1 and β 2 are the default values of the hyperparameters with 0.9 and 0.99, the deviation correction values of m t and v t are m ^ t and v ^ t , and g t is the gradient of the loss function of the current sample.

3. Results

In order to verify the accuracy and efficiency of the proposed model, experiments were conducted on three datasets, and the model was evaluated using three evaluation criteria: overall accuracy (OA), average accuracy (AA), and Kappa coefficient. At the same time, this paper also studied the running time of each model to evaluate its efficiency.

3.1. The Description of Public HSI Datasets

In this paper, three widely used HSI datasets, including Pavia Center (PC) [33], WHU-Hi-LongKou (LK) [34], and HangZhou (HZ) [35], are employed in the experiments.
The Pavia Center dataset was captured by the ROSIS sensor during a flight campaign over Pavia, Northern Italy. It consists of 1906 × 715 pixels with a spatial resolution of 1.3m. After removing 13 bad bands, it has 102 bands(430~860nm). The ground truth contains nine classes representing a typical urban site. The WHU-Hi-LongKou dataset covers a simple agricultural area and was captured by an 8mm focal length steeple-wall Headwall Nano-HyperSpec sensor equipped with a receiver Matrix 600 Pro UAV platform with six kinds of crops. The image size was 550 × 400 pixels, with 270 bands ranging from 400 to 1000 nm. The Hangzhou dataset was obtained by the EO-1 Hyperion hyperspectral sensor, which kept 198 bands after removing 22 bad bands and 590 × 230 pixels. The false-color images and corresponding ground-truth maps of the three datasets can be seen in Table 1, Table 2 and Table 3.

3.2. Experimental Setting

The GPU server used in this article is manufactured by Finehoo Technology Co., Ltd., located in Shanghai, China. The Python version used is 3.7. The experimental environment was an Intel(R) Xeon(R) CPU E5-2620 v4 @ 2.10GHz processor, 128 GB running memory (RAM), and NVIDIA GeForce RTX 2080Ti GPU. In addition, the deep learning framework was Pytorch and Tensorflow. The initial learning rate was set to 0.00012, the Lion optimization algorithm was used for training models, and the number of iterations was set to 6000. At the same time, in order to reduce the randomness brought by training samples, each experiment was repeated 10 times and the average accuracy is taken as the final experimental result. In order to evaluate the effectiveness of the model in this paper, the model in this paper is compared with five other algorithms, including Depthwise separable neural network (DSNN) [36], 2DCNN [37], 3DCNN [38], SSRN [39], and the advanced classification network with noisy samples, dual channel residual network (DCRN) [30]. The learning rate of DSNN, 2DCNN, 3DCNN, SSRN, and DCRN was set to 0.001, the optimizer was AdamW, the early stop method was used to train the network, and the epoch was set to 4000.

3.3. Classification Results of Different Methods

Table 4, Table 5 and Table 6, respectively, show the classification results of different methods in 24 clean samples of each class for PC, LK, and HZ data. It can be found that the OA of WSFL is the highest, with 98.77%, 97.58%, and 80.14%, respectively. Taking HZ data as an example, WSFL increased by 0.62%, 0.95%, 3.74%, 5.63%, and 3.05% compared to DCRN, SSRN, 3DCNN, 2DCNN, and DSNN, respectively. In summary, it can be fully demonstrated that the WSFL model is still the best performing model without adding noise samples.

3.3.1. Results of PC Datasets with Different Numbers of Noise Samples

The results of using different methods to classify PC datasets are shown in Table 7. In the PC dataset, 24 samples were taken from each class of clean samples and four, eight, and 12 noisy samples were taken from each class to verify the processing ability of different deep learning models. It was found that WSFL had the best overall classification results, reaching 98.52%, 97.50%, and 96.77%, respectively. In addition, the number of training samples selected in this paper accounts for approximately 0.1944% of the total sample size. Compared with the approximately 3% training sample size required for other popular depth models, the sample size required in this paper is greatly reduced, which can also prove that the model proposed has a reduced dependence on labeled samples.
Table 7 shows that in 24 clean samples and four noisy samples, the OA of the proposed WSFL model reached 98.52%, which is 1.4%, 3.1%, and 9.43% higher than DCRN, SSRN, and 3DCNN, respectively. Among 24 clean samples and eight noisy samples, our OA reached 97.50%, which was 1.16%, 3.48%, and 17.52% higher than DCRN, SSRN, and 3DCNN, respectively. Among 24 clean samples and 12 noisy samples, our OA reached 96.77%, which was 1.33%, 6.57%, and 27.79% higher than DCRN, SSRN, and 3DCNN, respectively. In summary, it can be found that WSFL significantly improves OA under different noise sample sizes. Although the method proposed in this paper cannot achieve the best accuracy for each class, out of 24 clean samples and four noisy samples, seven classes are the best in this paper. Out of 24 clean samples and eight noise samples, six categories are the best category, and out of 24 clean samples and 12 noise samples, six categories are the best category. This can also prove that WSFL can better handle noisy samples compared to models such as DCRN, SSRN, and 3DCNN. As the number of noisy samples increases, the OA of WSFL decreases by only about 1%, which is acceptable as a multiple of the number of noisy samples. This fully demonstrates the effectiveness of WSFL in HSI classification tasks with noisy labels.
Finally, Figure 7 shows the pseudo-color images of the classification results of various classification methods in the PC dataset. False-color images, as a subjective evaluation indicator, can more intuitively display the classification effect. From Figure 7, it can be seen that WSFL has a significant improvement in classification performance compared to DCRN, SSRN, and 3DCNN. The area of misclassification is greatly reduced, and it is closer to the true distribution of ground objects.

3.3.2. Results of LK Datasets with Different Numbers of Noise Samples

The classification results of LK dataset are shown in Table 8. There are abundant annotated samples in each category of the LK dataset, so all methods perform well. Compared to other models, the proposed WSFL model achieved the best classification performance, showing significant improvements in most categories. Although the method proposed in this paper cannot achieve the best accuracy for each class, out of 24 clean samples and four noisy samples, as well as out of 24 clean samples and eight noisy samples, all six classes are considered the best class. Among the 24 clean samples and 12 noisy samples, seven were the optimal categories, which also proves that WSFL can better handle noisy samples compared to models such as DCRN, SSRN, and 3DCNN, fully demonstrating the effectiveness of WSFL in HSI classification tasks with noisy labels.
Finally, Figure 8 shows the pseudo-color images of various classification methods on LK dataset. Taking 24 clean + four noise as an example, the broad-leaf soybeans in the middle part have severe classification confusion. Compared with the DCRN, SSRN, and 3DCNN, the model proposed in this paper has fewer misclassification phenomena and is closer to the true distribution of ground objects.

3.3.3. Results of HZ Datasets with Different Numbers of Noise Samples

The HZ dataset has the characteristics of small inter-class differences and large intra-class differences. Therefore, the accuracy rates of various methods are relatively low, among which the indicators of WSFL are at the best, and OA reaches 79.44%, 72.90%, and 63.57%, respectively. The classification results of the HZ dataset are shown in Table 9.
Table 9 shows that in 24 clean samples and four noisy samples, the OA of this article reached 79.44%, which is 1.93%, 2.37%, and 4.72% higher than DCRN, SSRN, and 3DCNN, respectively. Among 24 clean samples and 12 noisy samples, our OA reached 63.57%, which was 1.26%, 2.39%, and 4.38% higher than DCRN, SSRN, and 3DCNN, respectively. For WSFL, out of 24 clean samples and four noisy samples, one class is the best class in this article. Among 24 clean samples and eight noise samples, as well as 24 clean samples and 12 noise samples, there are two optimal categories.
Finally, Figure 9 shows the pseudo-color images of the classification results on the HZ dataset. Compared to DCRN, SSRN, and 3DCNN, WSFL has the smallest staggered area of water.

3.4. The Numbers of Clean and Noisy Samples

In the classification with noisy labels, the number of clean samples is crucial, and even when there are few available clean samples, the proposed WSFL framework can perform well in classification. As shown in Table 10, Table 11 and Table 12, when the number of clean samples in each category is limited, the proposed WSFL still has relatively good performance. Compared with DCRN, SSRN, and 3DCNN, the WSFL model has a more robust network structure, resulting in a significant improvement in performance. In addition, as the number of noise samples increases, WSFL exhibits a slow performance decline. Therefore, compared to other methods, WSFL has higher stability.
Table 10, Table 11 and Table 12 show the impact of different numbers of clean and noisy samples on the PC, LK, and HZ datasets. Taking the PC dataset as an example, the OA of the WSFL model reached 98.52%, 97.50%, and 96.77%, respectively. Although the OA of WSFL is also decreasing, compared to other models this model has higher performance. When the number of noise samples is fixed, the value of OA continuously increases as the number of clean samples increases. By comparing different clean sample numbers and noise sample numbers, it can be found that when the clean sample number is 24 and the noise sample number is four, WSFL has the optimal indicator. In summary, the model in this paper has better processing ability when dealing with noisy samples, which can also prove that WSFL has a more robust network structure and stronger feature learning ability.

3.5. Investigation of Running Time

Table 13 gave the computation time comparison for three HSI datasets. In addition, compared to complex models such as DCRN and SSRN, WSFL is 4.47s faster than DCRN in PC datasets and 27.4s slower than SSRN. In summary, compared to single model neural networks such as DSNN, 2DCNN, 3DCNN, etc., WSFL has a significant improvement in performance despite being slower in time, and the model has stronger anti-interference ability against noise labels. Secondly, compared to complex models such as DCRN and SSRN, it ranks second on the PC and LK datasets and third on the HZ dataset. Considering the balance between accuracy and efficiency, WSFL as proposed in this paper is optimal.

4. Discussion

4.1. Effectiveness of the Attention Model

In order to verify the effectiveness of MGRSAM and MRSAM on WSFL, this paper conducted ablation experiments and compared the OA, AA, and Kappa coefficients of MGRSAM, MRSAM, and MGRSAM + MRSAM as shown in Table 14. It can be seen that the simultaneous presence of MGRSAM and MRSAM has indeed improved OA, AA, and Kappa on the three datasets. However, on the WHU-Hi-LongKou dataset, the AA value of MRSAM is 0.26% higher than that of MGRSAM and MRSAM, as the accuracy of a certain class of MRSAM is slightly higher than that of the final method. Although the AA value of MRSAM has increased on the WHU-Hi-LongKou dataset, compared by OA, Kappa, and overall, the coexistence of MGRSAM and MRSAM is superior.

4.2. Effectiveness of the MLP Model

In order to verify the effectiveness of the MLP model in the model proposed in this paper, ablation experiments were conducted as shown in Table 15. It can be observed that after adding the MLP model, OA, AA and Kappa show a significant increase.

4.3. Effectiveness of the Number of Groups on the Model

In this section, in order to verify the impact of the number of groups in MGRSAM on the WSFL model, this paper selects two, three, four, six, and eight groups for comparison. From Figure 10, it can be seen that when the number of groups is three, OA, AA, and Kappa reach their highest values. Although the AA of three groups in the PC dataset is slightly lower than the AA of two groups, the improvement in OA, Kappa, and the value of three in the LK and HZ data are all in the optimal solution, which is acceptable. In addition, when the group value is changed from three to six, the overall indicator shows a decline phenomenon, which is because the features are too scattered in the later training of the spectral dimension. Although the model’s ability to fit noisy labels is significantly reduced, it also reduces the ability to fit clean samples. Therefore, this paper selects three groups to train our model, in order to reduce the fitting of noisy samples while retaining the ability to fit clean samples.

5. Conclusions

In this article, we propose WSFL, a novel weakly supervised feature learning architecture with the core goal of exploring the robustness of the model to different noise levels. The uniqueness of WSFL lies in its specific feature learning strategy for noisy labels, where it can adaptively learn features through multi-model attention adaptive feature learning without removing noisy samples. This preserves the diversity of features and reduces the influence of noisy samples on the model.
In addition, different architectures have been designed based on the characteristics of hyperspectral data in spectral, spatial, and spectral-spatial dimensions. Compared with other methods, WSFL can effectively capture information in hyperspectral data and transform it into discriminative feature representations. Specifically, multiple sets of residual spectral attention models were carefully designed in the spectral dimension, which differentiated features through multiple sets of spectral feature spaces to avoid excessive concentration of single layer spectral features and memory of noisy samples. Secondly, more clean spectral features were learned in the spectral attention space. In addition, a multi-granularity residual spatial attention model has been carefully designed in the spatial dimension. In the spatial feature attention space, the similarity between samples is calculated to reduce the weight of noisy samples and improve the influence of clean samples. Then, the spatial features are refined in the multi-granularity space to obtain more discriminative spatial features, improving the quality of capturing spatial features and enhancing the model’s constraint on noise samples. Finally, the MLP model is introduced to eliminate the adverse effects of local connectivity in the model, obtaining more spatial structure information from the HSI dataset.
A large number of experimental results indicate that the framework proposed in this article surpasses state-of-the-art algorithms and can still achieve good accuracy even in the presence of a large number of noisy samples. Therefore, the architecture of this article is more suitable for HSI classification with noisy labels. The future work direction of this article is to apply the proposed framework to other hyperspectral images, rather than just processing the aforementioned open-source datasets, in order to enhance the universality of the model in practice.

Author Contributions

Conceptualization, C.L., H.W. and L.Z.; methodology, software, validation, C.L.; writing—review and editing, H.W., L.Z. and C.L.; supervision, H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the High-end Foreign Experts Introduction Program (G2022012010L) and Heilongjiang Natural Science Foundation Project (LH2023F034) and Reserved Leaders of Heilongjiang Provincial Leading Talent Echelon (2021).

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Paoletti, M.E.; Moreno-Álvarez, S.; Xue, Y.; Haut, J.M.; Plaza, A. AAtt-CNN: Automatic Attention-Based Convolutional Neural Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5511118. [Google Scholar] [CrossRef]
  2. Qi, W.; Huang, C.; Wang, Y.; Zhang, X.; Sun, W.; Zhang, L. Global–Local 3-D Convolutional Transformer Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5510820. [Google Scholar] [CrossRef]
  3. Tu, B.; Ren, Q.; Li, Q.; He, W.; He, W. Hyperspectral Image Classification Using a Superpixel–Pixel–Subpixel Multilevel Network. IEEE Trans. Geosci. Remote Sens. 2023, 72, 5013616. [Google Scholar] [CrossRef]
  4. Weber, C.; Aguejdad, R.; Briottet, X.; Avala, J.; Fabre, S.; Demuynck, J.; Zenou, E.; Deville, Y.; Karoui, M.S.; Benhalouche, F.Z.; et al. Hyperspectral Imagery for Environmental Urban Planning. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1628–1631. [Google Scholar]
  5. Yang, X.; Yu, Y. Estimating Soil Salinity Under Various Moisture Conditions: An Experimental Study. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2525–2533. [Google Scholar] [CrossRef]
  6. Jiao, Q.; Zhang, B.; Liu, J.; Liu, L. A novel two-step method for winter wheat-leaf chlorophyll content estimation using a hyperspectral vegetation index. Int. J. Remote Sens. 2014, 35, 7363–7375. [Google Scholar] [CrossRef]
  7. Liang, L.; Di, L.; Zhang, L.; Deng, M.; Qin, Z.; Zhao, S.; Lin, H. Estimation of crop LAI using hyperspectral vegetation indices and a hybrid inversion method. Remote Sens. Environ. 2015, 165, 123–134. [Google Scholar] [CrossRef]
  8. Yue, J.; Fang, L.; Ghamisi, P.; Xie, W.; Li, J.; Chanussot, J.; Plaza, A. Optical Remote Sensing Image Understanding with Weak Supervision: Concepts, methods, and perspectives. IEEE Geosci. Remote Sens. Mag. 2022, 10, 250–269. [Google Scholar] [CrossRef]
  9. Pal, M.; Foody, G.M. Feature Selection for Classification of Hyperspectral Data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef]
  10. Cui, M.; Prasad, S. Class-Dependent Sparse Representation Classifier for Robust Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2683–2695. [Google Scholar] [CrossRef]
  11. Zheng, J.; Feng, Y.; Bai, C.; Zhang, J. Hyperspectral Image Classification Using Mixed Convolutions and Covariance Pooling. IEEE Trans. Geosci. Remote Sens. 2021, 59, 522–534. [Google Scholar] [CrossRef]
  12. Gu, Y.; Liu, T.; Jia, X.; Benediktsson, J.A.; Chanussot, J. Nonlinear Multiple Kernel Learning With Multiple-Structure-Element Extended Morphological Profiles for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3235–3247. [Google Scholar] [CrossRef]
  13. Liu, T.; Gu, Y.; Chanussot, J.; Dalla Mura, M. Multimorphological Superpixel Model for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6950–6963. [Google Scholar] [CrossRef]
  14. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  15. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  16. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  17. Fang, L.; Liu, G.; Li, S.; Ghamisi, P.; Benediktsson, J.A. Hyperspectral image classification with squeeze multibias network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1291–1301. [Google Scholar] [CrossRef]
  18. Feng, J.; Chen, J.; Sun, Q.; Shang, R.; Cao, X.; Zhang, X.; Jiao, L. Convolutional Neural Network Based on Bandwise-Independent Convolution and Hard Thresholding for Hyperspectral Band Selection. IEEE Trans. Cybern. 2021, 51, 4414–4428. [Google Scholar] [CrossRef]
  19. Kanthi, M.; Sarma, T.H.; Bindu, C.S. A 3d-Deep CNN Based Feature Extraction and Hyperspectral Image Classification. In Proceedings of the 2020 IEEE India Geoscience and Remote Sensing Symposium (InGARSS), Virtual, 1–4 December 2020; pp. 229–232. [Google Scholar]
  20. Wang, L.; Zhu, T.; Kumar, N.; Li, Z.; Wu, C.; Zhang, P. Attentive-Adaptive Network for Hyperspectral Images Classification With Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5505514. [Google Scholar] [CrossRef]
  21. Li, Y.F.; Guo, L.Z.; Zhou, Z.H. Towards Safe Weakly Supervised Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 334–346. [Google Scholar] [CrossRef]
  22. Kang, X.; Duan, P.; Xiang, X.; Li, S.; Benediktsson, J.A. Detection and correction of mislabeled training samples for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5673–5686. [Google Scholar] [CrossRef]
  23. Tu, B.; Zhang, X.; Kang, X.; Wang, J.; Benediktsson, J.A. Spatial density peak clustering for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5085–5097. [Google Scholar] [CrossRef]
  24. Tu, B.; Zhou, C.; He, D.; Huang, S.; Plaza, A. Hyperspectral classification with noisy label detection via superpixel-to-pixel weighting distance. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4116–4131. [Google Scholar] [CrossRef]
  25. Fang, Z.; Yang, Y.; Li, Z.; Li, W.; Chen, Y.; Ma, L.; Du, Q. Confident Learning-Based Domain Adaptation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5527116. [Google Scholar] [CrossRef]
  26. Algan, G.; Ulusoy, I. Image classification with deep learning in the presence of noisy labels: A survey. Knowl.-Based Syst. 2021, 215, 106771–106790. [Google Scholar] [CrossRef]
  27. Roy, S.K.; Hong, D.; Kar, P.; Wu, X.; Liu, X.; Zhao, D. Lightweight heterogeneous kernel convolution for hyperspectral image classification with noisy labels. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5509705. [Google Scholar] [CrossRef]
  28. Sukhbaatar, S.; Bruna, J.; Paluri, M.; Bourdev, L.; Fergus, R. Training convolutional networks with noisy labels. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1406–1427. [Google Scholar]
  29. Jiang, L.; Zhou, Z.; Leung, T.; Li, L.-J.; Li, F.-F. Mentornet:Learning data-driven curriculum for very deep neural networks on corrupted labels. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 2304–2313. [Google Scholar]
  30. Xu, Y.; Li, Z.; Li, W.; Du, Q.; Liu, C.; Fang, Z.; Zhai, L. Dual-Channel Residual Network for Hyperspectral Image Classification with Noisy Labels. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502511. [Google Scholar] [CrossRef]
  31. Potghan, S.; Rajamenakshi, R.; Bhise, A. Multi-Layer Perceptron Based Lung Tumor Classification. In Proceedings of the 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 29–31 March 2018; pp. 499–502. [Google Scholar]
  32. Chen, X.; Liang, C.; Huang, D.; Real, E.; Wang, K.; Liu, Y.; Pham, H.; Dong, X.; Luong, T.; Hsieh, C.; et al. Symbolic Discovery of Optimization Algorithms. arXiv 2023, arXiv:2302.06675v4. [Google Scholar]
  33. Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Li, W.; Du, Q. Deep Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501618. [Google Scholar] [CrossRef]
  34. Zhong, Y.; Hu, X.; Luo, C.; Wang, X.; Zhao, J.; Zhang, L. WHU-Hi: UAV-borne hyperspectral with high spatial resolution (H2) benchmark datasets and classifier for precise crop identification based on deep convolutional neural network with CRF. Remote Sens. Environ. 2020, 250, 112012. [Google Scholar] [CrossRef]
  35. Deng, B.; Jia, S.; Shi, D. Deep Metric Learning-Based Feature Embedding for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1422–1435. [Google Scholar] [CrossRef]
  36. Gao, H.; Yang, Y.; Li, C.; Gao, L.; Zhang, B. Multiscale Residual Network With Mixed Depthwise Convolution for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3396–3408. [Google Scholar] [CrossRef]
  37. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  38. Ying, L.; Haokui, Z.; Qiang, S. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar]
  39. Zhong, Z.; Li, J.; .Luo, Z.; Chapman, M. Spectral–Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2018, 56, 847–858. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed WSFL for HSI classification.
Figure 1. Framework of the proposed WSFL for HSI classification.
Remotesensing 15 04994 g001
Figure 2. Multiple groups residual spectral attention model.
Figure 2. Multiple groups residual spectral attention model.
Remotesensing 15 04994 g002
Figure 3. Multiple groups spectral feature space.
Figure 3. Multiple groups spectral feature space.
Remotesensing 15 04994 g003
Figure 4. Multi-granularity residual spatial attention model.
Figure 4. Multi-granularity residual spatial attention model.
Remotesensing 15 04994 g004
Figure 5. Multi-granularity space.
Figure 5. Multi-granularity space.
Remotesensing 15 04994 g005
Figure 6. Multilayer perceptrons.
Figure 6. Multilayer perceptrons.
Remotesensing 15 04994 g006
Figure 7. The classification results of PC dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Figure 7. The classification results of PC dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Remotesensing 15 04994 g007aRemotesensing 15 04994 g007b
Figure 8. The classification results of the LK dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Figure 8. The classification results of the LK dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Remotesensing 15 04994 g008
Figure 9. The classification results of the HZ dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Figure 9. The classification results of the HZ dataset. (a) 3DCNN; (b) SSRN; (c) DCRN; (d) WSFL.
Remotesensing 15 04994 g009aRemotesensing 15 04994 g009b
Figure 10. The effect of group on classification accuracy. (a) OA; (b) AA; (c) Kappa.
Figure 10. The effect of group on classification accuracy. (a) OA; (b) AA; (c) Kappa.
Remotesensing 15 04994 g010
Table 1. The Number of Samples of the PC Dataset.
Table 1. The Number of Samples of the PC Dataset.
ClassClass NameSamplesColorFalse-Color MapGround-Truth Map
C1Water65,971Remotesensing 15 04994 i001Remotesensing 15 04994 i002Remotesensing 15 04994 i003
C2Trees7598Remotesensing 15 04994 i004
C3Asphalt3090Remotesensing 15 04994 i005
C4Self-blocking Bricks2685Remotesensing 15 04994 i006
C5Bitumen6584Remotesensing 15 04994 i007
C6Tiles9248Remotesensing 15 04994 i008
C7Shadows7287Remotesensing 15 04994 i009
C8Meadows42,826Remotesensing 15 04994 i010
C9Bare soil2863Remotesensing 15 04994 i011
Background Remotesensing 15 04994 i012
Total148,152
Table 2. The Number of Samples of the LK Dataset.
Table 2. The Number of Samples of the LK Dataset.
ClassNameSamplesColorFalse-Color MapGround-Truth Map
C1Corn34,511Remotesensing 15 04994 i013Remotesensing 15 04994 i014Remotesensing 15 04994 i015
C2Cotton8374Remotesensing 15 04994 i016
C3Sesame3031Remotesensing 15 04994 i017
C4Broad-leaf soybean63,212Remotesensing 15 04994 i018
C5Narrow-leaf soybean4151Remotesensing 15 04994 i019
C6Rice11,854Remotesensing 15 04994 i020
C7Water67,056Remotesensing 15 04994 i021
C8Roads and houses7124Remotesensing 15 04994 i022
C9Mixed weed5229Remotesensing 15 04994 i023
Background Remotesensing 15 04994 i024
Total204,542
Table 3. The Number of Samples of the HZ Dataset.
Table 3. The Number of Samples of the HZ Dataset.
ClassNameSamplesColorFalse-Color MapGround-Truth Map
C1Water18,043Remotesensing 15 04994 i025Remotesensing 15 04994 i026Remotesensing 15 04994 i027
C2Land/building77,450Remotesensing 15 04994 i028
C3Plants40,207Remotesensing 15 04994 i029
Total135,700
Table 4. The classification results of the PC dataset by different methods.
Table 4. The classification results of the PC dataset by different methods.
ClassDSNN2DCNN3DCNNSSRNDCRNWSFL
C199.71 ± 0.0999.60 ± 0.0199.27 ± 0.4999.73 ± 0.1499.48 ± 0.1099.81 ± 0.14
C290.19 ± 9.4482.21 ± 3.7092.25 ± 1.8094.51 ± 4.9092.51 ± 1.9791.82 ± 0.95
C359.90 ± 12.2595.21 ± 0.8290.24 ± 3.4498.19 ± 0.5696.46 ± 0.3597.92 ± 0.77
C473.76 ± 9.6073.91 ± 9.5286.52 ± 7.7599.77 ± 0.0998.82 ± 0.1799.71 ± 0.06
C573.71 ± 10.3386.72 ± 7.8690.37 ± 2.9485.84 ± 3.1093.07 ± 1.8397.59 ± 1.78
C667.40 ± 12.4199.22 ± 0.0690.95 ± 4.5999.21 ± 0.2099.86 ± 0.1098.52 ± 0.39
C777.62 ± 5.9367.45 ± 9.4486.45 ± 3.7997.36 ± 1.6793.54 ± 5.0395.39 ± 2.80
C896.33 ± 0.9899.07 ± 0.0696.56 ± 2.1699.46 ± 0.3499.74 ± 0.1199.51 ± 0.16
C999.96 ± 0.09100.00 ± 0.0095.51 ± 0.98100 ± 0.0099.93 ± 0.0198.42 ± 0.32
OA (%)92.68 ± 1.8195.87 ± 0.2896.10 ± 0.4398.60 ± 0.6198.58 ± 0.2798.84 ± 0.40
AA (%)82.06 ± 3.2689.26 ± 1.0992.01 ± 0.5497.11 ± 1.3297.04 ± 0.6697.75 ± 0.55
Kappa89.56 ± 2.5394.33 ± 0.3994.54 ± 0.5998.01 ± 0.8698.01 ± 0.3898.36 ± 0.32
Table 5. The classification results of the LK dataset by different methods.
Table 5. The classification results of the LK dataset by different methods.
ClassDSNN2DCNN3DCNNSSRNDCRNWSFL
C196.97 ± 2.5887.62 ± 6.1495.94 ± 2.5197.51 ± 1.0298.94 ± 0.0992.62 ± 0.35
C283.04 ± 10.0868.48 ± 9.3663.42 ± 8.8494.97 ± 4.0698.96 ± 0.9299.27 ± 0.66
C368.86 ± 11.8373.83 ± 7.9085.10 ± 3.9099.54 ± 0.1099.84 ± 0.1299.92 ± 0.06
C482.64 ± 10.3070.87 ± 4.0775.54 ± 1.5286.71 ± 6.2893.95 ± 2.2398.12 ± 0.98
C555.38 ± 16.3056.25 ± 10.3074.73 ± 7.8697.63 ± 2.0699.37 ± 0.5097.82 ± 1.16
C674.41 ± 6.9390.11 ± 3.7394.94 ± 4.0899.07 ± 0.2099.74 ± 0.0899.13 ± 0.22
C799.93 ± 0.0199.92 ± 0.0299.84 ± 0.0799.72 ± 0.1899.41 ± 0.1599.56 ± 0.30
C880.67 ± 12.5689.15 ± 3.9279.41 ± 4.4895.38 ± 3.8793.02 ± 2.9296.97 ± 2.08
C955.37 ± 13.3972.96 ± 8.3083.30 ± 2.5395.44 ± 2.0896.51 ± 1.5396.04 ± 1.25
OA (%)88.79 ± 1.2684.67 ± 2.7288.05 ± 1.6194.77 ± 1.5097.34 ± 0.5997.69 ± 0.26
AA (%)77.47 ± 6.9278.79 ± 6.6283.58 ± 3.3496.21 ± 2.3697.74 ± 0.3997.71 ± 0.51
Kappa85.39 ± 1.4780.70 ± 3.3784.71 ± 2.0693.25 ± 1.8996.52 ± 0.7796.86 ± 0.33
Table 6. The classification results of the HZ dataset by different methods.
Table 6. The classification results of the HZ dataset by different methods.
ClassDSNN2DCNN3DCNNSSRNDCRNWSFL
C181.10 ± 4.1688.68 ± 2.6593.92 ± 2.8588.62 ± 1.7389.30 ± 1.1591.79 ± 1.29
C264.96 ± 10.7171.94 ± 5.0870.71 ± 2.6774.35 ± 3.4776.35 ± 4.6377.25 ± 4.87
C398.65 ± 0.6073.11 ± 10.5279.57 ± 3.4784.20 ± 1.0981.32 ± 6.3180.79 ± 2.98
OA (%)77.07 ± 7.9974.54 ± 4.8076.41 ± 1.4879.17 ± 2.0079.52 ± 1.2180.23 ± 1.82
AA (%)81.57 ± 3.7677.91 ± 1.3981.40 ± 1.0682.39 ± 1.0182.32 ± 0.9583.27 ± 0.80
Kappa62.51 ± 10.4059.19 ± 4.7560.66 ± 2.3264.95 ± 3.0565.41 ± 1.7766.51 ± 2.19
Table 7. The classification results of the PC dataset with 24 clean + 4/8/12 noisy samples.
Table 7. The classification results of the PC dataset with 24 clean + 4/8/12 noisy samples.
ClassThe Number of Clean and Noisy Training Samples
24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
C1
Std
98.33
0.84
98.88
0.54
98.35
0.53
99.87
0.12
94.07
3.70
97.32
1.39
97.49
1.31
99.53
0.15
86.15
3.39
94.56
4.10
98.14
0.12
98.23
0.10
C2
Std
82.22
1.62
94.75
2.46
93.13
2.84
96.00
2.47
70.41
3.62
91.17
4.99
92.51
3.03
91.07
2.81
54.34
10.35
89.30
6.78
91.00
4.53
92.83
2.39
C3
Std
82.15
2.41
89.77
6.92
91.70
0.48
90.74
1.46
74.38
5.95
87.84
6.90
94.12
1.98
94.22
1.30
62.42
12.65
63.44
4.71
93.89
2.30
86.32
3.50
C4
Std
79.01
3.59
95.62
1.63
96.28
3.21
98.81
1.08
52.66
3.15
88.30
7.29
91.01
4.08
97.99
1.42
44.74
1.73
91.00
8.79
94.50
2.29
92.34
3.25
C5
Std
75.53
6.29
92.55
3.30
96.29
2.67
96.91
2.33
62.45
10.71
92.59
2.33
89.34
1.74
91.02
2.27
47.53
10.01
85.03
8.64
92.03
4.32
92.04
2.91
C6
Std
81.60
7.80
94.42
2.05
96.04
1.06
98.21
0.74
62.99
8.02
89.92
2.37
98.96
0.76
98.28
1.25
50.35
3.66
81.70
1.94
95.73
2.62
98.33
1.22
C7
Std
72.25
2.43
94.70
3.28
94.78
1.07
92.42
1.95
61.85
5.57
91.19
4.77
93.62
1.07
90.81
1.92
52.28
9.03
77.83
5.02
93.59
2.06
91.84
2.37
C8
Std
83.54
6.46
91.24
0.61
97.06
1.25
98.77
0.19
68.04
4.14
91.69
1.59
96.48
2.09
97.54
0.32
56.75
4.26
90.38
2.47
93.28
5.06
97.49
1.00
C9
Std
91.95
7.20
97.60
1.46
98.39
0.32
98.67
0.58
79.55
8.09
95.94
3.72
99.57
0.13
99.62
0.07
75.85
6.19
87.68
7.12
91.16
5.91
96.36
2.13
OA(%)
Std
89.09
2.79
95.42
0.28
97.12
0.80
98.52
0.45
78.98
3.93
94.02
0.95
96.34
1.95
97.50
0.12
68.98
2.00
90.20
2.56
95.44
3.82
96.77
0.71
AA(%)
Std
82.95
2.53
94.39
1.31
95.78
1.60
96.71
1.34
69.60
3.50
91.77
1.21
94.79
0.79
95.57
0.40
58.93
1.31
84.55
0.70
93.70
4.42
93.97
1.81
Kappa
Std
84.82
3.68
93.69
0.40
95.82
1.13
97.91
0.64
71.44
4.85
91.68
1.31
94.87
2.61
96.48
0.17
58.64
1.21
86.40
3.33
93.62
5.99
95.48
1.00
Table 8. The classification results of the LK dataset with 24 clean + 4/8/12 noisy samples.
Table 8. The classification results of the LK dataset with 24 clean + 4/8/12 noisy samples.
ClassThe Number of Clean and Noisy Training Samples
24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
C1
Std
79.28
9.23
94.03
1.45
95.79
2.22
94.97
1.13
60.90
9.68
88.49
4.86
90.38
4.84
94.18
3.62
50.76
7.42
80.23
1.41
87.94
4.02
92.46
3.20
C2
Std
47.53
7.56
92.53
4.32
89.40
5.61
92.86
2.45
44.77
7.54
79.26
11.0
90.30
7.06
81.54
2.97
46.53
3.98
74.26
7.27
86.31
8.84
82.57
4.42
C3
Std
67.64
8.18
91.49
1.60
98.63
0.81
97.73
1.93
55.33
10.02
90.50
5.67
89.52
5.90
90.73
3.41
50.34
7.55
83.88
4.73
86.63
9.83
91.29
5.49
C4
Std
61.17
6.26
89.50
5.24
91.42
4.13
93.33
5.07
48.21
6.79
83.85
5.38
81.86
6.45
82.05
4.49
42.85
3.18
79.11
5.57
77.52
3.02
76.35
2.35
C5
Std
64.15
9.81
89.74
3.60
89.20
4.23
94.42
1.27
44.40
6.80
83.49
8.73
86.42
6.79
91.07
3.85
45.92
10.05
75.83
7.76
81.88
8.45
88.31
7.01
C6
Std
82.62
7.59
96.51
1.60
98.55
1.06
94.13
1.20
62.05
6.20
92.57
5.34
91.83
3.24
93.86
1.17
57.35
8.55
89.79
5.14
80.19
6.15
82.72
4.55
C7
Std
98.63
0.49
98.20
1.06
98.38
0.66
99.70
0.02
96.11
0.83
97.11
1.98
98.23
0.55
99.57
0.18
92.59
1.20
95.01
4.96
96.64
2.31
99.32
0.29
C8
Std
71.44
3.38
87.52
4.91
91.45
2.54
92.34
1.91
58.44
7.05
77.55
7.59
82.76
7.60
83.93
3.60
52.18
2.46
70.39
8.34
76.85
11.07
78.51
8.09
C9
Std
57.81
10.65
88.09
3.83
89.36
1.06
89.52
2.71
47.52
6.75
77.32
3.51
82.75
4.12
80.71
2.50
42.77
6.20
70.33
5.61
73.01
6.31
75.50
2.55
OA(%)
Std
77.63
3.92
93.58
2.00
94.78
1.42
95.68
0.71
67.09
3.80
89.01
2.85
89.85
3.01
90.85
1.58
61.99
2.87
84.42
2.06
86.15
3.73
87.74
1.85
AA(%)
Std
70.03
4.46
91.96
1.07
93.58
1.25
94.33
1.06
57.52
2.50
85.57
2.82
88.23
3.87
88.63
2.26
53.48
4.25
79.87
0.32
83.00
4.42
85.23
2.78
Kappa
Std
71.73
4.72
91.67
2.53
93.22
1.82
94.36
0.88
58.94
4.47
85.82
3.62
86.89
3.78
88.18
2.01
53.07
3.48
79.98
2.49
82.23
4.64
84.28
2.39
Table 9. The classification results of the HZ dataset with 24 clean + 4/8/12 noisy samples.
Table 9. The classification results of the HZ dataset with 24 clean + 4/8/12 noisy samples.
ClassThe Number of Clean and Noisy Training Samples
24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
C1
Std
86.43
4.55
81.07
9.46
81.43
4.13
88.37
3.86
77.53
5.09
71.54
8.98
73.39
5.81
83.05
3.34
77.06
4.86
75.88
8.70
69.43
7.41
81.33
3.20
C2
Std
71.20
5.98
79.61
5.96
76.85
5.35
78.97
2.24
60.06
8.90
58.05
5.60
64.80
3.49
73.07
2.39
51.85
8.49
57.98
9.17
59.52
8.70
62.45
4.57
C3
Std
76.26
2.13
70.40
8.89
77.03
7.67
76.35
3.26
62.12
7.36
75.63
5.30
67.95
7.68
68.03
3.82
65.33
8.46
60.75
9.60
64.51
8.10
57.76
4.37
OA(%)
Std
74.72
3.90
77.07
5.51
77.51
2.87
79.44
1.19
62.99
6.17
65.05
3.31
66.88
2.58
72.90
1.96
59.19
6.37
61.18
3.62
62.31
3.42
63.57
3.28
AA(%)
Std
77.96
2.78
77.03
6.04
78.44
2.90
81.23
1.86
66.57
4.28
68.41
1.70
68.71
1.40
74.72
0.94
64.75
4.90
64.87
1.52
64.49
2.04
67.18
2.02
Kappa
Std
57.61
5.95
60.30
8.62
61.49
4.39
64.84
2.29
40.81
8.50
44.08
4.15
68.71
3.52
54.69
2.39
36.00
8.62
37.42
2.94
39.46
3.90
40.32
3.01
Table 10. The classification results of the PC dataset with different numbers of clean samples and noisy samples.
Table 10. The classification results of the PC dataset with different numbers of clean samples and noisy samples.
ClassThe Numbers of Clean and Noisy Training Samples
8(clean) + 4(noisy)8 (clean) + 8(noisy)8(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
67.60
3.91
84.09
4.80
90.96
4.61
94.02
3.03
52.78
4.96
79.55
3.04
82.05
5.41
85.92
4.61
41.50
5.11
61.68
4.97
67.08
8.72
71.35
7.16
AA(%)
Std
57.81
7.67
79.92
7.97
84.36
3.06
86.78
3.21
42.80
5.97
70.72
3.57
67.39
3.63
85.26
2.79
32.65
5.08
55.81
1.63
57.90
6.96
64.29
6.02
Kappa
Std
57.05
5.15
79.01
6.47
87.35
6.21
91.53
4.05
40.60
4.71
72.31
3.36
75.21
7.14
80.74
5.14
27.63
5.22
50.26
5.45
56.60
10.37
62.51
8.15
20(clean) + 4(noisy)20(clean) + 8(noisy)20(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
82.16
2.50
94.41
0.81
96.76
0.61
97.61
0.94
73.48
0.46
93.36
1.55
93.92
3.77
96.63
2.05
62.58
3.92
86.14
1.06
91.82
6.25
96.38
2.43
AA(%)
Std
73.83
2.56
92.68
1.85
95.17
0.86
94.04
1.22
63.30
1.01
89.16
1.44
89.89
3.49
94.65
1.77
54.09
1.67
79.85
1.88
88.01
6.64
93.49
3.01
Kappa
Std
75.58
3.44
92.15
1.15
95.43
0.87
96.61
1.32
64.24
0.30
90.69
2.10
91.51
5.12
95.25
2.80
51.74
4.43
81.00
1.44
88.69
8.54
94.89
3.39
24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
89.09
2.79
95.42
0.28
97.12
0.80
98.52
0.45
78.98
3.93
94.02
0.95
96.34
1.95
97.50
0.12
68.98
2.00
90.20
2.56
95.44
3.82
96.77
0.71
AA(%)
Std
82.95
2.53
94.39
1.31
95.78
1.60
96.71
1.34
69.60
3.50
91.77
1.21
94.79
0.79
95.57
0.40
58.93
1.31
84.55
0.70
93.70
4.42
93.97
1.81
Kappa
Std
84.82
3.68
93.69
0.40
95.82
1.13
97.91
0.64
71.44
4.85
91.68
1.31
94.87
2.61
96.48
0.17
58.64
1.21
86.40
3.33
93.62
5.99
95.48
1.00
Table 11. The classification results of the LK dataset with different numbers of clean samples and noisy samples.
Table 11. The classification results of the LK dataset with different numbers of clean samples and noisy samples.
MetricsThe Numbers of Clean and Noisy Training Samples
8(clean) + 4(noisy)8 (clean) + 8(noisy)8(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
58.65
2.84
80.42
3.72
80.58
5.35
84.63
3.39
51.42
5.28
66.12
7.23
66.40
4.74
68.17
3.79
45.25
4.54
55.71
5.47
56.87
3.71
60.93
3.54
AA(%)
Std
46.53
2.73
74.66
2.33
78.51
1.67
78.45
2.86
36.50
4.83
60.84
4.15
65.69
3.18
66.81
4.05
32.91
2.73
47.92
4.81
51.97
3.09
52.57
2.62
Kappa
Std
49.13
2.96
75.17
4.38
75.55
6.37
80.39
4.02
40.69
6.01
58.16
8.04
58.63
5.56
60.50
4.02
34.25
4.61
46.18
5.94
47.02
3.93
52.10
3.74
Metrics20(clean) + 4(noisy)20(clean) + 8(noisy)20(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
69.08
2.20
87.52
4.74
92.12
2.60
94.83
1.74
66.98
3.09
79.70
3.76
87.63
6.29
89.54
2.79
55.47
5.70
74.68
6.44
81.10
5.51
85.21
2.52
AA(%)
Std
61.24
3.14
86.80
1.20
91.58
1.19
92.19
1.06
55.51
1.02
77.96
3.49
85.63
4.33
88.54
2.15
45.17
5.82
70.92
3.57
79.17
6.98
80.63
2.97
Kappa
Std
61.44
2.56
84.03
5.58
89.82
3.31
93.27
2.22
58.50
3.37
74.53
4.69
84.10
7.87
86.41
3.57
45.17
6.63
68.13
7.64
76.02
7.04
81.07
3.06
Metrics24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
77.63
3.92
93.58
2.00
94.78
1.42
95.68
0.71
67.09
3.80
89.01
2.85
89.85
3.01
90.85
1.58
61.99
2.87
84.42
2.06
86.15
3.73
87.74
1.85
AA(%)
Std
70.03
4.46
91.96
1.07
93.58
1.25
94.33
1.06
57.52
2.50
85.57
2.82
88.23
3.87
88.63
2.26
53.48
4.25
79.87
0.32
83.00
4.42
85.23
2.78
Kappa
Std
71.73
4.72
91.67
2.53
93.22
1.82
94.36
0.88
58.94
4.47
85.82
3.62
86.89
3.78
88.18
2.01
53.07
3.48
79.98
2.49
82.23
4.64
84.28
2.39
Table 12. The classification results of the HZ dataset with different numbers of clean samples and noisy samples.
Table 12. The classification results of the HZ dataset with different numbers of clean samples and noisy samples.
MetricsThe Numbers of Clean and Noisy Training Samples
8(clean) + 4(noisy)8(clean) + 8(noisy)8(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
58.03
4.59
62.36
5.29
62.82
7.45
66.26
5.06
41.66
5.94
44.89
4.76
48.66
8.29
54.97
3.30
35.53
8.79
35.68
4.51
36.60
6.92
43.34
5.26
AA(%)
Std
63.96
3.85
62.26
3.84
67.65
7.73
65.43
5.74
46.75
6.77
47.44
7.91
50.83
10.2
63.48
3.62
38.35
6.02
36.71
5.41
40.64
7.52
40.23
4.79
Kappa
Std
34.77
6.43
38.11
6.51
42.32
10.5
44.69
6.12
10.06
8.20
16.30
7.94
21.04
11.3
32.64
4.89
5.82
4.46
6.44
4.95
7.36
8.21
11.18
4.47
Metrics20(clean) + 4(noisy)20(clean) + 8(noisy)20(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
68.35
4.25
71.09
6.25
73.21
5.03
77.64
4.71
58.25
3.25
64.82
4.67
65.64
5.48
66.62
4.94
58.03
3.72
59.73
4.73
61.70
4.47
61.96
3.27
AA(%)
Std
71.75
2.43
75.48
3.69
75.65
2.70
80.65
2.55
63.30
3.40
68.48
2.11
70.13
3.96
70.96
3.09
62.53
2.70
61.21
2.43
62.39
4.54
65.68
3.97
Kappa
Std
49.23
6.00
52.84
7.93
55.43
6.27
62.26
5.71
33.69
4.90
43.07
5.91
70.13
7.71
45.84
5.49
33.84
5.36
34.69
5.33
37.66
6.36
38.13
4.82
Metrics24(clean) + 4(noisy)24(clean) + 8(noisy)24(clean) + 12(noisy)
3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL3DCNNSSRNDCRNWSFL
OA(%)
Std
74.72
3.90
77.07
5.51
77.51
2.87
79.44
1.19
62.99
6.17
65.05
3.31
66.88
2.58
72.90
1.96
59.19
6.37
61.18
3.62
62.31
3.42
63.57
3.28
AA(%)
Std
77.96
2.78
77.03
6.04
78.44
2.90
81.23
1.86
66.57
4.28
68.41
1.70
68.71
1.40
74.72
0.94
64.75
4.90
64.87
1.52
64.49
2.04
67.18
2.02
Kappa
Std
57.61
5.95
60.30
8.62
61.49
4.39
64.84
2.29
40.81
8.50
44.08
4.15
68.71
3.52
54.69
2.39
36.00
8.62
37.42
2.94
39.46
3.90
40.32
2.39
Table 13. Computation time comparison for three HSI datasets(s).
Table 13. Computation time comparison for three HSI datasets(s).
AlgorithmDSNN2DCNN3DCNNSSRNDCRNWSFL
Dataset
Pavia Center203.60207.69161.21241.27273.14268.67
WHU-Hi-LongKou89.7484.97298.79390.17330.76375.81
HangZhou56.0655.83118.26151.64151.92158.27
Table 14. The classification effectiveness of different attention models.
Table 14. The classification effectiveness of different attention models.
AlgorithmIndexMGRSAMMRSAMMGRSAM + MRSAM
Dataset
Pavia CenterOA98.2798.2698.52
AA96.4796.5496.71
Kappa97.5597.5497.91
WHU-Hi-LongKouOA95.3495.2895.68
AA91.0294.5994.33
Kappa93.9393.8694.36
HangZhouOA77.7878.9179.44
AA79.1480.1381.23
Kappa62.0181.2364.84
Table 15. The effectiveness of the MLP model on classification results.
Table 15. The effectiveness of the MLP model on classification results.
AlgorithmIndexWithout MLPWith MLP
Dataset
Pavia CenterOA98.2398.52
AA96.3496.71
Kappa97.5097.91
WHU-Hi-LongKouOA95.0195.68
AA94.1194.33
Kappa93.2194.36
HangZhouOA78.4879.44
AA78.3881.23
Kappa62.0664.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, C.; Zhao, L.; Wu, H. Hyperspectral Images Weakly Supervised Classification with Noisy Labels. Remote Sens. 2023, 15, 4994. https://doi.org/10.3390/rs15204994

AMA Style

Liu C, Zhao L, Wu H. Hyperspectral Images Weakly Supervised Classification with Noisy Labels. Remote Sensing. 2023; 15(20):4994. https://doi.org/10.3390/rs15204994

Chicago/Turabian Style

Liu, Chengyang, Lin Zhao, and Haibin Wu. 2023. "Hyperspectral Images Weakly Supervised Classification with Noisy Labels" Remote Sensing 15, no. 20: 4994. https://doi.org/10.3390/rs15204994

APA Style

Liu, C., Zhao, L., & Wu, H. (2023). Hyperspectral Images Weakly Supervised Classification with Noisy Labels. Remote Sensing, 15(20), 4994. https://doi.org/10.3390/rs15204994

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop