Next Article in Journal
Identification of Spatial Distribution of Afforestation, Reforestation, and Deforestation and Their Impacts on Local Land Surface Temperature in Yangtze River Delta and Pearl River Delta Urban Agglomerations of China
Previous Article in Journal
A Novel Framework for Spatiotemporal Susceptibility Prediction of Rainfall-Induced Landslides: A Case Study in Western Pennsylvania
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Weighted Contrastive Prototype Network for Few-Shot Hyperspectral Image Classification with Noisy Labels

by
Dan Zhang
1,2,
Yiyuan Ren
3,
Chun Liu
3,
Zhigang Han
4 and
Jiayao Wang
4,*
1
The College of Surveying and Mapping Engineering, Yellow River Conservancy Technical Institute, Kaifeng 475004, China
2
Henan Province Surveying and Mapping Real Scene 3D Technology Engineering Research Center, Yellow River Conservancy Technical Institute, Kaifeng 475004, China
3
School of Computer and Information Engineering, Henan University, Zhengzhou 450046, China
4
College of Geography and Environmental Science, Henan University, Zhengzhou 450046, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(18), 3527; https://doi.org/10.3390/rs16183527
Submission received: 2 August 2024 / Revised: 9 September 2024 / Accepted: 11 September 2024 / Published: 23 September 2024

Abstract

:
Few-shot hyperspectral image classification aims to develop the ability of classifying image pixels by using relatively few labeled pixels per class. However, due to the inaccuracy of the localization system and the bias of the ground survey, the potential noisy labels in the training data pose a very significant challenge to few-shot hyperspectral image classification. To solve this problem, this paper proposes a weighted contrastive prototype network (WCPN) for few-shot hyperspectral image classification with noisy labels. WCPN first utilizes a similarity metric to generate the weights of the samples from the same classes, and applies them to calibrate the class prototypes of support and query sets. Then the weighted prototype network will minimize the distance between features and prototypes to train the network. WCPN also incorporates a weighted contrastive regularization function that uses the sample weights as gates to filter the fake positive samples whose labels are incorrect to further improve the discriminative power of the prototypes. We conduct experiments on multiple hyperspectral image datasets with artificially generated noisy labels, and the results show that the WCPN has excellent performance that can sufficiently mitigate the impact of noisy labels.

1. Introduction

Hyperspectral images (HSIs) are three-dimensional cubes composed of hundreds of spectral channels. Each pixel of the images is a high-dimensional spectral vector. With rich spatial and spectral information, HSIs can reflect the radiometric properties and spatial geometric relationships of ground targets simultaneously. This results in HSIs that can capture more subtle differences between ground targets, and have great application values in various fields such as environmental monitoring [1,2], precision agriculture [3,4], and military applications [5,6].
HSI classification is to predict the classes of each pixel in the images, which is one of the important tasks in remote sensing applications. With the wide use of deep learning, significant improvements have been made in the accuracy of HSI classification [7,8]. In general, existing HSI classification methods can be divided into two categories, i.e., pixel-wise methods and spectral–spatial joint methods. The former focus on the spectral vectors behind each pixel, and apply the classifiers directly to them to predict their classes. In contrast, the latter consider the spectral information contained in the vectors but also the local spatial information around each pixel. For example, by cutting the fixed-size (e.g., 9 × 9 ) patches centering around each pixel from HSIs and taking them as samples to be classified, the spectral–spatial joint methods can capture the spectral–spatial fusion features and have shown great advantages in HSI classification.
As it is usually expensive and time-consuming to obtain a large number of labeled samples, which are often required by current deep learning methods, few-shot learning methods have been widely used for HSI classification in recent years. Few-shot learning aims to develop the ability of identifying new samples of some classes by using only few labeled samples per class. To achieve this goal, current few-shot learning methods usually train a model with available source datasets which already consist of abundant labeled samples and then fine-tune the model with these few-shot labeled samples in the target dataset. Based on the classical few-shot learning methods such as prototype network [9] and relation network [10], a series of few-shot methods have been designed for HSI classification. And considering that the source and target datasets are from different domains, many cross-domain few-shot methods including DCFSL [11], Gia-FSL [12], and RPCL-FSL [13] have also been proposed.
However, existing few-shot methods for HSI classification have assumed that the labels of all training samples are correct. These methods will suffer from performance degradation if there are noisy labels in the training data. Noisy labels are the labels that are mistakenly assigned to samples. For HSIs, accurate annotation may be difficult to achieve due to the inaccuracy of the localization system and the bias of the ground survey [14,15]. Also, the annotator may make mistakes when confronted with certain ambiguous samples. All these make it difficult to avoid noisy labels in HSI classification in practice.
The noisy labels problem has been investigated by many works in computer vision applications [16,17]. There are two kinds of solutions: one is to detect the noisy labels and clean the samples before training [18], and the other is to train a robust model directly on the corrupted dataset [19]. Due to the differences between hyperspectral images and natural images, these solutions cannot be directly applied to HSI classification. But by following the above two routes to address noisy labels, some methods have been proposed for HSI classification with noisy labels. For example, Tu et al. [20] proposed the density peak noisy label detection method to identify noisy labels and eliminate anomalous samples from training datasets. In contrast, Xu et al. [15] presented the dual-channel residual network to deal with noisy labels at the model level. But how to address the noisy labels for few-shot HSI classification remains the issue to be resolved.
To address HSI classification with noisy labels under the few-shot setting, this paper proposes a Weighted Contrastive Prototype Network (WCPN). While combining few-shot learning and contrastive learning to solve few-shot HSI classification, the main idea of the proposed method is to identify the potential noisy samples and assign different weights to them to balance their impact. All samples are passed through an adaptive mapping module and a deep 3D residual network to extract the embedding features, and then the proposed method employs a metric function to estimate the weights of each sample by utilizing the similarities between the samples. These weights are then used for the computation of the prototypes of both support and query sets, to calibrate the prototypes at the presence of noisy labels. In order to further improve the self-calibration capability of the model to reduce the impact of noisy labels, we design a regularized contrastive loss function that uses the estimated weights as gates to constrain the model. With these improvements, we aim to make the clean samples closer, and keep clean and noisy samples farther apart. The main contributions of the proposed WCPN are as follows.
  • To the best of our knowledge, a prototype network for handling noisy labels under few-shot setting has been introduced to the HSI classification for the first time. We provide better anti-noise performance than existing methods for few-shot HSI classification.
  • Our proposed weighted prototype network utilizes the weights calculated from similarities between samples to calibrate the prototypes and bring the samples closer to their clean prototypes at the presence of noisy labels.
  • We make full use of the noise information contained in the similarity weights of different samples and propose a new contrastive regularization function. This function can further constrain the model to reduce the impact of noisy samples and learn a clean feature representation.

2. Background and Related Work

The work in this paper involves few-shot learning for HSI classification, HSI classification with noisy labels, and contrastive learning with noisy labels. In this section, we give a brief introduction about these related works.

2.1. Few-Shot Learning for HSI Classification

As mentioned earlier, few-shot learning aims to solve the tasks where relatively few labeled samples are available. It has aroused much attention in a wide range of applications [21,22], and this research focuses only on its use for HSI classification.
Current works for few-shot HSI classification mainly follow the metric-based learning method which strives to learn a distance function that can measure the similarity between samples. Such a distance function is expected to have good generalization ability so that it can be well migrated to these samples in the target tasks. The classes of unlabeled samples in the target tasks can then be predicted based on their similarity with the given few labeled samples or their prototypes. By following the classical metric-based few-shot learning models, many methods have been presented for few-shot HSI classification. For example, based on the twin network [23], which receives a pair of samples and generates their similarity, ASSP-SCNN replaced the feature extractor with 3D CNN for HSI classification [24]. Based on the prototype network which predicts the sample classes according to their distance to different prototypes [9], DFSL [25] took a deep 3D residual network to learn the metric space in order to maximize class discriminability and separability; SSPN [26] used the local pattern coding to extract better features for HSI classification; HSEMD-Net [27] used the Earth Mover’s distance instead of the Euclidean distance to improve the model performance. Based on relational network, which is an improved version of prototype network [10], RN-FSC [28] and RL-Net [29] follow the meta-learning to train the model for few-shot HSI classification.
The above works for few-shot HSI classification usually take the source datasets which already consist of lots of labeled samples to train the model and learn a suitable distance function, and then transfer it to the target dataset. They often assume that the source datasets and the target datasets are from the same domains. To relax this assumption and address the cross-domain few-shot HSI classification, DCFSL [11] followed the adversarial way to train the model with the aim to obtain domain-independent features; SSFT [30] adopted a feature-wise transformation module to extract more generalized features; CMFSL [31] used the modules of spectral prior-based refinement and a lightweight cross-scale convolution to increase the feature extraction ability under few-shot setting; Gia-CFSL [12] used a domain alignment strategy to suppress domain bias; GPN [32] used a global prototype strategy to train the network, where the global prototypes are continuously updated during iterative training; RPCL-FSL [13] integrated supervised contrastive learning into few-shot HSI classification, in order to obtain better prototypes and improve the classification accuracy. Liu et al. [33] incorporated contrastive learning and a transformer-based cross-attention module into few-shot HSI classification to enhance multi-level sample relations and improve classification performance.
Different from these works, the proposed method in this paper aims to explore the similarities between the samples themselves, and study how to utilize them to identify the noisy samples and reduce their impact on cross-domain few-shot HSI classification.

2.2. HSI Classification with Noisy Labels

The majority of existing research approaches for learning with noisy labels in the literature reduce the impacts of noise by creating robust loss functions [34,35], a noise transition matrix [36,37], confident examples selection [38,39], reweighting examples [40,41], introducing regularization [42,43], and generating pseudo labels [44,45]. Specifically for few-shot learning, the existing methods focus on selecting credible samples from the model level. Mazumder et al. proposed RNNP [46] which combined data augmentation with k-means to produce refined prototypes. RapNets [47] utilized a BiLSTM-based attention module to overcome representation with noisy labels. Liang et al. proposed TraNFS [48] utilizing the Transformer’s attention mechanism to trade-off between noisy samples and correct samples. These methods can mitigate the effects of noisy labels somewhat, but mainly focus on the noisy samples in the support set.
In the field of HSI, many methods have been also proposed to address the noisy label problem by detecting and removing noisy labels before training or directly training robust models with noisy data. Among the methods for detecting and removing noisy labels, Jiang et al. [49] proposed RLPA to construct a spectral–spatial probabilistic transferring matrix, and utilized hyperpixel-constrained random labels to reduce the noisy labels. Tu et al. proposed an SDP method [50] based on spatial density peak clustering and a SPWD method [14] incorporating super-pixel weighted distance to detect mislabeled samples in the training set. Among the methods for training robust models on noisy data, Jiang et al. proposed MSSAs [51] that used spectral and spatial similarity to construct affinity graphs to regularize the process of noisy labels cleaning, thus transforming noisy labels cleaning into an optimization problem with graph constraints. Xu et al. proposed a two-channel residual network DCRN [15] to reduce the impact of noisy labels by utilizing a noise robust loss function to detect and reject these abnormal samples. All of the above methods mitigate the impact of noisy labels in HSIs to a certain extent. Different from these works, the proposed method in this paper focuses on how to address the noisy labels and train a robust model for HSI classification in the few-shot setting.

2.3. Contrastive Learning with Noisy Labels

Contrastive learning [52,53] has demonstrated impressive results not only in representation learning but also in various downstream tasks, where the learned representations have shown strong generalization capabilities. The essence of contrastive learning lies in learning to differentiate between similar and dissimilar samples by contrasting their representations. In particular, by maximizing the agreement between positive samples and minimizing it for negative samples, the model is able to learn rich, meaningful feature embeddings from data. One specific variant of contrastive learning is supervised contrastive learning [54], which extends this concept by incorporating class labels. The objective of supervised contrastive learning is to ensure that samples from the same class are pulled closer together in the representation space, while samples from different classes are pushed apart. This label-guided approach further enhances the discriminative power of the learned representations, making it particularly effective in scenarios where class distinctions are crucial.
The noisy labels problem also has much impact on the supervised contrast learning. To address this issue, there are some methods proposed by now. For example, the methods of ProtoMix [55] and NGC [56] perform pseudo-label generation and supervised contrastive learning against noisy labels. Sel-CL proposed by Li et al. [57] identifies plausible examples for constructing confidence pairs by measuring the consistency between learning and given labels; Yi et al. [58] combined the non-robustness of cross-entropy loss with a novel contrast regularization function so that noisy data do not dominate the learning of features. Our work takes full advantage of contrast learning by considering the different weights of different samples in the presence of noisy samples, thereby providing better regularization of features.

3. The Proposed Method

The proposed WCPN solves the task of few-shot HSI classification with noisy labels. The framework of the proposed method is shown in Figure 1. In an episodic manner of meta-learning, firstly, the support set and query set of a task are constructed from the source dataset and target dataset. Considering the presence of noisy labels, there are some noisy samples in the constructed support and query set, which are indicated by the red squares in Figure 1. Then, two trainable adaptive mapping modules are used to unify the spectral dimensions of the samples from the source dataset and target dataset, and a 3D residual module is taken to extract the features of all the samples. Following that, the weights of these samples in the support and query set are calculated, respectively, and the class prototypes of the support and query set are obtained accordingly. Finally, the weighted few-shot learning loss and weighted contrastive loss are derived to update the model. The model is trained alternately with the tasks constructed from the source and target datasets. When testing, the KNN classifier is used once the features of the support and query samples are obtained.

3.1. Meta-Learning and Feature Extraction

For few-shot learning, there are two datasets given: the source dataset D s with C s classes and the target dataset D t with C t classes. It is worth noting that there are only few labeled samples in D t and all the unlabeled samples are to be classified. The C t is also smaller than C s . In the training phase, a task which consists of the support and query sets is constructed from the dataset in each episode. There are N-way ( K + M ) -shot samples randomly selected from the dataset for constructing the task. The N-way refers to the number of classes selected, which is often set to the number of classes of the dataset. K is the number of support samples per class and M is the number of query samples per class. That is, there are N × K samples to form the support set S = { ( x i , y i ) } i = 1 N × K , and N × M samples to form the query set Q = { ( x j , y j ) } j = 1 N × M . In this paper, K is set to 5 and M is set to 15.
To alleviate the problem of domain shift between D s and D t , we train the model alternately with labeled samples in D s and D t . However, one problem with alternate training is that the spectral dimensions of the samples from source and target datasets are different. In order to unify the dimensionality, two mapping modules are used which are implemented with a two-dimensional convolutional layer consisting of d convolutional kernels of size 1 × 1 × c h , where c h and d are the dimensionality of the input and output data, respectively. Given different patches with the size of 9 × 9 × c h , the mapping modules transform them into the new patches with the size of 9 × 9 × d . After that, the spatial–spectrial joint features are extracted by the feature extractor module which is implemented with a deep 3D residual network. The components of the feature extractor are shown in Figure 2, and for more details of its implementation, please refer to [59].

3.2. Weighting Samples with Noisy Labels

Prototype network [9] is a simple and efficient method for few-shot learning. It weights all the samples in the support set equally and generates the class prototypes by means of mean aggregation. With the class prototypes, it predicts the classes of samples according to their distance to the prototypes. However, when noisy labels are present, the generated prototypes will deviate from the true ones and the classification performance is affected. An obvious solution is to remove the noisy samples [57], but this way tends to remove the useful information brought by them at the same time. This paper proposes to weight the samples by using the similarity information among them. While suppressing the noisy samples and amplifying the clean samples, it makes full use of the noisy samples as much as possible.
To calculate the similarity among samples, the cosine metric which performs well for high-dimensional data is used in this paper. With the cosine metric, the average similarity between one sample and all other samples of the same class is computed by Equation (1).
a i ( c ) = 1 K 1 i j h i ( c ) · h j ( c ) h i ( c ) h j ( c )
where K is the number of samples per class, and a i ( c ) is the average similarity between the ith sample to the other samples of class c. The larger a i ( c ) is, the closer the sample is to other similar samples.
Then, the average similarities a i ( c ) of the samples from the same classes are normalized by the softmax function shown in Equation (2), and the weights w i ( c ) of each sample are generated.
w i ( c ) = exp a i ( c ) / T j exp a j ( c ) / T
where T is the temperature term that controls softmax diffusion. When T 0 , the sample with the smallest distance to the other samples will be selected as the class prototype, and when T , it is equal to the mean operation.
The similarity-based way to weight samples can measure the noise distribution of the sample labels by using their inherent characteristics. The smaller w is, the less similar the sample is to the other samples of the same classes, and the higher the confidence that the sample is considered a noisy sample. In contrast, if w is larger, the sample is cleaner.

3.3. Weighted Prototype Network

The prototype network trains the feature extractor f θ by minimizing the distance between the query set samples and the class prototypes of the support set. As shown in the left part of Figure 3, the class prototypes are computed as the mean of the features of the samples from the same classes, which can be also seen in Equation (3).
P c = 1 K x i S c f θ ( x i )
where P c represents the prototype of class c, S c denotes the set of samples belonging to class c in the support set, and K is the number of support samples per class.
After obtaining the support prototypes, the embedded features of the query samples are also extracted by the feature extractor. For the query samples x j Q with true label c j , the prototype network predicts the probability that sample x j belongs to c j by following Equation (4).
p ( y j = c j | x j ) = exp d f θ ( x j ) , P s c j c = 1 N exp d f θ ( x j ) , P s c
where d ( · ) denotes the Euclidean distance used to calculate the distance between features and prototypes. Following this, the few-shot learning loss can be calculated by Equation (5).
L fsl = 1 n q j = 1 n q log p ( y j = c j | x j )
From Equations (3) and (4), it can be seen that in the presence of noisy support samples (e.g., red circles and red triangles shown in the left part of Figure 3), the dirty support prototypes will be generated, which will greatly affect the model’s performance. In particular, there may also be noisy samples in the query set, which will also affect the training effect. To address this problem, our idea is to weight the samples in both the support set and query set, and use the weights to balance the impacts of noisy samples, as shown in the right part of Figure 3.
Specifically, both the support prototypes and the query prototypes are first computed and calibrated by using the weights obtained by Equation (2). The clean prototypes obtained after calibration are shown in Equation (6) where S c is the set of samples of class c in either the support set or query set.
P c = 1 K x i S c f θ ( x i ) · w i
Second, with the calibrated support prototypes and the weights of query samples, the original few-shot learning loss function of prototype network is updated so that the noisy samples do not dominate the calculation of the loss. The updated loss function is shown in Equation (7), where P s c denotes the class prototypes in support set and w j q denotes the weights of query samples.
L f s l q = 1 N × M j = 1 N × M log exp d f θ ( x j ) , P s c j · w j q c = 1 N exp d f θ ( x j ) , P s c
Third, in order to make full use of the information contained in the samples to ehance the training effect, the prediction loss of taking the query prototypes as anchors to predict the classes of the support samples is also utilized in this paper, which is similar with RPCL [13]. By using the calibrated query prototypes and the weights of support samples, the function for calculating such a loss is defined in Equation (8), where P q c denotes the class prototypes in the query set and w j s denote the weights of support samples.
L f s l s = 1 N × K j = 1 N × K log exp d f θ ( x j ) , P q c j · w j s c = 1 N exp d f θ ( x j ) , P q c
Finally, the weighted few-shot learning loss function of the prototype network is defined as Equation (9), where L f s l s denotes the loss of support set weight calibration and L f s l q denotes the loss of query set calibration.
L w f s l = L f s l s + L f s l q

3.4. Weighted Contrastive Regularization Function

To further enhance the model’s performance, this paper also augments the prototype network with a contrastive regularization function. It is expected to constrain the model so that samples within classes are closer together and samples between classes are farther apart. By following the supervised contrastive learning, the positive samples of a sample in the support set are the other samples of the same class, and all the remaining N × ( K 1 ) are the negative samples. For a positive sample pair, the contrastive loss item can be derived by Equation (10) after obtaining the sample features.
L c l i , j = log exp sim z i , z j / τ k = 1 N × ( K 1 ) exp sim z i , z k / τ
where z i and z j are the feature representations of the samples in the pair, and sim z i , z j measures the cosine similarity between them. τ is a temperature coefficient set to 0.5 in our experiment. And K is the number of samples per class in the support set.
Considering the presence of noisy labels, when there are noisy samples in the same class, these noisy samples are actually the negative samples instead of the positive samples. Thus, it is reasonable to exclude them from the set of positive samples. Inspired by the work of Yi et al. [58], this paper uses an indicator function to filter these fake positive sample pairs, as shown by the weighted contrastive regularization function of Equation (11).
L w c l = 1 N 2 × K 2 i = 1 N × K j = 1 , j i N × K L c l i , j · 1 { w i ( 1 β ) M s }
where 1 { w i ( 1 β ) K } is the indicator function whose output is equal to 1 when w i ( 1 β ) K and 0 when w i < ( 1 β ) K . β denotes a tolerance hyperparameter. Our idea behind the indicator function is that when all the samples of a class are clean and the same, the weights of the samples should be 1 / K . When there are noisy samples, for a sample with the weight w s , if w s > 1 / K , it means that the sample is closer to the other samples than the average case and has a higher probability of being a clean sample. Otherwise, if w s < 1 / K , it indicates that the sample is further away from the other samples than the average case and has a higher probability of being a noisy sample. Therefore, 1 / K is taken as the threshold which is further balanced by the hyperparameter β .
Finally, the overall loss function of the proposed model is defined as Equation (12) where λ 1 and λ 2 are the weights given to the two kinds of losses. Both λ 1 and λ 2 are set to be 1 in this paper.
L = λ 1 L w f s l + λ 2 L w c l

4. Experiments

To validate the proposed method, several experiments have been conducted. In this section, we describe the experimental setup and results.

4.1. Datasets Description

In order to fairly evaluate the performance of the proposed method, we have selected datasets that are widely used in related works for training and testing. There are four datasets used, including Indian Pines (IP), Pavia University (PU), Salians (SA), and Chikusei. In a few-shot setting [25], Chikusei is selected as the source dataset, and the other three datasets are selected as the target datasets.
(a)
Chikusei dataset: Acquired on 29 July 2014, in Chikusei, Ibaraki, Japan, the dataset was gathered using Hyperspectral Visible/Near-Infrared Cameras (Hyperspec-VNIRC). It encompasses 19 distinct classes and spans 2517 × 2335 pixels, maintaining a spatial resolution of 2.5 m per pixel. Comprising 128 spectral bands spanning from 363 to 1018 nm, this dataset provides comprehensive spectral information. The pseudo-color composite image and the ground-truth map of Chikusei are depicted in Figure 4. Table 1 shows the land cover classes and the corresponding numbers of samples in the Chikusei dataset.
(b)
IP dataset: Acquired via the airborne visible infrared imaging spectrometer (AVIRIS) sensor at the Indiana Pine test site situated in northwestern Indiana, this dataset comprises 145 × 145 pixels and initially included 224 spectral bands. However, 20 defective bands were removed, resulting in the utilization of the remaining 200 spectral bands for the experiment. With a spatial resolution of 20 m, this dataset encompasses 16 distinct classes. The pseudo-color composite image and the ground-truth map of IP are depicted in Figure 5. Table 2 shows the land cover classes and the corresponding numbers of samples in the IP dataset.
(c)
SA dataset: Collected using a 224-band AVIRIS sensor over the scenic Salinas Valley in California, USA, the Salinas Valley dataset boasts high spatial resolution at 3.7 m per pixel. Featuring an image encompassing 512 lines by 217 samples, this dataset underwent the removal of 20 abundant spectral bands, leaving behind 204 bands that effectively represent 16 diverse classes. The pseudo-color composite image and the ground-truth map of SA are depicted in Figure 6. Table 3 shows the land cover classes and the corresponding numbers of samples in the SA dataset.
(d)
PU dataset: Captured over the Pavia University, Italy, utilizing the reflective optics system imaging spectrometer (ROSIS-3), this image spans a size of 610 × 340 pixels and maintains a spatial resolution of 1.3 m per pixel across 115 spectral bands. Following the removal of 12 noisy bands, subsequent experiments were carried out using the remaining 103 bands. The pseudo-color composite image and the ground-truth map of PU are depicted in Figure 7. Table 4 shows the land cover classes and the corresponding numbers of samples in the PU dataset.

4.2. Experiment Setting

Since the research in this paper addresses the few-shot HSI classification with noisy labels, two types of state-of-the-art methods are compared in this paper. First, two latest few-shot HSI classification methods, Gia-FSL [12] and RPCL [13], are selected for comparison to indicate the impact of the noisy labels on few-shot HSI classification and the outperformance of the proposed method under such impact. Gia-FSL [12] combined graph information aggregation-based FSL with domain alignment to address domain bias. RPCL [13] integrated supervised contrastive learning into few-shot HSI classification to obtain better prototypes. Second, two typical methods which can handle HSI classification with noisy labels, SSRN [59] and DCRN [15], are also selected to indicate the superiority of the proposed method to address noisy labels in the task of few-shot HSI classification. SSRN designed an end-to-end spectral–spatial residual network for performing hyperspectral image classification. DCRN [15] employed a two-channel residual network structure and an anti-noise loss function to enhance the robustness of the model to noisy labels.
In the comparison, 200 labeled samples per class from the source dataset were randomly selected for training. For the target datasets, only five labeled samples per class were selected for training. These were augmented to 200 samples per class by cropping and restoration. And the rest of the samples in the target datasets were used for testing. In each episode of the training process, a N-way K-shot task was constructed to form the support set, where N is the number of classes in the dataset (e.g., N = 16 for IP and SA, and 9 for PU) and K is set to 5. At the same time, the number of samples per class composing the query set is set to 15, i.e., M = 15.
For the ablation experiment’s purpose, we name the method that removes the weighted contrastive regularization function from the proposed method as WCPN-CL, to validate the contribution of the weighted contrastive regularization function.
Since the datasets used for evaluation do not contain noise themselves, we need to set the noisy labels manually to validate the effectiveness of the proposed method. To simulate the real-world noise situation, we choose the paired label swap method [60] which swaps the labels of two samples selected randomly. This is in accordance with the premise that the number of classes in the datasets is determined. We selected the support set and query set from the dataset to which noise has been added, and the number of noise samples for each task has randomness due to random sampling, which is more in line with real noise-construction scenarios. For the proportion of samples over the datasets, we set it consistent with most of the methods [46,48], including 0% (i.e., no noise), 20%, 40%, and 60%, so as to better verify the model’s robustness at different noise levels. Excessive noise such as noise rate greater than 80% is often discarded in real scenarios, so we do not take it into account.
In the meta-testing phase, the testing samples from the target datasets are fed into the trained mapping module and feature extractor to obtain discriminative features, and then a KNN classifier is used to classify the unlabeled samples. Overall accuracy (OA), average accuracy (AA), and kappa coefficient were used to evaluate the performance of different methods. We randomly ran each experiment 10 times and report the average results. All experiments were conducted on a Pytorch computer with a 56 MHz CPU, Tesla T4 GPU, and 64 GB of RAM. Adam was used as the optimizer and the number of training iterations was set to 3000. The learning rate was set to 0.001. The window size of the input patches was set to 9 × 9.

4.3. Comparing with Other Methods

Table 5 and Table 6 show the classification performance of the different methods on the IP dataset with different noise settings. For the few-shot methods for HSI classification, Gia-FSL and RPCL are very sensitive to noise. As the noise rate goes from 0% to 60%, the performance of Gia-FSL and RPCL decreases by 8.79% and 24.9%, respectively. At the same time, the performance of SSRN and DCRN are relatively stable with increasing noise rates. However, their accuracy rates are lower than WPCN, which shows the effect of the calibration brought by weighted samples. For WCPN and WCPN-CL, they perform relatively well at all noise rates, especially at noise levels of 20% and 40%. The WPCN exhibits better robustness, and the changes in noise do not have a significant effect on it. Figure 8a visualizes the classification effect of different methods on the IP dataset, and the advantages of the WCPN-CL and WCPN become more obvious as the noise rate increases.
Table 7 and Table 8 show the classification performance of the different methods on the SA dataset. Compared with the performance on IP dataset, all methods are less affected by noise on this dataset. This may be because the HSI images of this dataset are easy to classify (the OA accuracy of most methods exceed 90%). WPCN only achieved an advantage with the noise rate of 0% and 20%, indicating that the role of sample weights for decreasing the impact of noisy labels is not significant for datasets that are easy to classify. It may be because a lot of useful information is filtered out when using the weights for calibration. Figure 8b visualizes the classification effectiveness of the different methods on the SA dataset, which also confirms the superiority of the proposed method.
Table 9 and Table 10 show the classification performance of the different methods on the PU dataset. WPCN achieves optimal performance over different levels of noise rate, outperforming the mainstream method DCRN by about 1.5% on average. Its noise immunity is also well demonstrated on the PU dataset. Although the increase in noise rate significantly affects the classification accuracy, the presence of weighted contrastive learning suppresses this effect to some extent due to the large differences between the noisy samples of different classes. Figure 8c visualizes the classification effect of different methods on the PU dataset. And the advantages of WCPN-CL and WCPN can be also found from the results in this figure.

4.4. Classification Visualization

In order to further demonstrate the noise immunity and the accuracy of the proposed method in the presence of increasing noise, we show the classification maps and 2D feature visualization of the different methods on the IP, SA, and PU datasets. However, due to space constraints, we only show RPCL [13] which is the latest few-shot method for HSI classification, and our WCPN-CL and WCPN methods. Figure 9, Figure 10 and Figure 11 are the classification maps, and Figure 12, Figure 13 and Figure 14 are the 2D feature visualizations. The results in these figures also confirm the advantages of the proposed method.
Figure 9 shows the classification map on the IP dataset. For the different methods in the presence of a gradually increasing noise rate, the degree of confusion of the classification also gradually increases. However, it can be clearly seen that RPCL is much more disorganized in the presence of noise than our WPCN-CL and WCPN methods. The WCPN achieves a good noise-resistant classification for both class 7 (Grass-pasture-mowed) on the left and class 8 (Hay-windrowed) on the right. Figure 12 shows the 2D feature visualization on the IP dataset. The clustering of each class keeps getting worse with the increase in noise, and our WPCN method shows significant advantages when noise rate is greater than 40%. Figure 10 shows the classification map on the SA dataset, where the WCPN achieves a better classification for the two classes located at the top left (Brocoli-green-weeds 2), even similar to the noiseless case. Figure 13 shows the 2D feature visualization on the SA dataset; the clustering of each class does not differ much when the noise rate increases, but there is an indistinguishable case in RPCL in orange color (class 8) and blue color (class 15), which coincides with the results in Table 7 and Table 8. Figure 11 shows the classification map on the PU dataset, where the WCPN achieves a better classification for class 8 (Bricks) located in the center. Figure 14 shows the 2D feature visualization on PU dataset, where the class 8 in orange color also gets better 2D clustering.

4.5. Sensitive Analysis of Parameters

As shown in Equation (11) above, there is a hyperparameter β in the weighted contrastive regularization function, which indicates how tolerant the regularization function is to the noise weights below the mean. In our experiments, we set β [ 0 , 0.3 ] , and changed it with a step size of 0.05 to see how sensitive the proposed method is to the change of β .
Figure 15 shows the classification results of the WCPN when changing the value of β under different noise rates. From the sub-figures (a–c), it can be seen that the sensitivity of the proposed method to the hyperparameter increases when the noise rate is greater than 40% on different datasets. When the noise rate is less than 20%, the change in the hyperparameter β does not have a significant impact on the proposed method. The reasons are that when β = 0 , the criteria for determining whether it is a clean or a noisy positive sample are strict, and when β is larger, the criteria are more lenient. When the noise rate is large, there are many noisy samples, and a smaller β will filter more noisy samples and improve the credibility of positive samples. There will be better classification performance. Meanwhile, when the noise rate is small, there are fewer noisy samples, and either harsh or lenient criteria will have less impact on the classification performance.

5. Discussion

In this study, we propose WPCN to address the challenges posed by noisy labels in few-shot hyperspectral image classification. While our method demonstrates promising performance in mitigating the impact of noisy labels, several points warrant further discussion.
Firstly, in our experiments, noise was introduced by swapping the labels of part of the data in one dataset with those of another part. This method simulates label errors but does not necessarily reflect the arbitrary nature of real-world label noise, which might not be readily identifiable. Future work could explore more realistic noise models to better simulate the variability and complexity of label errors encountered in practical scenarios. Secondly, we observed that WPCN shows higher noise resistance in datasets that are relatively easy to classify. However, its effectiveness is less pronounced in more challenging datasets. This suggests that our method may need further refinement to handle complex data more robustly. Finally, incorporating contrastive learning into the WPCN results in increased training time. While contrastive learning enhances feature representation, its computational cost is non-trivial. Future research should address how to optimize the training process to reduce computational expenses while maintaining or improving the method’s efficacy.

6. Conclusions

In this paper, we propose a WCPN to address the challenges posed by noisy labels in few-shot HSI classification. By utilizing the similarity between the samples from same classes to obtain the sample weights and applying it to the prototype calculation of support and query sets, the WCPN demonstrates some self-calibration ability in the presence of noisy samples. In addition, we introduce a weighted contrastive regularization function to enhance the degree of data aggregation and improve the differentiation of prototypes. Experiments on artificially generated noisy labels on several HSI datasets are conducted, and the results show that the WCPN has excellent performance in mitigating the impact of noisy labels. This shows that for HSI classification, the proposed method can better adapt to noisy environments where there are small amounts of labeled data available. But it is worth noting that the proposed method will cost more time because of the computation of the weights of the samples.

Author Contributions

Methodology, C.L.; Validation, C.L. and Z.H.; Formal analysis, D.Z., Y.R. and C.L.; Investigation, D.Z. and Z.H.; Resources, Z.H.; Data curation, Y.R. and Z.H.; Writing—original draft, Y.R.; Writing—review & editing, C.L. and J.W.; Visualization, Y.R.; Supervision, D.Z., C.L. and J.W.; Project administration, D.Z.; Funding acquisition, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by National Science and Technology Major Project for High Resolution Earth Observation System grant number 80-Y50G19-9001-22/23, and Henan Province Key Research and Development Special Project grant number 241111210300.

Data Availability Statement

All the codes are available at github https://github.com/12ian/WCPN-FSL. The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Boggs, J.L.; Tsegaye, T.; Coleman, T.; Reddy, K.; Fahsi, A. Relationship between hyperspectral reflectance, soil nitrate-nitrogen, cotton leaf chlorophyll, and cotton yield: A step toward precision agriculture. J. Sustain. Agric. 2003, 22, 5–16. [Google Scholar] [CrossRef]
  2. Lee, M.A.; Huang, Y.; Yao, H.; Thomson, S.J.; Bruce, L.M. Determining the effects of storage on cotton and soybean leaf samples for hyperspectral analysis. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2562–2570. [Google Scholar] [CrossRef]
  3. Pontius, J.; Martin, M.; Plourde, L.; Hallett, R. Ash decline assessment in emerald ash borer-infested regions: A test of tree-level, hyperspectral technologies. Remote Sens. Environ. 2008, 112, 2665–2676. [Google Scholar] [CrossRef]
  4. Dalponte, M.; ørka, H.O.; Gobakken, T.; Gianelle, D.; Næsset, E. Tree species classification in boreal forests with hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2012, 51, 2632–2645. [Google Scholar] [CrossRef]
  5. Chi, J.; Crawford, M.M. Spectral unmixing-based crop residue estimation using hyperspectral remote sensing data: A case study at Purdue university. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2014, 7, 2531–2539. [Google Scholar] [CrossRef]
  6. Yuan, Y.; Wang, Q.; Zhu, G. Fast hyperspectral anomaly detection via high-order 2-D crossing filter. IEEE Trans. Geosci. Remote Sens. 2014, 53, 620–630. [Google Scholar] [CrossRef]
  7. Paoletti, M.; Haut, J.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  8. Kang, J.; Zhang, Y.; Liu, X.; Cheng, Z. Hyperspectral Image Classification Using Spectral–Spatial Double-Branch Attention Mechanism. Remote Sens. 2024, 16, 193. [Google Scholar] [CrossRef]
  9. Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  10. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to compare: Relation network for few-shot learning. In Proceedings of the IEEE Conference On Computer Vision And Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1199–1208. [Google Scholar]
  11. Li, Z.; Liu, M.; Chen, Y.; Xu, Y.; Du, Q. Deep Cross-Domain Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5501618. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph information aggregation cross-domain few-shot learning for hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  13. Liu, Q.; Peng, J.; Ning, Y.; Chen, N.; Sun, W.; Du, Q.; Zhou, Y. Refined Prototypical Contrastive Learning for Few-Shot Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5506214. [Google Scholar] [CrossRef]
  14. Tu, B.; Zhou, C.; He, D.; Huang, S.; Plaza, A. Hyperspectral classification with noisy label detection via superpixel-to-pixel weighting distance. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4116–4131. [Google Scholar] [CrossRef]
  15. Xu, Y.; Li, Z.; Li, W.; Du, Q.; Liu, C.; Fang, Z.; Zhai, L. Dual-channel residual network for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5502511. [Google Scholar] [CrossRef]
  16. Lee, K.H.; He, X.; Zhang, L.; Yang, L. Cleannet: Transfer learning for scalable image classifier training with label noise. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5447–5456. [Google Scholar]
  17. Zhang, W.; Wang, Y.; Qiao, Y. MetaCleaner: Learning to Hallucinate Clean Representations for Noisy-Labeled Visual Recognition. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–19 June 2019. [Google Scholar]
  18. Han, J.; Luo, P.; Wang, X. Deep self-learning from noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–19 June 2019; pp. 5138–5147. [Google Scholar]
  19. Ma, X.; Huang, H.; Wang, Y.; Romano, S.; Erfani, S.; Bailey, J. Normalized loss functions for deep learning with noisy labels. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 6543–6553. [Google Scholar]
  20. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Li, S. Density peak-based noisy label detection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1573–1584. [Google Scholar] [CrossRef]
  21. Bateni, P.; Goyal, R.; Masrani, V.; Wood, F.; Sigal, L. Improved few-shot visual classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14493–14502. [Google Scholar]
  22. Bendre, N.; Marín, H.T.; Najafirad, P. Learning from few samples: A survey. arXiv 2020, arXiv:2007.15484. [Google Scholar]
  23. Koch, G.; Zemel, R.; Salakhutdinov, R. Siamese neural networks for one-shot image recognition. In Proceedings of the ICML Deep Learning Workshop, Lille, France, 6–11 July 2015; Volume 2. [Google Scholar]
  24. Rao, M.; Tang, P.; Zhang, Z. A Developed Siamese CNN with 3D Adaptive Spatial-Spectral Pyramid Pooling for Hyperspectral Image Classification. Remote Sens. 2020, 12, 1964. [Google Scholar] [CrossRef]
  25. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep few-shot learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 2290–2304. [Google Scholar] [CrossRef]
  26. Tang, H.; Li, Y.; Han, X.; Huang, Q.; Xie, W. A spatial–spectral prototypical network for hyperspectral remote sensing image. IEEE Geosci. Remote Sens. Lett. 2019, 17, 167–171. [Google Scholar] [CrossRef]
  27. Sun, J.; Shen, X.; Sun, Q. Hyperspectral Image Few-Shot Classification Network Based on the Earth Mover’s Distance. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  28. Gao, K.; Liu, B.; Yu, X.; Qin, J.; Zhang, P.; Tan, X. Deep relation network for hyperspectral image few-shot classification. Remote Sens. 2020, 12, 923. [Google Scholar] [CrossRef]
  29. Ma, X.; Ji, S.; Wang, J.; Geng, J.; Wang, H. Hyperspectral image classification based on two-phase relation learning network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10398–10409. [Google Scholar] [CrossRef]
  30. Bai, J.; Huang, S.; Xiao, Z.; Li, X.; Zhu, Y.; Regan, A.C.; Jiao, L. Few-shot hyperspectral image classification based on adaptive subspaces and feature transformation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  31. Xi, B.; Li, J.; Li, Y.; Song, R.; Hong, D.; Chanussot, J. Few-shot learning with class-covariance metric for hyperspectral image classification. IEEE Trans. Image Process. 2022, 31, 5079–5092. [Google Scholar] [CrossRef] [PubMed]
  32. Zhang, C.; Yue, J.; Qin, Q. Global prototypical network for few-shot hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2020, 13, 4748–4759. [Google Scholar] [CrossRef]
  33. Liu, C.; Yang, L.; Li, Z.; Yang, W.; Han, Z.; Guo, J.; Yu, J. Multi-level relation learning for cross-domain few-shot hyperspectral image classification. Appl. Intell. 2024, 54, 4392–4410. [Google Scholar] [CrossRef]
  34. Cheng, H.; Zhu, Z.; Li, X.; Gong, Y.; Sun, X.; Liu, Y. Learning with instance-dependent label noise: A sample sieve approach. arXiv 2020, arXiv:2010.02347. [Google Scholar]
  35. Wei, T.; Shi, J.X.; Tu, W.W.; Li, Y.F. Robust long-tailed learning under label noise. arXiv 2021, arXiv:2108.11569. [Google Scholar]
  36. Cheng, D.; Liu, T.; Ning, Y.; Wang, N.; Han, B.; Niu, G.; Gao, X.; Sugiyama, M. Instance-dependent label-noise learning with manifold-regularized transition matrix estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16630–16639. [Google Scholar]
  37. Xia, X.; Liu, T.; Han, B.; Wang, N.; Gong, M.; Liu, H.; Niu, G.; Tao, D.; Sugiyama, M. Part-dependent label noise: Towards instance-dependent label noise. Adv. Neural Inf. Process. Syst. 2020, 33, 7597–7610. [Google Scholar]
  38. Chen, L.H.; Li, H.; Zhang, W.; Huang, J.; Ma, X.; Cui, J.; Li, N.; Yoo, J. Anomman: Detect anomaly on multi-view attributed networks. arXiv 2022, arXiv:2201.02822. [Google Scholar] [CrossRef]
  39. Li, S.; Ge, S.; Hua, Y.; Zhang, C.; Wen, H.; Liu, T.; Wang, W. Coupled-view deep classifier learning from multiple noisy annotators. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4667–4674. [Google Scholar]
  40. Shu, J.; Xie, Q.; Yi, L.; Zhao, Q.; Zhou, S.; Xu, Z.; Meng, D. Meta-weight-net: Learning an explicit mapping for sample weighting. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  41. Wang, R.; Mou, S.; Wang, X.; Xiao, W.; Ju, Q.; Shi, C.; Xie, X. Graph structure estimation neural networks. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 342–353. [Google Scholar]
  42. Chen, P.; Chen, G.; Ye, J.; Zhao, J.; Heng, P.A. Noise against noise: Stochastic label noise helps combat inherent label noise. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  43. Hu, W.; Li, Z.; Yu, D. Simple and effective regularization methods for training on noisily labeled data with generalization guarantee. arXiv 2019, arXiv:1905.11368. [Google Scholar]
  44. Li, S.; Liu, T.; Tan, J.; Zeng, D.; Ge, S. Trustable Co-Label Learning From Multiple Noisy Annotators. IEEE Trans. Multimed. 2023, 25, 1045–1057. [Google Scholar] [CrossRef]
  45. Zhang, Y.; Zheng, S.; Wu, P.; Goswami, M.; Chen, C. Learning with feature-dependent label noise: A progressive approach. arXiv 2021, arXiv:2103.07756. [Google Scholar]
  46. Mazumder, P.; Singh, P.; Namboodiri, V.P. Rnnp: A robust few-shot learning approach. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 2664–2673. [Google Scholar]
  47. Lu, J.; Jin, S.; Liang, J.; Zhang, C. Robust few-shot learning for user-provided data. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1433–1447. [Google Scholar] [CrossRef] [PubMed]
  48. Liang, K.J.; Rangrej, S.B.; Petrovic, V.; Hassner, T. Few-shot learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 9089–9098. [Google Scholar]
  49. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral image classification in the presence of noisy labels. IEEE Trans. Geosci. Remote Sens. 2018, 57, 851–865. [Google Scholar] [CrossRef]
  50. Tu, B.; Zhang, X.; Kang, X.; Wang, J.; Benediktsson, J.A. Spatial density peak clustering for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5085–5097. [Google Scholar] [CrossRef]
  51. Jiang, J.; Ma, J.; Liu, X. Multilayer spectral–spatial graphs for label noisy robust hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 839–852. [Google Scholar] [CrossRef]
  52. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Online, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  53. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  54. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised contrastive learning. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
  55. Li, J.; Xiong, C.; Hoi, S.C. Learning from noisy data with robust representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 9485–9494. [Google Scholar]
  56. Wu, Z.F.; Wei, T.; Jiang, J.; Mao, C.; Tang, M.; Li, Y.F. Ngc: A unified framework for learning with open-world noisy data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 62–71. [Google Scholar]
  57. Li, S.; Xia, X.; Ge, S.; Liu, T. Selective-supervised contrastive learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 316–325. [Google Scholar]
  58. Yi, L.; Liu, S.; She, Q.; McLeod, A.I.; Wang, B. On learning contrastive representations for learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16682–16691. [Google Scholar]
  59. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  60. Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.; Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
Figure 1. Flowchart of the proposed WCPN, which consists of two meta-training processes running alternately on the source and target datasets. In the meta-training process, the support and query sets are first constructed, where the red squares are the noisy samples. Then two trainable adaptive mapping modules are used to unify the spectral dimensions of the source and target datasets and a feature extractor is adopted to extract the features of the support and query samples. After that, calibrated prototypes are obtained by calculating the weights of each sample in the support and query sets. Lastly, the weighted few-shot learning loss and weighted contrastive regularization loss are used to update the entire model.
Figure 1. Flowchart of the proposed WCPN, which consists of two meta-training processes running alternately on the source and target datasets. In the meta-training process, the support and query sets are first constructed, where the red squares are the noisy samples. Then two trainable adaptive mapping modules are used to unify the spectral dimensions of the source and target datasets and a feature extractor is adopted to extract the features of the support and query samples. After that, calibrated prototypes are obtained by calculating the weights of each sample in the support and query sets. Lastly, the weighted few-shot learning loss and weighted contrastive regularization loss are used to update the entire model.
Remotesensing 16 03527 g001
Figure 2. The architecture of the feature extractor.
Figure 2. The architecture of the feature extractor.
Remotesensing 16 03527 g002
Figure 3. Comparison of original prototype network with noisy labels and weighted contrastive prototype network with noisy labels. Red samples indicate noisy samples. The prototypes of the original prototype network are obtained from the mean of the support features and then the model is trained by minimizing the Euclidean distance between the query features and the prototypes. The WCPN calibrates the prototypes of support and query sets using the weights of the samples according to the similarities between the samples from the same classes, and updates the contrastive regularization function by taking the weights as gates to filter the potential noisy samples when calculating the contrastive loss.
Figure 3. Comparison of original prototype network with noisy labels and weighted contrastive prototype network with noisy labels. Red samples indicate noisy samples. The prototypes of the original prototype network are obtained from the mean of the support features and then the model is trained by minimizing the Euclidean distance between the query features and the prototypes. The WCPN calibrates the prototypes of support and query sets using the weights of the samples according to the similarities between the samples from the same classes, and updates the contrastive regularization function by taking the weights as gates to filter the potential noisy samples when calculating the contrastive loss.
Remotesensing 16 03527 g003
Figure 4. Pseudo-color composite image and ground-truth map of Chikusei. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Figure 4. Pseudo-color composite image and ground-truth map of Chikusei. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Remotesensing 16 03527 g004
Figure 5. Pseudo-color composite image and ground-truth map of IP. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Figure 5. Pseudo-color composite image and ground-truth map of IP. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Remotesensing 16 03527 g005
Figure 6. Pseudo-color composite image and ground-truth map of SA. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Figure 6. Pseudo-color composite image and ground-truth map of SA. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Remotesensing 16 03527 g006
Figure 7. Pseudo-color composite image and ground-truth map of PU. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Figure 7. Pseudo-color composite image and ground-truth map of PU. (a) False-color image. (b) Ground truth. (c) Labels illustration.
Remotesensing 16 03527 g007
Figure 8. OA results for GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods at different noise rates: (a) IP, (b) PU, and (c) SA.
Figure 8. OA results for GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods at different noise rates: (a) IP, (b) PU, and (c) SA.
Remotesensing 16 03527 g008
Figure 9. Classification map on the IP dataset. RPCL (0%), 74.64%; RPCL (20%), 66.37%; RPCL (40%), 58.2%; RPCL (60%), 49.74%; WCPN-CL (0%), 77.22%; WCPN-CL (20%), 76.92%; WCPN-CL (40%), 76.35%; WCPN-CL (60%), 73.97%; WCPN (0%), 77.4%; WCPN (20%), 76.96%; WCPN (40%), 76.45%; WCPN (60%), 75.2%.
Figure 9. Classification map on the IP dataset. RPCL (0%), 74.64%; RPCL (20%), 66.37%; RPCL (40%), 58.2%; RPCL (60%), 49.74%; WCPN-CL (0%), 77.22%; WCPN-CL (20%), 76.92%; WCPN-CL (40%), 76.35%; WCPN-CL (60%), 73.97%; WCPN (0%), 77.4%; WCPN (20%), 76.96%; WCPN (40%), 76.45%; WCPN (60%), 75.2%.
Remotesensing 16 03527 g009
Figure 10. Classification map on the SA dataset. RPCL (0%), 90.93%; RPCL (20%), 89.04%; RPCL (40%), 86.31%; RPCL (60%), 83.2%; WCPN-CL (0%), 92.62%; WCPN-CL (20%), 91.79%; WCPN-CL (40%), 91.34%; WCPN-CL (60%), 89.8%; WCPN (0%), 92.67%; WCPN (20%), 92.39%; WCPN (40%), 91.39%; WCPN (60%), 90.5%.
Figure 10. Classification map on the SA dataset. RPCL (0%), 90.93%; RPCL (20%), 89.04%; RPCL (40%), 86.31%; RPCL (60%), 83.2%; WCPN-CL (0%), 92.62%; WCPN-CL (20%), 91.79%; WCPN-CL (40%), 91.34%; WCPN-CL (60%), 89.8%; WCPN (0%), 92.67%; WCPN (20%), 92.39%; WCPN (40%), 91.39%; WCPN (60%), 90.5%.
Remotesensing 16 03527 g010
Figure 11. Classification map on the PU dataset. RPCL (0%), 82.72%; RPCL (20%), 78.1%; RPCL (40%), 74.55%; RPCL (60%), 70.77%; WCPN-CL (0%), 84.42%; WCPN-CL (20%), 85.62%; WCPN-CL (40%), 85.48%; WCPN-CL (60%), 83.19%; WCPN (0%), 85.15%; WCPN (20%), 86.75%; WCPN (40%), 85.78%; WCPN (60%), 83.53%.
Figure 11. Classification map on the PU dataset. RPCL (0%), 82.72%; RPCL (20%), 78.1%; RPCL (40%), 74.55%; RPCL (60%), 70.77%; WCPN-CL (0%), 84.42%; WCPN-CL (20%), 85.62%; WCPN-CL (40%), 85.48%; WCPN-CL (60%), 83.19%; WCPN (0%), 85.15%; WCPN (20%), 86.75%; WCPN (40%), 85.78%; WCPN (60%), 83.53%.
Remotesensing 16 03527 g011
Figure 12. 2-D feature visualization on IP.
Figure 12. 2-D feature visualization on IP.
Remotesensing 16 03527 g012
Figure 13. 2-D feature visualization on SA.
Figure 13. 2-D feature visualization on SA.
Remotesensing 16 03527 g013
Figure 14. 2-D feature visualization on PU.
Figure 14. 2-D feature visualization on PU.
Remotesensing 16 03527 g014
Figure 15. WCPN classification results for β at different noise rates: (a) IP, (b) PU, and (c) SA.
Figure 15. WCPN classification results for β at different noise rates: (a) IP, (b) PU, and (c) SA.
Remotesensing 16 03527 g015
Table 1. Land cover classes and numbers of sampless in Chikusei.
Table 1. Land cover classes and numbers of sampless in Chikusei.
ClassNamePixelsClassNamePixels
1Water234511Row crops5961
2Bare soil (school)285912Plastic house2193
3Bare soil (park)23613Manmade (non-dark)1220
4Bare soil (farmland)48,52514Manmade (dark)7664
5Natural plants429715Manmade (blue)431
6Weeds in farmland110816Manmade (red)222
7Forest20,51617Manmade grass1040
8Grass651518Asphalt801
9Rice field (grown)13,36919Paved ground145
10Rice field (first stage)1268Total: 77,592
Table 2. Land cover classes and numbers of samples in IP.
Table 2. Land cover classes and numbers of samples in IP.
ClassNamePixelsClassNamePixels
1Alfalfa4610Soybean-notill972
2Corn-notill142811Soybean-mintill2455
3Corn-mintill83012Soybean-clean593
4Corn23713Wheat205
5Grass-pasture48314Woods1265
6Grass-trees73015Buildings-Grass-Trees-Drives386
7Grass-pasture-mowed2816Stone-Steel-Towers93
8Hay-windrowed478Total: 10,249
9Oats20
Table 3. Land cover classes and numbers of samples in SA.
Table 3. Land cover classes and numbers of samples in SA.
ClassNamePixelsClassNamePixels
1Brocoli_green_weeds 1200910Corn_senesced_green_weeds3278
2Brocoli_green_weeds 2372611Lettuce_romaine_4wk1068
3Fallow197612Lettuce_romaine_5wk1927
4Fallow_rough_plow139413Lettuce_romaine_6wk916
5Fallow_smooth267814Lettuce_romaine_7wk1070
6Stubble395915Vinyard_untrained7268
7Celery357916Vinyard_vertical_trellis1807
8Grapes_untrained11,271Total: 54,129
9Soil_vinyard_develop6203
Table 4. Land cover classes and numbers of samples in PU.
Table 4. Land cover classes and numbers of samples in PU.
ClassNamePixels
1Asphalt6631
2Meadows18,649
3Gravel2099
4Trees3064
5Metal sheets1345
6Bare soil5029
7Bitumen1330
8Bricks3682
9Shadows947
Total 42,776
Table 5. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the IP dataset with noise rate 0% and 20%.
Table 5. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the IP dataset with noise rate 0% and 20%.
ClassNoise Rate = 0%Noise Rate = 20%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
189.7699.7699.5199.7699.0299.7688.5496.8399.2799.7699.5198.54
241.3364.2765.0465.470.7168.433.4648.3860.8763.9469.6965.73
344.0661.9961.3167.8564.2764.3631.6151.0967.4161.6167.1566.81
472.4688.3690.5292.8489.6190.0448.3276.5588.9292.3390.2692.2
569.6779.8579.2580.1780.980.9464.3173.0877.279.8178.8379.29
676.2390.6989.8592.7990.7689.2669.6186.1991.2391.7791.1390.83
799.1310010010010010090.4399.1399.5710098.7100
888.4491.5496.2894.898.0898.8485.5889.9294.3194.3394.0694.29
99810010099.3310010096.67100100100100100
1057.3969.1970.5671.956872.0953.0764.9871.3468.8470.869.98
1158.8867.3666.2265.870.470.4254.1958.4662.8967.6867.9872.7
124460.4368.1168.5766.666.1232.6542.3361.8963.0865.9467.23
1396.959896.2597.796.5596.494.897.297.9596.598.698
1476.5889.2989.389.2589.991.870.7988.889090.3891.4888.03
1569.7185.7283.3990.5288.4887.0947.7472.0781.3688.1187.1783.96
1699.296.9397.3998.189595.6892.9597.3997.7398.398.397.39
OA61.62
±2.98
74.64
±2.88
75.05
±3.05
76.22
±3.15
77.22
±2.28
77.4
±2.83
54.24
±4.26
66.37
±3.01
73.79
±2.59
75.26
±3.66
76.92
±3.53
76.96
±2.38
AA73.86
±1.68
83.96
±1.63
84.56
±1.43
85.93
±1.7
85.52
±1.32
85.7
±1.55
65.92
±2.59
77.66
±1.72
83.87
±1.27
84.78
±2.0
85.6
±1.47
85.31
±1.62
Kappa0.56
±0.03
0.71
±0.03
0.71
±0.03
0.73
±0.03
0.74
±0.03
0.74
±0.03
0.48
±0.05
0.62
±0.03
0.70
±0.03
0.72
±0.04
0.74
±0.04
0.74
±0.03
Table 6. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the IP dataset with noise rate 40% and 60%.
Table 6. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the IP dataset with noise rate 40% and 60%.
ClassNoise Rate = 40%Noise Rate = 60%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
189.2792.299.5199.7699.2799.5184.3967.5698.2999.7698.5498.78
232.2437.2459.9562.7562.4265.4329.9926.4454.8854.4860.0262.45
329.2843.6654.1265.1461.8363.6130.2229.4960.5263.8164.7260.42
456.1649.1484.9191.2589.8789.2746.6426.0382.0787.8988.5886.38
561.9765.2176.3279.6280.0480.1357.4759.1471.4480.5280.0879.54
665.6776.9189.3492.6193.5691.5266.2373.3988.9492.8790.8792.8
796.5297.3910099.5710010092.6191.3100100100100
882.1481.2592.0195.6794.5995.9885.2969.7787.7692.1692.1894.84
996981001001001009890.67100100100100
1052.8356.5367.3870.7270.0271.8152.1550.3667.3164.5971.1972.7
1152.656.7669.0667.1971.2868.753.1852.1364.0268.3663.9866.98
1232.7731.1159.4966.4363.4465.3631.2421.959.8863.4266.0566.77
1396.1596.0598.498.1598.897.793.2590.2597.8597.89796.5
1474.3680.8492.1390.0892.6392.7573.2472.5288.6191.0291.0690.81
1547.3847.4585.8389.584.6584.1246.5932.4481.7886.0678.8583.94
1696.0284.298.9897.1698.1897.592.7368.1898.6497.9597.6197.16
OA53.61
±3.36
58.2
±3.67
73.64
±2.58
75.76
±2.64
76.35
±2.42
76.45
±2.97
52.83
±3.37
49.74
±3.07
71.12
±2.66
73.83
±3.76
73.97
±2.74
75.2
±2.64
AA66.34
±3.53
68.37
±2.51
82.96
±1.32
85.35
±1.74
85.04
±1.52
85.21
±1.92
64.58
±2.91
57.6
±3.56
81.37
±1.31
83.79
±2.2
83.8
±1.65
84.38
±1.82
Kappa0.48
±0.04
0.53
±0.04
0.70
±0.03
0.73
±0.03
0.73
±0.03
0.73
±0.03
0.47
±0.04
0.43
±0.03
0.68
±0.03
0.71
±0.04
0.71
±0.03
0.72
±0.03
Table 7. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the SA dataset with noise rates of 0% and 20%.
Table 7. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the SA dataset with noise rates of 0% and 20%.
ClassNoise Rate = 0%Noise Rate = 20%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
198.6299.3999.2497.7899.4799.7896.3598.2999.4499.3998.3698.7
299.4299.9499.3899.1799.9699.8797.7799.5599.4499.8299.9399.56
390.790.5990.5591.7394.1695.3187.9791.2292.294.6194.0490.63
498.8199.1899.0398.8198.8399.6399.1699.4899.5899.5999.5299.55
589.6290.9691.8392.693.1294.3492.5491.7194.4495.3594.5596.23
69998.8198.2498.8498.7798.7598.8898.9998.9199.1498.9298.51
798.5999.5596.7998.9899.5599.4398.5599.2397.3198.1499.1799.47
874.8580.3781.4380.4583.6984.8269.8572.9578.6876.6682.1281.03
997.4399.739999.1599.5299.997.8899.8499.1899.8599.4199.73
1079.9386.6686.0387.7191.0690.9372.1785.4689.1890.9388.2292.13
1196.6298.8998.5498.6799.1499.6395.9698.2899.5499.1899.4499.69
1299.1298.8298.8998.4696.4798.2297.8199.6199.8499.3798.6898.76
1398.1699.2599.0799.0899.1796.9899.4299.299.2499.5399.1799.34
1498.1497.9198.5199.4697.299.0398.2398.5798.4499.595.9797.43
1573.1879.2378.5381.4882.5979.5771.9176.7983.8282.4279.6682.15
1689.3391.5995.7793.6196.9496.4784.4292.3292.995.0696.4196.34
OA87.98
±2.34
90.93
±1.76
90.84
±2.34
91.28
±1.89
92.62
±1.37
92.67
±1.4
86
±2.49
89.04
±1.47
91.45
±1.88
91.33
±2.16
91.79
±1.82
92.15
±1.61
AA92.59
±1.35
94.43
±1.2
94.43
±1.61
94.75
±1.28
95.6
±0.95
95.79
±1.08
91.18
±1.37
93.84
±0.74
95.13
±1.13
95.54
±0.96
95.22
±1.15
95.58
±1.18
Kappa0.87
±0.03
0.90
±0.02
0.90
±0.03
0.90
±0.02
0.92
±0.02
0.92
±0.02
0.84
±0.03
0.88
±0.02
0.91
±0.02
0.90
±0.02
0.91
±0.02
0.91
±0.02
Table 8. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the SA dataset with noise rates of 40% and 60%.
Table 8. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the SA dataset with noise rates of 40% and 60%.
ClassNoise Rate = 40%Noise Rate = 60%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
198.1798.3198.8499.699.3498.8495.7893.4198.3299.4498.8299.5
296.1799.0499.299.5299.1198.497.8297.0299.1499.5499.4799.2
387.8490.5993.6492.5790.6793.2785.7887.5287.589.6488.2489.1
499.699.3499.6899.5399.5299.6999.4299.0199.7899.6299.5999.6
591.9391.2894.6195.7294.0792.9691.3790.6293.4994.993.9294.52
699.2798.8199.4699.5999.3199.399.1498.1398.9899.7599.4299.29
799.0698.4699.1498.3697.6493.8899.598.2499.4998.999.6497.42
868.7368.2978.3674.6280.7180.7670.8865.6873.1874.8475.1376.1
997.7299.7199.5399.7899.3999.5698.1797.3499.6699.599.0899.53
1075.3577.588991.0289.4790.9778.1769.1185.6289.7485.4187.54
1197.5997.3799.5399.4699.0699.0196.5892.898.6598.7699.0899.29
1299.5999.3199.6799.4399.8499.339995.8599.0199.3298.8599.83
1399.3198.9199.799.699.5998.9999.0998.8899.4299.8699.4599.77
1497.6297.1998.8599.5999.3498.7897.4296.0598.7899.398.8998.19
1574.1269.5781.981.9279.0178.1467.2163.3581.0981.5979.3581.62
1687.6589.7195.9995.9796.296.886.2584.6194.669491.6393.75
OA86.42
±1.24
86.31
±1.61
91.45
±1.77
90.85
±2.23
91.34
±1.75
91.05
±2.66
86
±2.41
83.2
±1.25
89.67
±2.69
90.56
±2.17
89.8
±3.41
90.5
±2.4
AA91.86
±1.35
92.09
±0.89
95.45
±1.09
95.39
±1.24
95.14
±0.92
94.92
±1.82
91.35
±1.66
89.23
±1.11
94.17
±2.1
94.92
±1.32
94.12
±2.55
94.64
±1.71
Kappa0.85
±0.01
0.85
±0.02
0.91
±0.02
0.90
±0.02
0.90
±0.02
0.90
±0.03
0.84
±0.03
0.81
±0.01
0.89
±0.03
0.90
±0.02
0.89
±0.04
0.89
±0.03
Table 9. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the PU dataset with noise rates of 0% and 20%.
Table 9. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the PU dataset with noise rates of 0% and 20%.
ClassNoise Rate = 0%Noise Rate = 20%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
179.8885.491.5189.9288.5986.8273.0680.587.7787.9388.9489.81
287.779.9979.6978.3581.4482.5780.1877.5783.5780.5483.4986.39
356.2465.5476.476.179.3780.655.9867.0978.5177.1382.3577.79
493.1191.899292.2786.9585.7289.691.4589.6490.7687.2188.17
598.8299.3998.5898.6999.1999.0994.1798.4499.7199.1999.699.72
674.8882.6875.2684.0782.1381.9262.2170.777.9183.8479.4779.64
775.4684.2992.8389.1592.1293.1874.2679.6994.3194.629391.09
872.383.4792.0889.6884.5590.8463.0168.5886.2888.8288.4387.84
997.2897.7596.8595.6896.6696.596.698.1296.1395.9998.1596.82
OA82.68
±2.88
82.72
±4.26
84.17
±4.75
84.03
±4.23
84.42
±4.5
85.15
±4.07
75.59
±4.24
78.1
±3.85
85.09
±3.19
84.71
±3.95
85.62
±4.11
86.75
±3.56
AA81.74
±1.21
85.6
±3.11
88.36
±2.2
88.21
±2.63
87.89
±2.88
88.58
±2.42
76.56
±1.74
81.35
±2.35
88.2
±2.84
88.76
±2.59
88.96
±2.57
88.58
±3.12
Kappa0.77
±0.03
0.78
±0.05
0.80
±0.05
0.80
±0.05
0.80
±0.05
0.81
±0.05
0.69
±0.05
0.72
±0.04
0.81
±0.04
0.80
±0.05
0.81
±0.05
0.83
±0.04
Table 10. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the PU dataset with noise rates of 40% and 60%.
Table 10. Classification performance of the GIAFSL, RPCL, SSRN, DCRN, WCPN-CL, and WCPN methods on the PU dataset with noise rates of 40% and 60%.
ClassNoise Rate = 40%Noise Rate = 60%
GIAFSLRPCLSSRNDCRNWCPN-CLWCPNGIAFSLRPCLSSRNDCRNWCPN-CLWCPN
171.6975.1389.2387.8890.1289.6670.5768.2586.3186.9987.9686.45
284.6476.1581.779.3983.3282.9382.0477.5477.9579.3679.4780.5
360.356.9374.2273.2377.7578.2552.9455.6675.4568.2279.3874.28
489.2489.992.1491.5991.3889.6789.6173.4490.6792.6890.2990.05
595.5395.4999.4998.9799.1899.8195.7788.1199.7397.6899.5999.45
659.5364.4879.7785.5277.881.3554.4360.2375.4677.5278.5881.66
781.1579.995.0891.6895.7493.4179.8378.7492.3290.6992.9592.58
861.5861.1786.3886.586.2489.4361.8246.681.6481.1482.7183.05
996.9796.8697.194.797.8197.8796.6893.9797.4194.4398.0598.81
OA77.34
±3.48
74.55
±3.79
84.74
±3.23
83.94
±4.83
85.48
±3.47
85.78
±3.89
75.08
±2.88
70.77
±2.72
81.62
±4.21
82.14
±5.52
83.19
±4.81
83.53
±4.09
AA77.85
±1.61
77.34
±1.31
88.35
±1.75
87.72
±2.14
88.81
±1.82
89.15
±1.83
75.97
±1.54
71.39
±2.42
86.33
±2.74
85.41
±1.99
87.67
±2.08
87.43
±1.77
Kappa0.70
±0.04
0.67
±0.04
0.80
±0.04
0.80
±0.05
0.81
±0.04
0.82
±0.05
0.68
±0.03
0.62
±0.03
0.77
±0.05
0.77
±0.06
0.79
±0.05
0.79
±0.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, D.; Ren, Y.; Liu, C.; Han, Z.; Wang, J. Weighted Contrastive Prototype Network for Few-Shot Hyperspectral Image Classification with Noisy Labels. Remote Sens. 2024, 16, 3527. https://doi.org/10.3390/rs16183527

AMA Style

Zhang D, Ren Y, Liu C, Han Z, Wang J. Weighted Contrastive Prototype Network for Few-Shot Hyperspectral Image Classification with Noisy Labels. Remote Sensing. 2024; 16(18):3527. https://doi.org/10.3390/rs16183527

Chicago/Turabian Style

Zhang, Dan, Yiyuan Ren, Chun Liu, Zhigang Han, and Jiayao Wang. 2024. "Weighted Contrastive Prototype Network for Few-Shot Hyperspectral Image Classification with Noisy Labels" Remote Sensing 16, no. 18: 3527. https://doi.org/10.3390/rs16183527

APA Style

Zhang, D., Ren, Y., Liu, C., Han, Z., & Wang, J. (2024). Weighted Contrastive Prototype Network for Few-Shot Hyperspectral Image Classification with Noisy Labels. Remote Sensing, 16(18), 3527. https://doi.org/10.3390/rs16183527

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop