Next Article in Journal
Challenges and Opportunities in Remote Sensing for Soil Salinization Mapping and Monitoring: A Review
Previous Article in Journal
Exploring the Spatial and Temporal Characteristics of China’s Four Major Urban Agglomerations in the Luminous Remote Sensing Perspective
Previous Article in Special Issue
Camouflaged Object Detection Based on Ternary Cascade Perception
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Agreement and Disagreement-Based Co-Learning with Dual Network for Hyperspectral Image Classification with Noisy Labels

1
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
2
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
3
School of Computer Science and Technology, Tiangong University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(10), 2543; https://doi.org/10.3390/rs15102543
Submission received: 14 April 2023 / Revised: 9 May 2023 / Accepted: 10 May 2023 / Published: 12 May 2023

Abstract

:
Deep learning-based label noise learning methods provide promising solutions for hyperspectral image (HSI) classification with noisy labels. Currently, label noise learning methods based on deep learning improve their performance by modifying one aspect, such as designing a robust loss function, revamping the network structure, or adding a noise adaptation layer. However, these methods face difficulties in coping with relatively high noise situations. To address this issue, this paper proposes a unified label noise learning framework with a dual-network structure. The goal is to enhance the model’s robustness to label noise by utilizing two networks to guide each other. Specifically, to avoid the degeneration of the dual-network training into self-training, the “disagreement” strategy is incorporated with co-learning. Then, the “agreement” strategy is introduced into the model to ensure that the model iterates in the right direction under high noise conditions. To this end, an agreement and disagreement-based co-learning (ADCL) framework is proposed for HSI classification with noisy labels. In addition, a joint loss function consisting of a supervision loss of two networks and a relative loss between two networks is designed for the dual-network structure. Extensive experiments are conducted on three public HSI datasets to demonstrate the robustness of the proposed method to label noise. Specifically, our method obtains the highest overall accuracy of 98.62%, 90.89%, and 99.02% on the three datasets, respectively, which represents an improvement of 2.58%, 2.27%, and 0.86% compared to the second-best method. In future research, the authors suggest using more networks as backbones to implement the ADCL framework.

1. Introduction

With the development of spectral imaging technology, hyperspectral image (HSI) has been widely used in various fields such as agricultural monitoring [1], food quality inspection [2], urban ground object recognition and classification [3], post-disaster change detection [4], and soil heavy metal detection [5]. The classification task is essential for these hyperspectral remote sensing applications. In the past few years, traditional machine learning methods such as support vector machine [6], random forests [7], extreme learning machine [8], and sparse representation classifier [9] have played an important role in HSI classification. Recently, deep learning has brought new prosperity to HSI classification with its powerful representation learning ability [10,11]. In general, both traditional and deep learning-based methods use a certain number of accurately labeled samples to train reliable models. However, in real situations, the training sets that are available for training usually contain mislabeled or wrong samples, which is called the label noise problem. Label noise is not conducive to training effective models.
Many attempts have been made to address the label noise problem in HSI classification. Representative works of traditional methods include the following. Given that noisy labels are usually located in low-density regions, Refs. [12,13,14] investigated a series of density peak-based noisy label detection methods. To increase the detection accuracy of noisy labels, Ref. [15] proposed a random label propagation algorithm (RLPA) to detect label noise. The main idea of RLPA is to randomly divide the data and perform label propagation several times, and then ensemble the outcomes of multiple label propagations for noisy label detection. To overcome the shortcoming that RLPA is sensitive to superpixel segmentation scale, a multi-scale superpixel segmentation method and a new similarity graph construction approach were proposed [16]. In order to overcome the influence of random noise and edge noise on label information, the spectral–spatial sparse graph was introduced into RLPA to construct an adaptive label propagation algorithm [17]. Since ensemble learning can enhance the robustness of the model, Ref. [18] proposed an adapted random forest that can consider mislabeled training labels.
Compared to traditional label noise learning methods, deep learning-based label noise learning methods have more advantages owing to the powerful discriminative feature representation learning abilities of deep neural networks. Recently, researchers have investigated deep learning methods for HSI classification in the presence of label noise. For instance, an entropic optimal transport loss was designed for end-to-end style deep neural networks to improve their robustness to label noise [19]. In order to enhance the robustness of the classification model, Ref. [20] investigated a novel dual-channel network structure and a noise-robust loss function. Ref. [21] designed a superpixel-guided sample network framework with end-to-end training style for handling label noise, comprising two stages: sample selection and sample correction. Ref. [22] proposed a lightweight heterogeneous kernel convolution (HetConv3D) to improve the robustness of the network to label noise. HetConv3D used two different types of convolutional kernels. Ref. [23] employed both labeled and unlabeled data to build a unified deep learning network, which was shown to be robust to noisy labels. To handle label noise and limited samples simultaneously, Ref. [24] presented a novel dual-level deep spatial manifold representation (SMR) network for HSI classification, embedding SMR-based feature extraction and classifier blocks into one framework. Ref. [25] investigated the robustness of several loss functions to convolutional neural networks and proposed an HSI pixel-to-image sampling method to prevent overfitting on label noise. To address the inaccurate supervision caused by label noise, selective complementary learning was introduced into convolutional neural networks for HSI classification with noisy labels [26].
The above methods have made positive contributions to HSI classification in the presence of label noise. Traditional methods typically detect and remove mislabeled samples before constructing classification models, while deep learning-based methods do not require this step. Instead, deep learning-based methods consider the effect of label noise on model construction and design robust loss functions or specific network structures to improve the learning ability of label noise. However, current deep learning-based methods for HSI classification with noisy labels still have limitations. For example, they often focus on improving one aspect, such as designing a robust loss function, revamping the network structure, or adding a noise adaptation layer, which may not be sufficient to handle relatively high noise rates. This issue deserves further study.
To address the above issue, we propose a unified label noise learning framework that can be adapted to various deep neural networks. Inspired by collaborative learning, we design a disagreement-based co-learning (DCL) framework with a dual-network structure, in which the “disagreement” strategy is incorporated with co-learning. In DCL, the two networks attempt to cross-propagate their own losses to the peer network through the “disagreement” strategy, which can avoid the dual-network training degenerating into self-training. However, the “disagreement” strategy can only select a subset of training samples, which are not guaranteed to have real labels especially with high label noise. Therefore, we introduce the idea of “agreement” in co-training into DCL and propose an agreement and disagreement-based co-learning (ADCL) framework for HSI classification with noisy labels. Additionally, a joint loss function is designed for our dual-network framework. The designed loss consists of a supervision loss of two networks and a relative loss between two networks.
The remainder of this paper is organized as follows. Related work and contributions are described in Section 2. The detailed description of ADCL is shown in Section 3. Experimental results and analysis are reported in Section 4. Section 5 shows the discussion. The conclusions of this work are shown in Section 6.

2. Related Work and Contributions

2.1. Label Noise Learning Based on Deep Learning

Deep learning-based label noise learning has been extensively studied in the fields of machine learning and computer vision. Some surveys on this topic can be found in [27,28,29]. Deep learning-based label noise learning approaches can be roughly divided into five categories:
(1)
Robust network architecture: Adding a noise adaptation layer [30] or designing a specific architecture [31] to improve the reliability of estimating label transition probabilities and to mimic the label transition behavior in deep network learning. The goal of the specific architecture is to improve the reliability of estimating label transition probabilities.
(2)
Robust loss function: Developing a loss function that is robust to label noise [32,33]. Generally, robust loss functions attempt to achieve a small risk on the training set with label noise. Current studies of robust loss function mainly rely on the basis of mean absolute error loss and cross entropy loss.
(3)
Robust regularization: Adding a regularization term into optimization objective to alleviate the overfitting of deep learning on training samples with label noise. Regularization techniques include explicit regularization (such as weight decay [34] and dropout [35]) and implicit regularization (such as mini-batch stochastic gradient descent [36] and data augmentation [37]).
(4)
Loss adjustment: Adjusting the loss of all training samples to reduce the effects of label noise. Unlike robust loss functions, loss adjustment adjusts update rules to minimize the negative effects of label noise. Loss adjustment includes loss correction [38], loss reweighting [39], and label refurbishment [40].
(5)
Sample selection: Selecting true-labeled samples from the training set with noisy labels. The aim of sample selection is to update deep neural networks for the selected clean samples. Sample selection generally includes multi-network collaborative learning [21,41], multi-round iterative learning [42], and the combination with other learning paradigms [43].

2.2. Deep Neural Network-Based Label Noise Learning in Remote Sensing

In the remote sensing (RS) field, various deep neural network-based label noise learning methods have been investigated, including:
(1)
Specific network architecture [20,22,23,24,44,45,46,47]. For synthetic aperture radar images, Ref. [44] designed a noise-tolerant network based on layer attention. The developed layer attention module adaptively weights the features of different convolution layers. To handle the noisy label data for building extraction, Ref. [46] proposed a general deep neural network model that is adaptive to label noise, which consists of a base network and an additional probability transition module. To suppress the impact of label noise on the semantic segmentation of RS images, Ref. [47] constructed a general network framework by combining an attention mechanism and a noise robust loss function.
(2)
Robust loss function [19,20,25,47,48,49,50]. Ref. [47] added two hyperparameters into the symmetric cross-entropy loss function for label noise learning. Refs. [48,49] proposed two novel loss functions for deep learning, the first being the robust normalized softmax loss used for the characterization of RS images based on deep metric learning, and the second being the noise-tolerant deep neighborhood embedding, which accurately encodes the semantic relationships among RS scenes. Ref. [50] constructed a joint loss consisting of a cross-entropy loss with the updated label and a cross-entropy loss with the original noisy label.
(3)
Label correction [45,50,51,52,53,54]. For road extraction from RS images, Ref. [45] introduced label probability sequence into sequence deep learning framework for correcting error labels. Ref. [50] utilized the information entropy to measure the uncertainty of the prediction, which served as a basis for label correction. Ref. [51] adopted unsupervised clustering to recognize the sample’s label, and the network trained on augmented samples with clean labels was used to correct noisy labels further. Similarly, for object detection in aerial images, Ref. [52] designed a new noise filter named probability differential to recognize and correct mislabeled labels. Ref. [53] used the initial pixel-level labels to train an under-trained initial network that was treated as starting training for network updating and initial label correction. In addition, Ref. [54] proposed a novel adaptive multi-feature collaborative representation classifier to correct the labels of uncertain samples.
(4)
Hybrid approach [21,23,26,51,55,56,57,58]. Both [51] and [55] introduced unsupervised method into label noise leaning. In [55], an unsupervised method was combined with domain adaptation for HSI classification. In addition, complementary learning was combined with deep learning for HSI classification [26] and RS scene classification [56]. Recently, Ref. [57] incorporated knowledge distillation with label noise learning to improve building extraction. To obtain datasets containing less noise, Ref. [58] introduced semisupervised learning into the objective learning framework to produce a low-noise dataset.

2.3. Co-Training in Remote Sensing

Co-training, originally proposed by Blum and Mitchell [59], uses two sufficiently redundant and conditionally independent views to improve the generalization performance of the model. In the past few years, researchers have studied the theory of co-training and developed various variations of co-training. Recently, some studies incorporated co-training with deep learning for label noise learning [60,61,62,63].
In the study of the remote sensing field, the idea of co-training has been introduced into several tasks such as land cover mapping, image segmentation, image classification and recognition, and so on [64,65,66,67,68,69]. For example, Ref. [64] proposed an improved co-training method for semisupervised HSI classification, which used spectral features and two-dimensional Gabor features as two different views to train collaboratively. Similarly, Ref. [65] implemented a co-training paradigm with the P-N learning method, in which the P-expert assumes that adjacent pixels in space have the same class label, while the N-expert believes that the pixels with similar spectra have the same class label. Then, co-training was combined with a deep stacked autoencoder for semisupervised HSI classification [66]. In the application of RS, Ref. [67] proposed conditional co-training method and applied it to RS image segmentation in coastal areas. Refs. [68,69] proposed novel co-training methods for land cover mapping and crop mapping, respectively.

2.4. Contributions

Compared to previous work, we improve the label noise robustness of the model by addressing both network structure and loss function. Specifically, we construct a unified dual-network structure that leverages the mutual information between the two networks to guide each other. In addition, we design a more robust loss function for the specific network structure. The main contributions of our work are as follows:
(1)
A new framework incorporating “disagreement” strategy into co-learning, named DCL, is proposed for HSI classification with noisy labels.
(2)
A stronger framework that introduces an “agreement” strategy into DCL, termed ADCL, is designed.
(3)
A joint loss function is proposed for the dual-network structure.
(4)
Extensive experiments on public HSI data sets demonstrate the effectiveness of the proposed method.

3. Proposed ADCL Method

The loss function is an essential component of deep neural networks, and a noise robust loss function can significantly improve their performance. In the proposed ADCL framework, two networks with the same structure are used, and a joint loss function is designed to make the framework more robust. The loss function takes into account the supervision information and the mutual guidance information between the two networks.
The main idea of ADCL is to have the two networks guide each other in learning. To achieve this goal, in the training process of ADCL, two networks predict all samples, and the samples with inconsistent predictions constitute the disagreement data. At the end of forward propagation, each deep network selects data with small loss from disagreement data to minimize the loss of the network. In the back propagation process, each network uses the data with small loss from the peer network to update the weight parameters. To make the model more powerful, in addition to selecting its own small loss data from disagreement data, each network also adds the data with the same classification on two networks into the peer network for back propagation. This design makes full use of the mutual information between the two networks, enhancing the noise tolerance of the model.
Taking CNN as the backbone for our ADCL, Figure 1 shows the overall framework of ADCL. Firstly, the training data with noisy labels is fed into two CNNs: A and B . Secondly, after training one mini-batch, disagreement and consistent data predictions will be generated by the two networks. Thirdly, the two networks select their own small loss data according to the designed loss. Fourthly, each network uses the small loss data from the peer network and consistent data from two networks to update its own convolution kernels. Finally, after multiple epochs of training and updating, the two trained networks are combined to generate the classification map.
In the next subsections, we will introduce the designed joint loss, then detail the proposed ADCL framework, and finally show the formula analysis of the proposed method.

3.1. Joint Loss

In the case of the dual network, the most straightforward way to construct a loss function is to apply independent regularization when training each individual network. Although regularization can improve generalization performance by promoting consistency between two networks, it can still be influenced by the memory effect of label noise [35]. Therefore, we adopt a joint loss function based on regularization techniques in this work.
Let T = x i , y i i = 1 N be the training set with N samples, where x i represents the ith sample, and its corresponding observation label is y i 1 , , C . The joint loss function is designed as
l x i = β l S x i , y i + 1 β l R x i ,
where l S represents the supervision loss under two networks, l R represents the relative loss, and parameter β is utilized to balance the supervision loss and relative loss.
The symmetric cross-entropy (SCE) adds reverse cross-entropy (RCE) to cross-entropy (CE), so as to have a certain robustness to label noise [70]. This paper adopts SCE to construct supervision loss l S . Before introducing SCE, the relationship between CE and Kullback–Leibler (KL) divergence is analyzed first, and then, the definition of SCE is introduced. Based on SCE, the supervision loss l S is constructed.
For each sample x i , the class predictive distribution predicted by a classifier is denoted as p ( c | x i ) , and q c | x i is used to represent the ground-truth distribution of the sample x i on the observation label. The CE loss is defined as
l c e = c = 1 C q c | x i log p c | x i = H q , p ,
when c = y i , q c | x i = 1 ; otherwise, q c | x i = 0 . The relationship between cross-entropy H q , p and KL divergence can be written as
K L ( q | | p ) = H q , p H q ,
generally, H q is a constant for a given ground-truth distribution, so it can be omitted from Formula (3) to obtain Formula (2). From the perspective of KL divergence, the essence of classification is to learn a prediction distribution p c | x i that is close to the ground-truth distribution q c | x i , which minimizes the KL divergence between the two distributions. In the case of label noise, q c | x i as the ground-truth distribution, it does not represent a real class probability distribution. On the contrary,   p c | x i to a certain extent reflects the true distribution. Therefore, in addition to q c | x i as a ground truth, we also need to consider the KL divergence in the other direction, namely K L ( p | | q ) . Thus, the symmetric KL divergence is written as
S K L = K L ( q | | p ) + K L ( p | | q ) .
According to the relationship between KL divergence and CE, the SCE and its corresponding loss can be written as follows:
S C E = C E + R C E = H q , p + H p , q .
l s c e x i = c = 1 C q c | x i log p c | x i c = 1 C p c | x i log q c | x i .
For the two networks A and B , the supervision loss l S constructed with SCE loss is defined as
l S x i = l s c e A x i + l s c e B x i .
Generally speaking, the two networks can filter out the errors caused by noisy labels due to their different learning abilities, enabling the model to iterate forward stably. As can be seen from Formula (7), the supervision loss l S comes from the loss combinations under two networks, and each individual network uses SCE loss with good noise resistance, which enables l S to optimize the model in the right direction.
In addition to the supervision loss l S on the two networks allowing the model to be more stable, the relative loss l R between the two networks is also useful in identifying noisy labels. According to the principle of consistency maximization, different models will make an agreement on the correct labels for most samples, while it is unlikely to agree on the wrong labels. Suppose the predictive distributions of sample x i on two networks A and B are p A and p B , respectively; we use R-Drop [71] to regularize the model predictions by minimizing the bidirectional KL divergence between these two predictive distributions for the sample x i . The R-Drop-based relative loss is defined as
l R x i = 1 2 D KL p A p B + D KL p B p A ,
where D KL p A p B = c = 1 C p A c x i log p A c x i p B c x i , D KL p B p A = c = 1 C p B c x i log p B c x i p A c x i . It can be seen from Formula (8) that the relative loss between the two networks is paired, and the relative loss is only related to the predictive distributions. The relative loss reflects the degree of consistency discrimination of two networks for the same sample.
Through the above analysis, the supervision loss and relative loss are obtained by Formulas (7) and (8), respectively. To this end, the joint loss induced by Formula (1) can be written as
l x i = β l s c e A x i + l s c e B x i + 1 β 2 D KL p A p B + D KL p B p A .

3.2. Agreement and Disagreement-Based Co-Learning Framework

Different learners utilize their own unique structures to learn decision boundaries and thus enjoy distinct learning abilities. Therefore, they are desired to exhibit distinct abilities to filter label noise when learning data with noisy labels. In this work, we propose an HSI classification method based on co-learning with a dual network, so that two networks can exchange and select samples with small losses, that is, update network A (corresponding to B) with mini-batch data selected from B (corresponding to A). If the selected sample is not completely “clean”, the two networks adapt to correct the peer-to-peer training errors. This is similar to cross-validation; since errors from one network are not propagated directly back to itself, it can be expected that the method based on co-learning with a dual network can handle higher noise.
As the number of iterations increases, the two networks reach an agreement, and the co-learning function decays into two self-training networks. To make the learning more robust, we incorporate “disagreement” strategy into the co-learning and put forward a more robust learning paradigm, namely, disagreement-based co-learning (DCL). The training process of DCL includes two update steps: data update and parameter update. First, in the data update stage, the two deep networks predict all samples in mini-batch and retain the data with inconsistent prediction results of the two deep networks, which maintains the divergence of two deep networks trained by the DCL. Then, in the parameter update stage, each deep network chooses data with small loss from the disagreement data to minimize the loss of the deep network and utilize the data with small loss from the peer network to update its own weight parameters. However, the “disagreement” strategy cannot guarantee real supervision information. Therefore, we leverage the “agreement” strategy in co-training to improve the DCL and propose an agreement and disagreement-based co-learning (ADCL) framework for HSI classification. During the parameter update of ADCL, each network selects its own small loss data from the disagreement data and adds the data with the same classification results of the two networks into the peer network for back propagation.
Figure 2 shows detailed procedure of forward propagation and back propagation. For the t-th mini-batch, the two networks A and B predict the mini-batch according to the parameters w A and w B , respectively. The disagreement data D t are determined by Formula (10), i.e., the samples with inconsistent predictions from the two networks.
D t = x i , y i : y i A y i B ,
In the end of forward propagation, D A t and D B t are determined by Formula (11) so that the losses of network A and B are minimized.
D A t = arg m i n D A : D A λ e D t l D A , w A D B t = arg m i n D B : D B λ e D t l D B , w B ,
where λ e is used to control how much small loss data should be chosen in every e p o c h e . Because of the memory effect, the deep network will first match data without noise and then gradually fit data with label noise. Formula (12) relates the noise rate r and the parameter λ e , which controls the amount of small loss data to be chosen in each epoch.
λ e = 1 1 + e E k E m a x E k r ,
where E k and E m a x represent a constant and the biggest e p o c h value, respectively. As can be seen from Formula (12), λ e is large at the beginning of the training phase, which would maintain more data with small loss. As the e p o c h increase, less data with small loss are retained in each mini-batch. The gradual decrease in λ e alleviates the overfitting on noisy data of deep networks to a great extent.
w A = w A η l D B t + C t , w A w B = w B η l D A t + C t , w B ,
In back propagation, we use Formula (13) to update the network weight parameters, which can ensure that there are real distributions playing a role in the training under the condition of high noise rate. It can be seen that when updating the weight parameters of networks A and B , not only the disagreement data are used, but also the consistent data are added to calculate the loss. The consistent data C t in Formula (13) can be obtained by Formula (14).
C t = x i , y i : y i A = y i , y i B = y i .

3.3. Formula Analysis

In this subsection, we use formulas to analyze the main procedure of the proposed method. We use the 2D convolutional neural network as the backbone to describe the method.
The symbols A and B represent two convolutional neural networks with initial convolutional kernels wA and wB, l represents the l-th layer, and δ means the gradient. Suppose that the data are divided into m mini-batch. For the i-th mini-batch data D, the training process can be described as follows:
Forward Propagation:
1. Assign the input data x to the input neurons a A 1 and a B 1 , a A 1 = x , a B 1 = x .
2. For the second layer to the L 1 layer, perform forward propagation calculations according to the following three cases:
2.1. If the current layer is a convolutional layer, then we have a A l = σ z A l = σ w A * a A l 1 + b A l , a B l = σ z B l = σ w B * a B l 1 + b B l .
2.2. If the current layer is a pooling layer, then we have a A l = p o o l i n g a A l 1 , a B l = p o o l i n g a B l 1 .
2.3. If the current layer is a fully connected layer, then we have a A l = σ z A l = σ w A a A l 1 + b A l , a B l = σ z B l = σ w B a B l 1 + b B l .
3. Output layer: a A L = s o f t m a x z A L = s o f t m a x w A a A L 1 + b A L , a B L = s o f t m a x z B L = s o f t m a x w B a B L 1 + b B L .
4. Obtain the small loss data D A and D B through Formula (11). Obtain the consistent data C by Formula (14).
Back Propagation:
1. Compute the gradient of output layer δ A L D B + C and δ B L D A + C .
2. For the L 1 layer to the second layer, perform backward propagation according to the following three cases:
2.1. If the current layer is a fully connected layer, then we have δ A l = w A l + 1 T δ A l + 1 D B + C σ z A l D B + C , δ B l = w B l + 1 T δ B l + 1 D A + C σ z B l D A + C .
2.2. If the previous layer is a pooling layer, then we have δ A l = u p s a m p l i n g δ A l + 1 D B + C , δ B l = u p s a m p l i n g δ B l + 1 D A + C .
2.3. If previous layer is a convolutional layer, then we have δ A l = δ A l + 1 * r o t 180 w A l + 1 σ z A l D B + C , δ B l = δ B l + 1 * r o t 180 w B l + 1 σ z B l D A + C .
3. Then, we can update the model parameters:
3.1. If the current layer is a fully connected layer, we have w A l = w A l η m δ A l a D B + C l 1 T , w B l = w B l η m δ B l a D A + C l 1 T .
3.2. If the current layer is the convolutional layer, we have w A l = w A l η m δ B l * r o t 90 a D B + C l 1 , 2 , w A l = w A l η m δ A l * r o t 90 a D A + C l 1 , 2 .
The above description is the main steps of the proposed method. It can be seen that the dual-network structure uses the information of the peer network to guide each other in the training process, and the network structure is easy to implement.

4. Experimental Results and Analysis

4.1. HSI Data Sets

To demonstrate the effectiveness of our proposed method, we conducted experiments on three publicly available HSI data sets. The detailed descriptions of the three data sets are provided below:
(1)
Salinas Valley (SV) [72]: The SV data set was collected by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor over the agricultural area described as Salinas Valley in California, USA, in 1998. The data set contains 512 × 217 pixels characterized by 224 spectral bands. A total of 204 bands were used for experiments after removing 20 redundant ones. The spatial resolution of SV is 3.7 m per pixel, and the land cover contains 16 classes. The three-band pseudocolor image of the SV and its corresponding reference map are illustrated in Figure 3.
(2)
Houston (HOU) [73]: The HOU data set was obtained by the ITRES CASI-1500 sensor and provided by the 2013 IEEE GRSS Data Fusion Competition. The data set contains 349 × 1905 pixels characterized by 144 spectral bands ranging from 364 to 1046 nm. The spatial resolution of HOU is 2.5 m per pixel, and the land cover includes 15 classes. The three-band pseudocolor image of the HOU and its corresponding reference map are shown in Figure 4.
(3)
Kennedy Space Center (KSC) [72]: The KSC data set was acquired by the AVIRIS sensor over the KSC, Florida, on 23 March 1996. The data set contains 512 × 614 pixels characterized by 224 spectral bands. A total of 176 bands were retained for our experiment after removing water absorption bands and low signal-to-noise ratio bands. The spatial resolution of KSC is 3.7 m per pixel, and the land cover includes 13 classes. The three-band pseudocolor image of the KSC and its corresponding reference map are illustrated in Figure 5.
For each data set, we randomly selected 10% of the samples as the training set, and the remaining 90% of the samples were treated as the testing set. Detailed descriptions of the three HIS data sets are given in Table 1, Table 2 and Table 3.

4.2. Experiment Settings

In our experiments, we used a 2D CNN as the backbone to implement our ADCL. For simplicity, we denoted the 2D CNN-based ADCL as 2D-ADCL. In 2D-ADCL, the Adam optimizer was adopted to dominate the training process. We set the learning rate to 0.001 with 150 epochs. We implemented 2D-ADCL in PyTorch 1.8.1, and a single NVIDIA RTX 3070 GPU with CUDA 11.1 was used to boost the training process.
To generate training sets with different label noise levels, we randomly selected a portion of samples from the training set and uniformly assigned any other class labels to them. We set different noise ratios r to obtain training sets with different levels of label noise. We used several metrics, including overall accuracy (OA), average accuracy (AA), and Kappa coefficient (k), to evaluate the classification performance of the proposed method. Detailed calculations of these metrics are given below.
The overall accuracy can be calculated by Formula (15).
OA = i = 1 C M i M ,
where C represents the number of class, and M i represents the number of correctly classified samples in the i-th class.
The calculation of average accuracy is shown in Formula (16).
AA = i = 1 C UA i C ,
where UA i = M i i / j = 1 C M i j is the ratio of the number of correctly classified samples in the i-th class to the total number of samples in the i-th class, and M i j represents the number of samples of the i-th class that are classified as the j-th class.
The kappa coefficient can be obtained by Formula (17).
Kappa = N i = 1 C M i i i = 1 C j = 1 C M i j j = 1 C M j i N 2 i = 1 C j = 1 C M i j j = 1 C M j i ,

4.3. Evaluation of the Joint Loss Function

To evaluate the performance of the designed joint loss function, we set the noise ratio r to 0.3 for the experiments, i.e., 30% of training samples were randomly assigned with wrong labels. The proposed joint loss function has a parameter β that is used to balance the supervision loss and relative loss in the joint loss. We conducted experiments with different values of β varying from 0.05 to 0.95 with a step of 0.05. The OA curves of 2D-ADCL on the three HSI data sets were illustrated in Figure 6. As can be seen from Figure 6, a relatively small β can obtain better performance than a large β , which means that the relative loss should obtain more attention.
The designed loss is related to CE loss, SCE loss, and R-Drop loss. We compared the proposed joint loss with CE loss, SCE loss and R-Drop loss on three HSI data sets, where β was set to 0.15 for the proposed joint loss. The classification accuracies of 2D-ADCL with different loss functions on the three data sets are shown in Table 4. The results in Table 4 demonstrate that the proposed joint loss obtains the best performance, R-Drop loss achieves suboptimal performance, and the other two loss functions perform less well.

4.4. Comparison with State-of-the-Art Methods

We compared 2D-ADCL with state-of-the-art methods to demonstrate its performance. The comparison methods include RLPA [15], DPNLD [12], SALP [17], DCRN [20], and S3Net [21]. The detailed settings for the above methods were consistent with their corresponding references. It should be noted that the SVM classifier was used as the base classifier for RLPA, DPNLD, and SALP, while 2D-CNN was adopted as the backbone network for S3Net. The noise rate r was set to 0.3. We repeated each algorithm ten times to obtain the average results. Table 5, Table 6 and Table 7 show the OAs, Aas, kappa coefficients, and class-specific accuracies of the comparison methods on the SV, HOU, and KSC data sets, respectively. The best classification accuracies of different methods are highlighted in bold. The classification result maps of different methods on the SV, HOU, and KSC data sets are illustrated in Figure 7, Figure 8 and Figure 9, respectively.
Several results can be observed from Table 5, Table 6 and Table 7 and Figure 7, Figure 8 and Figure 9. Firstly, 2D-ADCL achieves higher class-specific accuracy than other methods in most cases. Specifically, 2D-ADCL attains 13, 11, and 10 best class-specific accuracies on SV, BOT, and KSC data sets, respectively. Secondly, 2D-ADCL achieves the best OAs, AAs, and kappa coefficients on all data sets. The average OA of 2D-ADCL is more than two percentage points higher than the second-place and more than 10 percentage points higher than the last place. Thirdly, the accuracies of deep learning-based methods (DCRN, S3Net, and 2D-ADCL) are significantly higher than those of traditional methods (RLPA, DPNLD, and SALP). Fourthly, the classification maps of different methods on the three data sets demonstrate that 2D-ADCL achieves satisfactory classification results.

4.5. Performance Evaluation under Different Noise Rates

In order to study the effect of noise rate on classification performance, we set different noise rates for experiments, in which the noise rate ranged from 0.1 to 0.7 with a step of 0.05. First, when the noise rate was equal to 0.1, the training set contained only 10% of noisy samples. Then, the number of noisy samples in the training set increased gradually with the increase in noise rate. Finally, when the noise rate was equal to 0.7, the number of noisy samples in the training set reached 70%. We ran all methods ten times to obtain average results. The OA curves of different methods on SV, HOU, and KSC data sets are plotted in Figure 10.
As seen from Figure 10, for each data set, the classification result of 2D-ADCL is consistently better than that of the other methods in terms of OA. The average OA of all comparison methods decreases with the increase in noise rate. It can be seen that the deep learning-based methods (DCRN, S3Net, and 2D-ADCL) show a lower attenuation speed on OA than the traditional methods (RLPA, DPNLD, and SALP). When the noise rate increases from 0.1 to 0.7, the OA attenuation values of 2D-ADCL on SV, HOU, and KSC data sets are approximately within 5%, 8% and 4%, respectively. This indicates that ADCL is robust to high noise rates.

4.6. Computational Cost

Computational cost is also an important metric to evaluate classification algorithms. We set the noise rate to 0.5 to compare the running times of different methods, including training time and test time. Table 8 displays the running times of different methods on SV, HOU, and KSC data sets.
The results from Table 8 show that the running times for comparison methods range from tens of seconds to hundreds of seconds, and none of them require too much running time. Another finding is that the running times of deep learning-based methods are longer than those of traditional methods, since deep learning methods need more time for training models. For the three deep learning methods (DCRN, S3Net, 2D-ADCL), the running time of 2D-ADCL is slightly longer than that of the other two methods, which is still within the acceptable range.

4.7. Further Analysis

In order to investigate the role of the “agreement” strategy in ADCL, we took ADCL and DCL for comparison, where DCL only adopted the data with a small loss from the peer network to update the weight parameters. The other settings for DCL are the same as ADCL. Table 9 shows the classification results of the two methods when the noise rate was 0.5.
As seen from the results in Table 9, the average OAs, AAs, and kappa coefficients of 2D-ADCL on the three data sets are higher than those of 2D-DCL, indicating that the “agreement” strategy plays an important role in learning from label noise.

5. Discussion

Previous experiments revealed some important findings that require further discussion.
As demonstrated in Section 4.3, compared with related loss functions, the proposed joint loss function has better performance because it makes full use of both networks’ own supervision information and mutual information between the two networks. Additionally, the experimental results illustrate that the relative loss in the joint loss plays a more important role, because the supervision information from the peer network is more effective than its own supervision information in the presence of label noise.
As shown in Section 4.4 and Section 4.5, compared with state-of-the-art methods, 2D-ADCL obtains better classification performance in terms of OA, AA, and kappa coefficient. In addition, 2D-ADCL has better robustness to high noise rate. This can be attributed to several factors, including the unified framework with a dual network that leverages the mutual guidance abilities of the two networks, the “disagreement” and “agreement” strategies that enhance the model’s discrimination ability, and the designed loss function that improves the model’s robustness to label noise.
The experimental results from Section 4.6 indicate that the running time of 2D-ADCL is acceptable. The main reason for this result is that the proposed framework is simple, and it does not have complicated network structures. The results in Section 4.7 suggest that the “agreement” strategy is crucial in label noise learning, particularly in the presence of a high noise ratio, as the agreement data playa a significant role.

6. Conclusions

In this paper, we proposed an ADCL framework for HSI classification with noisy labels. The proposed ADCL adopted a unified framework with a dual-network structure for label noise learning. The experimental results demonstrated the effectiveness of the proposed method. Previous results and analysis can be summarized in the following four conclusions:
  • The proposed framework, based on a dual-network structure, proved to be robust to label noise, and it can achieve good classification performance even in the case of a high noise rate.
  • The designed joint loss function, composed of the supervision loss and relative loss, demonstrated good robustness to label noise. This is because when there is label noise, the self-supervised information of each network may not be completely accurate, but the mutual supervised information from both networks will help to correct and improve the accuracy of the predictions.
  • In terms of time efficiency, the proposed method is acceptable because we do not use a complex network except for a dual-network structure.
  • The “agreement” strategy plays an important role in improving the classification accuracy, as it helps mitigate the problem of difficult convergence of neural networks when there is a high ratio of label noise.
The limitation of this work is that ADCL requires estimating the noise rate to determine the small loss data, which may not be feasible in some scenarios. Future research could explore small loss data selection methods that are independent of the noise rate. Additionally, this work only used a 2D-CNN as the backbone for ADCL, but other advanced neural networks could be adopted to further improve the performance of the proposed framework.

Author Contributions

Conceptualization, Y.Z. and J.S.; formal analysis, Y.Z., Z.G. and H.S.; funding acquisition, Y.Z., J.S. and G.C.; methodology, Y.Z., Q.Y. and X.L.; validation, Z.G. and X.L.; writing—original draft, Y.Z.; writing—review and editing, H.S., Z.G., X.L. and Q.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grants 62201282 and 62203231, in part by the Natural Science Foundation of Jiangsu Province under grants BK20200763 and BK20191284, in part by the Natural Science Research Project of Jiangsu Higher Education Institutions under grants 19KJB510052 and 22KJB510037, in part by the State Key Laboratory of Ocean Engineering (Shanghai Jiao Tong University) under grant GKZD010084, in part by the China Postdoctoral Science Foundation under grant 2020M681685, in part by the Postdoctoral Research Funding Project of Jiangsu Province under grant 2021K161B, and in part by the Research Start Foundation of Nanjing University of Posts and Telecommunications under grant NY220157.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  2. Huang, H.; Liu, L.; Ngadi, M.O. Recent developments in hyperspectral imaging for assessment of food quality and safety. Sensors 2014, 14, 7248–7276. [Google Scholar] [CrossRef] [PubMed]
  3. Cruz-Ramos, C.; Garcia-Salgado, B.P.; Reyes-Reyes, R.; Ponomaryov, V.; Sadovnychiy, S. Gabor features extraction and land-cover classification of urban hyperspectral images for remote sensing applications. Remote Sens. 2021, 13, 2914. [Google Scholar] [CrossRef]
  4. Ye, C.; Li, Y.; Cui, P.; Liang, L.; Pirasteh, S.; Marcato, J.; Goncalves, W.N.; Li, J. Landslide detection of hyperspectral remote sensing data based on deep learning with constrains. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 5047–5060. [Google Scholar] [CrossRef]
  5. Wang, F.; Gao, J.; Zha, Y. Hyperspectral sensing of heavy metals in soil and vegetation: Feasibility and challenges. ISPRS J. Photogramm. Remote Sens. 2018, 136, 73–84. [Google Scholar] [CrossRef]
  6. Okwuashi, O.; Ndehedehe, C.E. Deep support vector machine for hyperspectral image classification. Pattern Recognit. 2020, 103, 107298. [Google Scholar] [CrossRef]
  7. Zhang, Y.; Cao, G.; Li, X.; Wang, B.; Fu, P. Active semi-supervised random forest for hyperspectral image classification. Remote Sens. 2019, 11, 2974. [Google Scholar] [CrossRef]
  8. Yu, X.; Feng, Y.; Gao, Y.; Jia, Y.; Mei, S. Dual-weighted kernel extreme learning machine for hyperspectral imagery classification. Remote Sens. 2021, 13, 508. [Google Scholar] [CrossRef]
  9. Peng, J.; Sun, W.; Li, H.; Li, W.; Meng, X.; Ge, C.; Du, Q. Low-rank and sparse representation for hyperspectral image processing: A review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 10–43. [Google Scholar] [CrossRef]
  10. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  11. Vali, A.; Comai, S.; Matteucci, M. Deep learning for land use and land cover classification based on hyperspectral and multispectral earth observation data: A review. Remote Sens. 2020, 12, 2495. [Google Scholar] [CrossRef]
  12. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Li, S. Density peak-based noisy label detection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1573–1584. [Google Scholar] [CrossRef]
  13. Tu, B.; Zhang, X.; Kang, X.; Wang, J.; Benediktsson, J.A. Spatial density peak clustering for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5085–5097. [Google Scholar] [CrossRef]
  14. Tu, B.; Zhou, C.; He, D.; Huang, S.; Plaza, A. Hyperspectral classification with noisy label detection via superpixel-to-pixel weighting distance. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4116–4131. [Google Scholar] [CrossRef]
  15. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral image classification in the presence of noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 851–865. [Google Scholar] [CrossRef]
  16. Jiang, J.; Ma, J.; Liu, X. Multilayer spectral-spatial graphs for label noisy robust hyperspectral image classification. IEEE Trans. Neural Networks Learn. Syst. 2022, 33, 839–852. [Google Scholar] [CrossRef] [PubMed]
  17. Leng, Q.; Yang, H.; Jiang, J. Label noise cleansing with sparse graph for hyperspectral image classification. Remote Sens. 2019, 11, 1116. [Google Scholar] [CrossRef]
  18. Maas, A.E.; Rottensteiner, F.; Heipke, C. A label noise tolerant random forest for the classification of remote sensing data based on outdated maps for training. Comput. Vis. Image Underst. 2019, 188, 102782. [Google Scholar] [CrossRef]
  19. Damodaran, B.B.; Flamary, R.; Seguy, V.; Courty, N. An entropic optimal transport loss for learning deep neural networks under label noise in remote sensing images. Comput. Vis. Image Underst. 2020, 191, 102863. [Google Scholar] [CrossRef]
  20. Xu, Y.; Li, Z.; Li, W.; Du, Q.; Liu, C.; Fang, Z.; Zhai, L. Dual-channel residual network for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502511. [Google Scholar] [CrossRef]
  21. Xu, H.; Zhang, H.; Zhang, L. A superpixel guided sample selection neural network for handling noisy labels in hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9486–9503. [Google Scholar] [CrossRef]
  22. Roy, S.K.; Hong, D.; Kar, P.; Wu, X.; Liu, X.; Zhao, D. Lightweight heterogeneous kernel convolution for hyperspectral image classification with noisy labels. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5509705. [Google Scholar] [CrossRef]
  23. Wei, W.; Xu, S.; Zhang, L.; Zhang, J.; Zhang, Y. Boosting hyperspectral image classification with unsupervised feature learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502315. [Google Scholar] [CrossRef]
  24. Wang, C.; Zhang, L.; Wei, W.; Zhang, Y. Toward effective hyperspectral image classification using dual-level deep spatial manifold representation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5505614. [Google Scholar] [CrossRef]
  25. Ghafari, S.; Ghobadi Tarnik, M.; Sadoghi Yazdi, H. Robustness of convolutional neural network models in hyperspectral noisy datasets with loss functions. Comput. Electr. Eng. 2021, 90, 107009. [Google Scholar] [CrossRef]
  26. Huang, L.; Chen, Y.; He, X. Weakly supervised classification of hyperspectral image based on complementary learning. Remote Sens. 2021, 13, 5009. [Google Scholar] [CrossRef]
  27. Song, H.; Kim, M.; Park, D.; Shin, Y.; Lee, J.G. Learning from noisy labels with deep neural networks: A survey. IEEE Trans. Neural Networks Learn. Syst. 2022; in press. [Google Scholar] [CrossRef]
  28. Algan, G.; Ulusoy, I. Image classification with deep learning in the presence of noisy labels: A survey. Knowledge-Based Syst. 2021, 215, 106771. [Google Scholar] [CrossRef]
  29. Karimi, D.; Dou, H.; Warfield, S.K.; Gholipour, A. Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis. Med. Image Anal. 2020, 65, 101759. [Google Scholar] [CrossRef]
  30. Goldberger, J.; Ben-Reuven, E. Training deep neural-networks using a noise adaptation layer. In Proceedings of the 5th International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017; pp. 1–9. [Google Scholar]
  31. Yao, J.; Wang, J.; Tsang, I.W.; Zhang, Y.; Sun, J.; Zhang, C.; Zhang, R. Deep learning from noisy image labels with quality embedding. IEEE Trans. Image Process. 2019, 28, 1909–1922. [Google Scholar] [CrossRef]
  32. Ghosh, A.; Kumar, H.; Sastry, P.S. Robust loss functions under label noise for deep neural networks. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 1919–1925. [Google Scholar]
  33. Englesson, E.; Azizpour, H. Generalized jensen-shannon divergence loss for learning with noisy labels. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, Online, 6–14 December 2021; pp. 30284–30297. [Google Scholar]
  34. Gupta, A.; Lam, S.M. Weight decay backpropagation for noisy data. Neural Networks 1998, 11, 1127–1138. [Google Scholar] [CrossRef] [PubMed]
  35. Arplt, D.; Jastrzȩbskl, S.; Bailas, N.; Krueger, D.; Bengio, E.; Kanwal, M.S.; Maharaj, T.; Fischer, A.; Courville, A.; Bengio, Y.; et al. A closer look at memorization in deep networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), Sydney, NSW, Australia, 6–11 August 2017; pp. 233–242. [Google Scholar]
  36. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. Mixup: Beyond empirical risk minimization. In Proceedings of the 6th International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–13. [Google Scholar]
  37. Nishi, K.; Ding, Y.; Rich, A.; Höllerer, T. Augmentation strategies for learning with noisy labels. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, Online, 19–25 June 2021; pp. 8022–8031. [Google Scholar]
  38. Patrini, G.; Rozza, A.; Menon, A.K.; Nock, R.; Qu, L. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1944–1952. [Google Scholar]
  39. Liu, T.; Tao, D. Classification with noisy labels by importance reweighting. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 447–461. [Google Scholar] [CrossRef] [PubMed]
  40. Song, H.; Kim, M.; Lee, J.G. SELFIE: Refurbishing unclean samples for robust deep learning. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019; pp. 5907–5915. [Google Scholar]
  41. Ye, M.; Li, H.; Du, B.; Shen, J.; Shao, L.; Hoi, S.C.H. Collaborative refining for person re-identification with label noise. IEEE Trans. Image Process. 2022, 31, 379–391. [Google Scholar] [CrossRef] [PubMed]
  42. Shen, Y.; Sanghavi, S. Learning with bad training data via iterative trimmed loss minimization. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019; pp. 5739–5748. [Google Scholar]
  43. Yi, R.; Huang, Y.; Guan, Q.; Pu, M.; Zhang, R. Learning from pixel-level label noise: A new perspective for semi-supervised semantic segmentation. IEEE Trans. Image Process. 2022, 31, 623–635. [Google Scholar] [CrossRef] [PubMed]
  44. Meng, D.; Gao, F.; Dong, J.; Du, Q.; Li, H.C. Synthetic aperture radar image change detection via layer attention-based noise-tolerant network. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4026505. [Google Scholar] [CrossRef]
  45. Li, P.; He, X.; Qiao, M.; Cheng, X.; Li, J.; Guo, X.; Zhou, T.; Song, D.; Chen, M.; Miao, D.; et al. Exploring label probability sequence to robustly learn deep convolutional neural networks for road extraction with noisy datasets. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5614018. [Google Scholar] [CrossRef]
  46. Zhang, Z.; Guo, W.; Li, M.; Yu, W. GIS-supervised building extraction with label noise-adaptive fully convolutional neural network. IEEE Geosci. Remote Sens. Lett. 2020, 17, 2135–2139. [Google Scholar] [CrossRef]
  47. Xi, M.; Li, J.; He, Z.; Yu, M.; Qin, F. NRN-RSSEG: A deep neural network model for combating label noise in semantic segmentation of remote sensing images. Remote Sens. 2023, 15, 108. [Google Scholar] [CrossRef]
  48. Kang, J.; Fernandez-Beltran, R.; Kang, X.; Ni, J.; Plaza, A. Noise-tolerant deep neighborhood embedding for remotely sensed images with label noise. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 4, 2551–2562. [Google Scholar] [CrossRef]
  49. Kang, J.; Fernandez-Beltran, R.; Duan, P.; Kang, X.; Plaza, A.J. Robust normalized softmax loss for deep metric learning-based characterization of remote sensing images with label noise. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8798–8811. [Google Scholar] [CrossRef]
  50. Dong, R.; Fang, W.; Fu, H.; Gan, L.; Wang, J.; Gong, P. High-resolution land cover mapping through learning with noise correction. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4402013. [Google Scholar] [CrossRef]
  51. Wang, C.; Shi, J.; Zhou, Y.; Li, L.; Yang, X.; Zhang, T.; Wei, S.; Zhang, X.; Tao, C. Label noise modeling and correction via loss curve fitting for SAR ATR. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5216210. [Google Scholar] [CrossRef]
  52. Hu, Z.; Gao, K.; Zhang, X.; Wang, J.; Wang, H.; Han, J. Probability differential-based class label noise purification for object detection in aerial images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6509705. [Google Scholar] [CrossRef]
  53. Cao, Y.; Huang, X. A coarse-to-fine weakly supervised learning method for green plastic cover segmentation using high-resolution remote sensing images. ISPRS J. Photogramm. Remote Sens. 2022, 188, 157–176. [Google Scholar] [CrossRef]
  54. Li, Y.; Zhang, Y.; Zhu, Z. Error-tolerant deep Learning for remote sensing image scene classification. IEEE Trans. Cybern. 2021, 51, 1756–1768. [Google Scholar] [CrossRef]
  55. Wei, W.; Li, W.; Zhang, L.; Wang, C.; Zhang, P.; Zhang, Y. Robust hyperspectral image domain adaptation with noisy labels. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1135–1139. [Google Scholar] [CrossRef]
  56. Li, Q.; Chen, Y.; Ghamisi, P. Complementary learning-based scene classification of remote sensing images with noisy labels. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8021105. [Google Scholar] [CrossRef]
  57. Xu, G.; Deng, M.; Sun, G.; Guo, Y.; Chen, J. Improving building extraction by using knowledge distillation to reduce the impact of label noise. Remote Sens. 2022, 14, 5645. [Google Scholar] [CrossRef]
  58. Xu, G.; Fang, Y.; Deng, M.; Sun, G.; Chen, J. Remote sensing mapping of build-up land with noisy label via fault-tolerant learning. Remote Sens. 2022, 14, 2263. [Google Scholar] [CrossRef]
  59. Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the Annual Conference on Computational Learning Theory (COLT), Madison, WI, USA, 24–26 July 1998; pp. 92–100. [Google Scholar]
  60. Malach, E.; Shalev-Shwartz, S. Decoupling “when to update” from “how to update”. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 961–971. [Google Scholar]
  61. Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.W.; Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montréal, QC, Canada, 3–8 December 2018; pp. 8536–8546. [Google Scholar]
  62. Yu, X.; Han, B.; Yao, J.; Niu, G.; Tsang, I.W.; Sugiyama, M. How does disagreement help generalization against label corruption? In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 10–15 June 2019; pp. 7164–7173. [Google Scholar]
  63. Wei, H.; Feng, L.; Chen, X.; An, B. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–18 June 2020; pp. 13726–13735. [Google Scholar]
  64. Zhang, X.; Song, Q.; Liu, R.; Wang, W.; Jiao, L. Modified co-training with spectral and spatial views for semisupervised hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2044–2055. [Google Scholar] [CrossRef]
  65. Romaszewski, M.; Głomb, P.; Cholewa, M. Semi-supervised hyperspectral classification from a small number of training samples using a co-training approach. ISPRS J. Photogramm. Remote Sens. 2016, 121, 60–76. [Google Scholar] [CrossRef]
  66. Zhou, S.; Xue, Z.; Du, P. Semisupervised stacked autoencoder with cotraining for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3813–3826. [Google Scholar] [CrossRef]
  67. Fang, B.; Chen, G.; Chen, J.; Ouyang, G.; Kou, R.; Wang, L. CCT: Conditional co-training for truly unsupervised remote sensing image segmentation in coastal areas. Remote Sens. 2021, 13, 3521. [Google Scholar] [CrossRef]
  68. Hu, T.; Huang, X.; Li, J.; Zhang, L. A novel co-training approach for urban land cover mapping with unclear landsat time series imagery. Remote Sens. Environ. 2018, 217, 144–157. [Google Scholar] [CrossRef]
  69. Jia, D.; Gao, P.; Cheng, C.; Ye, S. Multiple-feature-driven co-training method for crop mapping based on remote sensing time series imagery. Int. J. Remote Sens. 2020, 41, 8096–8120. [Google Scholar] [CrossRef]
  70. Wang, Y.; Ma, X.; Chen, Z.; Luo, Y.; Yi, J.; Bailey, J. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 322–330. [Google Scholar]
  71. Liang, X.; Wu, L.; Li, J.; Wang, Y.; Meng, Q.; Qin, T.; Chen, W.; Zhang, M.; Liu, T.Y. R-Drop: Regularized dropout for neural networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, Online, 6–14 December 2021; pp. 10890–10905. [Google Scholar]
  72. Grupo de Inteligencia Computacional (GIC). Available online: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes (accessed on 28 February 2020).
  73. 2013 IEEE GRSS Data Fusion Contestest. Available online: https://hyperspectral.ee.uh.edu/?page_id=459 (accessed on 31 May 2013).
Figure 1. The framework of ADCL. The data set with noisy labels is fed into two convolutional neural networks A and B firstly. Then, each network uses the small loss data from the peer network and consistent data from two networks to update its own parameters. At last, the trained networks A and B are fused to classify the data.
Figure 1. The framework of ADCL. The data set with noisy labels is fed into two convolutional neural networks A and B firstly. Then, each network uses the small loss data from the peer network and consistent data from two networks to update its own parameters. At last, the trained networks A and B are fused to classify the data.
Remotesensing 15 02543 g001
Figure 2. Detailed procedure of forward propagation and back propagation.
Figure 2. Detailed procedure of forward propagation and back propagation.
Remotesensing 15 02543 g002
Figure 3. Pseudocolor image and reference map of the SV. (a) Three-band pseudocolor image. The image is generated by using bands 180, 27 and 17 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Figure 3. Pseudocolor image and reference map of the SV. (a) Three-band pseudocolor image. The image is generated by using bands 180, 27 and 17 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Remotesensing 15 02543 g003
Figure 4. Pseudocolor image and reference map of the HOU. (a) Three-band pseudocolor image. The image is generated by using bands 70, 50 and 20 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Figure 4. Pseudocolor image and reference map of the HOU. (a) Three-band pseudocolor image. The image is generated by using bands 70, 50 and 20 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Remotesensing 15 02543 g004
Figure 5. Pseudocolor image and reference map of the KSC. (a) Three-band pseudocolor image. The image is generated by using bands 28, 19 and 10 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Figure 5. Pseudocolor image and reference map of the KSC. (a) Three-band pseudocolor image. The image is generated by using bands 28, 19 and 10 as the R, G, and B channels, respectively. (b) Reference map. The number represents the class number, where 0 represents the background.
Remotesensing 15 02543 g005
Figure 6. Classification accuracies (%) of 2D-ADCL on three data sets under different values of β, where β ranges from 0.05 to 0.95 with step size of 0.05.
Figure 6. Classification accuracies (%) of 2D-ADCL on three data sets under different values of β, where β ranges from 0.05 to 0.95 with step size of 0.05.
Remotesensing 15 02543 g006
Figure 7. Classification maps for the SV image with 30% noisy labels in the training set. (a) RLPA: OA = 87.75%. (b) DPNLD: OA = 90.03. (c) SALP: OA = 91.15%. (d) DCRN: OA = 94.33%. (e) S3Net: OA = 96.04%. (f) 2D-ADCL: 98.63%.
Figure 7. Classification maps for the SV image with 30% noisy labels in the training set. (a) RLPA: OA = 87.75%. (b) DPNLD: OA = 90.03. (c) SALP: OA = 91.15%. (d) DCRN: OA = 94.33%. (e) S3Net: OA = 96.04%. (f) 2D-ADCL: 98.63%.
Remotesensing 15 02543 g007
Figure 8. Classification maps for the HOU image with 30% noisy labels in the training set. (a) RLPA: OA = 76.65%. (b) DPNLD: OA = 78.68%. (c) SALP: OA = 80.86%. (d) DCRN: OA = 86.66%. (e) S3Net: OA = 88.63%. (f) 2D-ADCL: OA = 90.88%.
Figure 8. Classification maps for the HOU image with 30% noisy labels in the training set. (a) RLPA: OA = 76.65%. (b) DPNLD: OA = 78.68%. (c) SALP: OA = 80.86%. (d) DCRN: OA = 86.66%. (e) S3Net: OA = 88.63%. (f) 2D-ADCL: OA = 90.88%.
Remotesensing 15 02543 g008
Figure 9. Classification maps for the KSC image with 30% noisy labels in the training set. (a) RLPA: OA = 89.22%. (b) DPNLD: OA = 91.86%. (c) SALP: OA = 95.55%. (d) DCRN: OA = 97.36%. (e) S3Net: OA = 98.16%. (f) 2D-ADCL: OA = 99.02%.
Figure 9. Classification maps for the KSC image with 30% noisy labels in the training set. (a) RLPA: OA = 89.22%. (b) DPNLD: OA = 91.86%. (c) SALP: OA = 95.55%. (d) DCRN: OA = 97.36%. (e) S3Net: OA = 98.16%. (f) 2D-ADCL: OA = 99.02%.
Remotesensing 15 02543 g009
Figure 10. The influence of noise rate (r) on the classification accuracy. The horizontal axis represents the noise rate ranging from 0.1 to 0.7, and the vertical axis represents the OA changes of RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL. (a) SV. (b) HOU. (c) KSC.
Figure 10. The influence of noise rate (r) on the classification accuracy. The horizontal axis represents the noise rate ranging from 0.1 to 0.7, and the vertical axis represents the OA changes of RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL. (a) SV. (b) HOU. (c) KSC.
Remotesensing 15 02543 g010
Table 1. The class information and data partition of SV data set.
Table 1. The class information and data partition of SV data set.
Class No.Class NameTrainTestTotal
1Brocoli_green_weeds_120118082009
2Brocoli_green_weeds_237333533726
3Fallow19817781976
4Fallow_rough_plow13912551394
5Fallow_smooth26824102678
6Stubble39635633959
7Celery35832213579
8Grapes_untrained11271014411271
9Soil_vinyard_develop62055836203
10Corn_senesced_green_weeds32829503278
11Lettuce_romaine_4wk1079611068
12Lettuce_romaine_5wk19317341927
13Lettuce_romaine_6wk92824916
14Lettuce_romaine_7wk1079631070
15Vinyard_untrained72765417268
16Vinyard_vertical_trellis18116261807
Table 2. The class information and data partition of HOU data set.
Table 2. The class information and data partition of HOU data set.
Class No.Class NameTrainTestTotal
1Healthy grass12511261251
2Stressed grass12511291254
3Stressed grass70627697
4Trees12411201244
5Soil12411181242
6Water33292325
7Residential12711411268
8Commercial12411201244
9Road12511271252
10Highway12311041227
11Railway12411111235
12Parking Lot 112311101233
13Parking Lot 247422469
14Tennis Court43385428
15Running Track66594660
Table 3. The class information and data partition of KSC data set.
Table 3. The class information and data partition of KSC data set.
Class No.Class NameTrainTestTotal
1Scrub76685761
2Willow swamp24219243
3Cabbage palm hammock26230256
4Cabbage palm/oak hammock25227252
5Slash pine16145161
6Oak/broadleafhammock23206229
7Hardwood swamp1194105
8Graminoid marsh43388431
9Spartina marsh52468520
10Cattail marsh40364404
11Salt marsh42377419
12Mud flats50453503
13Water93834927
Table 4. Classification accuracy of different loss functions on three data sets. The comparison loss functions include CE, SCE, and R-Drop. The Oas, Aas, and kappas of different methods are reported.
Table 4. Classification accuracy of different loss functions on three data sets. The comparison loss functions include CE, SCE, and R-Drop. The Oas, Aas, and kappas of different methods are reported.
Data SetsSVHOUKSC
LossCESCER-DropProposedCESCER-DropProposedCESCER-DropProposed
OA93.2195.2695.3898.6284.7586.3587.1190.8995.5897.2897.4599.02
AA92.3595.1495.7898.7384.5485.9286.8790.9495.4396.8797.1498.50
k × 10091.9595.0895.4698.4784.3986.1186.9490.1195.1797.0597.2398.91
Table 5. Classification accuracy (in %) including class-specific accuracy, OA, AA, and kappa on SV data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
Table 5. Classification accuracy (in %) including class-specific accuracy, OA, AA, and kappa on SV data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
ClassRLPA [15]DPNLD [12]SALP [17]DCRN [20]S3Net [21]2D-ADCL
197.7598.4899.8399.7099.5099.99
299.1499.7599.8099.7899.3599.86
396.5697.0799.5592.7398.3199.78
489.8899.6999.5399.6499.6498.86
594.2896.8496.6188.2396.7397.18
697.4798.7497.8398.9494.5497.19
798.1199.2099.4499.7899.8099.80
872.8879.3978.5088.4594.0698.20
997.3599.0298.7995.3294.2899.43
1078.9187.2292.6096.3393.2897.19
1190.9491.8496.8196.0395.3598.34
1297.4399.6299.0294.6291.5399.89
1397.3297.4398.3397.4698.8898.52
1493.1794.9594.1793.3096.1197.36
1574.9969.9277.7191.7696.6998.03
1694.1598.3098.5999.9899.39100
OA87.7790.0191.1694.3196.0498.62
AA91.9094.2294.5895.7696.7298.73
k × 10086.4188.8790.1793.6895.5998.47
Table 6. Classification accuracy (in %) including class-specific accuracy, OA, AA, and kappa on HOU data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
Table 6. Classification accuracy (in %) including class-specific accuracy, OA, AA, and kappa on HOU data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
ClassRLPA [15]DPNLD [12]SALP [17]DCRN [20]S3Net [21]2D-ADCL
190.2193.4790.6695.0995.1698.05
296.9998.5391.7898.5497.6898.83
392.6788.4297.7798.0297.2497.69
491.7292.4691.8096.2796.8597.16
593.7695.8095.4096.5398.0998.12
683.6594.4492.4795.2991.5998.59
782.9276.6380.7584.4785.5092.22
868.3663.9268.6878.6882.2086.47
973.1979.9577.8177.1984.8783.25
1071.7678.3374.3587.7185.2989.84
1168.7958.3070.1978.1884.7481.70
1227.7945.4558.0871.9081.5979.11
1325.8823.1638.3044.7336.5766.27
1492.0596.1089.3797.6298.1398.35
1582.7695.1697.6096.8595.5598.32
OA76.6378.6980.8686.6788.6290.89
AA76.1778.6881.0086.4787.4090.94
k × 10074.7476.9579.3085.5687.6990.11
Table 7. Classification accuracy (%) including class-specific accuracy, OA, AA, and kappa on KSC data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
Table 7. Classification accuracy (%) including class-specific accuracy, OA, AA, and kappa on KSC data set. Classification accuracy obtained by RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL with 30% noisy labels in the training set.
ClassRLPA [15]DPNLD [12]SALP [17]DCRN [20]S3Net [21]2D-ADCL
197.5096.9897.9098.8299.4799.21
284.7791.7793.4295.4797.9497.94
390.3390.2394.1496.8897.66100
470.2476.9887.3088.1094.4497.62
559.6365.2277.0286.9091.3093.79
651.9764.6380.3595.2488.6594.76
792.3874.2995.2499.07100100
884.6990.2295.1398.8598.6198.84
991.9293.4697.5099.0199.0499.81
1086.1495.4396.7898.9797.7799.01
1197.1898.5799.2899.0599.6999.76
1290.6693.8496.8298.4199.8199.80
1399.7599.7699.8999.89100100
OA89.2091.8695.5397.3598.1699.02
AA84.4087.0193.1495.7397.1998.50
k × 10087.9590.9395.0297.0597.9598.91
Table 8. Running times (s) on three data sets. Running time for RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL, where the running time consists of training time and testing time.
Table 8. Running times (s) on three data sets. Running time for RLPA, DPNLD, SALP, DCRN, S3Net, and 2D-ADCL, where the running time consists of training time and testing time.
Time (s)SVHOUKSC
RLPA [15]85.855.319.5
DPNLD [12]65.541.326.5
SALP [17]113.646.528.5
DCRN [20]148.891.445.3
S3Net [21]161.7116.848.7
2D-ADCL178.1130.455.4
Table 9. Classification results in terms of OA, AA, and kappa on three data sets. Classification accuracy obtained by 2D-DCL and 2D-ADCL with 50% noisy labels in the training set.
Table 9. Classification results in terms of OA, AA, and kappa on three data sets. Classification accuracy obtained by 2D-DCL and 2D-ADCL with 50% noisy labels in the training set.
Data SetsSVHOUKSC
Method2D-DCL2D-ADCL2D-DCL2D-ADCL2D-DCL2D-ADCL
OA95.6497.2086.1588.5196.2597.82
AA95.8297.3486.4488.4395.7497.14
k × 10095.2397.1186.2988.3195.8297.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Sun, J.; Shi, H.; Ge, Z.; Yu, Q.; Cao, G.; Li, X. Agreement and Disagreement-Based Co-Learning with Dual Network for Hyperspectral Image Classification with Noisy Labels. Remote Sens. 2023, 15, 2543. https://doi.org/10.3390/rs15102543

AMA Style

Zhang Y, Sun J, Shi H, Ge Z, Yu Q, Cao G, Li X. Agreement and Disagreement-Based Co-Learning with Dual Network for Hyperspectral Image Classification with Noisy Labels. Remote Sensing. 2023; 15(10):2543. https://doi.org/10.3390/rs15102543

Chicago/Turabian Style

Zhang, Youqiang, Jin Sun, Hao Shi, Zixian Ge, Qiqiong Yu, Guo Cao, and Xuesong Li. 2023. "Agreement and Disagreement-Based Co-Learning with Dual Network for Hyperspectral Image Classification with Noisy Labels" Remote Sensing 15, no. 10: 2543. https://doi.org/10.3390/rs15102543

APA Style

Zhang, Y., Sun, J., Shi, H., Ge, Z., Yu, Q., Cao, G., & Li, X. (2023). Agreement and Disagreement-Based Co-Learning with Dual Network for Hyperspectral Image Classification with Noisy Labels. Remote Sensing, 15(10), 2543. https://doi.org/10.3390/rs15102543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop