Next Article in Journal
The Influence of Hydroxyapatite Crystals on the Viscoelastic Behavior of Poly(vinyl alcohol) Braid Systems
Previous Article in Journal
The Pine Cone Optimization Algorithm (PCOA)
Previous Article in Special Issue
Bioinspired Artificial Intelligence Applications 2023
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency

1
School of Artificial Intelligence, Hebei University of Technology, Tianjin 300401, China
2
Hebei Province Key Laboratory of Big Data Computing, Hebei University of Technology, Tianjin 300401, China
3
Hebei Engineering Research Center of Data-Driven Industrial Intelligent, Hebei University of Technology, Tianjin 300401, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(2), 92; https://doi.org/10.3390/biomimetics9020092
Submission received: 18 December 2023 / Revised: 30 January 2024 / Accepted: 31 January 2024 / Published: 3 February 2024
(This article belongs to the Special Issue Bioinspired Artificial Intelligence Applications)

Abstract

:
Inspired by the biological nervous system, deep neural networks (DNNs) are able to achieve remarkable performance in various tasks. However, they struggle to handle label noise, which can poison the memorization effects of DNNs. Co-teaching-based methods are popular in learning with noisy labels. These methods cross-train two DNNs based on the small-loss criterion and employ a strategy using either “disagreement” or “consistency” to obtain the divergence of the two networks. However, these methods are sample-inefficient for generalization in noisy scenarios. In this paper, we propose CoDC, a novel Co-teaching-basedmethod for accurate learning with label noise via both Disagreement and Consistency strategies. Specifically, CoDC maintains disagreement at the feature level and consistency at the prediction level using a balanced loss function. Additionally, a weighted cross-entropy loss is proposed based on information derived from the historical training process. Moreover, the valuable knowledge involved in “large-loss” samples is further developed and utilized by assigning pseudo-labels. Comprehensive experiments were conducted on both synthetic and real-world noise and under various noise types. CoDC achieved 72.81% accuracy on the Clothing1M dataset and 76.96% (Top1) accuracy on the WebVision1.0 dataset. These superior results demonstrate the effectiveness and robustness of learning with noisy labels.

1. Introduction

Inspired by the biological nervous system, deep neural networks (DNNs) have been successfully utilized in various tasks [1,2,3,4,5,6], particularly in computer vision [7,8,9,10,11,12]. This is made possible by large-scale datasets with accurate labels, although collecting them can be challenging and costly, especially in certain professional fields that require personnel with relevant professional knowledge to label samples. As a result, researchers often seek cost-effective solutions such as crowdsourcing [13], web-crawling [14], and search engine queries [15] to build datasets. However, these methods can introduce noise to the labels, which can seriously affect the generalization of DNNs during the training process [16].
Learning with noisy labels has become a popular research topic; finding ways to reduce the impact of noisy labels on networks is key to solving the problem. Existing methods include designing robust loss functions and sample selection techniques, which are often combined to create a more effective model. Co-teaching [17], a recent approach, trains two networks that each treat small-loss samples as clean and then uses them to update each other’s parameters. By starting with different random parameters, the networks are able to filter different types of errors, resulting in a more robust model. Other approaches, such as Co-teaching+ [18] and CoDis [19], have introduced disagreement strategies for sample selection and training, based on the view that the differences between the two networks are beneficial to their overall robustness. JoCoR, on the other hand, uses joint training and co-regularization to select samples with more consistent predictions and trains both networks simultaneously, arguing that disagreement strategies may not always select truly clean samples. The MLC [20] provides a detailed analysis of the significance of maintaining disagreement.
However, these methods constrain DNNs to one of a “disagreement” or “consistency” perspective, meaning that valuable knowledge involved in “high-loss” samples is ignored. Thus, we propose a new method called CoDC (Accurate Learning with Noisy Labels via Disagreement and Consistency). Specifically, we train two networks with disagreement loss at the feature level to prevent the two networks from becoming a self-training network. At the prediction level, we apply consistency loss to ensure that the predicted results of the network remain consistent. The joint application of disagreement loss and consistency loss can constrain the networks and the sample selection, enabling them to maintain a balance between disagreement and consistency and enhancing the generalization ability of the resulting network.
Furthermore, cross-entropy loss functions have been shown to exhibit an overfitting phenomenon [13] and DNNs have strong learning ability, making it inevitable that they will fit part of the noise from labels during training. By analyzing the historical training situation, we re-weight the cross-entropy loss to obtain the modified classification loss, which reduces the overfitting phenomenon of the network. Instead of discarding noisy samples directly, we use a peer network to guide learning, further improving the model’s robustness. The main contributions of this work can be summarized as follows:
  • A weighted cross-entropy loss is proposed to remit the overfitting phenomenon, and the loss weights are learned directly from information derived from the historical training process.
  • We propose a novel method called CoDC to alleviate the negative effect of label noise. CoDC maintains disagreement at the feature level and consistency at the prediction level using a balance loss function, which can effectively improve the generalization of networks.

2. Related Work

Robust loss. Robust loss functions have been widely explored in supervised learning to address the issue of overfitting caused by noisy labels. Several loss functions have been proposed to improve robustness against noisy labels, including Mean Absolute Error (MAE) [13], Reverse Cross-Entropy (RCE) [21], Generalized Cross-Entropy (GCE) [22], Active Passive Loss (APL) [23], Negative Learning+ (NL+), and Positive Learning+ (PL+) [24]. Ghosh et al. [13] demonstrated the effectiveness of symmetric loss in handling noisy labels and introduced the MAE loss. Wang et al. [21] proposed RCE, which performs well in multi-classification tasks but may result in underfitting. Zhang et al. [22] designed GCE by combining MAE and CE, further enhancing robustness against noisy labels and mitigating underfitting issues. Ma et al. [23] introduced the APL loss to improve the convergence of the loss function. Kim et al. [24] proposed NL+ and PL+ to address the underfitting problem of robust loss functions by analyzing the loss gradient.
Label correction. Label correction is a crucial technique for improving the performance of deep learning models trained on noisy datasets. Label Smoothing [25], a commonly used method, is known to act as a regularization technique that helps the model parameters converge to a small norm solution [26]. Generating pseudo-labels for noisy samples is another approach that can make full use of all samples’ information. The Meta Pseudo-Labels [27] method treats suspected noise samples in training data as unlabeled data for semi-supervised learning. PENCIL [28] uses given labels to learn from the network and updates and corrects noisy labels through the prediction results of the network, thereby improving the prediction ability of the model. However, the effectiveness of label correction methods is highly dependent on the model’s accuracy. Erroneous labels may be generated constantly, resulting in the accumulation of errors if the model’s accuracy is poor. Co-LDL [29] is a method that trains on high-loss samples using label distribution learning and enhances the learned representations using a self-supervised module, which further boosts model performance and the use of training data.
Small-loss samples. The existing methods mainly apply small-loss criteria for selection. They consider the small-loss samples as clean samples and the large-loss samples as noisy samples, then adopt different strategies for learning. MentorNet [30] pretrains a teacher network to select clean samples, which then guides the student network. However, this method has the disadvantage of cumulative error caused by sample selection bias. Co-teaching [17] trains two networks; each network selects small loss samples to send to the peer network for learning. However, due to continuous peer-to-peer learning, the parameters of the two networks will gradually become consistent. Therefore, Co-teaching+ [18] only chooses the samples with divergent predictions between the two networks for learning. Similarly, CoDis [19] applies the regular variance term and selects samples with large variance to send to the peer network for learning in order to avoid consistency between the two networks. However, JoCoR [31] assumes that few of the samples with divergent predictions are likely to be clean. It is recommended that the two networks be trained jointly using samples with predictive consistency. As a result, the two networks will easily converge to a consensus or even produce the same prediction error. Liu et al. [20] analyzed the existence of appropriate “differences” between the two networks in Co-teaching, which is beneficial to the improvement of model robustness. In contrast, our proposed approach takes into account both the differences and consistency of the two networks.
In addition, DivideMix [32] fits the loss of all samples into a Gaussian mixture model (GMM), based on the idea that the loss distribution of all samples will conform to two Gaussian distributions. The clean samples belong to the component with a lower central value, while the noisy samples belong to the component with a higher central value, allowing the noisy and clean samples to be divided. LongReMix [33] selects clean samples with high confidence through a two-stage selection method. Kong et al. [34] proposed a penalty label and introduced new sample selection criteria to screen clean samples. However, these methods essentially apply the small-loss sample selection principle.

3. Materials and Methods

3.1. Preliminary

In this paper, we consider a classification task with C classes. The dataset contains noisy labels. D = x i , y ˜ i i = 1 n , where n is the sample size, x i is the i-th sample, and y ˜ i [ C ] is the label with noise. For y ˜ i , its true and unobservable label is represented by y i . As in Co-teaching, two networks are trained, denoted as f 1 = f x , θ 1 and f 2 = f x , θ 2 . The two networks have the same architecture and are initialized by random initialization, meaning that they have different network parameters. In this method, p 1 = p 1 1 , p 1 2 , , p 1 C and p 2 = p 2 1 , p 2 2 , , p 2 C represent the respective prediction probabilities of the two networks for the samples. The aim is to learn a robust classifier that can be trained on D to obtain correct predictions.

3.2. Training with Disagreement and Consistency

In this section, we mainly introduce the training functions used in our method, including the modified classification loss, balance loss, and peer network guiding learning.
Modified classification loss. During the initial training phase, Cross-Entropy (CE) is utilized as the classification loss. Considering that the two networks possess distinct learning capabilities, the cross-entropies of both networks are summed to minimize the distance between them in terms of predicting clean samples and their labels. This approach helps to minimize erroneous information. The classification loss can be represented as follows.
L c x i , y ˜ i = CE 1 x i , y ˜ i + CE 2 x i , y ˜ i = i = 1 N m = 1 C y ˜ i log p 1 m x i i = 1 N m = 1 C y ˜ i log p 2 m x i
In previous research, it has been discovered that CE can overfit noise label learning [13]. The small-loss sample selection strategy can result in incorrect selection of clean samples. As the number of epochs increases, CE may cause the network to fit noise labels, which can negatively impact the model’s generalization. To address this issue, we propose a modified loss function that optimizes the weight of CE for each sample and reduces the fitting noise by combining the historical law of sample loss. Specifically, when the model separates clean and noisy samples, the classification loss of each sample should be maintained at its original value and should not fluctuate greatly during the learning process. Due to the overfitting problem, the loss of noisy samples with large losses may gradually decrease, resulting in the classifier’s failure to correctly distinguish between clean and noisy samples.
Therefore, we have modified the classification loss to stabilize the sample loss after the network separates clean and noisy samples. We define the set L s = L c t 1 , L c t 2 , , L c t s to record the value of the classification loss of sample x in each epoch of training, where t is the current number of training epochs and s is the length of the set. We define L history as follows.
L history = 1 s i = t s t 1 L c i
The difference between the loss of the current sample and the loss of the training history, denoted by L c t L history , can indicate the model’s training stability. To ensure stable training and prevent the model from overfitting on noisy samples, we utilized the modified classification loss L mcl to train the model. Minimizing L mcl enhances the stability of the training process.
L mcl = L c t L history
Balance loss. Co-teaching [17] has been successful in the field of learning with noisy labels. The approach involves training two networks simultaneously, with each network feeding what it deems to be clean samples to its peer network for learning. Because the initial random parameters of the two networks are different, they can filter different types of errors, which improves the generalization ability of the networks. However, as the two networks learn from each other their parameters gradually converge, resulting in decreased ability to filter different errors [18,20].
To control the disagreement between the dual networks at the feature level and encourage them to learn more feature knowledge, we introduce the disagreement loss L dis . The cosine similarity is used to calculate the output features of the two networks in order to measure such feature differences. The equation is as follows:
L dis = C o s i n e S i m i l a r i t y ( z 1 x i , z 2 x i ) = i = 1 N m = 1 K z 1 m x i · z 2 m x i z 1 m x i × z 2 m x i
where z 1 and z 2 are the feature representations of the sample in network one and network two, respectively, and K is the dimension of the feature. We apply L dis in the final loss function to control the difference between the two networks, encourage the networks to learn more at the feature level, and inhibit the convergence of the two networks.
In combination with the theory summarized above, we encourage the two networks to maintain differentiation at the feature level, preventing the two networks from reaching a consensus to ensure that more knowledge is learned. However, from the view of agreement maximization principles [35], different networks will agree on the prediction of clean samples. Thus, although the differentiation in features is encouraged, the two networks should be consistent in the prediction of clean samples. We designed the consistency loss L con to minimize the two networks’ predictions p 1 and p 2 . Specifically, we apply the Jensen–Shannon (JS) Divergence, which measures the difference between two probability distributions, to capture the consistency of two networks’ predictions:
L con = D J S p 1 p 2 = 1 2 D K L p 1 p 1 + p 2 2 + 1 2 D K L p 2 p 1 + p 2 2
where
D K L p 1 p 1 + p 2 2 = i = 1 N m = 1 C p 1 m x i log 2 p 1 m x i p 1 m x i + p 2 m x i
D K L p 2 p 1 + p 2 2 = i = 1 N m = 1 C p 2 m x i log 2 p 2 m x i p 1 m x i + p 2 m x i .
We weight the L dis and L con to obtain L balance .
L balance = β L dis + γ L con
The significance of Equation (6) is that we aim for the two networks to maintain differences at the feature level while making the same predictions.
Peer network guided learning. Many previous works have simply discarded noisy samples, resulting in insufficient training due to the decrease in training samples. Semi-supervised learning strategies can be used to generate pseudo-labels to replace the given labels for robust learning. However, the network’s generalization ability affects the quality of the pseudo-labels. Moreover, the network tends to fall into self-cognitive errors as the number of network iterations increases, causing the self-generated pseudo-labels to accumulate errors. Therefore, we used a peer network to guide learning. The peer network prediction p is used for pseudo-label learning, and the cross-entropy loss can be formulated as follows:
p l 1 = CE p 1 x i , ε ( p 2 x i , τ ) ,
p l 2 = CE p 2 x i , ε ( p 1 x i , τ ) ,
where
ε ( p , T ) = exp ( p / T ) m = 1 C exp p m x i / T .
The sharpening function ε ( p , T ) is utilized to enhance the guidance capabilities of the peer network, where T is the temperature parameter. To align with Jo-SRC [36], we set τ = 0.1. However, there may be label errors due to the network’s limited generalization ability. To prevent the network from fitting to these errors, a small weight λ is applied; in our experiments, λ was typically set to 0.5. Additionally, we employed weak data augmentation techniques such as random cropping and random symmetry flipping to further enhance the guidance capabilities of the peer network. Peer-to-peer guided learning is described as follows.
L p l = λ p l 1 + p l 2
Training loss. Finally, the training loss is constructed as follows.
L = α L mcl + L balance + L p l
Unlike Co-teaching, Equation (9) is used to train the two networks simultaneously.
Small-loss samples are more likely to be clean samples; therefore, we are more inclined to select these as clean samples for learning. As mentioned above, we aim to add more samples that exhibit different features in the two networks while maintaining consistent prediction results for learning. Such samples are also considered cleaner. Therefore, we define the loss of sample selection in the training process as follows.
L select = α L mcl + L balance
As in Equation (6), the L balance is used to ensure a balance between feature differences and prediction consistency, with α controlling the weight of the modified classification loss.
Equation (10) is used to select clean samples for learning, while the remaining samples are considered to be noise samples. The noise samples are learned using the peer network guided learning method to enhance the network’s robustness.
Considering the ability of deep networks to learn clean samples first, we warmed up the network at the beginning of training using all samples and employed the cross-entropy loss as the training loss. After the warm-up period, we selected the top 1 τ small-loss samples as clean samples and considered the rest to be noise samples, where τ represents the noise rate in the dataset. Specifically, our approach is as follows:
D ˜ n = arg min D n : D n ( 1 τ ) D n L select D n .
Our sample selection method is depicted in Figure 1. The pseudocode is presented in Algorithm 1.
Algorithm 1 CoDC algorithm
Input: two networks with weights θ 1 and θ 2 , learning rate η , noise rate τ , epoch T w a r m u p and T m a x , iteration t m a x ;
  1:
for  T = 1 T m a x do
  2:
   Shuffle training dataset D ;
  3:
   if  T < T w a r m u p  then
  4:
     training by Cross-Entropy;
  5:
   else
  6:
     for  t = 1 t m a x  do
  7:
        Fetch mini-batch D n from D;
  8:
        Obtain  D ˜ n by L select ;
  9:
        Update  ( θ 1 , θ 2 ) = ( θ 1 , θ 2 ) η L ( D ˜ n ) ;
10:
     end for
11:
   end if
12:
end for
Output: θ 1 and θ 2

4. Results

4.1. Datasets and Implementation Details

Datasets and noise types. We used four benchmark datasets with noise (F-MNIST [37] (http://fashion-mnist.s3-website.eu-central-1.amazonaws.com/, accessed on 12 November 2023), SVHN [38] (http://ufldl.stanford.edu/housenumbers/, accessed on 12 November 2023), CIFAR-10 [39] (https://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz, accessed on 12 November 2023), and CIFAR100 [39] (https://www.cs.toronto.edu/~kriz/cifar-100-python.tar.gz, accessed on 12 November 2023)) and two real-world datasets (Clothing1M [40] (https://opendatalab.com/OpenDataLab/Clothing1M, accessed on 12 November 2023) and WebVision1.0 [41] (https://data.vision.ee.ethz.ch/cvl/webvision/dataset2017.html, accessed on 12 November 2023)) to evaluate our method. F-MNIST (Fashion-MNIST) contains ten classes of grayscale images, each sized 28 × 28, with 60,000 training images and 10,000 test images. SVHN is derived from Google Street View House Number, including ten classes of color images, each sized 32 × 32, with 73,257 training images and 26,032 test images. Both the CIFAR-10 and CIFAR-100 datasets contain 60,000 training images and 10,000 test images, each sized 32 × 32. CIFAR-10 comprises ten classes, while CIFAR-100 comprises 100 classes.
Clothing1M is made up of one million training images collected from online shopping sites and contains fourteen classes with labels generated using surrounding text. WebVision1.0 is a large dataset with real-world noise labels. It has taken over 2.4 million images from the internet, including 1000 classes contained in ImageNet ILSVRC2012 [42]. In order to facilitate comparison with other methods, we followed the previous work and obtained the first 50 classes of images from Google images for both training and testing.
For SVHN datasets, the number of samples in each class is unbalanced. Specifically, the number of samples in each class is 13,861, 10,585, 8497, 7458, 6882, 5727, 5595, 5045, 4659, and 4948 respectively. For F-MNIST, CIFAR-10, CIFAR-100, Clothing1M, and WebVision1.0 datasets, the distribution of samples in each class is uniform.
Clothing1M and WebVision1.0 are real-world datasets that contain noisy labels. All of the benchmark datasets are clean; thus, we set up the label transition matrix [25] and manually added noise to test the validity of the method. Our experiment mainly followed [43], and we conducted experiments with a variety of noise types: symmetric noise, pairflip noise, tridiagonal noise, and asymmetric noise. Noise rates of 20% and 40% were tested for each type of noise. It is necessary to ensure that the number of clean samples is greater than the number of noisy samples; otherwise, it will be impossible to distinguish the real class of samples [23]. The construction of each noise type was as follows:
(1)
Symmetric noise: the clean labels in each class are uniformly flipped to labels of other wrong classes.
(2)
Asymmetric noise: considers the visual similarity in the flipping process, which is closer to real-world noise; for example, the labels of cats and dogs will be reversed, and the labels of planes and birds will be reversed. Asymmetric noise is an unbalanced type of noise.
(3)
Pairflip noise: this is realized by flipping clean labels in each class to adjacent classes.
(4)
Tridiagonal noise: realized by two consecutive pairflips of two classes in opposite directions.
The label transition matrix is shown in Figure 2, taking six categories and a 40% noise rate as an example.
In addition, we conducted experiments on noisy long-tailed datasets [19]. We first built a long-tailed distribution dataset. Specifically, we reduced the number of samples in different classes such that the dataset formed a long-tailed distribution with class imbalance [44]. In this paper, we use two ways of simulating the long-tailed distribution: exponential simulation (exp) and linear simulation (line), as shown in Figure 3. Taking SVHN as an example, the two different simulation appraoches are named SVHN-EXP and SVHN-line, respectively. We further added asymmetric noise to the long-tailed dataset and tested four different noise intensities.
Models and parameters. Following previous works [31], we used a nine-layer CNN for F-MNIST, SVHN, CIFAR-10, and CIFAR-100. We used a SGD optimizer with momentum set at 0.9, an initial learning rate of 0.02, and a weight decay of 0.0005. The learning rate was reduced by a factor of 10 at the 100th and 150th epochs. The batch size was set to 128, and the total number of training epochs was 200.
In the real-world datasets, for the Clothing1M dataset the training used ResNet-18 [45] pretrained on ImageNet. For WebVision, the training used inception-resnet v2 [46]. We trained the network for 100 epochs using a SGD optimizer with an initial learning rate of 0.002. The learning rate was reduced by a factor of 10 at the 30th and 60th epochs. The batch size was set to 64; the other settings were the same as those for the benchmark datasets.
We implemented all the methods with default parameters using PyTorch 1.13.1 and conducted all the experiments on a NVIDIA RTX3090 GPU.

4.2. Comparison with SOTA Methods

We compared our method with the Standard CE baseline and more recent SOTA methods, including:
(1)
Standard: trains a single network and uses only standard cross-entropy loss.
(2)
Co-teaching [17]: trains two networks simultaneously; the two networks guide each other during learning.
(3)
Co-teaching-plus [18]: trains two networks simultaneously while considering the small loss samples of the two networks’ diverging samples.
(4)
JoCoR [31]: trains two networks simultaneously and applies the co-regularization method to maximize the consistency between the two networks.
(5)
Co-learning [47]: a simple and effective method for learning with noisy labels; it combines supervised and self-supervised learning to regularize the network and improve generalization performance.
(6)
Co-LDL [29]: an end-to-end framework proposed to train high-loss samples using label distribution learning and enhance the learned representations by a self-supervised module, further boosting model performance and the use of training data.
(7)
CoDis [19]: trains two networks simultaneously and applies the covariance regularization method to maintain the divergence between the two networks.
(8)
Bare [48]: proposes an adaptive sample selection strategy to provide robustness against label noise.
Results on benchmark datasets. Table 1 shows the experimental results under four benchmark datasets with four noise types and two noise rates. The test accuracy consists of the mean and standard deviation (%) calculated from the last ten epochs of training. Based on the experimental results, we can conclude that our method is superior to state-of-the-art methods on the benchmark datasets. For example, on the CIFAR-10 dataset with symmetric noise and a 40% noise rate, our method scores 1.26% higher than Co-LDL and 10.75% higher than JoCoR. For pairflip noise with a 40% noise rate, our method scores 4.43% higher than Co-LDL and 21.17% higher than JoCoR. Because JoCoR only considers the consistency of the two networks, it is extremely easy for errors to accumulate with asymmetric and pairflip noise, which ultimately affects the generalization of the networks. This experiment proves the superiority of our method on the benchmark datasets.
Results on noisy long-tailed datasets. We tested the model on noisy long-tailed datasets with four noise rates. It can be seen that our method achieves the best performance in most cases. Although the accuracy is not higher than all advanced methods, our method remains competitive. Table 2 shows the results on the noisy long-tailed datasets. Figure 4 shows the test accuracy comparison for the number of epochs.
For the results shown in Figure 4, each experiment was repeated three times; the curve represents the average accuracy of the three experiments, while the shaded part represents the error bar for the standard deviation. From the curve, it can be seen that our method has excellent performance under both high and low noise rates. The small area of the shaded part indicates that our method has good statistical significance and is less affected by random factors.
Results on real-world datasets. Table 3 and Table 4 show the results on the real-world datasets Clothing1M and WebVision. In addition, we evaluated the generalization ability of these methods on ImageNet ILSVRC12. Our method achieves the best performance on real-world datasets among the compared methods.

4.3. Ablation Study

We conducted ablation experiments on the CIFAR-100 dataset under symmetric noise at a 40% noise rate to emphasize our design featuring the modified classification loss, balance loss, and peer network guiding learning. The accuracy when the networks lack a sample selection support strategy is only 33.20%, which is due to the influence of the noisy label environment. When the sample selection strategy is added, the accuracy increase to 64.74%. The modified classification loss, which can alleviate overfitting caused by CE, improves the accuracy by 1.17%. When the balance loss is added, the accuracy improves by 2.47%, which confirms our assumption that maintaining a balance between disagreement and consistency in the two networks can improve their learning performance. Finally, the addition of peer network guided learning further improves the accuracy of the networks. The experimental results are shown in Table 5.

4.4. Comparison of Running Time

We compared the running time on the CIFAR-10 dataset under symmetric noise at a 40% noise rate. All the methods were trained for 200 epochs in the same experimental environment. The results are shown in Table 6. Our method outperforms all the compared methods, with less time consumption than Co-LDL and Bare and similar time consumption to Co-teaching and Co-teaching-plus.

5. Conclusions

The existing co-teaching methods train two networks simultaneously; however, as iteration progresses, the parameters of the two networks gradually converge and they degenerate into a self-training network. To address this problem, we propose using the disagreement loss and consistency loss to balance the relationship between the two networks. Our ablation experiments confirm that balancing the relationship between the two networks can improve the generalization ability of the resulting model. In addition, to alleviate the over-fitting phenomenon when using cross-entropy, we propose a modified classification loss based on analyzing the change in the historical classification loss. Finally, we use a peer network to guide the learning of noisy samples to ensure that all samples are utilized for training. A large number of experiments demonstrate the effectiveness of the proposed method, providing a new research direction for co-teaching. In our future work, we intend to further explore the relationship between the two networks based on the idea of co-teaching in order to enhance their performance when learning with noisy labels; for example, more reasonable methods could be employed to balance the disagreement and consistency between the two networks, and more accurate pseudo-label correction strategies could be utilized.

Author Contributions

Resources, Y.D.; investigation, W.J. and J.L.; methodology, Z.W.; formal analysis, Z.W.; writing—original draft preparation, J.L.; visualization, J.L.; writing—review and editing, Z.W.; validation, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (NSFC) under Grant 62306103 and 62376194, and in part by the Higher Education Science and Technology Research Project of Hebei Province under Grant QN2023262.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Basheri, M. Intelligent Breast Mass Classification Approach Using Archimedes Optimization Algorithm with Deep Learning on Digital Mammograms. Biomimetics 2023, 8, 463. [Google Scholar] [CrossRef] [PubMed]
  2. Albraikan, A.A.; Maray, M.; Alotaibi, F.A.; Alnfiai, M.M.; Kumar, A.; Sayed, A. Bio-Inspired Artificial Intelligence with Natural Language Processing Based on Deceptive Content Detection in Social Networking. Biomimetics 2023, 8, 449. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, T.; Yin, M.; Zhu, X.; Zhu, X.; Sun, J.; Meng, C.; Bei, S. A Fast and Robust Lane Detection via Online Re-Parameterization and Hybrid Attention. Sensors 2023, 23, 8285. [Google Scholar] [CrossRef] [PubMed]
  4. Tewes, T.J.; Welle, M.C.; Hetjens, B.T.; Tipatet, K.S.; Pavlov, S.; Platte, F.; Bockmühl, D.P. Understanding Raman Spectral Based Classifications with Convolutional Neural Networks Using Practical Examples of Fungal Spores and Carotenoid-Pigmented Microorganisms. AI 2023, 4, 114–127. [Google Scholar] [CrossRef]
  5. Lee, H.U.; Chun, C.J.; Kang, J.M. Causality-Driven Efficient Feature Selection for Deep-Learning-Based Surface Roughness Prediction in Milling Machines. Mathematics 2023, 11, 4682. [Google Scholar] [CrossRef]
  6. Pinto, J.; Ramos, J.R.C.; Costa, R.S.; Oliveira, R. A General Hybrid Modeling Framework for Systems Biology Applications: Combining Mechanistic Knowledge with Deep Neural Networks under the SBML Standard. AI 2023, 4, 303–318. [Google Scholar] [CrossRef]
  7. Liao, H.; Zhu, W. YOLO-DRS: A Bioinspired Object Detection Algorithm for Remote Sensing Images Incorporating a Multi-Scale Efficient Lightweight Attention Mechanism. Biomimetics 2023, 8, 458. [Google Scholar] [CrossRef] [PubMed]
  8. Song, F.; Li, P. YOLOv5-MS: Real-time multi-surveillance pedestrian target detection model for smart cities. Biomimetics 2023, 8, 480. [Google Scholar] [CrossRef] [PubMed]
  9. Zhu, P.; Yao, X.; Wang, Y.; Cao, M.; Hui, B.; Zhao, S.; Hu, Q. Latent heterogeneous graph network for incomplete multi-view learning. IEEE Trans. Multimed. 2022, 25, 3033–3045. [Google Scholar] [CrossRef]
  10. Liu, B.; Feng, L.; Zhao, Q.; Li, G.; Chen, Y. Improving the Accuracy of Lane Detection by Enhancing the Long-Range Dependence. Electronics 2023, 12, 2518. [Google Scholar] [CrossRef]
  11. Chen, T.; Xie, G.S.; Yao, Y.; Wang, Q.; Shen, F.; Tang, Z.; Zhang, J. Semantically meaningful class prototype learning for one-shot image segmentation. IEEE Trans. Multimed. 2021, 24, 968–980. [Google Scholar] [CrossRef]
  12. Nanni, L.; Loreggia, A.; Brahnam, S. Comparison of Different Methods for Building Ensembles of Convolutional Neural Networks. Electronics 2023, 12, 4428. [Google Scholar] [CrossRef]
  13. Ghosh, A.; Kumar, H.; Sastry, P.S. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  14. Mahajan, D.; Girshick, R.; Ramanathan, V.; He, K.; Paluri, M.; Li, Y.; Bharambe, A.; Van Der Maaten, L. Exploring the limits of weakly supervised pretraining. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 181–196. [Google Scholar]
  15. Yan, Y.; Rosales, R.; Fung, G.; Subramanian, R.; Dy, J. Learning from multiple annotators with varying expertise. Mach. Learn. 2014, 95, 291–327. [Google Scholar] [CrossRef]
  16. Zhang, C.; Bengio, S.; Hardt, M.; Recht, B.; Vinyals, O. Understanding deep learning (still) requires rethinking generalization. Commun. ACM 2021, 64, 107–115. [Google Scholar] [CrossRef]
  17. Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.; Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Proceedings of the 32nd Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Advances in Neural Information Processing Systems. Volume 31. [Google Scholar]
  18. Yu, X.; Han, B.; Yao, J.; Niu, G.; Tsang, I.; Sugiyama, M. How does disagreement help generalization against label corruption? In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7164–7173. [Google Scholar]
  19. Xia, X.; Han, B.; Zhan, Y.; Yu, J.; Gong, M.; Gong, C.; Liu, T. Combating Noisy Labels with Sample Selection by Mining High-Discrepancy Examples. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 1833–1843. [Google Scholar]
  20. Liu, J.; Jiang, D.; Yang, Y.; Li, R. Agreement or Disagreement in Noise-tolerant Mutual Learning? In Proceedings of the 26th International Conference on Pattern Recognition, Montreal, QC, Canada, 21–25 August 2022; pp. 4801–4807. [Google Scholar]
  21. Wang, Y.; Ma, X.; Chen, Z.; Luo, Y.; Yi, J.; Bailey, J. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–3 November 2019; pp. 322–330.
  22. Zhang, Z.; Sabuncu, M. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  23. Ma, X.; Huang, H.; Wang, Y.; Romano, S.; Erfani, S.; Bailey, J. Normalized loss functions for deep learning with noisy labels. In Proceedings of the International Conference on Machine Learning, Online, 13–18 July 2020; pp. 6543–6553. [Google Scholar]
  24. Kim, Y.; Yun, J.; Shon, H.; Kim, J. Joint negative and positive learning for noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 9442–9451. [Google Scholar]
  25. Patrini, G.; Rozza, A.; Krishna Menon, A.; Nock, R.; Qu, L. Joint negative and positive learning for noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1944–1952. [Google Scholar]
  26. Lukasik, M.; Bhojanapalli, S.; Menon, A.; Kumar, S. Does label smoothing mitigate label noise? In Proceedings of the International Conference on Machine Learning, Online, 13–18 July 2020; pp. 6448–6458. [Google Scholar]
  27. Pham, H.; Dai, Z.; Xie, Q.; Le, Q.V. Meta pseudo labels? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 11557–11568. [Google Scholar]
  28. Yi, K.; Wu, J. Probabilistic end-to-end noise correction for learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 7017–7025. [Google Scholar]
  29. Sun, Z.; Liu, H.; Wang, Q.; Zhou, T.; Wu, Q.; Tang, Z. Co-ldl: A co-training-based label distribution learning method for tackling label noise. IEEE Trans. Multimed. 2021, 24, 1093–1104. [Google Scholar] [CrossRef]
  30. Liu, J.; Zhou, Z.; Leung, T.; Li, L.J.; Li, F.-F. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 2304–2313. [Google Scholar]
  31. Wei, H.; Feng, L.; Chen, X.; An, B. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 13726–13735. [Google Scholar]
  32. Li, J.; Socher, R.; Hoi, S.C. Dividemix: Learning with noisy labels as semi-supervised learning. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  33. Cordeiro, F.R.; Sachdeva, R.; Belagiannis, V.; Reid, I.; Carneiro, G. Longremix: Robust learning with high confidence samples in a noisy label environment. Pattern Recognit. 2023, 133, 109013. [Google Scholar] [CrossRef]
  34. Kong, K.; Lee, J.; Kwak, Y.; Cho, Y.; Kim, S.; Song, W. Penalty based robust learning with noisy labels. Neurocomputing 2022, 489, 112–127. [Google Scholar] [CrossRef]
  35. Sindhwani, V.; Niyogi, P.; Belkin, M. A co-regularization approach to semi-supervised learning with multiple views. In Proceedings of the International Conference on Machine Learning Workshop on Learning with Multiple Views, Bonn, Germany, 7–11 August 2005; pp. 74–79. [Google Scholar]
  36. Yao, Y.; Sun, Z.; Zhang, C.; Shen, F.; Wu, Q.; Zhang, J.; Tang, Z. Jo-src: A contrastive approach for combating noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 5192–5201. [Google Scholar]
  37. Xiao, H.; Rasul, K.; Vollgraf, R. Fashion-Mnist: A Novel Image Dataset for Benchmarking Machine Learning Algorithms. Available online: https://arxiv.org/pdf/1708.07747.pdf (accessed on 25 August 2017).
  38. Netzer, Y.; Wang, T.; Coates, A.; Bissacco, A.; Wu, B.; Ng, A.Y. Reading digits in natural images with unsupervised feature learning. In Proceedings of the 25th Conference on Neural Information Processing Systems (NeurIPS) Workshop on Deep Learning and Unsupervised Feature Learning, Granada, Spain, 12 December 2011; pp. 462–471. [Google Scholar]
  39. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical report; University of Toronto: Toronto, ON, Canada, 9 April 2009. [Google Scholar]
  40. Xiao, T.; Xia, T.; Yang, Y.; Huang, C.; Wang, X. Learning from massive noisy labeled data for image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2691–2699. [Google Scholar]
  41. Li, W.; Wang, L.; Li, W.; Agustsson, E.; Van, G.L. Webvision Database: Visual Learning and Understanding from Web Data. Available online: https://arxiv.org/pdf/1708.02862.pdf (accessed on 9 August 2017).
  42. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  43. Zhang, Y.; Niu, G.; Sugiyama, M. Imagenet: Learning noise transition matrix from only noisy labels via total variation regularization. In Proceedings of the Conference on Machine Learning, Online, 18–24 July 2021; pp. 12501–12512. [Google Scholar]
  44. Yang, Y.; Xu, Z. Rethinking the value of labels for improving class-imbalanced learning. In Proceedings of the Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Volume 33, pp. 19290–19301. [Google Scholar]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vegas, NV, USA, 26–30 June 2016; pp. 770–778. [Google Scholar]
  46. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  47. Tan, C.; Xia, J.; Wu, L.; Li, S.Z. Co-learning: Learning from noisy labels with self-supervision. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 1405–1413. [Google Scholar]
  48. Patel, D.; Sastry, P.S. Adaptive sample selection for robust learning under label noise. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 3932–3942. [Google Scholar]
Figure 1. Sample selection of CoDC. For each sample, the classification loss, disagreement loss, and consistency loss are calculated and the small-loss samples are taken as the clean sample. In general, the essence of this method is to maintain a balance of disagreement and consistency between the two networks to achieve the best performance.
Figure 1. Sample selection of CoDC. For each sample, the classification loss, disagreement loss, and consistency loss are calculated and the small-loss samples are taken as the clean sample. In general, the essence of this method is to maintain a balance of disagreement and consistency between the two networks to achieve the best performance.
Biomimetics 09 00092 g001
Figure 2. Label transition matrix.
Figure 2. Label transition matrix.
Biomimetics 09 00092 g002
Figure 3. Construction of two long-tailed distributed datasets.
Figure 3. Construction of two long-tailed distributed datasets.
Biomimetics 09 00092 g003
Figure 4. Test accuracy comparison for the number of epochs on the four long-tailed noisy datasets. Each experiment was repeated three times. The error bar for the standard deviation in each figure has been shaded.
Figure 4. Test accuracy comparison for the number of epochs on the four long-tailed noisy datasets. Each experiment was repeated three times. The error bar for the standard deviation in each figure has been shaded.
Biomimetics 09 00092 g004
Table 1. Mean and standard deviations of test accuracy (%) on four benchmark datasets with four noise types and two noise rates (20%, 40%) for each noise type. The test accuracy is calculated using the last ten epochs. The best mean results are in bold.
Table 1. Mean and standard deviations of test accuracy (%) on four benchmark datasets with four noise types and two noise rates (20%, 40%) for each noise type. The test accuracy is calculated using the last ten epochs. The best mean results are in bold.
Noise TypeSymAsymPairTrid
Setting20%40%20%40%20%40%20%40%
Standard87.46 ± 0.0972.46 ± 0.6889.34 ± 0.1577.73 ± 0.1984.23 ± 0.2361.46 ± 0.3285.86 ± 0.466.44 ± 0.28
Co-teaching91.20 ± 0.0388.06 ± 0.0691.29 ± 0.1686.32 ± 0.0989.97 ± 0.0984.99 ± 0.1190.54 ± 0.0586.29 ± 0.09
Co-teaching-plus92.85 ± 0.0391.42 ± 0.0393.42 ± 0.0289.38 ± 0.1792.90 ± 0.0388.38 ± 0.1692.75 ± 0.0890.40 ± 0.04
JoCoR94.11 ± 0.0293.24 ± 0.0194.24 ± 0.0291.18 ± 0.1693.90 ± 0.0191.37 ± 0.2194.03 ± 0.0192.93 ± 0.01
F-MNISTCo-learning90.03 ± 0.2990.03 ± 0.2190.88 ± 0.2288.67 ± 0.4891.08 ± 0.1887.59 ± 0.3590.98 ± 0.1690.27 ± 0.24
Co-LDL94.68 ± 0.1093.97 ± 0.1495.02 ± 0.1393.71 ± 0.5794.84 ± 0.0893.66 ± 0.1194.88 ± 0.1194.26 ± 0.09
CoDis90.97 ± 0.0587.92 ± 0.1091.55 ± 0.0885.77 ± 0.3590.23 ± 0.0483.92 ± 0.0890.33 ± 0.0286.10 ± 0.09
Bare93.59 ± 0.1692.79 ± 0.1393.79 ± 0.1293.47 ± 0.1093.60 ± 0.1492.75 ± 0.1393.64 ± 0.1193.02 ± 0.14
CoDC94.87 ± 0.0694.05 ± 0.0894.66 ± 0.0893.63 ± 0.0494.64 ± 0.0993.67 ± 0.0994.89 ± 0.0694.09 ± 0.08
Standard87.21 ± 0.1070.97 ± 0.3489.55 ± 0.0577.79 ± 0.2184.84 ± 0.1061.39 ± 0.3586.37 ± 0.267.25 ± 0.44
Co-teaching91.98 ± 0.0489.21 ± 0.2391.91 ± 0.3588.00 ± 0.4391.24 ± 0.2785.94 ± 0.0692.03 ± 0.0487.94 ± 0.04
Co-teaching-plus94.71 ± 0.0192.86 ± 0.0394.37 ± 0.0388.14 ± 0.0694.15 ± 0.0288.68 ± 0.1494.74 ± 0.0191.88 ± 0.06
JoCoR96.41 ± 0.0295.70 ± 0.0196.17 ± 0.0293.89 ± 0.3796.29 ± 0.0193.88 ± 0.2696.54 ± 0.0195.29 ± 0.09
SVHNCo-learning93.18 ± 0.1092.08 ± 0.1192.74 ± 0.1888.07 ± 0.4393.39 ± 0.1189.07 ± 0.1693.48 ± 0.0792.21 ± 0.08
Co-LDL96.66 ± 0.0595.51 ± 0.3396.20 ± 0.0791.80 ± 0.3096.29 ± 0.0791.56 ± 0.2396.42 ± 0.0795.03 ± 0.06
CoDis91.67 ± 0.0489.19 ± 0.0892.10 ± 0.0687.27 ± 0.1891.25 ± 0.0785.09 ± 0.1191.83 ± 0.0388.34 ± 0.01
Bare96.20 ± 0.0694.92 ± 0.1095.83 ± 0.1292.44 ± 0.2595.84 ± 0.1891.66 ± 0.3095.98 ± 0.0794.46 ± 0.13
CoDC96.64 ± 0.0395.89 ± 0.0396.26 ± 0.0694.60 ± 0.0896.26 ± 0.0495.13 ± 0.0496.48 ± 0.0495.18 ± 0.05
Standard75.74 ± 0.3358.64 ± 0.8182.32 ± 0.1472.19 ± 0.3276.38 ± 0.3555.03 ± 0.2776.26 ± 0.2658.72 ± 0.26
Co-teaching82.24 ± 0.1877.16 ± 0.1080.76 ± 0.1172.85 ± 0.1182.55 ± 0.1075.74 ± 0.1482.50 ± 0.1276.28 ± 0.12
Co-teaching-plus81.96 ± 0.1271.49 ± 0.3379.68 ± 0.1370.96 ± 0.6979.71 ± 0.1458.39 ± 0.7681.15 ± 0.0564.79 ± 0.47
JoCoR85.23 ± 0.0979.77 ± 0.1885.62 ± 0.2173.50 ± 1.2581.75 ± 0.2668.29 ± 0.3082.78 ± 0.1173.53 ± 0.28
CIFAR-10Co-learning88.82 ± 0.2385.99 ± 0.2288.58 ± 0.2882.66 ± 0.4489.13 ± 0.1179.68 ± 0.2789.47 ± 0.1986.40 ± 0.23
Co-LDL91.57 ± 0.1588.89 ± 0.2690.89 ± 0.2284.21 ± 0.2691.10 ± 0.1785.03 ± 0.1591.12 ± 0.2287.62 ± 0.19
CoDis82.36 ± 0.2477.04 ± 0.0984.58 ± 0.0575.30 ± 0.3282.53 ± 0.2370.86 ± 0.2282.69 ± 0.0774.59 ± 0.05
Bare85.30 ± 0.6178.90 ± 0.7086.44 ± 0.3981.20 ± 0.4684.98 ± 0.5774.53 ± 0.2885.53 ± 0.4577.40 ± 0.59
CoDC92.35 ± 0.0790.15 ± 0.1391.53 ± 0.1084.25 ± 0.1292.03 ± 0.0589.46 ± 0.1592.59 ± 0.1589.70 ± 0.18
Standard46.10 ± 0.5333.20 ± 0.5846.11 ± 0.5333.20 ± 0.5846.01 ± 0.3333.16 ± 0.3246.30 ± 0.2133.29 ± 0.31
Co-teaching50.87 ± 0.3143.38 ± 0.3548.88 ± 0.2935.92 ± 0.3449.57 ± 0.3535.11 ± 0.4550.14 ± 0.3139.77 ± 0.38
Co-teaching-plus51.72 ± 0.3344.31 ± 0.6751.48 ± 0.2834.20 ± 0.6450.71 ± 0.7734.29 ± 0.2751.54 ± 0.2941.36 ± 0.29
JoCoR51.61 ± 0.3742.78 ± 0.2651.21 ± 0.0942.68 ± 0.2351.46 ± 0.3242.01 ± 0.2151.58 ± 0.0642.77 ± 0.34
CIFAR-100Co-learning62.43 ± 0.3157.18 ± 0.3563.04 ± 0.3649.69 ± 0.3162.53 ± 0.3149.29 ± 0.4163.26 ± 0.3559.57 ± 0.43
Co-LDL66.60 ± 0.2461.51 ± 0.3466.85 ± 0.2559.98 ± 0.5066.59 ± 0.2958.87 ± 0.3566.53 ± 0.2660.87 ± 0.30
CoDis50.65 ± 0.3543.44 ± 0.2750.14 ± 0.3835.43 ± 0.3850.80 ± 0.4135.15 ± 0.2951.50 ± 0.4041.43 ± 0.45
Bare63.32 ± 0.3255.33 ± 0.6261.62 ± 0.2640.88 ± 1.1761.22 ± 0.5140.92 ± 1.4062.88 ± 0.2848.82 ± 0.51
CoDC72.42 ± 0.2268.76 ± 0.1670.94 ± 0.1157.37 ± 0.3370.78 ± 0.2756.14 ± 0.1970.78 ± 0.1761.67 ± 0.16
Table 2. Mean and standard deviations of test accuracy (%) on four noisy long-tailed datasets with four different noise rates (20%, 30%, 40%, 45%). The test accuracy is calculated using the last ten epochs. The best mean results are in bold.
Table 2. Mean and standard deviations of test accuracy (%) on four noisy long-tailed datasets with four different noise rates (20%, 30%, 40%, 45%). The test accuracy is calculated using the last ten epochs. The best mean results are in bold.
Noise TypeAsym.20%Asym.30%Asym.40%Asym.45%
Standard85.77 ± 0.1780.99 ± 0.4574.90 ± 0.2570.90 ± 0.37
Co-teaching89.41 ± 0.1386.83 ± 0.1684.21 ± 0.1282.27 ± 0.21
Co-teaching-plus90.00 ± 0.1788.60 ± 0.2082.36 ± 0.2676.19 ± 0.57
JoCoR91.00 ± 0.1088.12 ± 0.1284.10 ± 0.1778.41 ± 0.29
F-MNIST-expCo-learning86.70 ± 0.2586.95 ± 0.2986.78 ± 0.4183.02 ± 0.66
Co-LDL92.16 ± 0.1191.40 ± 0.1691.28 ± 0.2189.31 ± 0.20
CoDis89.11 ± 0.1787.11 ± 0.1683.84 ± 0.2680.88 ± 0.13
Bare91.43 ± 0.1590.89 ± 0.1888.45 ± 1.3186.41 ± 1.15
CoDC92.45 ± 0.0892.24 ± 0.1091.37 ± 0.0690.20 ± 0.11
Standard86.94 ± 0.2281.38 ± 0.3375.26 ± 0.3971.68 ± 0.32
Co-teaching89.79 ± 0.1087.89 ± 0.1585.24 ± 0.2683.43 ± 0.14
Co-teaching-plus90.61 ± 0.1487.79 ± 0.2082.83 ± 0.3577.28 ± 0.58
JoCoR91.12 ± 0.1589.02 ± 0.1785.29 ± 0.2684.01 ± 0.26
F-MNIST-lineCo-learning87.84 ± 0.2386.22 ± 0.4185.66 ± 0.4383.09 ± 1.04
Co-LDL92.73 ± 0.2191.47 ± 0.2489.98 ± 0.3888.39 ± 0.47
CoDis90.13 ± 0.1087.79 ± 0.1684.77 ± 0.2681.46 ± 0.13
Bare92.11 ± 0.1391.79 ± 0.0990.55 ± 0.2889.92 ± 0.61
CoDC93.04 ± 0.0992.72 ± 0.0691.39 ± 0.1590.59 ± 0.05
Standard89.02 ± 0.2183.54 ± 0.2176.37 ± 0.4073.09 ± 0.27
Co-teaching93.08 ± 0.1691.22 ± 0.2488.49 ± 0.1485.21 ± 0.28
Co-teaching-plus93.10 ± 0.1389.57 ± 0.1983.58 ± 0.4373.28 ± 0.69
JoCoR95.01 ± 0.0894.62 ± 0.0892.65 ± 0.1189.95 ± 0.07
SVHN-expCo-learning91.67 ± 0.1890.60 ± 0.2386.52 ± 0.4081.97 ± 0.58
Co-LDL95.33 ± 0.2793.77 ± 0.3589.58 ± 0.6182.08 ± 1.56
CoDis92.80 ± 0.1791.09 ± 0.1986.89 ± 0.1181.25 ± 0.13
Bare95.42 ± 0.1492.51 ± 0.2389.48 ± 0.3383.31 ± 0.65
CoDC95.68 ± 0.0495.28 ± 0.0793.97 ± 0.0593.47 ± 0.10
Standard88.90 ± 0.1583.33 ± 0.2076.92 ± 0.2972.57 ± 0.24
Co-teaching92.92 ± 0.1091.17 ± 0.1088.33 ± 0.2585.77 ± 0.26
Co-teaching-plus92.73 ± 0.1389.41 ± 0.2082.28 ± 0.3773.94 ± 0.47
JoCoR95.34 ± 0.0494.88 ± 0.0793.56 ± 0.0691.00 ± 0.07
SVHN-lineCo-learning91.90 ± 0.0990.70 ± 0.1688.08 ± 0.2581.80 ± 0.46
Co-LDL95.28 ± 0.1293.40 ± 0.1988.79 ± 0.2583.66 ± 0.40
CoDis92.94 ± 0.1790.78 ± 0.1986.71 ± 0.1180.93 ± 0.13
Bare95.40 ± 0.1193.38 ± 0.2289.62 ± 0.1686.35 ± 0.72
CoDC95.96 ± 0.0695.77 ± 0.0695.01 ± 0.0693.56 ± 0.06
Table 3. Test accuracy (%) on Clothing1M. The best results are in bold.
Table 3. Test accuracy (%) on Clothing1M. The best results are in bold.
MethodAcc
Standard67.22
Co-teaching69.21
Co-teaching-plus59.32
JoCoR70.30
Co-learning68.72
Co-LDL71.10
CoDis71.60
Bare70.32
CoDC72.81
Table 4. Top-1 and Top-5 test accuracy (%) on the WebVision1.0 and ILSVRC12 datasets. The best results are in bold.
Table 4. Top-1 and Top-5 test accuracy (%) on the WebVision1.0 and ILSVRC12 datasets. The best results are in bold.
DatasetWebVisionILSVRC12
Method top1top5top1top5
Co-teaching63.5885.2061.4884.70
Co-teaching-plus68.5686.6465.6086.60
JoCoR61.8483.7259.1684.16
Co-LDL69.7484.2668.6384.61
CoDis70.5287.8866.8887.20
Bare69.6088.8466.4888.76
CoDC76.9691.5673.4492.08
Table 5. Ablation study on CIFAR-100 under symmetric noise at a 40% noise rate.
Table 5. Ablation study on CIFAR-100 under symmetric noise at a 40% noise rate.
ModulesAcc
L select L mlc L balance L pl
33.20
64.74
65.91
68.38
68.76
Table 6. Comparison of running time between the proposed method and four co-teaching-based methods.
Table 6. Comparison of running time between the proposed method and four co-teaching-based methods.
MethodAcc (%)Time (h)
Co-teaching77.161.34
Co-teaching-plus71.491.62
Co-LDL88.891.87
Bare78.904.02
CoDC90.151.76
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, Y.; Li, J.; Wang, Z.; Jia, W. CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency. Biomimetics 2024, 9, 92. https://doi.org/10.3390/biomimetics9020092

AMA Style

Dong Y, Li J, Wang Z, Jia W. CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency. Biomimetics. 2024; 9(2):92. https://doi.org/10.3390/biomimetics9020092

Chicago/Turabian Style

Dong, Yongfeng, Jiawei Li, Zhen Wang, and Wenyu Jia. 2024. "CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency" Biomimetics 9, no. 2: 92. https://doi.org/10.3390/biomimetics9020092

APA Style

Dong, Y., Li, J., Wang, Z., & Jia, W. (2024). CoDC: Accurate Learning with Noisy Labels via Disagreement and Consistency. Biomimetics, 9(2), 92. https://doi.org/10.3390/biomimetics9020092

Article Metrics

Back to TopTop