Next Article in Journal
Atmospheric Boundary Layer Stability in Urban Beijing: Insights from Meteorological Tower and Doppler Wind Lidar
Next Article in Special Issue
Underutilized Feature Extraction Methods for Burn Severity Mapping: A Comprehensive Evaluation
Previous Article in Journal
Classification of Ship Type from Combination of HMM–DNN–CNN Models Based on Ship Trajectory Features
Previous Article in Special Issue
Large-Kernel Central Block Masked Convolution and Channel Attention-Based Reconstruction Network for Anomaly Detection of High-Resolution Hyperspectral Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Network-Based Distillation for Hyperspectral Image Classification in the Presence of Label Noise

1
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
2
Tianjin Key Laboratory of Autonomous Intelligence Technology and Systems, Tiangong University, Tianjin 300387, China
3
Portland Institute, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
4
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
5
School of Computer Science and Technology, Tiangong University, Tianjin 300387, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(22), 4247; https://doi.org/10.3390/rs16224247
Submission received: 14 September 2024 / Revised: 6 November 2024 / Accepted: 12 November 2024 / Published: 14 November 2024

Abstract

:
Deep learning has made remarkable strides in hyperspectral image (HSI) classification, significantly improving classification performance. However, the challenge of obtaining accurately labeled training samples persists, primarily due to the subjectivity of human annotators and their limited domain knowledge. This often results in erroneous labels, commonly referred to as label noise. Such noisy labels can critically impair the performance of deep learning models, making it essential to address this issue. While previous studies focused on label noise filtering and label correction, these approaches often require estimating noise rates and may inadvertently propagate noisy labels to clean labels, especially in scenarios with high noise levels. In this study, we introduce an ensemble network-based distillation (END) method specifically designed to address the challenges posed by label noise in HSI classification. The core idea is to leverage multiple base neural networks to generate an estimated label distribution from the training data. This estimated distribution is then used alongside the ground-truth labels to train the target network effectively. Moreover, we propose a parameter-adaptive loss function that balances the impact of both the estimated and ground-truth label distributions during the training process. Our approach not only simplifies architectural requirements but also integrates seamlessly into existing deep learning frameworks. Comparative experiments on four hyperspectral datasets demonstrate the effectiveness of our method, highlighting its competitive performance in the presence of label noise.

Graphical Abstract

1. Introduction

Hyperspectral imaging provides rich spectral information, enabling fine-grained classification of ground objects with greater accuracy and detail. This capability is of immense importance across diverse domains including agricultural monitoring [1], forest inventory control [2], and urban planning and management [3]. For instance, Yuan et al. [4] used GaoFen-5 hyperspectral satellite imagery to assess and analyze the classification of complex urban functional zones located in the heart of Wuhan, China. Similarly, Rajamani et al. [5] utilized a convolutional neural network to analyze building footprints and detect roads. Fine-grained classification is essential for the effective use of hyperspectral remote sensing. As computer technology has advanced, methods for processing hyperspectral data have steadily improved.
In recent years, machine learning methods, particularly deep learning, have emerged as promising approaches for HSI classification tasks [6]. As a supervised method, deep learning heavily depends on accurate labels to train reliable models. Accurate labels are crucial because incorrect or noisy labels can cause the model to learn incorrect patterns, leading to poor performance on unseen data. Labeling complex data like HSIs is often time-consuming and challenging, typically requiring human annotators to manually assign labels to each data point. This process can be subjective and prone to errors. Despite best efforts, label noise is unavoidable due to the limitations of expert knowledge and the inherent subjectivity of human annotation [7].
In the early stages, researchers proposed several traditional methods to address label noise in HSI classification [8,9,10,11]. More recently, significant efforts have focused on tackling label noise using deep learning methodologies. For example, Ghafari et al. [12] investigated the robustness of convolutional neural networks by evaluating the performance of different loss functions, including cross-entropy, pseudo-Huber, and correntropy, on noisy hyperspectral data. Xu et al. [13] developed a novel dual-channel residual network to mitigate the effects of noisy labels. Roy et al. [14] integrated spectral and spatial domain convolutional kernels within a heterogeneous kernel convolution framework designed specifically for HSI classification with label noise. Wang et al. [15] investigated attention mechanisms and introduced an end-to-end attentive-adaptive network architecture, combined with a noise-resistant loss function to enhance overall efficiency. Zhang et al. [16] proposed a triple contrastive learning framework that explored cluster-, instance-, and structure-level representations of HSI data. To address the challenges posed by label noise, Ma et al. [17] designed a noise-tolerant learning algorithm within a spatial pooling transformer network. Xu et al. [18] introduced a superpixel-guided sample selection network for HSI classification with noisy labels, aiming to prevent the spread of label noise and correct noisy labels using a superpixel technique. Wei et al. [19] developed a unified deep learning network that effectively leverages both labeled and unlabeled data, addressing the twin challenges of limited samples and label noise. Similarly, Wang et al. [20] introduced a dual-level deep spatial manifold representation network, specifically designed for HSI classification scenarios where training samples are scarce or corrupted by noise.
Deep learning-based methods for handling label noise in HSI classification offer promising solutions. These approaches generally avoid filtering out samples with label noise, instead directly utilizing noisy labeled data to train models. However, previous methods often rely on designing complex network architectures or employing multiple techniques to effectively manage label noise. Moreover, some approaches may unintentionally exacerbate the spread of label noise to correctly labeled samples, particularly when using hard label corrections. This issue becomes even more pronounced in high-noise scenarios.
To address the aforementioned problem, we propose an innovative approach called the ensemble network-based distillation (END) method, specifically designed for HSI classification. The main idea behind END is to leverage an ensemble network to estimate the label distribution of training data affected by label noise. This estimated label distribution is then combined with the ground-truth label distribution to train a student network. Unlike previous methods that rely on corrected samples with hard labels for training, our END method generates soft labels by distilling label distribution information from the ensemble network. During the training student network’s training phase, we introduce a robust loss function that intelligently incorporates both the soft labels from the ensemble teacher and the original ground-truth labels. To effectively balance the influence of these two label sources, we include an adaptive parameter within the loss function.
The main contributions of our research are outlined as follows:
(1)
We introduce an ensemble-based strategy that revolutionizes the way hyperspectral data’s label distribution is evaluated, offering a more comprehensive and accurate perspective;
(2)
The knowledge distillation technique is utilized to train the classification network, considering both the estimated label distribution and the ground-truth label distribution;
(3)
A robust loss function with an adaptive parameter is designed for the classification network, avoiding the need to estimate the noise rate;
(4)
Extensive experiments on real hyperspectral datasets demonstrate that our END method achieves competitive results.
The rest of this paper is organized as follows: Section 2 reviews related work. The proposed END method is introduced in Section 3. Section 4 reports the experimental results and analysis. Section 5 provides a discussion, and finally, Section 6 presents the conclusions of this work.

2. Related Work

2.1. Traditional Methods for HSI Classification

Over the past few decades, researchers have developed a variety of traditional methods for HSI classification. These methods primarily focus on extracting and analyzing spectral and spatial features and can be broadly categorized into three main groups:
(1)
Spectral feature analysis methods: These methods focus on the spectral information of each pixel by analyzing reflectance or absorption characteristics across different bands for data classification. Common techniques include spectral representation and band selection [21,22], minimum distance classifiers [23], maximum likelihood classifiers [24], discriminant analysis [25], random forests [26,27], and support vector machines [27,28]. These approaches utilize statistical and machine learning models to leverage the spectral features of HSI data for identification and classification;
(2)
Spatial feature analysis methods: These methods leverage spatial information to improve classification accuracy. Typical approaches include texture analysis-based classification techniques [29], Markov random fields [30], and morphological filtering techniques [31]. These approaches utilize spatial structural features of the image, such as edges, shapes, and textures, to assist in classification and enhance the spatial consistency and accuracy of the results;
(3)
Spectral–spatial joint analysis methods: To further enhance classification performance, some methods integrate both spectral and spatial features for joint analysis. An illustrative example is object-based image analysis (OBIA) [32], which segments the image into distinct objects and utilizes both their spectral and spatial attributes for classification, providing a comprehensive and nuanced approach to image interpretation. Additionally, sparse representation and multi-scale segmentation methods [33,34] are commonly employed in spectral–spatial HSI classification;
Traditional HSI classification methods have strengths in handling high-dimensional data and capturing subtle differences, but they also face challenges, such as high computational complexity. With the advancement of deep learning technologies, there is a growing focus on applying deep learning methods to HSI classification to address the limitations of traditional approaches.

2.2. Deep Learning-Based Methods for HSI Classification

In recent years, the rapid advancement of deep learning technology has led to its increased application in HSI classification. Leveraging powerful feature extraction and learning capabilities, deep learning methods have significantly improved HSI classification performance. These deep learning-based methods for HSI classification can be broadly categorized as follows:
(1)
Convolutional neural networks (CNNs): Early 2D-CNN methods [35,36,37] primarily used two-dimensional convolution operations to extract either spectral or spatial features. While effective, these methods often addressed only one dimension at a time, potentially missing the high-dimensional characteristics of hyperspectral data. To overcome this limitation, 3D-CNN methods [38,39] employ three-dimensional convolution operations, allowing for the simultaneous processing of both spectral and spatial features. This approach significantly enhances classification accuracy by leveraging the full potential of hyperspectral data;
(2)
Recurrent neural networks (RNNs) and long short-term memory networks (LSTMs): RNNs are well-suited for processing sequential data and capturing long-range dependencies in the spectral dimension. For instance, Mou et al. [40] were among the first to apply RNNs to HSI classification, effectively modeling spectral sequence information. LSTMs, a specialized type of RNN, handle long-range dependencies more effectively and mitigate issues like gradient vanishing. As a result, they are widely used in HSI classification to extract spectral dependencies [41,42];
(3)
Deep belief networks (DBNs) and sparse auto-encoders (SAEs): DBNs exploit layer-wise unsupervised training methods to carefully extract complex features from HSIs. For example, Zhong et al. [43] introduced a novel diversified DBN framework specifically designed for HSI classification, demonstrating the versatility and efficacy of this approach. Auto-encoders perform feature extraction and data reconstruction through encoding and decoding processes. SAEs enhance these capabilities by improving feature compression and dimensionality reduction [44];
(4)
Generative adversarial networks (GANs): GANs utilize adversarial training between a generator and a discriminator to create realistic HSI data for data augmentation and sample expansion. They hold significant potential for addressing challenges related to HSI data labeling [45,46];
(5)
Advanced models: Attention mechanisms, Transformers, and Mamba models have been widely applied in HSI classification. For instance, Sun et al. [47] introduced a spectral–spatial attention network that enhances performance by integrating an attention module to extract key features from hyperspectral images. Liu et al. [48] analyzed the properties of HSI and developed a scaled dot-product central attention mechanism for spectral–spatial feature extraction, leading to the creation of a central attention network. Scheibenreif et al. [49] proposed a spatial-spectral factorization for Transformers, which reduces computational load while improving performance on hyperspectral data through self-supervised learning. More recently, Mamba models [50] have been investigated in HSI classification due to their strong long-distance modeling capabilities and linear computational complexity. Xu et al. [51] combined orientational learning with CNN to develop an orientational clustering method for open-set HSI classification;
(6)
Hybrid models: Hybrid models integrate deep learning techniques with other methodologies, such as active learning, semi-supervised learning, and transfer learning. For example, Di et al. [52] incorporated active learning into a Siamese network to reduce labeling costs. Wu et al. [53] merged semi-supervised learning with deep learning to leverage both limited labeled data and abundant unlabeled data for training effective deep neural networks. Zhong et al. [54] applied deep transfer learning for cross-scene HSI classification.
In summary, deep learning technology offers powerful tools and methods for HSI classification. Its continued development and optimization are anticipated to drive further advancements and expand applications in this field.

2.3. Label Noise Learning in HSI Classification

Label noise refers to errors or inconsistencies in the labeling of training samples, which can significantly degrade the performance of classification models. This issue has been widely investigated in the fields of machine learning and computer vision. Recent surveys on label noise learning can be found in [55,56,57]. Several studies related to our research have emerged in recent years. For example, Li et al. [58] introduced a unified distillation framework that leverages auxiliary information, such as a small clean dataset and label relationships within a knowledge graph, to mitigate the risks associated with learning from noisy labels. Lukov et al. [59] proposed soft label smoothing (SLS) to address noisy labels by adjusting high-confidence class probabilities and assigning lower probabilities to low-confidence classes. Algan and Ulusoy [60] developed MetaLabelNet (MLN), a label-noise-robust algorithm that trains soft labels using meta-objectives, optimizing gradients to minimize meta-data loss through a single-layer perceptron. Similarly, Wu et al. [61] proposed a meta-learning model that automatically estimates soft labels using meta-gradient descent, adapting label corrections iteratively based on a small amount of noise-free meta-data.
In HSI, label noise can arise from several sources [7]. Firstly, human annotators may introduce labeling errors due to limited experience. Secondly, confusion between similar spectral signatures of different classes can result in incorrect labeling. Finally, variability in sensor performance and environmental conditions during data acquisition can also contribute to label noise. To address the issue of label noise in HSI classification, researchers have employed various strategies, which can be broadly categorized into two main approaches.
(1)
Label noise detection and cleaning methods: These methods focus on identifying and correcting noisy labels before training. Common techniques include using cross-validation to detect outliers and applying majority voting among multiple classifiers. For example, Tu et al. [62,63,64] developed a series of outlier detection methods specifically for label noise. Kang et al. [65] introduced an innovative method based on constrained energy minimization to identify and correct mislabeled training samples, thereby improving data quality. Leng et al. [66] proposed a spectral–spatial sparse graph-based adaptive label propagation technique that facilitates the recognition and iterative correction of “polluted” samples, refining the dataset for better classification results. Bahraini et al. [67] suggested a modified mean-shift method to detect and remove mislabeled samples from the training set;
(2)
Label noise robust classification models: These methods aim to enhance classification performance by developing models that are resilient to label noise. Key approaches include designing noise-tolerant loss functions, applying regularization techniques, and developing novel learning architectures. For example, Kang et al. [68] explored deep metric learning and introduced a robust normalized softmax loss function specifically for remote sensing images. Damodaran et al. [69] developed a loss function based on optimal transport theory to improve deep learning under label noise. Zhang et al. [70] implemented a co-learning strategy using a dual network architecture to address the challenges of HSI classification with noisy labels. Liao et al. [71] proposed a meta-learning framework that employs joint positive and negative learning to adaptively reweight samples, enhancing classification robustness. Fang et al. [72] introduced a deep reinforcement learning method to address label noise in HSI classification.
In summary, label noise robust classification models provide several advantages over label noise detection and cleaning methods. Firstly, they utilize the entire training data without discarding noisy samples. Secondly, robust classification models, especially those incorporating deep learning techniques, can achieve superior performance by extracting more complex features for improved modeling.

2.4. Knowledge Distillation in HSI Classification

Knowledge distillation is a model compression technique designed to transfer knowledge from a large, complex model (often referred to as the teacher) to a smaller, more efficient model (the student). This process involves two key stages. First, the teacher model is trained on labeled data, generating soft labels—probability distributions over classes—that capture class similarities. These soft labels provide richer information than hard labels alone. In the second stage, the student model learns by mimicking the teacher’s soft labels in conjunction with the original hard labels, enabling it to understand both the ground truth and the nuanced relationships learned by the teacher. This dual-learning approach enhances the student model’s performance and generalization ability.
In recent years, researchers have investigated various knowledge distillation methods for HSI classification. Notable contributions include the following: Due to the limited availability of labeled HSI samples, deep learning methods have remained underutilized. To address this issue, Yue et al. [73] proposed a self-supervised learning method with adaptive distillation that leverages abundant unlabeled samples for training. Zhao et al. [74] introduced a lifelong learning strategy to develop universal HSI classification models, continuously updating model weights through spectral–spatial feature distillation. To tackle classification problems with a limited number of samples while maximizing the use of unlabeled data, Chi et al. [75] proposed a novel self-supervised learning method that incorporates knowledge distillation for HSI classification. Additionally, Feng et al. [76] developed a method to enhance cross-domain learning for HSI classification by treating meta-knowledge extraction and source domain debiasing as a synergistic process through decoupled knowledge distillation.
The aforementioned knowledge distillation methods for HSI classification primarily focus on limited labeled samples, domain adaptation, and universal models. Our method first employs knowledge distillation to address the issue of label noise in HSI classification. While similar to these methods in utilizing knowledge distillation to generate soft labels, our approach differs by integrating knowledge distillation within an ensemble learning framework. In this framework, the resampling strategy and out-of-bag (OOB) error estimation are used to produce more robust soft labels. Our goal is to enhance the model’s robustness under label noise conditions, ensuring that the student model can still effectively distinguish categories in its final output.

3. Proposed Method

In this paper, we propose an ensemble network-based distillation (END) method for HSI classification under label noise conditions. The overall framework of the END method is depicted in Figure 1. The approach begins by applying ensemble learning to construct multiple base networks, each trained on distinct subsets of the data. These base networks are then used to estimate the labels of the training samples while accounting for label noise. By aggregating the predictions from these base networks, we generate an estimated distribution (ED) that captures the collective consensus.
This ED is subsequently combined with the ground-truth distribution (GD) to train a student network. The student network learns from both the ED and GD, harnessing the benefits of ensemble learning while mitigating the effects of label noise. To balance the influence of the ED and GD during training student networks, we introduce a robust loss function with an adaptive parameter. This parameter dynamically adjusts the weighting between the ED and GD, enabling the student network to learn accurate representations despite the presence of noisy labels.
By integrating ensemble learning with knowledge distillation, the proposed END method not only strengthens the model’s robustness against label noise but also enhances its overall classification accuracy, delivering more reliable and precise results even in challenging scenarios. Additionally, the adaptive nature of the loss function ensures the method’s effectiveness across different noise levels and data distributions, making it a versatile and powerful tool for HSI classification tasks.
To ensure that the effectiveness of our method in handling label noise is not merely a consequence of using advanced base network models, we initially implemented our approach with a 2D-CNN as the base network. In subsequent experiments, we will also validate our method using a 3D network, specifically the spectral–spatial residual network (SSRN) [38].
In the subsequent subsections, we delve into a detailed exploration of the ensemble network-based distillation process and the robust loss function, providing a comprehensive understanding of their complexities and contributions to the proposed framework.

3.1. Ensemble Network-Based Distillation

Ensemble learning has demonstrated strong performance in HSI classification [27,77], with research indicating its robustness in the presence of label noise, making it particularly effective for handling noisy data [78,79]. To estimate the label distribution of the training data, we construct an ensemble network by integrating multiple deep networks, thereby leveraging their diverse strengths to improve classification accuracy and stability.
To facilitate this, we employ the bootstrap sampling method, which is well-suited for ensemble learning. Bootstrap sampling generates multiple subsets of the training data by randomly sampling with replacement. This approach fosters the creation of diverse training sets and naturally produces a set of out-of-bag (OOB) samples—data points excluded from their respective bootstrap samples. These OOB samples provide additional valuable information, as the OOB error, derived from predictions made on these samples, closely approximates the model’s generalization error [80,81]. This makes OOB samples particularly useful for evaluating the ensemble network’s performance and reliability.
In our method, bootstrap sampling is used to generate multiple training subsets, each assigned to a different base network within the ensemble. Once trained, the predictions on the OOB samples are used to estimate the label distribution of the training data. This OOB-based estimation offers a more reliable and accurate assessment of the label distribution, which is critical for effectively managing noisy labels in subsequent stages of our method. By combining these predictions, we construct an estimated label distribution that serves as a robust foundation for refining the classification model through the ensemble network-based distillation approach.
The following part of this subsection details the steps involved in obtaining the estimated distribution based on the ensemble network.
(1) Given the hyperspectral dataset D = { ( x i , y i ) } i = 1 N , where x i is a sample described by spectral bands, and the label corresponding to x i is y i { 1 , , C } ;
(2) Set 0 matrices E D N × C , P N × C , and Q N × C , where E D , P , and Q represents the estimated distribution, prediction matrix, and sample count matrix, respectively;
(3) Obtain T subsets D 1 , D 2 , , D T of dataset D by employing a bootstrap sampling strategy;
(4) Train T base neural networks b 1 , b 2 , b T on subsets D 1 , D 2 , , D T ;
(5) For each set D D t , every OOB sample x j in D D t is classified by its corresponding neural network b i . Update p and Q through the following equations:
P x j = P x j + b ( x j ) ,
Q x j = Q x j + 1 .
(6) Compute estimated distribution through the following formula:
E D = P . / Q ,
where the symbol “./” denotes element-wise division operation.
Rather than categorizing samples as simply correct or incorrect and discarding those deemed incorrect, we adopt a more nuanced strategy. Using the estimated distribution (ED), we assign probability scores to all possible classes for each sample, offering a deeper insight into the ensemble’s decision-making process. By labeling each sample with a probability distribution across all classes, rather than using a binary correct/incorrect label, we capture the uncertainty and variability inherent in the ensemble’s predictions. This approach ensures that no potentially valuable data are discarded, allowing the model to learn from the full spectrum of available information and enhancing its overall robustness during training.

3.2. Noise Robust Loss Function with an Adaptive Parameter

After deriving the ED distilled from the ensemble network, we combine it with the GD to train the student network. To effectively guide this training, we require a loss function that accommodates both estimated and ground-truth labels. While cross-entropy (CE) loss is widely recognized for its efficacy in model convergence, previous studies [82,83] have shown that it lacks robustness in the presence of label noise. In contrast, reverse cross-entropy (RCE) has been proven to be more resistant to noisy labels [83]. This suggests that combining CE and RCE could result in a noise-robust loss function.
Building on these findings, we design a hybrid loss function that merges CE and RCE to improve noise robustness. To ensure a balanced influence between the soft labels (derived from the ED) and the ground-truth labels, we incorporate an adaptive parameter that dynamically adjusts this balance throughout the training process. Furthermore, we apply a normalized version of CE (NCE) loss to the estimated label distribution, as normalization has been shown to further enhance robustness against label noise.
The following content will provide a detailed procedure for the design and implementation of this loss function.
For a sample x i , its predicted label distribution from a classifier is denoted as p ( c | x i ) , and the ground-truth label distribution over observed labels is denoted as q ( c | x i ) . The CE loss for sample x i is as follows:
l C E ( x i ) = c = 1 C q ( c | x i ) log p ( c | x i ) .
The NCE is written as follows.
l N C E ( x i ) = c = 1 C q ( c | x i ) log p ( c | x i ) j = 1 C c = 1 C q ( y = j | x i ) log p ( c | x i ) .
RCE is the reverse version of CE, the RCE loss is as follows:
l R C E   x i = c = 1 C p ( c | x i ) log q ( c | x i ) .
Thus, we define the sample-wise mixed loss for the student network (S) as follows:
E N D M I X x i = 1 2 l N C E S x i , 1 λ E D x i + λ y i + 1 2 l R C E [ S x i , 1 λ E D x i + λ y i ]
In Equation (7), S x i represents the prediction made by the student network for the sample x i . The parameter λ is defined as follows:
λ = i = 1 N 𝟙 ( y i = = a r g  max  c E D ( c | x i ) ) N ,
where N is the total number of samples, y i is the ground-truth label, a r g  max  c E D ( c | x i ) is the class label with the highest probability in the estimated distribution for the sample x i , and the indicator function 𝟙 ( · ) is a binary operator that evaluates a given condition. It returns a value of 1 if the condition is true and 0 if the condition is false. Essentially, λ represents the proportion of samples for which the predicted label from the ensemble network matches the ground-truth label.
The parameter λ plays a crucial role in determining how the student network is trained in relation to the ensemble model’s confidence in a given sample. Specifically, when the ensemble model exhibits low confidence, indicated by a lower value of λ, the student network is guided predominantly by the ground-truth labels during training. This reduces the reliance on the estimated distribution, ensuring that the model aligns more closely with the true class labels. On the other hand, when the ensemble model shows high confidence in a sample, as indicated by a higher value of λ, the estimated distribution takes precedence in the training process, even if it contradicts the ground-truth label. This strategy allows the student model to focus more on the information provided by the ensemble model’s probability estimates, which is believed to reflect a more reliable and confident prediction. In this way, the system dynamically adjusts its reliance on either the ground-truth labels or the estimated distribution, depending on the confidence level of the ensemble model.
The above analysis provides key insights into the END method, and Algorithm 1 presents the pseudocodes of the END method.
Algorithm 1 Ensemble Network-Based Distillation with Robust Loss Function
Input:   Ensemble   size :   T ;   Training   dataset   D = { ( x i , y i ) } i = 1 N .
Initialization: E D N × C = 0 ,   P N × C = 0 ,   Q N × C = 0 .
1 :   G D = g r o u n d t r u t h ( D ) .  //obtain the ground-truth distribution
2: For t = 1 to T do
3 :      D t = R e s a m p l i n g ( D ) .  //bootstrap resampling
4 :      b t = T r a i n ( D t ) .  //train base network
5 :      For   x j D D t  do
6 :              P x j = P x j + b ( x j ) .    / /   the   cumulative   prediction   results   for   x j
7 :              Q x j = Q x j + 1 .  //frequency count for x j
8:      End
9: End
10 :   E D = P . / Q .  //compute the estimated distribution
11 :   S = T r a i n D , G D + E D , E N D M I X .   //train the student network S through (7)
Output: The student network S.

4. Experimental Results and Analysis

4.1. Hyperspectral Datasets and Experimental Settings

We evaluated the proposed END method against several state-of-the-art approaches using the Salinas Valley, Houston, and Pavia University datasets to demonstrate its effectiveness. A detailed description of these datasets is provided below.
(1)
Salinas Valley (SV) dataset: Collected in 1998 over Salinas Valley, CA, USA, this dataset consists of 512 × 217 pixels and contains 224 spectral bands, covering a wavelength range from 400 to 2500 nm. After removing 20 water absorption bands, 204 spectral bands remain. With a spatial resolution of 3.7 m, the dataset covers 16 land cover classes;
(2)
Houston (HOU) dataset: Acquired by the ITRES CASI-1500 sensor, this dataset was part of the 2013 IEEE GRSS Data Fusion Competition. It consists of 349 × 1905 pixels and includes 144 spectral bands, spanning wavelengths from 364 to 1046 nm. The HOU dataset has a spatial resolution of 2.5 m and covers 15 land cover classes, providing a comprehensive representation of urban and suburban landscapes;
(3)
Pavia University (PU) dataset: Acquired by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor over the University of Pavia in northern Italy, this dataset contains an image matrix of 610 pixels in width and 340 pixels in height. It includes 103 spectral bands, covering a wavelength range from 430 to 860 nm.
Table 1 provides a comprehensive overview of the class names along with the corresponding number of labeled samples for each of the three datasets. Figure 2, Figure 3 and Figure 4 illustrate the false-color images of these HSI datasets along with their respective reference maps.
To ensure a rigorous quantitative evaluation of our experimental results, we employed three key metrics: overall accuracy (OA), average accuracy (AA), and the Kappa coefficient (κ). Each metric serves as a crucial indicator to assess the performance and accuracy of our classification outcomes. Label noise was simulated using a uniform distribution, where randomly selected samples were reassigned to different class labels with a random probability. To assess the effect of varying noise levels on classification performance, we conducted experiments using training data with multiple levels of label noise.
Our proposed END method was built on a 2D CNN as the base network. The training process was optimized using the Adam optimizer with a learning rate of 0.001. The base networks in the ensemble were trained for 30 epochs, while the student network was trained for 100 epochs. The implementation was performed using PyTorch, leveraging an NVIDIA RTX 3070 GPU equipped with CUDA 11, which significantly accelerated the training process. Following preliminary experiments, the ensemble size was set to 30 for optimal performance.

4.2. Comparison with the State-of-the-Art Methods

We conducted a comparative evaluation of our proposed END method against several state-of-the-art approaches, including MSSG [10], DCRN [13], AAN [15], TCRL [16], SLS [59], and MLN [60]. Additionally, we compared the proposed method with the base network 2D-CNN and a traditional ensemble method, random forests (RF) [80]. For consistency, we adhered to the parameter settings specified in their respective publications. In the case of MSSG, the segmentation scale parameter was optimized to 0, replicating the best-performing configuration from the original experiments. To ensure standardization of the training data, we randomly sampled 50 instances from each class and ran each method ten times to obtain average performance metrics. With a fixed noise rate (r) of 0.3, each algorithm was executed 10 times, and the average results were used for comparison. Table 2, Table 3 and Table 4 provide a detailed analysis of the classification performance on the SV, HOU, and PU datasets. Figure 5, Figure 6 and Figure 7 display the classification maps for the three HSI datasets after a single run of each method.
Several key insights can be drawn from Table 2, Table 3 and Table 4 and Figure 5, Figure 6 and Figure 7. First, the OA and Kappa coefficients of these methods on the SV and PU datasets demonstrate a clear ranking: END, TCRL, AAN, DCRN, MLN, MSSG, SLS, 2D-CNN, and RF. In contrast, for the HOU dataset, the only difference is that the OA and Kappa coefficients of MSSG are lower than those of SLS. The AA ordering of DCRN, TCRL, and MLN is inconsistent across the three datasets in the comparison methods. However, the AA ordering for the other methods is consistent and follows this sequence: END, AAN, MSSG, SLS, 2D-CNN, and RF. The above ranking underscores the END method’s superior ability to manage label noise effectively while achieving high classification accuracy. Second, the END method consistently delivers the highest class-specific accuracy across a broader range of classes compared to the other methods, further affirming its robustness. Notably, it performs exceptionally well in challenging scenarios. For example, on the SV dataset, the END method significantly improves accuracy for class 8 by approximately 4% and for class 15 by about 14% compared to the second-best performing method. Finally, the visual results presented in Figure 5, Figure 6 and Figure 7 further validate the effectiveness of the END method. The classification maps generated by the END approach offer more precise and detailed representations of the underlying land cover classes compared to those produced by other methods, highlighting its superior performance.

4.3. Classification Performance Across Varying Noise Levels

In this section, we delve into experiments that varied the level of label noise, with the noise rate (r) systematically increasing from 0.1 to 0.5 in steps of 0.1, while maintaining consistency in all other parameters as established in the previous section. Table 5, Table 6 and Table 7 comprehensively present the impact of these varying noise levels on the performance of the comparison methods, reporting the OA, AA, and Kappa coefficients for the SV, HOU, and PU datasets, respectively.
Several key observations can be drawn from Table 5, Table 6 and Table 7. First, across all three HSI datasets, there is a noticeable decline in OA, AA, and Kappa coefficients as the noise rate increases, underscoring the significant challenge that label noise presents in HSI classification. Second, the proposed END method consistently outperforms other approaches in terms of classification accuracy, regardless of the noise level. This demonstrates the robustness of the END method in effectively managing label noise compared to alternative techniques. Lastly, the END method exhibits low sensitivity to increasing noise levels, with only minimal reductions in performance observed. This adaptability highlights its effectiveness in handling various noise conditions, making it particularly well-suited for real-world HSI classification tasks, where label noise is often unavoidable. Overall, the robustness and flexibility of the END method position it as a highly reliable solution for classification tasks under noisy conditions.

4.4. Evaluation of the Loss Function

To assess the effectiveness of the proposed loss function, we conducted comparisons with CE loss, NCE loss, and RCE loss. The loss functions for the proposed method can be written as follows:
E N D C E x i =   l C E S x i , 1 λ E D x i + λ y i ,
E N D N C E x i = l N C E S x i , 1 λ E D x i + λ y i ,
E N D R C E x i = l R C E S x i , 1 λ E D x i + λ y i ,
where the parameter λ is set the same as in (8), acting as an adaptive parameter, and S x i represents the prediction of the student network.
In this experiment, we formed training sets of various sizes by randomly selecting 10 to 80 samples for each class, while maintaining a fixed noise rate r = 0.3. Figure 8 illustrates the OA curves of the END method with different versions of loss functions plotted against the number of training samples.
The results in Figure 8 indicate, first, that the number of training samples has a notable influence on classification accuracy regardless of the chosen loss function. Specifically, under a fixed noise rate (r = 0.3), the OA tends to increase as the number of training samples grows. Second, the E N D M I X loss significantly improves the model’s classification accuracy in the presence of label noise. This effectiveness can be attributed to two factors. On the one hand, the designed loss function simultaneously leverages the complementary advantages of cross-entropy and reverse cross-entropy, allowing it to fit both noisy and clean data more effectively. On the other hand, the adaptive parameter λ effectively balances the roles of ED and GD in the target network during training. This adaptability addresses varying levels of noise, enabling the model to achieve strong classification performance across different noise scenarios.

4.5. Ablation Study

To fully investigate the significance of each component in the proposed method, we removed various elements of the END method for our experiments. Detailed information is provided in Table 8. In this section, we conducted experiments using a 2D-CNN as the base network as well as a 3D network, specifically the spectral–spatial residual network (SSRN) [38]. The parameter settings for the SSRN adhere to those outlined in the original paper, while the parameters for the other methods remain consistent with our previous experiments. Table 9 presents the OA of three methods applied to the SV, HOU, and PU datasets.
Table 9 presents several important findings. First, SSRN consistently outperforms 2D-CNN in classification accuracy across all three datasets, due to the enhanced feature extraction capabilities of the 3D network. Second, while Bagging does improve classification accuracy in the presence of label noise, the degree of improvement remains relatively modest. Finally, the END method demonstrates the most substantial enhancement in accuracy, primarily because it utilizes knowledge distillation from the ensemble network. In other words, END employs ensemble learning not directly for classification, but to estimate a label distribution, which is then combined with the ground-truth distribution to effectively train the classification network.

4.6. Analysis of the Ensemble Size

In the previous experiment, we used an ensemble size of 30. To explore the influence of ensemble size on classification performance and determine the optimal size, we conducted experiments with ensemble sizes ranging from 5 to 50 in increments of 5. Figure 9 illustrates the relationship between classification accuracy and ensemble size.
The results in Figure 9 indicate that classification accuracy increases rapidly with the rise in ensemble size before stabilizing. As shown in Figure 9, setting the ensemble size to around 20 or more yields satisfactory classification results. This suggests that the proposed method does not require constructing a large number of base networks, which effectively reduces computational costs.

4.7. Analysis of the Training Epoch for Base Network

To obtain a relatively accurate estimated distribution distilled from the ensemble, it is crucial to determine an appropriate epoch for terminating the training of the base networks early, in order to prevent overfitting to noisy samples. In our experiment, we set the maximum epoch to 100. Starting from the 10th epoch, at the end of each subsequent epoch, we use the resulting ED and ground-truth distribution to train the student network. The classification accuracy is shown in Figure 10.
As shown in Figure 10, the OA exhibits oscillations over time. Generally, the OA initially increases, stabilizes, and then decreases as the number of epochs progresses. This trend can be attributed to the model’s ability to learn more effectively from “clean” samples in the early stages, leading to more accurate estimated distributions. However, as the model begins to incorporate noisy data, convergence becomes more difficult, resulting in less accurate estimated distributions. Therefore, it is recommended to terminate the training early, ideally between 20 and 50 epochs, to prevent overfitting to noisy samples.

4.8. Experimental Results on Toulouse Dataset

To evaluate the robustness to label noise on a large dataset, we conducted experiments using the Toulouse HSI dataset [84]. This dataset was captured by an AisaFENIX 1K camera over the city of Toulouse, France. It contains approximately 380,000 labeled pixels and includes 310 spectral bands, spanning wavelengths from 0.4 to 2.5 μm. The Toulouse dataset has a spatial resolution of 1 m and covers 32 land cover classes. The ground truth of the Toulouse data consists of eight spatially disjoint splits.
We adopted the original data division in [84] for our experiments, and the experimental settings were consistent with previous studies. Table 10 reports the classification accuracy of the comparison methods on the Toulouse dataset. The results presented in Table 10 demonstrate that our proposed method achieves the highest classification accuracy among all comparison methods. Furthermore, it outperforms the second-place method by approximately 2 percent, indicating that our approach is effective for disjoint HSI classification with noisy labels.

5. Discussion

Some important findings from previous experimental results warrant further discussion.
The experimental results in Section 4.2 indicate that the END method outperforms other approaches, particularly when classifying challenging classes. When the labels of these challenging classes are contaminated by noise, it becomes increasingly difficult to accurately classify such samples. The superior performance of the END method in handling noisy labels suggests that its ensemble-based approach effectively mitigates the impact of label noise. In future research, it would be valuable to focus on improving classification accuracy specifically for these challenging classes. Designing methods that better address noisy labels in difficult classes could significantly enhance overall performance in HSI classification tasks.
The experimental results in Section 4.3 demonstrate that other approaches, compared to the END method, are more sensitive to high levels of label noise. The proposed END method addresses this issue by using ensemble learning and knowledge distillation to estimate the label distribution and an adaptive loss function to balance the estimated distribution with the ground-truth distribution. This allows the END method to handle high levels of noise more robustly. Future research should focus on further exploring techniques and methods to enhance performance in high-noise environments.
As shown in Section 4.4, classification accuracy drops significantly across all loss functions when fewer than 40 labeled samples per class are available. This indicates that, under conditions of limited labeled samples, label noise has a more pronounced negative impact on HSI classification. The primary challenge lies in the difficulty of extracting the true label distribution when there are already too few labeled samples. Future research should address both the scarcity of labeled samples and the presence of label noise. For instance, exploring unsupervised methods to infer the underlying label distribution could prove beneficial. This approach could leverage unlabeled data to enhance the model’s robustness in noisy environments, improving overall classification performance.
The ablation study in Section 4.5 demonstrates that while the direct performance improvement from ensemble learning alone is limited, the knowledge distilled from the ensemble model can provide valuable guidance for training the target network. This knowledge distillation approach enhances the model’s robustness, especially in noisy label scenarios, by effectively transferring ensemble-derived insights. Future research can further explore techniques to maximize the effectiveness of ensemble learning in distillation, aiming to distill richer and more targeted knowledge that can boost classification accuracy and generalization in various contexts.

6. Conclusions

In this paper, we introduce an ensemble network-based distillation method for HSI classification. The proposed END method leverages ensemble learning to effectively estimate the distribution of training data. Additionally, a novel loss function, more robust to label noise, is designed for training the target network. Experimental results demonstrate that the proposed method can effectively alleviate the influence of label noise, even at high noise rates. Several key conclusions can be drawn from the experimental results and analysis:
  • While ensemble learning offers limited performance gains for HSI classification with noisy labels, the knowledge distilled from the ensemble model can serve as valuable guidance for training the target network;
  • The predictions for OOB samples generated through resampling effectively estimate the label distribution of the training data. By building multiple base networks and aggregating their outputs for OOB samples, this method improves noise robustness and provides a refined label distribution for subsequent learning;
  • The tailored loss function enhances the model’s resilience against label noise, further bolstering its robustness. By dynamically balancing the importance of the estimated label distribution and the ground-truth label distribution, the loss function enables the student network to learn accurate representations.
In future research, several avenues can be explored to further refine the END method. First, incorporating resampling strategies, such as boosting, could help mitigate the impact of noisy labels by focusing on difficult-to-classify samples. Moreover, future studies could explore applying advanced deep learning architectures, such as transformers or graph neural networks, in conjunction with the proposed method. This would open new possibilities for improving the model’s adaptability and scalability in various HSI classification tasks.

Author Contributions

Conceptualization, R.D.; methodology, Y.Z. and H.S.; software, J.L. and Q.Y.; validation, R.D. and J.L.; formal analysis, Q.Y. and X.L.; writing—original draft preparation, Y.Z.; writing—review and editing, R.D. and X.L.; visualization, H.S.; supervision, Y.Z. and X.L.; funding acquisition, Y.Z. and G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded in part by the National Natural Science Foundation of China under grants 62201282 and 62203231, in part by the Open Project of Tianjin Key Laboratory of Autonomous Intelligence Technology and Systems under grant TJKL-AITS-20241003, in part by the Natural Science Foundation of Jiangsu Province under grant BK20231456, and in part by the Basic Research Project of the Natural Science Foundation of the Jiangsu Higher Education Institutions under grants 22KJB510037 and 22KJB520006.

Data Availability Statement

The Salinas Valley and Pavia University datasets used in this study are accessible at [https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes] (accessed on 10 June 2023). The Houston dataset is available at [https://hyperspectral.ee.uh.edu/?page_id=459] (accessed on 8 May 2024). The Toulouse dataset is available at [https://www.toulouse-hyperspectral-data-set.com/] (accessed on 29 October 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  2. Anderson, J.E.; Plourde, L.C.; Martin, M.E.; Braswell, B.H.; Smith, M.L.; Dubayah, R.O.; Hofton, M.A.; Blair, J.B. Integrating waveform lidar with hyperspectral imagery for inventory of a northern temperate forest. Remote Sens. Environ. 2008, 112, 1856–1870. [Google Scholar] [CrossRef]
  3. Heiden, U.; Heldens, W.; Roessner, S.; Segl, K.; Esch, T.; Mueller, A. Urban structure type characterization using hyperspectral remote sensing and height information. Landsc. Urban Plann. 2012, 10, 361–375. [Google Scholar] [CrossRef]
  4. Yuan, J.; Wang, S.; Wu, C.; Xu, Y. Fine-grained classification of urban functional zones and landscape pattern analysis using hyperspectral satellite imagery: A case study of Wuhan. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3972–3991. [Google Scholar] [CrossRef]
  5. Rajamani, T.; Sevugan, P.; Ragupathi, S. Automatic building footprint extraction and road detection from hyperspectral imagery. J. Electron. Imaging. 2023, 32, 011005. [Google Scholar] [CrossRef]
  6. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J.; Khan, A.M.; Mazzara, M.; Distefano, S.; Chanussot, J. Hyperspectral image classification—Traditional to deep models: A survey for future prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 968–999. [Google Scholar] [CrossRef]
  7. Jiang, J.; Ma, J.; Wang, Z.; Chen, C.; Liu, X. Hyperspectral image classification in the presence of noisy labels. IEEE Trans. Geosci. Remote Sens. 2019, 57, 851–865. [Google Scholar] [CrossRef]
  8. Tu, B.; Zhou, C.; Liao, X.; Xu, Z.; Peng, Y.; Ou, X. Hierarchical structure-based noisy labels detection for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2183–2199. [Google Scholar] [CrossRef]
  9. Li, Z.; Yang, X.; Meng, D.; Cao, X. An adaptive noisy label-correction method based on selective loss for hyperspectral image-classification problem. Remote Sens. 2024, 16, 2499. [Google Scholar] [CrossRef]
  10. Jiang, J.; Ma, J.; Liu, X. Multilayer spectral–spatial graphs for label noisy robust hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 839–852. [Google Scholar] [CrossRef]
  11. Yang, S.; Jia, Y.; Ding, Y.; Wu, X.; Hong, D. Unlabeled data guided partial label learning for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5503405. [Google Scholar] [CrossRef]
  12. Ghafari, S.; Tarnik, M.G.; Yazdi, H.S. Robustness of convolutional neural network models in hyperspectral noisy datasets with loss functions. Comput. Electr. Eng. 2021, 90, 107009. [Google Scholar] [CrossRef]
  13. Xu, Y.; Li, Z.; Li, W.; Du, Q.; Liu, C.; Fang, Z.; Zhai, L. Dual-channel residual network for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502511. [Google Scholar] [CrossRef]
  14. Roy, S.K.; Hong, D.; Kar, P.; Wu, X.; Liu, X.; Zhao, D. Lightweight heterogeneous kernel convolution for hyperspectral image classification with noisy labels. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5509705. [Google Scholar] [CrossRef]
  15. Wang, L.; Zhu, T.; Kumar, N.; Li, Z.; Wu, C.; Zhang, P. Attentive-adaptive network for hyperspectral images classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5505514. [Google Scholar] [CrossRef]
  16. Zhang, X.; Yang, S.; Feng, Z.; Song, L.; Wei, Y.; Jiao, L. Triple contrastive representation learning for hyperspectral image classification with noisy labels. IEEE Trans. Geosci. Remote Sens. 2023, 61, 500116. [Google Scholar] [CrossRef]
  17. Ma, J.; Zou, Y.; Tang, X.; Zhang, X.; Liu, F.; Jiao, L. Spatial pooling transformer network and noise-tolerant learning for noisy hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5509719. [Google Scholar] [CrossRef]
  18. Xu, H.; Zhang, H.; Zhang, L. A superpixel guided sample selection neural network for handling noisy labels in hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9486–9503. [Google Scholar] [CrossRef]
  19. Wei, W.; Xu, S.; Zhang, L.; Zhang, J.; Zhang, Y. Boosting hyperspectral image classification with unsupervised feature learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5502315. [Google Scholar] [CrossRef]
  20. Wang, C.; Zhang, L.; Wei, W.; Zhang, Y. Toward effective hyperspectral image classification using dual-level deep spatial manifold representation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5505614. [Google Scholar] [CrossRef]
  21. Kang, X.; Zhu, Y.; Duan, P.; Li, S. Two dimensional spectral representation. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5502809. [Google Scholar] [CrossRef]
  22. Sun, W.; Du, Q. Hyperspectral band selection: A review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 118–139. [Google Scholar] [CrossRef]
  23. Zhao, Q.; Jia, S.; Li, Y. Hyperspectral remote sensing image classification based on tighter random projection with minimal intra-class variance algorithm. Pattern Recognit. 2021, 111, 107635. [Google Scholar] [CrossRef]
  24. Peng, J.; Li, L.; Tang, Y.Y. Maximum likelihood estimation-based joint sparse representation for the classification of hyperspectral remote sensing images. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1790–1802. [Google Scholar] [CrossRef]
  25. Li, X.; Zhang, L.; You, J. Locally weighted discriminant analysis for hyperspectral image classification. Remote Sens. 2019, 11, 109. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Cao, G.; Li, X.; Wang, B.; Fu, P. Active semi-supervised random forest for hyperspectral image classification. Remote Sens. 2019, 11, 2974. [Google Scholar] [CrossRef]
  27. Sheykhmousa, M.; Mahdianpari, M.; Ghanbari, H.; Mohammadimanesh, F.; Ghamisi, P.; Homayouni, S. Support vector machine versus random forest for remote sensing image classification: A meta-analysis and systematic review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6308–6325. [Google Scholar] [CrossRef]
  28. Chen, Y.N.; Thaipisutikul, T.; Han, C.C.; Liu, T.J.; Fan, K.C. Feature line embedding based on support vector machine for hyperspectral image classification. Remote Sens. 2021, 13, 130. [Google Scholar] [CrossRef]
  29. Mirzapour, F.; Ghassemian, H. Improving hyperspectral image classification by combining spectral, texture, and shape features. Int. J. Remote Sens. 2015, 36, 1070–1096. [Google Scholar] [CrossRef]
  30. Li, W.; Prasad, S.; Fowler, J.E. Hyperspectral image classification using Gaussian mixture models and Markov random fields. IEEE Geosci. Remote Sens. Lett. 2013, 11, 153–157. [Google Scholar] [CrossRef]
  31. Samat, A.; Li, E.; Wang, W.; Liu, S.; Lin, C.; Abuduwaili, J. Meta-XGBoost for hyperspectral image classification using extended MSER-guided morphological profiles. Remote Sens. 2020, 12, 1973. [Google Scholar] [CrossRef]
  32. Hossain, M.D.; Chen, D. Segmentation for object-based image analysis (OBIA): A review of algorithms and challenges from remote sensing perspective. ISPRS J. Photogramm. Remote Sens. 2019, 150, 115–134. [Google Scholar] [CrossRef]
  33. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral–spatial hyperspectral image classification via multiscale adaptive sparse representation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7738–7749. [Google Scholar] [CrossRef]
  34. Li, S.; Ni, L.; Jia, X.; Gao, L.; Zhang, B.; Peng, M. Multi-scale superpixel spectral–spatial classification of hyperspectral images. Int. J. Remote Sens. 2016, 37, 4905–4922. [Google Scholar] [CrossRef]
  35. Yu, S.; Jia, S.; Xu, C. Convolutional neural networks for hyperspectral image classification. Neurocomputing 2017, 219, 88–98. [Google Scholar] [CrossRef]
  36. Gao, Q.; Lim, S.; Jia, X. Hyperspectral image classification using convolutional neural networks and multiple feature learning. Remote Sens. 2018, 10, 299. [Google Scholar] [CrossRef]
  37. Zhang, M.; Li, W.; Du, Q. Diverse region-based CNN for hyperspectral image classification. IEEE Trans. Image Process. 2018, 27, 2623–2634. [Google Scholar] [CrossRef]
  38. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  39. Zhou, J.; Zeng, S.; Xiao, Z.; Zhou, J.; Li, H.; Kang, Z. An enhanced spectral fusion 3D CNN model for hyperspectral image classification. Remote Sens. 2022, 14, 5334. [Google Scholar] [CrossRef]
  40. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  41. Liu, Q.; Zhou, F.; Hang, R.; Yuan, X. Bidirectional-convolutional LSTM based spectral-spatial feature learning for hyperspectral image classification. Remote Sens. 2017, 9, 1330. [Google Scholar] [CrossRef]
  42. Mei, S.; Li, X.; Liu, X.; Cai, H.; Du, Q. Hyperspectral image classification using attention-based bidirectional long short-term memory network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5509612. [Google Scholar] [CrossRef]
  43. Zhong, P.; Gong, Z.; Li, S.; Schönlieb, C.B. Learning to diversify deep belief networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3516–3530. [Google Scholar] [CrossRef]
  44. Tao, C.; Pan, H.; Li, Y.; Zou, Z. Unsupervised spectral–spatial feature learning with stacked sparse autoencoder for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2438–2442. [Google Scholar]
  45. Feng, J.; Feng, X.; Chen, J.; Cao, X.; Zhang, X.; Jiao, L.; Yu, T. Generative adversarial networks based on collaborative learning and attention mechanism for hyperspectral image classification. Remote Sens. 2020, 12, 1149. [Google Scholar] [CrossRef]
  46. Zhu, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Generative adversarial networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5046–5063. [Google Scholar] [CrossRef]
  47. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 323–3245. [Google Scholar] [CrossRef]
  48. Liu, H.; Li, W.; Xia, X.; Zhang, M.; Gao, C.; Tao, R. Central attention network for hyperspectral imagery classification. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 8989–9003. [Google Scholar] [CrossRef]
  49. Scheibenreif, L.; Mommert, M.; Borth, D. Masked vision transformers for hyperspectral image classification. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition Workshops (CVPRW), Vancouver, BC, Canada, 17–24 June 2023; pp. 2166–2176. [Google Scholar]
  50. Li, Y.; Luo, Y.; Zhang, L.; Wang, Z.; Du, B. MambaHSI: Spatial-spectral mamba for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5524216. [Google Scholar] [CrossRef]
  51. Xu, H.; Chen, W.; Tan, C.; Ning, H.; Sun, H.; Xie, W. Orientational clustering learning for open-set hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5508605. [Google Scholar] [CrossRef]
  52. Di, X.; Xue, Z.; Zhang, M. Active learning-driven siamese network for hyperspectral image classification. Remote Sens. 2023, 15, 752. [Google Scholar] [CrossRef]
  53. Wu, H.; Prasad, S. Semi-supervised deep learning using pseudo labels for hyperspectral image classification. IEEE Trans. Image Process. 2017, 27, 1259–1270. [Google Scholar] [CrossRef] [PubMed]
  54. Zhong, C.; Zhang, J.; Wu, S.; Zhang, Y. Cross-scene deep transfer learning with spectral feature adaptation for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2861–2873. [Google Scholar] [CrossRef]
  55. Song, H.; Kim, M.; Park, D.; Shin, Y.; Lee, J.G. Learning from noisy labels with deep neural networks: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 8135–8153. [Google Scholar] [CrossRef] [PubMed]
  56. Shi, J.; Zhang, K.; Guo, C.; Yang, Y.; Xu, Y.; Wu, J. A survey of label-noise deep learning for medical image analysis. Med. Image Anal. 2024, 95, 103166. [Google Scholar] [CrossRef]
  57. Shin, J.; Won, J.; Lee, H.S.; Lee, J.W. A review on label cleaning techniques for learning with noisy labels. ICT Express 2024, in press. [Google Scholar] [CrossRef]
  58. Li, Y.; Yang, J.; Song, Y.; Cao, L.; Luo, J.; Li, L.J. Learning from noisy labels with distillation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1910–1918. [Google Scholar]
  59. Lukov, T.; Zhao, N.; Lee, G.H.; Lim, S.N. Teaching with soft label smoothing for mitigating noisy labels in facial expressions. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 648–665. [Google Scholar]
  60. Algan, G.; Ulusoy, I. MetaLabelNet: Learning to generate soft-labels from noisy-labels. IEEE Trans. Image Process. 2022, 31, 4352–4362. [Google Scholar] [CrossRef]
  61. Wu, Y.; Shu, J.; Xie, Q.; Zhao, Q.; Meng, D. Learning to purify noisy labels via meta soft label corrector. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 10388–10396. [Google Scholar]
  62. Tu, B.; Zhou, C.; Kuang, W.; Guo, L.; Ou, X. Hyperspectral imagery noisy label detection by spectral angle local outlier factor. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1417–1421. [Google Scholar] [CrossRef]
  63. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Li, S. Density peak-based noisy label detection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1573–1584. [Google Scholar] [CrossRef]
  64. Tu, B.; Zhou, C.; He, D.; Huang, S.; Plaza, A. Hyperspectral classification with noisy label detection via superpixel-to-pixel weighting distance. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4116–4131. [Google Scholar] [CrossRef]
  65. Kang, X.; Duan, P.; Xiang, X.; Li, S.; Benediktsson, J.A. Detection and correction of mislabeled training samples for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5673–5686. [Google Scholar] [CrossRef]
  66. Leng, Q.; Yang, H.; Jiang, J. Label noise cleansing with sparse graph for hyperspectral image classification. Remote Sens. 2019, 11, 1116. [Google Scholar] [CrossRef]
  67. Bahraini, T.; Azimpour, P.; Yazdi, H.S. Modified-mean-shift-based noisy label detection for hyperspectral image classification. Comput. Geosci. 2021, 155, 104843. [Google Scholar] [CrossRef]
  68. Kang, J.; Fernandez-Beltran, R.; Duan, P.; Kang, X.; Plaza, A.J. Robust normalized softmax loss for deep metric learning-based characterization of remote sensing images with label noise. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8798–8811. [Google Scholar] [CrossRef]
  69. Damodaran, B.B.; Flamary, R.; Seguy, V.; Courty, N. An entropic optimal transport loss for learning deep neural networks under label noise in remote sensing images. Comput. Vis. Image Underst. 2020, 191, 102863. [Google Scholar] [CrossRef]
  70. Zhang, Y.; Sun, J.; Shi, H.; Ge, Z.; Yu, Q.; Cao, G.; Li, X. Agreement and disagreement-based co-learning with dual network for hyperspectral image classification with noisy labels. Remote Sens. 2023, 15, 2543. [Google Scholar] [CrossRef]
  71. Liao, Q.; Zhao, L.; Luo, W.; Li, X.; Zhang, G. Joint negative–positive-learning based sample reweighting for hyperspectral image classification with label noise. Pattern Recognit. Lett. 2024, 183, 98–103. [Google Scholar] [CrossRef]
  72. Fang, C.; Zhang, G.; Li, J.; Li, X.; Chen, T.; Zhao, L. Intelligent agent for hyperspectral image classification with noisy labels: A deep reinforcement learning framework. Int. J. Remote Sens. 2024, 45, 2939–2964. [Google Scholar] [CrossRef]
  73. Yue, J.; Fang, L.; Rahmani, H.; Ghamisi, P. Self-supervised learning with adaptive distillation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5501813. [Google Scholar] [CrossRef]
  74. Zhao, W.; Peng, R.; Wang, Q.; Cheng, C.; Emery, W.J.; Zhang, L. Life-long learning with continual spectral-spatial feature distillation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5544214. [Google Scholar] [CrossRef]
  75. Chi, Q.; Lv, G.; Zhao, G.; Dong, X. A novel knowledge distillation method for self-supervised hyperspectral image classification. Remote Sens. 2022, 14, 4523. [Google Scholar] [CrossRef]
  76. Feng, S.; Zhang, H.; Xi, B.; Zhao, C.; Li, Y.; Chanussot, J. Cross-domain few-shot learning based on decoupled knowledge distillation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5534414. [Google Scholar] [CrossRef]
  77. Ullah, F.; Ullah, I.; Khan, R.U.; Khan, S.; Khan, K.; Pau, G. Conventional to deep ensemble methods for hyperspectral image classification: A comprehensive survey. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 3878–3916. [Google Scholar] [CrossRef]
  78. Lu, Y.; Bo, Y.; He, W. An ensemble model for combating label noise. In Proceedings of the 15th ACM International Conference on Web Search and Data Mining (WSDM), Tempe, AZ, USA, 21–25 February 2022; pp. 608–617. [Google Scholar]
  79. Shao, H.C.; Wang, H.C.; Su, W.T.; Lin, C.W. Ensemble learning with manifold-based data splitting for noisy label correction. IEEE Trans. Multimed. 2022, 24, 1127–1140. [Google Scholar] [CrossRef]
  80. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  81. Bylander, T. Estimating generalization error on two-class datasets using out-of-bag estimates. Mach. Learn. 2002, 48, 287–297. [Google Scholar] [CrossRef]
  82. Zhang, Z.; Sabuncu, M.R. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QC, Canada, 3–8 December 2018; pp. 8792–8802. [Google Scholar]
  83. Wang, Y.; Ma, X.; Chen, Z.; Luo, Y.; Yi, J.; Bailey, J. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 322–330. [Google Scholar]
  84. Thoreau, R.; Risser, L.; Achard, V.; Berthelot, B.; Briottet, X. Toulouse Hyperspectral data set: A benchmark data set to assess semi-supervised spectral representation learning and pixel-wise classification techniques. ISPRS J. Photogramm. Remote Sens. 2024, 212, 323–337. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed END method. First, T-based neural networks are trained on resampling datasets. Next, the estimated label distribution of the training data is computed by predicting out-of-bag (OOB) samples. Finally, the estimated distribution (ED) is combined with the ground-truth distribution (GD) to train a student network (S).
Figure 1. The framework of the proposed END method. First, T-based neural networks are trained on resampling datasets. Next, the estimated label distribution of the training data is computed by predicting out-of-bag (OOB) samples. Finally, the estimated distribution (ED) is combined with the ground-truth distribution (GD) to train a student network (S).
Remotesensing 16 04247 g001
Figure 2. False-color image and reference map of the SV. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Figure 2. False-color image and reference map of the SV. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Remotesensing 16 04247 g002
Figure 3. False-color image and reference map of the HOU. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Figure 3. False-color image and reference map of the HOU. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Remotesensing 16 04247 g003
Figure 4. False-color image and reference map of the PU. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Figure 4. False-color image and reference map of the PU. (a) False-color image created from a combination of three spectral bands. (b) Reference map detailing the ground truth classification.
Remotesensing 16 04247 g004
Figure 5. Classification maps generated for the SV image utilizing various comparison methods. (a) RF: OA = 86.66. (b) 2D-CNN: OA = 87.29%. (c) SLS: OA = 89.71%. (d) MSSG: OA = 92.53%. (e) MLN: OA = 92.98%. (f) DCRN: OA = 94.36%. (g) AAN: OA = 95.14%. (h) TCRL: OA = 95.78%. (i) END: OA = 97.30%.
Figure 5. Classification maps generated for the SV image utilizing various comparison methods. (a) RF: OA = 86.66. (b) 2D-CNN: OA = 87.29%. (c) SLS: OA = 89.71%. (d) MSSG: OA = 92.53%. (e) MLN: OA = 92.98%. (f) DCRN: OA = 94.36%. (g) AAN: OA = 95.14%. (h) TCRL: OA = 95.78%. (i) END: OA = 97.30%.
Remotesensing 16 04247 g005
Figure 6. Classification maps generated for the HOU image utilizing various comparison methods. (a) RF: OA = 77.61%. (b) 2D-CNN: OA = 79.36%. (c) MSSG: OA = 82.12%. (d) SLS: OA = 83.02%. (e) MLN: OA = 84.24%. (f) DCRN: OA = 84.47%. (g) AAN: OA = 87.79%. (h) TCRL: OA = 88.47%. (i) END: OA = 90.77%.
Figure 6. Classification maps generated for the HOU image utilizing various comparison methods. (a) RF: OA = 77.61%. (b) 2D-CNN: OA = 79.36%. (c) MSSG: OA = 82.12%. (d) SLS: OA = 83.02%. (e) MLN: OA = 84.24%. (f) DCRN: OA = 84.47%. (g) AAN: OA = 87.79%. (h) TCRL: OA = 88.47%. (i) END: OA = 90.77%.
Remotesensing 16 04247 g006
Figure 7. Classification maps generated for the PU image utilizing various comparison methods. (a) RF: OA = 80.26%. (b) 2D-CNN: OA = 84.21%. (c) SLS: OA = 88.16%. (d) MSSG: OA = 89.65%. (e) MLN: OA = 90.36%. (f) DCRN: OA = 94.26%. (g) AAN: OA = 95.07%. (h) TCRL: OA = 95.98%. (i) END: OA = 96.94%.
Figure 7. Classification maps generated for the PU image utilizing various comparison methods. (a) RF: OA = 80.26%. (b) 2D-CNN: OA = 84.21%. (c) SLS: OA = 88.16%. (d) MSSG: OA = 89.65%. (e) MLN: OA = 90.36%. (f) DCRN: OA = 94.26%. (g) AAN: OA = 95.07%. (h) TCRL: OA = 95.98%. (i) END: OA = 96.94%.
Remotesensing 16 04247 g007
Figure 8. OA curves of the END method with different versions of loss functions plotted against the number of training samples. (a) SV; (b) HOU; (c) HOU.
Figure 8. OA curves of the END method with different versions of loss functions plotted against the number of training samples. (a) SV; (b) HOU; (c) HOU.
Remotesensing 16 04247 g008
Figure 9. OA curves of SV, HOU, and PU datasets under different ensemble sizes.
Figure 9. OA curves of SV, HOU, and PU datasets under different ensemble sizes.
Remotesensing 16 04247 g009
Figure 10. Classification accuracy curves of SV, HOU, and PU datasets with the increase in training epochs.
Figure 10. Classification accuracy curves of SV, HOU, and PU datasets with the increase in training epochs.
Remotesensing 16 04247 g010
Table 1. Class names along with the number of labeled samples for three hyperspectral datasets.
Table 1. Class names along with the number of labeled samples for three hyperspectral datasets.
No SVHOUPU
Class NameSamplesClass NameSamplesClass NameSamples
1Brocoli_green_weeds_12009Healthy grass1251Asphalt6631
2Brocoli_green_weeds_23726Stressed grass1254Meadows18,649
3Fallow1976Synthetic grass697Gravel2099
4Fallow_rough_plow1394Trees1244Trees3064
5Fallow_smooth2678Soil1242Painted metal sheets1345
6Stubble3959Water325Bare Soil5029
7Celery3579Residential1268Bitumen1330
8Grapes_untrained11,271Commercial1244Self-Blocking Bricks3682
9Soil_vinyard_develop6203Road1252Shadows947
10Corn_senesced_green_weeds3278Highway1227
11Lettuce_romaine_4wk1068Railway1235
12Lettuce_romaine_5wk1927Parking Lot 11233
13Lettuce_romaine_6wk916Parking Lot 2469
14Lettuce_romaine_7wk1070Tennis Court428
15Vinyard_untrained7268Running Track660
16Vinyard_vertical_trellis1807
Table 2. Classification accuracy (%) achieved by comparison methods on the SV dataset at a noise level of 0.3.
Table 2. Classification accuracy (%) achieved by comparison methods on the SV dataset at a noise level of 0.3.
Class2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
197.8698.5999.4899.6799.4599.4899.2596.8799.75
299.7399.5999.8199.6999.7799.7399.8099.6899.80
398.2992.3599.4798.9396.2898.3299.4297.3197.52
499.6396.4396.9099.1998.9099.6299.5293.6098.92
597.7798.6794.9299.1799.0797.3997.9495.5499.14
697.0899.7799.0099.6599.5999.6699.7897.3399.72
799.0999.2199.4699.5899.5898.7299.2198.8499.66
861.0289.1689.0684.8686.8777.2980.7265.5293.27
998.3499.7299.6199.4699.3998.5298.8797.1899.72
1087.8790.1596.0193.2594.1089.0690.5285.1594.87
1193.8193.7395.9798.1095.8194.4397.1988.8198.13
1298.9899.5899.8999.9499.9497.3199.4698.0599.94
1399.5497.9598.7698.0998.5798.1198.8097.8399.30
1493.5097.2592.4193.6593.9194.4097.5391.5497.28
1580.1669.3972.3476.8777.0570.5881.7575.6995.56
1695.5695.1197.0097.2598.3096.8398.0791.3398.62
OA87.2992.5494.3395.1495.7789.7192.9886.6697.28
AA93.6494.7995.6396.1096.0494.3496.1191.8998.20
κ × 10085.9291.7393.7794.6695.3688.5592.2285.1996.97
Table 3. Classification accuracy (%) achieved by comparison methods on the HOU dataset at a noise level of 0.3.
Table 3. Classification accuracy (%) achieved by comparison methods on the HOU dataset at a noise level of 0.3.
Class2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
182.7994.0894.4885.9392.3390.3096.8884.3795.36
295.4497.8598.0996.5798.5697.7095.0898.0698.33
395.1399.9910010010095.3596.0897.78100
494.3497.4398.0799.2096.3093.6493.7888.2497.75
597.3799.4499.1999.2897.5896.0395.5495.1798.07
684.1198.1597.2395.3894.7793.8495.7495.0899.69
779.2989.9188.8088.0178.9479.1284.3075.7281.70
857.7962.0668.6574.2076.2173.9575.5254.8278.38
965.3671.7377.9681.0783.7179.4468.0561.9387.14
1074.5460.8865.1282.0782.3172.8072.4973.2787.45
1172.4467.4571.3472.6382.7574.1473.0765.9884.37
1259.4259.6966.1078.5174.8663.1473.4861.7682.40
1348.3859.2862.0591.4797.2350.4965.7942.8295.10
1497.5397.4399.3099.3099.7795.4498.4895.5999.77
1597.1698.9498.9498.9499.2495.5193.1190.6399.55
OA79.3682.1184.4887.7888.4783.0284.2477.6190.78
AA80.0783.6285.6989.5090.3083.3985.1678.7792.34
κ × 10077.6980.6583.2286.8087.5481.6482.9675.8090.04
Table 4. Classification accuracy (%) achieved by comparison methods on the PU dataset at a noise level of 0.3.
Table 4. Classification accuracy (%) achieved by comparison methods on the PU dataset at a noise level of 0.3.
Class2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
180.1790.9192.2092.8493.1786.8387.1080.5693.45
282.7887.5598.1798.7699.0088.1390.9475.5299.23
383.5490.7195.0095.0596.1489.9289.4880.1297.05
493.2897.6896.0895.9296.1596.5497.7695.0397.39
599.5999.4899.4199.5599.7098.0898.6699.0199.48
683.1289.2482.0082.6087.7986.5287.1079.0992.07
791.2495.7194.3695.7196.2491.2396.1583.5997.89
884.4080.6189.7993.6493.8978.2086.2181.0095.65
999.9510010010099.7999.7799.1999.78100
OA84.2289.6494.2895.0795.9588.1690.3680.2596.93
AA88.6892.4394.1194.9095.7690.5792.5185.9796.91
κ × 10079.1386.3092.3893.4394.6183.9986.3574.6795.92
Table 5. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the SV dataset.
Table 5. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the SV dataset.
rMetric2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
0.1OA88.5692.9794.6895.9496.3290.5493.7587.9397.94
AA94.6895.8495.9296.8896.9795.2396.6492.8698.41
κ × 10086.8992.4693.9694.8395.8289.1592.7886.6397.76
0.2OA87.8292.8394.5595.7196.1190.1293.2387.1297.66
AA94.1195.5595.7796.7396.8894.8496.3592.3598.32
κ × 10086.3392.2193.8894.7595.4788.8992.4585.7597.39
0.3OA87.2992.5494.3395.1495.7789.7192.9886.6697.28
AA93.6494.7995.6396.1096.0494.3496.1191.8998.20
κ × 10085.9291.7393.7794.6695.3688.5592.2285.1996.97
0.4OA86.5491.8393.8494.5894.9189.0592.2285.8796.83
AA93.1593.9595.0695.7595.8293.8595.7791.2497.25
κ × 10085.0690.9692.5692.5193.6488.0691.5584.6696.07
0.5OA85.6291.6393.1293.6493.8588.1291.1885.0896.38
AA92.8693.7694.2294.6694.8793.2095.1390.4696.88
κ × 10083.8290.8491.8991.8492.9787.5491.0483.3395.72
Table 6. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the HOU dataset.
Table 6. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the HOU dataset.
rMetric2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
0.1OA81.0383.6786.1488.4689.5784.1285.2678.9391.38
AA80.9684.7887.0490.3190.9484.2686.1579.6793.45
κ × 10078.6584.1585.6287.9688.9183.6484.4676.8691.02
0.2OA80.1582.9885.3688.1389.0683.6884.8778.3491.06
AA80.4584.1186.2990.0890.7683.8785.6479.2593.04
κ × 10078.1483.9784.6787.8488.4682.7883.7676.4290.85
0.3OA79.3682.1184.4887.7888.4783.0284.2477.6190.78
AA80.0783.6285.6989.5090.3083.3985.1678.7792.34
κ × 10077.6980.6583.2286.8087.5481.6482.9675.8090.04
0.4OA78.1281.2683.6887.1287.7581.9683.4276.4590.23
AA79.3382.6284.5788.9389.3482.4584.4678.0291.35
κ × 10076.8779.9582.7686.1886.8980.3582.0374.9689.58
0.5OA76.8480.5382.4786.2886.8280.8482.2674.8289.48
AA78.1281.7483.3887.8487.4181.6583.1076.9890.81
κ × 10075.7279.1481.9785.4985.6779.2680.8773.7188.64
Table 7. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the PU dataset.
Table 7. Classification accuracy (in %) obtained by comparison methods with different noise rates (r) when applied to the PU dataset.
rMetric2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
0.1OA85.2690.8294.9395.5896.4790.0491.4581.1597.26
AA89.2193.6694.7195.2996.1391.5893.7086.6297.18
κ × 10080.0488.0692.8993.9795.2885.1187.7275.2896.21
0.2OA84.8190.3594.7495.4296.3288.6590.9780.7697.11
AA88.9593.1294.5895.1895.9891.1593.2486.3397.06
κ × 10079.6587.2292.7693.8394.9784.6686.9475.0596.08
0.3OA84.2289.6494.2895.0795.9588.1690.3680.2596.93
AA88.6892.4394.1194.9095.7690.5792.5185.9796.91
κ × 10079.1386.3092.3893.4394.6183.9986.0574.6795.92
0.4OA83.4188.6793.4194.3695.1487.2589.5279.6596.52
AA88.1291.5393.3494.2194.8289.6291.6285.4296.57
κ × 10078.6185.5491.6692.7494.0283.1485.2373.9595.57
0.5OA82.2587.2692.3393.1794.0685.9388.2578.7696.02
AA87.2687.5592.4493.2893.9588.4590.4884.6995.97
κ × 10077.5784.4890.6391.2593.0781.7983.9673.1495.03
Table 8. Description of the END method and its streamlined versions.
Table 8. Description of the END method and its streamlined versions.
MethodEnsembleDistillationDescription
SingleTrain a single network model and use it for classification
BaggingTrain an ensemble network model and use it for classification
ENDThe trained ensemble network is used to guide the training of the target network
Table 9. Classification accuracy (OA, %) of the three methods on the SV, HOU, and PU datasets.
Table 9. Classification accuracy (OA, %) of the three methods on the SV, HOU, and PU datasets.
MethodBase NetworkSVHOUPU
Single2D-CNN87.2979.3684.22
SSRN89.7584.4787.61
Bagging2D-CNN90.3585.7489.74
SSRN92.4288.5992.58
END2D-CNN97.2890.7896.93
SSRN98.1692.3397.85
Table 10. Classification accuracy (%) achieved by comparison methods on the Toulouse dataset at a noise level of 0.3.
Table 10. Classification accuracy (%) achieved by comparison methods on the Toulouse dataset at a noise level of 0.3.
Method2D-CNNMSSG [10]DCRN [13]AAN [15]TCRL [16]SLS [59]MLN [60]RF [80]END
OA73.4876.6879.2581.4582.9577.8578.6471.3585.64
AA74.1177.3280.0682.6383.4777.9679.2271.9886.23
κ × 10071.3275.4378.2480.5581.6875.9877.1370.0683.82
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Ding, R.; Shi, H.; Liu, J.; Yu, Q.; Cao, G.; Li, X. Ensemble Network-Based Distillation for Hyperspectral Image Classification in the Presence of Label Noise. Remote Sens. 2024, 16, 4247. https://doi.org/10.3390/rs16224247

AMA Style

Zhang Y, Ding R, Shi H, Liu J, Yu Q, Cao G, Li X. Ensemble Network-Based Distillation for Hyperspectral Image Classification in the Presence of Label Noise. Remote Sensing. 2024; 16(22):4247. https://doi.org/10.3390/rs16224247

Chicago/Turabian Style

Zhang, Youqiang, Ruihui Ding, Hao Shi, Jiaxi Liu, Qiqiong Yu, Guo Cao, and Xuesong Li. 2024. "Ensemble Network-Based Distillation for Hyperspectral Image Classification in the Presence of Label Noise" Remote Sensing 16, no. 22: 4247. https://doi.org/10.3390/rs16224247

APA Style

Zhang, Y., Ding, R., Shi, H., Liu, J., Yu, Q., Cao, G., & Li, X. (2024). Ensemble Network-Based Distillation for Hyperspectral Image Classification in the Presence of Label Noise. Remote Sensing, 16(22), 4247. https://doi.org/10.3390/rs16224247

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop