Next Article in Journal
Simultaneous Recognition and Detection of Adenosine Phosphates by Machine Learning Analysis for Surface-Enhanced Raman Scattering Spectral Data
Next Article in Special Issue
Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination
Previous Article in Journal
Phonocardiogram (PCG) Murmur Detection Based on the Mean Teacher Method
Previous Article in Special Issue
Hybrid Sparse Transformer and Wavelet Fusion-Based Deep Unfolding Network for Hyperspectral Snapshot Compressive Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme

Electronic and Information Engineering, Shenyang Aerospace University, Shenyang 110136, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(20), 6647; https://doi.org/10.3390/s24206647
Submission received: 31 July 2024 / Revised: 10 October 2024 / Accepted: 14 October 2024 / Published: 15 October 2024

Abstract

:
The cost of hyperspectral image (HSI) classification primarily stems from the annotation of image pixels. In real-world classification scenarios, the measurement and annotation process is both time-consuming and labor-intensive. Therefore, reducing the number of labeled pixels while maintaining classification accuracy is a key research focus in HSI classification. This paper introduces a multi-strategy triple network classifier (MSTNC) to address the issue of limited labeled data in HSI classification by improving learning strategies. First, we use the contrast learning strategy to design a lightweight triple network classifier (TNC) with low sample dependence. Due to the construction of triple sample pairs, the number of labeled samples can be increased, which is beneficial for extracting intra-class and inter-class features of pixels. Second, an active learning strategy is used to label the most valuable pixels, improving the quality of the labeled data. To address the difficulty of sampling effectively under extremely limited labeling budgets, we propose a new feature-mixed active learning (FMAL) method to query valuable samples. Fine-tuning is then used to help the MSTNC learn a more comprehensive feature distribution, reducing the model’s dependence on accuracy when querying samples. Therefore, the sample quality is improved. Finally, we propose an innovative dual-threshold pseudo-active learning (DSPAL) strategy, filtering out pseudo-label samples with both high confidence and uncertainty. Extending the training set without increasing the labeling cost further improves the classification accuracy of the model. Extensive experiments are conducted on three benchmark HSI datasets. Across various labeling ratios, the MSTNC outperforms several state-of-the-art methods. In particular, under extreme small-sample conditions (five samples per class), the overall accuracy reaches 82.97% (IP), 87.94% (PU), and 86.57% (WHU).

1. Introduction

Hyperspectral images (HSIs) present rich and unique spatial and spectral features in a three-dimensional cubic data structure. As one of the key technologies for observation and exploration in the 21st century, HSI methods have been applied in several crucial fields, such as environmental monitoring [1], land cover evaluation [2], and ocean monitoring [3]. In addition, HSIs also play a significant role in the military field, such as camouflage recognition [4], battlefield space information acquisition [5], differentiation of target and decoy [6], detection of weapons of mass destruction [7], and monitoring of international treaty compliance [8]. HSIs can also play a significant role at different frequency ranges, like X-ray, ultraviolet, visible and near-infrared, and terahertz [9,10,11,12,13].
In these applications, HSI’s multiband spectral resolution provides valuable spectral data support for various industries, significantly improving the accuracy of analysis and decision making. However, the classification of HSIs remains a primary challenge, making HSI classification a continuous research hotspot within the academic community. Especially in the case of limited training samples, classification performance can easily decrease. Meanwhile, the original HSI is often affected by the spectral changes caused by sensor noise and environmental conditions [14,15]. Therefore, how to intelligently construct a streamlined training set with limited training data and effectively handle the large-scale or complex feature space constitutes a key problem.
Earlier studies focused on the spectral features of the pixels, covering support vector machines (SVMs) [16], K-nearest neighbors [17], polynomial logistic regression [18], sparsity algorithms of object recognition [19], etc. However, since ground targets are usually spatially continuous, there is strong spatial correlation between adjacent pixels in hyperspectral images. Fu et al. [20] generated joint spectral–spatial features, while Mu et al. [21] performed feature extraction in the spectral and spatial domains separately with two-branch networks. The spectral and spatial processing in these methods is conducted independently, ignoring the joint dependence of spectral and spatial information. Fu et al. [22] introduced tensor singular spectral analysis (TensorSSA) to extract global and low-rank features of HSI, while Dai et al. [23] proposed a smart-HOSVD method for 3D feature extraction in HSI. These methods have yielded promising results in classifying HSIs with small sample sizes. However, the above methods require manual selection and design of key spectral–spatial features to train the model.
In recent years, the development and application of deep learning have propelled the widespread use of convolutional neural networks (CNNs) in areas such as image classification and semantic segmentation [24,25,26]. Compared to traditional methods, deep learning methods exhibit strong advantages that do not require complex hand-crafted feature engineering or much prior knowledge. Hu et al. [27] were pioneers in using convolutional neural networks for HSI classification. One-dimensional CNNs (1D-CNNs) can perform convolution operations along the spectral dimension, effectively capturing the correlations and features between bands. However, since ground targets are usually spatially continuous, there is a strong spatial correlation between adjacent pixels in hyperspectral images. However, the method only considers spectral information while ignoring the potential role of spatial information.
As research in deep learning progressed, researchers designed and adopted 2D-CNNs [28], which typically excel at capturing spatial variations between pixels. Nevertheless, 2D-CNNs do not fully leverage the abundant spectral information available in HSIs. To overcome this limitation, G. Cheng [29] developed a robust spectral–spatial feature representation by fusing spectral features with deep spatial features. Still, 3D-CNNs are more favored as they utilize 3D convolutional kernels to jointly capture spectral and spatial information. This innovative and effective method promptly garnered interest. The HSI cube data extraction approach, implemented using a 3D neural network, does not necessitate any pre-processing to extract its inherent spectral–spatial characteristics. Paoletti et al. [30] presented two separate CNN architectures: one designed for extracting spatial characteristics and the other for extracting spectral characteristics. Zhong et al. [31] introduced a spectral–spatial residual network (SSRN) for learning spectral–spatial representations and employed supervised 3D deep learning for HSI classification. This approach effectively addresses the issue of accuracy deterioration by utilizing consecutive blocks of spectral and spatial residuals. These methods successfully extracted spatial and spectral characteristics and achieved advancements in HSI classification. However, there are still hurdles in terms of effectively utilizing the information. This is because the spatial resolution and spectral resolution in HSI data are inconsistent, leading to potential inconsistencies in the perception capabilities of the 3D convolutional kernel at different scales, which affects classification performance. To alleviate this issue, Roy et al. [32] introduced a hybrid spectral CNN model called HybridSN. The HybridSN model integrates the advantages of 3D-CNNs and 2D-CNNs, effectively learning spatial and spectral features in hyperspectral images. The 3D-CNN is responsible for the initial extraction of spatial–spectral joint features, while the 2D-CNN further abstracts high-level spatial features. In this way, the classification performance of the hybrid convolutional neural network model is enhanced, and compared to using a 3D-CNN alone, the network complexity is reduced, and classification accuracy is improved. Yu et al. [33] introduced a contrastive GCN (ConGcn) that improves contrast learning performance by incorporating a spectral space prior. This enhances the expressive power of the generated representation and thus improves the classification results. Xue et al. [34] introduced a network architecture for a spectral–spatial Siamese network (S3-Net) involving feeding features of different sizes separately into the network. They employed a weighted loss function to enable the model to learn more features when faced with limited samples. While the above-mentioned methods have achieved excellent results, the imbalance between positive and negative sample pairs can impact both the convergence speed and classification performance of the model. Subsequently, Xue et al. introduced a differentiated-scale restricted graph convolutional network (DSR-GCN) [35], building upon S3-Net, to address the issue of missing pixel features caused by sample imbalance. The accuracy of DSR-GCN was remarkable.
Although these methods have achieved great success in hyperspectral image classification, their accuracy heavily depends on a large number of labeled samples.
For HSI classification, the dataset consists of a single hyperspectral image. A portion of the pixels is used as the training set, while the remaining pixels (excluding background and irrelevant areas) serve as the test set.
Pixel-level labeling of HSI involves complex calibration calculations and substantial manual labor, potentially damaging the original features of land cover during extensive labeling. Labeling specific classes requires expert knowledge for accuracy. Current HSI classification algorithms heavily rely on labeled data, and obtaining sufficient labeled data is costly, significantly limiting practical applications. Thus, researching HSI classification with small samples will greatly reduce application costs, lay the foundation for expanding hyperspectral image technology applications, and have significant practical implications.
Liu et al. [36] proposed a deep convolutional neural network model for classifying hyperspectral images with few samples. The training strategy of the model is to gradually increase the number of samples in each class by 5 samples at a time, with the total number of samples ranging from 5 to 25. This approach aims to enhance the learning capability of the model with small batches of data to address the classification challenges in the case of a limited number of samples. Sun et al. [37] introduced an adversarial representation module to extract spectral and spatial features, replacing feature fusion approaches with class consistency. Zhang et al. [38] proposed graph information aggregation cross-domain few-shot learning (Gia-CFSL) to solve the problems of differences in spatial and spectral resolutions of different sensors and differences in the same land cover category. It performs FSL and domain alignment under the condition that all labeled samples are available in the source data and a few labeled samples are available in the target data. These methods consider the issue of HSI small-sample classification from different perspectives and propose various solutions to address the problem.
In the case of small-sample classification, contrastive learning [39] and active learning [40] are considered by scholars to be important methods for improving classification accuracy. Common contrastive learning methods include Siamese networks [41], triplet networks [42], SimCLR [43] (simple framework for contrastive learning of visual representations), and MoCo [44] (momentum contrast). Zhao et al. [45] developed a two-branch concatenated network with common parameters to acquire knowledge about the distinctions among various attributes. Cao et al. [46] introduced a contrast learning approach for a 3D convolutional Siamese network to address the challenge of HSI classification when there is a lack of appropriate data. By integrating contrast learning with conventional label-based supervised learning, the labels are effectively utilized in conjunction with the inherent information present in the data. Jia et al. [47] introduced a semi-supervised Siamese network called 3DAES. This network combines an autoencoder module with a Siamese network to analyze the information contained in a substantial volume of unlabeled data, rectifying it using a small collection of labeled samples. Corrections can effectively mitigate numerous issues arising from limited training data.
Active learning is a method that involves selecting the most valuable samples from an unlabeled dataset for labeling. These samples are typically the ones about which the model is most uncertain, most representative, or most important to the classification boundary. By labeling these samples, the benefit of each labeling effort is maximized, improving the model’s learning efficiency and classification accuracy [48,49,50]. Active learning algorithms commonly employed in remote sensing include disconnection [51], normalized entropy [52], uncertainty sampling [53], and marginal sampling [54]. In HSI classification, a significant number of labeled samples may be redundant and unnecessary. Through active learning, it is possible to reduce annotation costs [55]. Haut [56] designed a Bayesian convolutional neural network (B-CNN); this method performs active learning on the proposed B-CNN based on a three-step training phase, thereby addressing the difficulties caused by high-dimensional data and overfitting.
In small-sample HSI classification, the limited size of the training sample set often restricts the classification model due to insufficient information, which leads to inaccurate model classification results [57]. Model training errors subsequently influence sample query strategies, which results in the selection of suboptimal candidate samples based on misclassification information. Lei et al. [58] introduced a novel rank learning loss function into an active learning model, forming an uncertainty predictor, and achieved end-to-end uncertainty learning. Wang et al. [59] introduced a dual-branch domain adaptation few-shot learning (DBDAFSL) method to transfer knowledge obtained from a source domain to a completely different target domain efficiently. Wang et al. [60] proposed a collaborative active learning (CAL) scheme that takes into account the uncertainty and diversity of actively selected samples, as well as the cost of expert annotations.
While the aforementioned models have achieved positive results in the field of HSI classification, the crucial challenge remains in leveraging a limited amount of training sample data to extract deep feature information effectively and enhance classification accuracy. Our goal is to address the issue of decreased prediction accuracy in scenarios with few or extremely few samples. We recognize that models struggle to learn complete features from a limited number of training samples, leading to prediction biases. Therefore, on the one hand, we leverage contrastive learning and pseudo-active learning to fully explore the type features of both labeled and unlabeled samples. On the other hand, we focus on the quality of training samples by employing active learning methods to query representative samples, helping the model learn comprehensive features. We designed a novel multi-strategy triplet network classifier (MSTNC) in this study. This multi-strategy learning method optimizes multiple related tasks simultaneously, leveraging shared information between tasks to improve the model’s generalization ability and classification performance. Inspired by [42], we first designed a triplet network classifier as the backbone, which includes contrastive and classification modules, with the TNC used for model fine-tuning and classification.
Secondly, we consider ways to improve the classification accuracy of the model without increasing the cost of sample labeling from the perspective of the training strategy. To address the limitation that traditional active learning only considers the sample values on one side of the decision boundary, inspired by [61], a feature-mixture-based active learning (FMAL) method is proposed. This approach aims to explore more valuable samples, enabling the model to learn richer and more detailed features.
Finally, a dual-strategy pseudo-active learning method is proposed to screen out samples with high accuracy and diversity, which can provide more feature information for the model and effectively compensate for the lack of samples in the training set. Regarding the pseudo-active learning method, we provide a detailed introduction in Section 3.4.
Through the joint training of these strategies, the small-sample hyperspectral image classification method can fully utilize the potential of data and models, improve classification accuracy and robustness in situations with extremely limited samples, and overcome the limitations of traditional methods in small-sample environments.
The main contributions of this article are summarized as follows:
(1)
Addressing the challenge of HSI classification with small samples and multiple classes, we employ a composite three-input neural network structure to enhance the capability of capturing intra-class and inter-class features. This method expands the triple sample pair to optimize the model’s classification performance. The triplet network composed of the three-input neural network, projection head, and classifier is used as the backbone for feature extraction and classification tasks.
(2)
In this paper, a novel active learning method based on feature mixing is proposed. By blending certain features of labeled samples into unlabeled samples, the predicted changes in the mixed samples can reveal new features in the unlabeled samples. With FMAL, the dependence of sample selection on the quality of the initial model is reduced; thus, representative samples with high information richness are selected. Analytical experiments verify that our method is more effective than the classical methods.
(3)
We propose a dual-strategy pseudo-active learning method. Two filters were used to select valuable samples in unannotated samples as pseudo-samples. The two filters adopt different filtering strategies, and the aim is to identify samples with both certain confidence and certain new features. Pseudo-samples are added to the training set, enriching it to improve the accuracy of the model without increasing the labeling costs. The results show that compared with several existing state-of-the-art methods, the MSTNC strategy enhances the overall accuracy in the three generic datasets by 8.34–28.22%, 2.34–12.19%, and 7.07–23.66%, respectively, under the condition of limited training samples (five samples per class).
The remainder of this paper is organized as follows. Section 2 introduces the basic concepts of the triplet network and features mixture-based active learning as well as pseudo-active learning methods. A detailed description of the proposed MSTNC model and the sample construction process is provided in Section 3. In Section 4, a series of elaborately designed experiments validate the effectiveness and necessity of the strategies employed in the proposed model. Section 5 provides a discussion about the proposed method. Finally, Section 6 summarizes the conclusions and provides prospects for future research.

2. Related Work

2.1. Triplet Network

Triplet network (TN) models [42] are primarily applied in situations where there are numerous (or uncertain) sample classes; at the same time, the number of samples in the training dataset is limited (this is exactly the problem facing HSI classification). The TN comprises three identical feed forward networks (with shared parameters). The input consists of a triplet composed of three different samples: a reference sample ( x ), a sample from the same class ( x ), and a sample from a different class ( x + ). Their relationship is represented by the Euclidean distance. The training parameters bring x closer to x + and away from x , thus enabling the classification task. Triplet loss is expressed as follows:
L = m a x d x , x + d x , x + m a r g i n , 0 ,
where d(·) represents the Euclidean distance.
This similarity-based approach allows for a more detailed characterization of differences between different classes without increasing the number of labeled samples. Consequently, it facilitates better capture of both intra-class and inter-class features. Florian Schroff et al. [62] used the triplet network for the face recognition system. Chengliang Liu et al. [63] used the triplet network for tactile grasp outcomes prediction.
In this study, we applied the TN to the field of HSI classification and enriched the TN structure by integrating the TN and the classifier into a framework that mainly consists of an encoder, a classifier, and a projection head. The data dimension of hyperspectral images is extremely high and contains a lot of redundant information. The encoder maps high-dimensional data to a lower dimensional latent feature space through the multi-layer structure of deep neural networks. This can not only effectively reduce the dimensionality of the data but also improve the performance of the classifier by learning to capture high-order feature representations in spectral and spatial information. The main purpose of the projection head is to map feature vectors from a high-dimensional feature space to another target space (usually a low-dimensional space). In the process of contrastive learning, the projection head can project the feature vector into a new space to perform more effective feature comparison in the training phase to improve, in turn, the model’s ability to extract potential features. A small number of labeled samples can be used to create a large number of triple training samples, expanding the training set. Projection head contrastive training and classifier classification training are performed alternately so that the model can learn richer feature representations so as to maximize the classification accuracy of the model.

2.2. Active Learning

Active learning is a method that involves selecting the most valuable samples from an unlabeled dataset for labeling. These samples are typically the ones about which the model is most uncertain, most representative, or most important to the classification boundary. By labeling these samples, the benefit of each labeling effort is maximized, improving the model’s learning efficiency and classification accuracy [64,65,66]. Ajay J et al. [67] proposed an uncertainty measure that generalizes margin-based uncertainty to the multi-class case and is easy to compute. Because this approach relies on both the best guess and the second-best guess, it is referred to as the best and suboptimal (BvSB) method. If the probability difference between the best and second-best categories for a sample is small, it means that the classifier is highly uncertain about this sample, indicating that it has a higher value for labeling. This is a commonly used classic active learning method.
Cao [68] integrated the best-versus-second-best (BvSB) method with a CNN into a framework, leveraging the CNN’s powerful feature extraction ability and the annotation efficiency of active learning.
Despite the effectiveness of the above method, it is still difficult when applied to deep neural networks, high-dimensional data, and low-data states. It remains to be explored whether active learning methods based on posterior probability can identify truly valuable samples under the premise of minimum labeling cost.
Amin Parvaneh et al. [61] introduced an active learning method called ALFA-Mix, which constructs interpolations between the representations of labelled and unlabeled samples and then examines the predicted labels. This method identifies unlabeled samples with sufficiently distinct features by seeking inconsistencies in predictions resulting from interventions on their representations.
Specifically, the following conclusions and methods are presented:
  • The characteristics of the latent space play a crucial role in identifying the most valuable samples to be labeled.
  • The model’s incorrect predictions mainly stem from novel “features” in the input that are not recognizable.
  • Interpolation between the representations of unlabeled and labeled instances is adopted to achieve sampling of new instances, without explicitly modeling the joint probability of labeled and unlabeled instances [69,70,71,72].
  • The model predicts the loss of the pseudo-label of an unlabeled instance at its interpolation with a labelled one. By utilizing losses, it is possible to calculate which features are novel to the model.
In small-sample HSI classification, the limited size of the training sample set often restricts the classification model due to insufficient information, which leads to inaccurate model classification results [73]. Model training errors subsequently influence sample query strategies, resulting in the selection of suboptimal candidate samples based on misclassification information. To address the challenges of small-sample HSI classification, this paper draws on the theories and methods and introduces a feature-mixing-based active learning (FMAL) method. We refer to the feature vectors output by the encoding layer as encoding-layer features. Encoding-layer features play a crucial role in identifying valuable unlabeled samples. The changes in the prediction values of the model are observed by mixing the encoding-layer features of the labeled samples with the encoding-layer features of the unlabeled samples. These changes in predicted values can assist the model in discovering new features within unlabeled samples that are overlooked or unrecognized by the model. Subsequently, valuable samples can be selected based on these discoveries. The samples queried by the FMAL method contain substantial information and strong representation, effectively improving the model’s accuracy while working with a small-sample dataset.

2.3. Pseudo-Active Learning

The pseudo-label model [74], as a simple and effective semi-supervised learning method, has two core ideas:
  • Using the trained model to give unlabeled data a pseudo label. The method is very straightforward: use the training model to predict unlabeled data and use the category with the highest probability as the pseudo label for unlabeled data.
  • Applying entropy regularization to transform unsupervised data into regularization terms of the objective function (loss).
Haofeng Zhang et al. [75] used pseudo-labelling for image retrieval. Wenying Zhu et al. [76] used pseudo-labelling for bearing fault diagnosis.
HSIs contain a significant number of unannotated pixels that can be used as candidate pseudo-label samples. In this study, we combine the above ideas into model training and propose a pseudo-active learning strategy.
Distinguishing from AL with oracle labeling, we label unlabeled samples based on the model’s predicted values. The predicted values of unlabeled samples generated by the classifier are referred to as pseudo-labels. The samples labeled with the predicted values of the classifier are called pseudo-labeled samples or pseudo-samples.
Using pseudo-label samples may decrease the model’s performance. This is mainly due to the following two reasons:
  • Noise caused by incorrect labeling of pseudo-label samples;
  • The lack of new features between the pseudo-label samples and the training set samples, leading to model overfitting.
To reduce noise interference, it is necessary to select samples with high confidence. Meanwhile, in order to avoid overfitting, it is necessary to select samples with high uncertainty. In order to meet the needs of both aspects, a new pseudo-label filtering method, the dual strategy dual threshold filtering method, is proposed.
After screening, the pseudo-labeled samples that meet the criteria are merged into the training set for retraining. This approach effectively expands the training set and improves model classification accuracy without increasing the cost of sample labeling.

3. Materials and Methods

3.1. The MSTNC Framework

This study proposes a multi-strategy learning method based on triplet networks. As shown in Figure 1, given an HSI dataset denoted as H   ϵ   R H × W × B , where H, W, and B represent the height, width, and spectral dimensions of the image, respectively, we first applied traditional principal component analysis (PCA) [77] to the spectral bands of the initial HSI data to reduce the number of spectral bands, resulting in dimensionality-reduced HSI data with spectral dimension C, denoted as I   ϵ   R H × W × C . This process retains the critical spectral information for classification while reducing the computational load. The height and width of the original data remain unchanged, such that it preserves the spatial information, which is crucial for recognizing any object. The HSI data cube (I) is divided into small overlapping 3D patches O   ϵ   R S × S × C , where S denotes the height and width, and the truth-label of each small patch is determined by the label of the center pixel. These small 3D patches cover the S × S window spatial extent and all C spectral bands and constitute the original training and test sets. The TNC serves as the backbone and is trained with samples from the original training set to obtain TNC1. Subsequently, FMAL utilizes the encoding-layer feature vectors generated by the encoder part of TNC1 to query the most valuable unlabeled samples in the test set for oracle labeling. The labeled samples are removed from the test set and added to the training set, and the TNC is retrained to obtain TNC2. TNC2 is used to predict the remaining samples in the test set, and the predicted values serve as labels for the sample. The high-quality samples that meet the requirements are selected by the PAL strategy and added to the training set. Then, the latest training set is used to train the TNC, resulting in TNC3. Finally, the softmax calculation is performed on the classifier results to output the classification results.

3.2. Triplet Network Classifier

As shown in Figure 2, the TNC in this study comprises three main components: an encoder f e , projection head f p , and classifier f c . The encoder f e consists of three cascaded 3D convolutional layers, followed by a 2D convolutional layer, and finally linked to a global average pooling layer (GAP). Both f p and f c consist of two fully connected (FC) layers.
The contrastive learning module of the TNC network is formed by f p and f e . The classification learning module is formed by f c and f e . The f c is connected to a softmax layer to output the classification results.
The TNC will be trained as a backbone in two stages. The first stage is the contrastive learning phase. Specifically, in a dataset containing samples of C l s classes of land cover, initially, each class selects n labeled samples x l to form the initial training dataset D t r a i n = x i l , y i i = 0 n × C l s , where y i is the label of the i-th sample. Let N be the total number of samples and the remaining unlabeled samples x u constitute the test set D t e s t = x i u i = 0 N n × C l s . For a sample x l , it forms a triplet sample with other samples x l + and x l in the contrastive training set, where x l + shares the same class as x l (positive pair), and x l belongs to a different class (negative pair). In this way, the total n × C l s samples of D t r a i n can form n × ( n 1 ) × n × ( C l s 1 ) triple samples. These triplet samples constitute a contrastive training set, which greatly enriches the number of training samples. To ensure training efficiency, not all samples in the contrastive training set will be used for contrastive learning; instead, a randomly selected number of samples, as a multiple of n × C l s , will be utilized. In the first stage, f e is used to process x l , x l + , and x l . After processing by f e , z = f e ( x ) , three D-dimensional encoding-layer feature vectors, z , z + , and z , are generated. Encoding-layer feature vectors are crucial for improving model accuracy. These feature vectors are then input into f p to generate three P-dimensional feature vectors: v , v + , and v . The model is fine-tuned and updated through the triplet loss function, which is described as
L t r i p = m a x d p o s 2 d n e g 2 + m a r g i n , 0 ,
d p o s 2 = f p z f p z + 2 2 ,
d n e g 2 = f p z f p z 2 2 .
The ultimate optimization goal is to minimize the distance between positive pairs and maximize the distance between negative pairs. Let θ e , θ c , and θ f represent the parameters of the encoder f e , classifier fc, and projection head fp in the TNC, which can be updated using Equation (4). During the contrastive learning phase, the parameters θ c of the classifier f c are not updated.
θ e , p = argmin θ e , p L t r i p f p f e x l , f p f e x l + , f p f e x l ; θ e , p
The second stage is the classification learning phase. As illustrated in Figure 2, the input x is processed by the updated f e to generate a D-dimensional feature representation vector z . Subsequently, it undergoes processing by f c to produce a C l s -dimensional output vector. Training is guided by computing the cross-entropy loss:
L c l a s s = i = 1 C l s y i l o g y ^ i ,
θ e , c = argmax θ e , c L c l a s s f c f e x i , y i ; θ e , c ,
where y i represents the value of the i-th class in the ground truth labels (0 or 1) and y ^ i denotes the probability predicted by the model for the i-th class. The parameters θ e , c of the encoder f e and classifier f c can be updated using Equation (6). During the classification learning phase, the parameters θ p of the projection head f p are not updated.
After fine-tuning and updating the parameters of the backbone, the final classification results are obtained through the softmax classifier. Joint training of the TNC through the two stages enables the encoding-layer feature vectors to be more conducive to classifying different classes. Simultaneously, this joint learning strategy can enable the model to achieve better classification performance.

3.3. Feature-Mixture-Based Active Learning

Active learning can enhance the classification ability of a model by acquiring the most valuable samples with the minimum annotation cost. In the small-sample scenario, the initial performance of the general classifier is poor. The active learning method based on posterior probability cannot guarantee that the samples queried by it have the strongest uncertainty, and the effect of adding them to the training set to improve the accuracy of the model is not ideal. Therefore, we actively select labeled samples to add to D t r a i n in a gradually released fashion, rather than extracting them all at once. As the accuracy of the model predictions continues to improve, the quality of the queried samples will also gradually improve.
Specifically, given the total number of samples M that can be chosen. S is the set of the number of target samples to be identified in each iteration. S = s 0 , s 1 , s 2 , s i , where s i = m i n { M k = 0 i 1 s k , k = 0 i 1 s k } , s 0 represents the number of samples in the initial training set, and s i denotes the number of samples selected in the i-th iteration. The iteration concludes when the number of samples in the training set reaches M . s 0 samples are initially randomly selected as the training set (and the rest as the test set) to train the model TNC. Then, the active learning method is used to select s 1 samples in the test set; these samples are removed from the test set and then labeled and added to the training set. Subsequently, the model is retrained using the updated training set. Then, the active learning method is used to select s 2 samples in the test set, and these samples are removed from the test set, labeled, and added to the training set. The model is then retrained using the updated training set. The above process is repeated until the number of training set samples reaches M. In this way, the M most valuable samples can be selected. The accuracy of the model trained with the training set composed of these samples is higher than that of the training set composed of randomly selected samples.
The iteration process is illustrated in Figure 3.
In the following, the sample selection method in each iteration is elaborated in detail. We refer to the feature vectors output by the encoding layer as encoding-layer features. Encoding-layer features play a crucial role in identifying valuable unlabeled samples. This is because the encoding layer, typically positioned in the middle of deep learning models, extracts high-level features from the input data. Compared to raw data or primary features, these high-level features can better capture the essence and complex patterns of the data. Therefore, encoding-layer feature vectors can provide richer and more meaningful representations than the original spectral data. Secondly, raw hyperspectral data typically have very high dimensions and may contain a lot of redundant information. The encoding layer, through the training process of the neural network, can effectively reduce the dimensionality, remove redundant information, and retain only the most important features that represent class differences. These main features after dimensionality reduction are more concise and clearer, highlighting the essential differences between different classes. Last, but not least, encoding-layer feature vectors can be used to calculate the similarity and differences between samples, as well as the model’s uncertainty in predicting a particular sample.
We draw on [50] for the feature mixing method by mixing the encoding-layer features of labeled samples with those of unlabeled samples to observe changes in the prediction values of the unlabeled samples.
If the model’s predictions remain unchanged, it indicates that the proportion of uncertain features in the sample is relatively low, which suggests that there are fewer novel features in the sample. Conversely, if the predictions change, it suggests that the sample contains a sufficient number of novel features, which meet the criteria for the samples we are seeking.
Specifically, we calculate the average value of encoding-layer features of each class of labeled samples in D t r a i n as the representative value for each class of labeled samples, which is referred to as the anchor value and denoted as z ¯ . z u represents the encoding-layer feature representation of the samples in D t e s t .
We mix each z ¯ of different classes with each z u in a certain proportion:
z ~ i = α i z ¯ i + 1 α i z u ,   α i 0 , 1 ) D ,   α i 2 ε ,   a r g m a x f c z u = k , i = 1,2 , k 1 , k + 1 , c l s ,
where z ¯ i represents the anchor value of the i-th class of samples and α i acts as a weight determining the proportion of features that are chosen to be mixed. α i is specific to z ¯ i and z u .
For each sample x u in D t e s t , we compute its encoding-layer feature z u (assuming its model prediction is k, i.e., a r g m a x f c z u = k ). Then, we mix z u with all anchor values except for class k using certain proportions ( α i ). Let z ~ i denote the mixed feature value. If the model predictions for all z ~ i remain unchanged and still belong to class k, this indicates high consistency between this sample and the k-class samples in the training set. However, if the prediction of any z ~ i changes, i.e., i ( a r g m a x f c z ~ i k ), this suggests that this sample may contain a sufficient number of new features. This is the sample with training value that we need to retrieve.
ε is the threshold controlling the mixing proportion, and it determines the uncertainty of the selected sample. If ε is smaller, then the selected samples have stronger uncertainty, which means a greater degree of novel features in the samples.
The selection of α i is specific to the input and determines the features to be chosen. Given the ε threshold, α i should be chosen to induce a change in the prediction of the model for z ~ . We study the impact of feature mixing on the prediction of z u by examining the changes in the loss function.
Using a first-order Taylor expansion with respect to z u , the loss of the model for predicting the class of z ~ i as k can be rewritten as
L c l s f c z ~ i ,   y k L c l s f c z u ,   y k + z ~ i z u T · g r a d z u L c l s f c z u , y k .
From Equations (7) and (8), we obtain
L c l s f c z ~ i , y k L c l s f c z u , y k ( α i z ¯ i z u ) T · g r a d z u L c l s f c z u ,   y k .
The choice of α i determines the features to be selected with the aim of maximizing the change in loss:
α i * = argmax α i 2 ε α i ( z ¯ i z u ) T · g r a d z u L c l s f c z u , y k .
The optimal value α i * is determined by satisfying the threshold constraint, which maximizes the change in the loss function. α i * is calculated as follows:
α i ( z ¯ i z u ) 2 α i 2 · z ¯ i z u 2 .
Substitute the constraint α i 2 ε .
α i ( z ¯ i z u ) 2 ε · z ¯ i z u 2   Let   α i = α i ( z ¯ i z u ) ,   ε = ε · z ¯ i z u 2 ,   and   g r a d = g r a d z u L c l s f c z u ,   y k .
Equation (10) can be rewritten as
( α i ) * = argmax α i 2 ε α i T · g r a d .
This is solved using the dual norm formula:
( α i ) * = ε ·   g r a d g r a d 2 .
Substitute α i , ε :
α i * = ε · z ¯ i z u 2 g r a d 2 · g r a d z ¯ i z u .
The size of ε determines the quantity of samples identified, and the required number of samples is found step by step by means of fine-tuning ε : = ε + ε . Our experiments demonstrate that using this method effectively controls the number of samples identified while maintaining low computational complexity.

3.4. Dual-Strategy Pseudo-Active Learning

After the training with the FMAL strategy, the model acquires a certain level of classification ability. At this point, a significant portion of the model’s predicted values for the samples in the test set are accurate, which means that the number of correctly predicted samples is much larger than the number of samples in the training set. Moreover, the correctly predicted samples often include new features distinct from those found in the training set.
The above analysis inspires us to consider the possibility of using a strategy to identify a certain proportion of high confidence samples and then using a strategy to identify semantically rich samples within these samples. The predicted values of the model are used as labels to label these samples, and then these samples are expanded to the training set. We refer to this novel method as dual-strategy pseudo-active learning (DSPAL). Unlike AL with oracle labeling, we label unlabeled samples using the model’s predicted values.
The goal of DSPAL is to expand the training set using pseudo-labeled samples and then retrain the model to further improve the model’s accuracy. Obviously, the quality of pseudo-label samples has a significant impact on the performance of the model, so it is necessary to design a high-quality pseudo-label sample selection strategy.
To enable the model to learn reliable representations of the data during the DSPAL phase, an iterative training approach is adopted. At each iteration, the filter adds filter-qualified pseudo-labeled samples to D t r a i n and then retrains the model with an updated D t r a i n . The DSPAL training process is illustrated in Figure 4.
The filter sets two thresholds, P 1 and P 2 , which represent the P 1 % of high-confidence samples and P 2 % of valuable samples selected for each class of unlabeled samples, respectively.
When selecting high confidence samples, we choose to use the BvSB [65] measure. The BvSB measure is specifically designed for multi-class classification problems. Its core idea is to determine the confidence level of the predicted value by comparing the difference in prediction probabilities between the best and second-best classes of the model. The greater the difference, the higher the confidence level; conversely, the smaller the difference, the lower the confidence level. The B v S B values for all samples in D t e s t are computed.
B v S B = P B x u P S B x u
Here, P B x u and P S B x u represent the conditional probabilities of sample x u belonging to the optimal and suboptimal classes, respectively.
A greater B v S B value indicates greater confidence in the f c classification results, which suggests more accurate predictions. Conversely, a smaller B v S B value signifies greater sample uncertainty, making it more likely to contain new features.
Q i = s o r t B v S B i x u
B v S B i denotes the B v S B value of the i-th class sample, and Q i denotes the queue with the B v S B values of the i-th class samples in descending order. Among them, the samples whose position is less than P 1 % in the queue are regarded as reliable samples.
We chose the FMAL’s feature mixing method as the filter’s strategy for selecting samples with new characteristics. Among the high-confidence samples selected above, the FMAL screening method is used to select the P 2 % of samples with novel features in each class. It should be noted that the anchor values used here are calculated from the updated training set.
The filter uses these two strategies to filter out samples from D t e s t that meet these criteria, yielding candidate samples with high reliability and strong uncertainty. These samples are then added to D t r a i n and used to train the TNC with the updated training set.
The above process is repeated to improve the model accuracy by continuously expanding the number of samples in the training set.
The complete procedures of our MSTNC are summarized in Algorithm 1.
Algorithm 1. MSTNC model.
Input: HSI patch x ∈  R S × S × C , The total number of samples is denoted as M.
Initialization: train-set:  D t r a i n , test-set:  D t e s t
1. Obtain: Triplet samples ( x i l , x i l + , x i l )
2. Train the TNC with D t r a i n , Obtain the model TNC1
3. while: D t r a i n Number of samples < M:
4.    Compute the anchor values z ¯ for D t r a i n
5.    Encoding-layer features mixing
z ~ i = α i z ¯ i + ( 1 α i ) z u
6.  Put all x u corresponding to z u that meet the condition a r g m a x f c ( z ~ ) a r g m a x f c ( z u ) into D t r a i n , update D t r a i n , D t e s t
7.    Train TNC with the updated D t r a i n to obtain the model TNC2
8. while: iterations < 4
9.    Calculate the BvSB values f D t e s t
10.  Sorting the BvSB values for each class of samples. Q i = s o r t ( B v S B i ( x u ) )
11.  Put the samples in Q between 0 and P 1 into candidate set D c a n d
12.  while: D t r a i n Number of samples < P 2 :
13.       Compute the anchor values z ¯ for D t r a i n
14.       Encoding-layer features mixing
z ~ i = α i z ¯ i + ( 1 α i ) z c a n d
15.          Put all x c a n d corresponding to z c a n d that meet the condition a r g m a x f c z ~ a r g m a x f c z u into D t r a i n , update D t r a i n
16.     Train TNC with the updated D t r a i n and obtain the model TNC3
17. Output: Predict test sample with the model TNC3

4. Results

4.1. Hyperspectral Datasets

(1) The Indian Pines (IP) dataset was acquired using an airborne visible and infrared imaging spectrometer (AVI-331 RIS). The image size is 145 × 145 pixels, with a wavelength range of 400–2000 nm, a spatial resolution of 20 m, and 200 spectral bands after excluding noisy bands. The dataset contains a total of 10,249 labeled pixels, which span 16 classes, including crops, wood, and other perennial plants. The false-color image and ground-truth map are illustrated in Figure 5.
(2) The Pavia University (PU) dataset was captured over the city of Pavia, Italy, using an airborne hyperspectral imaging spectrometer from Germany. The image has dimensions of 610 × 340 pixels, covers a wavelength range of 430–860 nm, has a spatial resolution of 1.3 m, and consists of 103 spectral bands.
The total number of labeled pixels in this dataset is 42,776, and there are nine classes of labeled samples, most of which are urban land cover, such as metal plates, roofs, and concrete pavements. The false-color image and ground-truth map are illustrated in Figure 6.
(3) The WHU-Hi-HanChuan (WHU) dataset covers the agricultural area of Hanchuan City, Hubei Province, China. The images was acquired by a Headwall nano-hyperspectral imaging sensor mounted on a UAV platform. The scene size is 1217 × 303 pixels with a spatial resolution of 0.109 m. The images were acquired by the Headwall Nano Hyperspectral Imaging Sensor (HNHIS). It consists of 274 bands with wavelengths ranging from 0.4 to 1.0 μm. The ground truth data consisted of 253,580 labeled pixels from 16 different crop type classes. The distribution of the false-color image and ground-truth map are illustrated in Figure 7.

4.2. Experimental Settings

All experiments in this paper are conducted on a server equipped with an Intel(R) Core (TM) i7-12700 CPU @ 1.7 GHz 2.19 GHz, 64 GB DDR5 RAM, and an RTX 4090-24 GB GPU. The specific programs were implemented on the Ubuntu operating system using PyTorch 2.0.0 deep learning framework and Python 3.8 compiler. Evaluation metrics include producer accuracy (PA), overall accuracy (OA), average accuracy (AA), and the Kappa statistic (Kappa). Higher values of these metrics indicate that the model is more capable of classifying. Moreover, train time (s) and test time (s) were used for complexity comparison. During the experiments, a batch size of 64 is set, the initial learning rate is set to 0.0001, and the Adam optimizer is employed to facilitate rapid convergence of the model.

4.3. Comparison of Classification Results

To verify the effectiveness of the MSTNC, we studied methods outlined in the literature and designed the following sets of comparative experiments. The choice of comparison methods takes into account its properties and the setup of the original experimental sample.
To ensure a fair comparison, we reproduced all of these methods, and all hyperparameters were also based on publicly available original code or papers. All experiments were conducted using the same hardware, and the ratio of training to test samples was the same. Each datum represents the average of 10 repetitions of the experiment. Table 1, Table 2 and Table 3 give per-class quantitative results for the three datasets.
  • Performance with very few training samples
To evaluate the superiority of the proposed method in a few labeled sample scenarios, we selected five samples for each class in the dataset to form the training set. In this way, the proportion of samples in the training set in the three datasets was 0.78% (IP), 0.1% (PU), and 0.03% (WHU). The methods 3D-CNN [20], DFSL-NN [36], DFSL-SVM [36], graph information aggregation cross-domain few-shot learning (Gia-CFSL) [38], dual-branch domain adaptation few-shot learning (DBDAFSL) [59], and CapsGLOM [60] were adopted for comparison with our proposed MSTNC. These HSI classification methods are representative state-of-the-art deep learning methods specifically designed for small sample classification.
The first five methods randomly select five samples per class for experimentation. CapsGLOM uses its original sample selection method; i.e., one sample per class is randomly selected and iterates four times according to the AL method it provided, resulting in the final selection of five samples per class. The MSTNC randomly selects three samples per class and then uses the AL method proposed in this study to select two more samples per class.
Table 1, Table 2 and Table 3 give per-class quantitative results for the three datasets.
From Table 1, Table 2 and Table 3, it can be observed that our proposed MSTNC consistently outperforms other methods in terms of OA, AA, and Kappa on the three datasets. OA is higher by 8.34–28.22%, 2.34–12.19%, and 7.07–23.66%, respectively, AA is higher by 7.31–25.06%, 1.95–16.36%, and 3.58–37.47%, while Kappa is higher by 10.10–28.49%, 3.35–16.64%, and 8.14–29.06%, respectively. Examining the classification accuracy for each class, it can be noted that the IP dataset exhibits its lowest accuracy for class 5 at 69.21%, which surpasses the lowest accuracy of other methods by 24.43–48.98%. For the PU dataset, the lowest accuracy is in class 4 at 71.06%, which surpasses the lowest accuracy of other methods by 16.65–51.6%. In the case of the WHU dataset, the lowest accuracy is in class 12 at 51.94%, which surpasses the lowest accuracy of other methods by 8.51–51.90%.
Figure 8, Figure 9 and Figure 10 show the classification maps of all the methods on the IP, PU, and WHU datasets. The maps demonstrate that our proposed model exhibits superior classification performance and clearer boundaries. For example, in Figure 8 (IP dataset), it can be observed that other models incorrectly label some pixels of class 2 (Corn—no-till) as class 11 (Soybean—no-till) or class 14 (Soybean—min-till); in Figure 9 (PU dataset), other models exhibit considerable confusion in class 8 (self-blocking bricks); and in Figure 10 (WHU dataset), it is apparent that other models incorrectly classify class 8 (trees) and class 10 (red roof) as other classes. However, the MSTNC achieves considerably high classification accuracy for these classes.
Conclusively, the proposed method shows advantages in improving the classification performance of all datasets. Compared to other methods, the proposed model demonstrates superior classification of background pixels, which affirms the effectiveness of this approach.
B.
Performance with limited training samples
To make a broader comparison with the representative and state-of-the-art classification methods under multiple technical routes, we performed the following experiments:
The training set represented 2% (IP), 1% (PU), and 0.5% (WHU) of the categories in each sample, respectively.
The spectral–spatial residual network (SSRN) [31], semi-supervised Siamese network (3DAES) [47], contrastive GCN (ConGcn) [33], TensorSSA [22], and SmaT-HOSVD [23] models were adopted for comparison with our proposed MSTNC.
The sample size selected for this experiment is slightly larger than that of the previous group. Especially for the PU dataset and WHU dataset, as these two datasets are relatively large, the absolute sample size of the training set is also large. Therefore, the accuracy is generally higher than the previous group.
Table 4 provides quantitative results for the three datasets.
As can be observed from Table 4, our proposed MSTNC method has significant advantages over other methods in terms of OA, AA, and Kappa coefficients on all three datasets. On the Indian Pines (IP) dataset, the MSTNC outperforms OA by 3.91–20.2%, AA by 4.4–14.77%, and Kappa coefficient by 2.84–23.36%. For the Pavia University (PU) dataset, the MSTNC achieved 100% for OA, AA, and Kappa, which were 3.12–8.73%, 5.55–11.5%, and 2.12–12% higher, respectively, than the other method. On the WHU dataset, the MSTNC also achieved 100% for OA, AA, and Kappa, improving by 1.99–10.09%, 1.33–8.79%, and 2.05–7.41%, respectively, compared to the other method. By specifically analyzing each dataset, it can be seen that the MSTNC method consistently outperforms other methods in all the metrics, especially in the PU and WHU datasets, where the MSTNC achieves 100% classification accuracy and Kappa coefficient, which further proves the excellent performance of the method in the hyperspectral image classification task.
C.
Performance with Different Training Percentages
To validate the generalization performance of the proposed MSTNC, we compared the classification results of different methods with different numbers of training samples. Here, we selected the seven most representative deep learning methods in experiments A and B. Considering the great disparity in the number of samples in different datasets, different percentages of training samples were selected. Experiments were conducted with selected samples of 1%, 2%, 5%, 10%, and 15% of the IP dataset; 0.1%, 0.2%, 0.5%, 1%, and 5% of the PU dataset; and 0.05%, 0.1%, 0.25%, 0.5, and 2.5% of the WHU dataset as the training set, with the rest of the samples as the test set. The experimental results are shown in Figure 11, Figure 12 and Figure 13. Our proposed MSTNC consistently achieves the highest accuracy across different sample quantities on the three datasets, which demonstrates its superiority in terms of generalization performance.
From Figure 11, Figure 12 and Figure 13, it can be observed that on the IP dataset, when the proportion of samples is 5%, the MSTNC method achieves more than 99% for all three metrics. It performs 5.11–9.52%, 5.95–16.68%, and 4.70–14.02% better than other methods. Other methods require a sample proportion of 15% to achieve a similar level of accuracy.
On the PU dataset, when the proportion of samples is 0.5%, the MSTNC method achieves over 98% for all three metrics. It performs 3.86–15.68%, 7.01–30.43%, and 5.07–23.21% better than other methods. Other methods require a sample proportion of 5% to achieve a similar level of accuracy.
On the WHU dataset, when the proportion of samples is 0.25%, the MSTNC method achieves over 98% for all three metrics. It performs 3.01–7.86%, 2.28–11.09%, and 3.11–15.15% better than other methods. Other methods require a sample proportion of 2.5% to achieve a similar level of accuracy.
D.
Complexity Analysis
In order to control the complexity of the model, we chose a lightweight model structure suitable for small samples to reduce the complexity of the model. GAP is used to replace the fully connected layer to reduce the number of parameters and the amount of computation. At the same time, we adopt the PCA dimensionality reduction technique to further reduce the dimension of the input data to reduce the computational overhead.
To balance model complexity and computational efficiency, we fine-tuned the model structure and hyperparameters through multiple experiments. This ensured that we could simplify the model as much as possible without compromising classification accuracy significantly. Taking the IP dataset as an example, the OA of the model is 82.97%. In the case of, we reduce the inference time of the model to 1.32 s by reducing the number of layers and parameters. Thus, a balance between accuracy and efficiency is achieved.
The training time and test time of different models on different datasets are reported in Figure 14.
From the figure, it is evident that TensorSSA and SmaT-HOSVD boast the shortest training time, significantly shorter than other deep-learning-based methods, thus exhibiting considerable advantages. However, during the testing phase, the disparity narrows considerably, with all methods falling below 12 s. Among the deep learning methods, CapsGLOM and MSTNC necessitate active learning techniques for sample querying, resulting in longer training durations. Nevertheless, their testing times are moderate. Notably, the MSTNC excels in testing efficiency by utilizing learning strategies to enhance model efficiency, coupled with a lightweight classifier.
Unlike ordinary images, a hyperspectral image of a scene often corresponds to a very wide area of space in the actual range. The pixel-level labeling of hyperspectral images involves many complicated works. Therefore, controlling the number of samples and improving the model accuracy is the primary goal of HSI classification. The MSTNC achieves high accuracy in extremely small sample scenarios with moderate time complexity.

4.4. Ablation Analysis

In order to eliminate the effect of different training samples on each ablation method, we unified the training samples using fixed random seeds.

4.4.1. Effectiveness of Triplet Contrastive Learning

To validate the effectiveness of triplet contrastive learning, experiments were conducted to compare the complete TNC with the TNC containing only f e and f c . In these experiments, three labeled samples per class were randomly selected for training. The results of the evaluation metrics are shown in Figure 15. On the three datasets, OA improved by 3.24%, 4.21%, and 3.86%, respectively. AA improved by 3.16%, 9.85%, and 2.44%, and Kappa improved by 4.20%, 4.50%, and 2.31%. Thus, the expansion of the triplet training set effectively extracts richer features and information by reducing the intra-class distance and increasing inter-class distance. This approach significantly enhances the model’s classification. The reason why this method is effective is that it uses triple sample pairs in training. A small number of training samples can form enough triple sample pairs, greatly expanding the training set.

4.4.2. Effectiveness of AL and PAL Strategies

To validate the effectiveness of AL and PAL strategies, three sets of experiments were conducted on the IP dataset.
In the first set, five labeled samples were randomly selected from each class of the IP dataset. In the second set, three samples were first randomly selected from each class, and then two samples were selected by our AL method to form an initial training set for further training. The experimental results in Figure 16 reveal that compared to training with randomly selected samples, the AL method resulted in improvements of 5.59%, 1.10%, and 5.48% in OA, AA, and Kappa, respectively. This proves that the proposed AL strategy is effective, indicating that its selected samples are more efficient. In the third set, the PAL strategy was additionally applied to the training process based on the second set. As shown in Figure 16, compared with AL alone, the OA, AA, and Kappa values increased by 4.26%, 2.95%, and 4.74%, respectively, after PAL application. The results highlight significant improvement in the predictive accuracy with PAL, showcasing its ability to enhance classification accuracy without increasing the cost of labeled samples. This underscores the superiority and feasibility of the proposed method, which confirms that the accuracy and semantic richness of pseudo-labeled samples align with our theoretical expectations. It also demonstrates that the joint application of these two strategies effectively enhances the feature representation of the network, detail capture, and discriminative abilities, particularly in scenarios with extremely limited initial samples.

5. Discussion

5.1. Parameter Analysis

Deep learning methods can be considered as end-to-end learning methods. End-to-end learning can automatically extract features and perform classification through models such as CNN. This process does not rely on artificial feature selection. This enables the model to adapt to the influence of noise and improves the generalization ability and robustness of the model.
Hyperspectral images are characterized by high dimensionality and severe redundancy. With dimension-reduction processing, important spectral information can be effectively retained, while reducing data redundancy and computational complexity. PCA is one of the most commonly used dimensionality reduction methods. Because it does not rely on sample labels, it not only effectively reduces the dimensionality of spectral bands but also maintains the integrity of spatial information. Therefore, this paper adopted PCA for dimensionality-reduction processing. The high dimensionality of spectral bands is beneficial for information preservation, but it increases redundancy, noise, and the complexity of model training. On the contrary, the dimensionality of spectral bands is low, which may lead to information loss but can help remove redundancy and noise.
After dimensionality reduction, the data are normalized along the spectral dimension to ensure that the mean is 0 and the unit variance is used as the standard. This standardization helps improve the performance and stability of deep learning models. HSI classification algorithms are often trained and classified based on hyperspectral image patches. Hyperspectral image patches are used as input samples and cropped around the central pixel. The selection of patch size primarily considers the extraction of spatial features and computational costs. Increasing the patch size provides a larger perceptual field, which is beneficial for the model. At the same time, it also brings more noise and chaos, which is harmful to the model.
For the MSTNC model, a patch size of 25 × 25 with 30 bands is used for the IP and WHU datasets. On the PU dataset, the patch size is set to 19 × 19, with 15 bands. The chosen patch values and band values are based on reference [47,59,60,78] and extensive experiments [79], and this article continues to build upon the conclusions drawn from previous work.
In the FMAL method, the threshold ( ε , ε ) affects the execution efficiency of FMAL. ε reflects the quality of the selected samples. The larger the value of ε, the fewer new features the selected sample contains, and vice versa. Therefore, (ε, ∆ε) determines the number of samples identified at each time.
In the DSPAL method, P 1 and P 2 are the thresholds of the confidence filter and uncertainty filter, respectively. The larger the P 1 , the lower the confidence of the selected sample and vice versa. The larger the P 2 , the higher the uncertainty of the selected sample, and vice versa.
For the MSTNC model, ε = 0.006, ε = 0.0005, P 1 = 40, and P 2 = 39.
The selection of the above four parameters in this article is mainly based on the analysis of experimental data. More scientific methods need to be further explored.

5.2. Analysis of Various MSTNC Strategies

We conducted extensive experiments to thoroughly validate the effectiveness of our proposed method.

5.2.1. Comparison of Three Different Sample Selection Methods

To validate the superiority of the FMAL method, we compared it with the classical BVSB method and random sample selection method. On the IP dataset, we initially randomly selected 3 samples from each class and then used these 4 methods to choose 3, 5, 7, 9, 11, and 13 samples for training in each class. The experimental results are shown in Figure 17.
The data clearly show that the FMAL method proposed in this paper has the best performance. Specifically, when three labeled samples were selected for each class, the OA, AA, and Kappa of the FMAL method were 2.36%, 3.15%, and 2.47% higher than those of the BvSB method, and 7.09%, 2.28%, and 8.23% higher than those of the random method.
The BvSB method’s OA and Kappa are also superior to the random method, as random sampling may lead to an imbalanced distribution of training samples among different classes, which results in insufficiently representative features for the selected samples. In Figure 17b, the AA values of the BvSB method and the random method repeatedly cross. This may be because the classification accuracy of some classes is low when using the initial model for classification, and the BvSB method heavily relies on the classification results of the initial model, which leads to lower-quality samples for these specific classes. In contrast, FMAL maintains good performance, which indicates that the method of mixing partially labeled sample characteristics reduces the dependence of sample selection on model accuracy.
Figure 18 shows the classification maps when seven samples per class were selected using three different sample selection methods, visualizing the classification effect of the different methods. It is evident that FMAL method exhibits the strongest purposefulness in sample selection, which results in the best classification performance. On the other hand, the random method lacks purpose, leading to it having the poorest performance. Specifically, as can be seen from the red boxed area in Figure 18, the sample selected by FMAL is more representative and the most efficient. Although the samples selected by BVSB show slightly higher relevance in some cases, the efficiency is still better than that of the random method. In addition, the blue boxed area in Figure 18c demonstrates that selecting too many samples at once results in smaller differences between some of the samples, highlighting the importance of iteration.

5.2.2. Analysis of AL Iterative Strategy

To validate the role of iteration in active learning, we conducted experiments using the IP dataset, which is the most representative and has the highest classification difficulty. Initially, 48 samples (3 samples per class) were randomly selected, which served as the initial training set. Then, the number of samples gradually increased using three different methods. AL0 denotes that all desired samples are randomly selected, AL1 denotes that all desired samples are selected using one active learning method, and ALn denotes that the desired samples are selected step by step through iterations.
From the results in Figure 19a–c, it is evident that the accuracy of the model using the Aln method is consistently higher than that of AL0 and AL1. This suggests that model accuracy has an impact on the quality of the samples selected. The iterative approach selects samples on the basis of a model with progressively higher accuracy and therefore achieves better results.
When the number of samples increases to 192, AL1 surpasses the AL0 method, which indicates that FMAL’s purposeful identification of uncertain samples is more effective than randomly selecting samples.
As the number of samples increases to 384, the accuracy improvement rate of AL1 significantly decreases, and during the increment, the accuracy values of AL1 and AL0 cross over, with AL0 subsequently surpassing AL1. The reason behind this phenomenon is that as the number of samples selected by the AL1 method increases, a phenomenon akin to “oversampling” emerges. This leads to high consistency among the chosen samples, which causes the model accuracy to fall even below that of randomly selected samples.
When the total number of samples in the training set reached 768, the classification accuracy of Aln almost reached 100%, maintaining the highest classification accuracy. Simultaneously, examining the change in model classification accuracy in Figure 19 reveals that the accuracy of AL0 approached that of Aln. This is because the coverage of the IP dataset increased significantly when too many samples were selected, leading to smaller and smaller differences in the samples selected. Similarly, the accuracy of AL1 did not significantly decrease for the same reason.

5.2.3. Analysis of the Number of DSPAL Iterations

One iteration of PAL can significantly increase the number of training samples without increasing the sample labeling cost. To analyze the impact of multiple iterations of DSPAL on model accuracy, eight iterations of DSPAL were conducted on models trained through active learning on three datasets. The filter thresholds P1 and P2 were set as [40, 50, 60, 70, 80, 80, 80, 80] and [40-1, 50-2, 60-2, 70-3, 80-3, 80-3, 80-3, 80-3], respectively.
The changes in three metrics during multiple iterations on the three datasets are shown in Figure 20a–c. It can be observed that the model’s accuracy gradually improved during the iterations and reached its peak at around the fourth or fifth iteration, after which it started to fluctuate. DSPAL improved OA from 77.85% to 82.44% on the IP dataset, from 79.43% to 88.95% on the PU dataset, and from 83.31% to 86.16% on the WHU dataset. Notably, the first four iterations led to a significant improvement in OA, AA, and Kappa. On the IP dataset, the OA reached its maximum value in the fourth iteration, while on the PU and WHU datasets, the OA reached its maximum in the fifth iteration. As the number of iterations increased, the differences in OA gradually decreased. After the fifth iteration, the OA did not continue to rise but showed a decline. This is because, after a certain number of iterations, it becomes challenging to find samples with new feature information from the remaining test set. This indicates that DSPAL can find samples with new features through iteration, which significantly improves the model. However, with the decrease in new feature Information in the selected samples, too many iterations can lead to overfitting, resulting in a decrease in accuracy.
The results demonstrate that samples selected through DSPAL iterations have new features, which contribute to higher network accuracy. However, blindly increasing the number of iterations does not necessarily continuously improve the classification ability of the network. The network reaches an optimal state within a certain number of iterations, after which the network accuracy shows little improvement. Therefore, based on experimental results, this study selected four iterations of pseudo-active learning to achieve effective accuracy improvements within an appropriate range.

6. Conclusions

The core issues of HSI classification are accuracy and cost. The main method for reducing costs is to limit the number of annotated samples. This study proposes a multi-strategy approach to address the small-sample HSI classification problem. On the one hand, it improves the quality of annotated samples through active learning by introducing FMAL to identify valuable samples precisely. It takes into consideration the impact of the initial model accuracy on the quality of identified samples. FMAL creates anchor values from the encoding-layer features of the original training set samples. By predicting the values of unlabeled samples mixed with a certain proportion of anchor values, the value of the samples can be determined. This effectively reduces the dependence of sample selection on the original model. An efficient implementation of the optimal solution for calculating the mixed proportional weights of features is given. On the other hand, the sample size of the training set is increased through contrast learning and pseudo-active learning. To establish a reliable initial model, we introduce triplet sample pairs to expand the training samples. On this basis, DSPAL was designed to generate pseudo-label samples. This method uses two strategies to evaluate the confidence and uncertainty of the samples separately. A large number of pseudo-label samples were filtered out to expand the training set. This may be considered a promising direction for solving small-sample classification problems. Comparative experiments on three datasets demonstrated that the MSTNC method outperforms several state-of-the-art methods in terms of accuracy.
Although we have obtained encouraging experimental results, we will continue our research, specifically focusing on the following aspects:
(1) The number of samples selected in each iteration is equal to the number of samples in the training set prior to that iteration in the feature-mixture-based active learning method employed in this study. The threshold ε , which controls the feature mixing ratio, and the model’s loss in predicting the mixed feature class are two important factors in determining the quality of the selected samples. We can consider using these two values to calculate the number of selected samples in each iteration scientifically and improve the active learning method. (2) The two thresholds of the filter in pseudo-active learning were selected based on extensive experimental evaluations, and future research should aim to identify more scientific and efficient methods. (3) Our proposed dual-strategy pseudo-active method selects samples based on the classification results of the model, allowing for flexibility in the model’s structure and making it applicable to classification tasks in other fields.

Author Contributions

Conceptualization, Y.B.; methodology, Y.B.; software, D.L.; validation, D.L.; formal analysis, Y.B. and D.L.; investigation, Y.B.; resources, Y.B. and D.L.; data curation, D.L.; writing-original draft preparation, D.L.; writing-review and editing, Y.B., D.L., L.Z., and H.W.; visualization, H.W.; supervision, Y.B.; project administration, Y.B.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Liaoning Province Education Administration under Grant LJKZ0174.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Indian Pines and Pavia University datasets are available at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 20 January 2023). The Wuhan Han Chuan dataset is available at http://rsidea.whu.edu.cn/resource_WHUHi_sharing.htm (accessed on 20 January 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

Nomenclatures and Notations

MSTNCMulti-strategy triplet network classifier
HSIHyperspectral Imaging
CNNConvolutional neural networks
conv3D3-dimensional convolutional neural networks
conv2D2-dimensional convolutional neural networks
PCAPrincipal component analysis
TNTriplet network
TNCTriplet networfk classifier
ALActive learning
MThe total number of samples that can be chosen
FMALFeature mixture-based active learning
DSPALDual- strategy pseudo-active learning
FCFully connected
GAPGlobal average pooling layer
BvSBBest vs. Second Best
C l s The classes of land cover
f e Encoder
f p Projection head
f c Classifier
x l Labeled sample
y i Label of the i-th sample
x u Unlabeled sample
D t e s t Test set
D t r a i n Train set
x l + The same class as x l (positive pair)
x l The different class as x l (negative pair)
z The output vector corresponding to x l
z + The output vector corresponding to x l +
z The output vector corresponding to x l
y i The value of the i-th class
y ^ i Projected value
θ p Projection head parameters
θ c Classifier parameters
z ¯ Anchor value
z u Encoding-layer feature
z ¯ i The anchor value of the i-th class of samples
z ~ i Mixed feature value

References

  1. Hong, D.; Yokoya, N.; Chanussot, J.; Zhu, X.X. CoSpace: Common subspace learning from hyperspectral-multispectral correspondences. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4349–4359. [Google Scholar] [CrossRef]
  2. Hong, D.; Yokoya, N.; Ge, N.; Chanussot, J.; Zhu, X.X. Learnable manifold alignment (LeMA): A semi-supervised cross-modality learning framework for land cover and land use classification. ISPRS J. Photogramm. Remote Sens. 2019, 147, 193–205. [Google Scholar] [CrossRef]
  3. Han, X.; Zhang, H.; Sun, W. Spectral anomaly detection based on dictionary learning for sea surfaces. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1502505. [Google Scholar] [CrossRef]
  4. Kumar, V.; Ghosh, J.K. Camouflage detection using MWIR hyperspectral images. J. Indian Soc. Remote Sens. 2017, 45, 139–145. [Google Scholar] [CrossRef]
  5. Shimoni, M.; Haelterman, R.; Perneel, C. Hypersectral imaging for military and security applications: Combining myriad processing and sensing techniques. IEEE Geosci. Remote Sens. Mag. 2019, 7, 101–117. [Google Scholar] [CrossRef]
  6. Briottet, X.; Boucher, Y.; Dimmeler, A. Military applications of hyperspectral imagery. In Targets and Backgrounds XII: Characterization and Representation; SPIE: Bellingham, DC, USA, 2006; Volume 6239, pp. 82–89. [Google Scholar]
  7. Ke, C. Military object detection using multiple information extracted from hyperspectral imagery. In Proceedings of the 2017 International Conference on Progress in Informatics and Computing (PIC), Nanjing, China, 15–17 December 2017; IEEE: New York, NY, USA, 2017; pp. 124–128. [Google Scholar]
  8. Jasani, B.; Stein, G. Commercial Satellite Imagery: A Tactic in Nuclear Weapon Deterrence; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  9. Carpenter, M.H.; Croce, M.P.; Baker, Z.K.; Batista, E.R.; Caffrey, M.P.; Fontes, C.J.; Koehler, K.E.; Kossmann, S.E.; McIntosh, K.G.; Rabin, M.W.; et al. Hyperspectral X-ray Imaging with TES Detectors for Nanoscale Chemical Speciation Mapping. J. Low Temp. Phys. 2020, 200, 437–444. [Google Scholar] [CrossRef]
  10. Al Ktash, M.; Stefanakis, M.; Englert, T.; Drechsel, M.S.L.; Stiedl, J.; Green, S.; Jacob, T.; Boldrini, B.; Ostertag, E.; Rebner, K.; et al. UV Hyperspectral Imaging as Process Analytical Tool for the Characterization of Oxide Layers and Copper States on Direct Bonded Copper. Sensors 2021, 21, 7332. [Google Scholar] [CrossRef] [PubMed]
  11. Batshev, V.I.; Krioukov, A.V.; Machikhin, A.S.; Zolotukhina, A.A. Multispectral video camera optical system. J. Opt. Technol. 2023, 90, 706–712. [Google Scholar] [CrossRef]
  12. Adesokan, M.; Alamu, E.O.; Otegbayo, B.; Maziya-Dixon, B. A Review of the Use of Near-Infrared Hyperspectral Imaging(NIR-HSI) Techniques for the Non-Destructive Quality Assessment of Root and Tuber Crops. Appl. Sci. 2023, 13, 5226. [Google Scholar] [CrossRef]
  13. Kulya, M.; Petrov, N.V.; Tsypkin, A.; Egiazarian, K.; Katkovnik, V. Hyperspectral data denoising for terahertz pulse time-domain holography. Opt. Express 2019, 27, 18456–18476. [Google Scholar] [CrossRef]
  14. Zare, A.; Ho, K.C. Endmember variability in hyperspectral analysis:Addressing spectral variability during spectral unmixing. IEEE Signal Process. Mag. 2014, 31, 95–104. [Google Scholar] [CrossRef]
  15. Mei, S.; Bi, Q.; Ji, J.; Hou, J.; Du, Q. Spectral variation alleviation bylow-rank matrix approximation for hyperspectral image analysis. IEEE Geosci. Remote Sens. Lett. 2016, 13, 796–800. [Google Scholar] [CrossRef]
  16. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  17. Zhao, W.; Du, S. Spectral-spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  18. Liu, W.; Tao, D. Multiview Hessian regularization for polynomial logistic regression in hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2013, 24, 1897–1909. [Google Scholar]
  19. Shevkunov, I.; Katkovnik, V.; Claus, D.; Pedrini, G.; Petrov, N.V.; Egiazarian, K. Spectral object recognition in hyperspectral holography with complex-domain denoising. Sensors 2019, 19, 5188. [Google Scholar] [CrossRef]
  20. Fu, H.; Sun, G.; Ren, J.; Zhang, A.; Jia, X. Fusion of PCA and segmented-PCA domain multiscale 2-D-SSA for effective spectral–spatial feature extraction and data classification in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5500214. [Google Scholar] [CrossRef]
  21. Mu, C.; Zeng, Q.; Liu, Y.; Qu, Y. A two-branch network combined with robust principal component analysis for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 2147–2151. [Google Scholar] [CrossRef]
  22. Fu, H.; Sun, G.; Zabalza, J.; Zhang, A.; Ren, J.; Jia, X. Tensor singular spectrum analysis for 3D feature extraction in hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5403914. [Google Scholar] [CrossRef]
  23. Dai, Q.; Ma, C.; Zhang, Q. Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction. Sensors 2024, 24, 4072. [Google Scholar] [CrossRef]
  24. Zheng, J.; Feng, Y.; Bai, C.; Zhang, J. Hyperspectral image classification using mixed convolutions and covariance pooling. IEEE Trans. Geosci. Remote Sens. 2020, 59, 522–534. [Google Scholar] [CrossRef]
  25. Wang, W.; Chen, Y.; He, X.; Li, Z. Soft augmentation-based Siamese CNN for hyperspectral image classification with limited training samples. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5508505. [Google Scholar] [CrossRef]
  26. Wu, X.; Hong, D.; Chanussot, J. Convolutional neural networks for multimodal remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5517010. [Google Scholar] [CrossRef]
  27. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  28. Makantasis, K.; Karantzalos, K.; Doulamis, A. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; IEEE: New York, NY, USA, 2015; pp. 4959–4962. [Google Scholar]
  29. Cheng, G.; Li, Z.; Han, J.; Yao, X.; Guo, L. Exploring hierarchical convolutional features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6712–6722. [Google Scholar] [CrossRef]
  30. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. A new deep convolutional neural network for fast hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2018, 145, 120–147. [Google Scholar] [CrossRef]
  31. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  32. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  33. Yu, W.; Wan, S.; Li, G.; Yang, J.; Gong, C. Hyperspectral image classification with contrastive graph convolutional network. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503015. [Google Scholar] [CrossRef]
  34. Xue, Z.; Zhou, Y.; Du, P. S3Net: Spectral–spatial Siamese network for few-shot hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531219. [Google Scholar] [CrossRef]
  35. Xue, Z.; Liu, Z.; Zhang, M. DSR-GCN: Differentiated-scale restricted graph convolutional network for few-shot hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5504918. [Google Scholar] [CrossRef]
  36. Liu, B.; Yu, X.; Yu, A.; Zhang, P.; Wan, G.; Wang, R. Deep few-shot learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2290–2304. [Google Scholar] [CrossRef]
  37. Sun, C.; Zhang, X.; Meng, H.; Cao, X.; Zhang, J.; Jiao, L. Dual-branch spectral-spatial adversarial representation learning for hyperspectral image classification with few labeled samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1–15. [Google Scholar] [CrossRef]
  38. Zhang, Y.; Li, W.; Zhang, M.; Wang, S.; Tao, R.; Du, Q. Graph information aggregation cross-domain few-shot learning for hyperspectral image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1912–1925. [Google Scholar] [CrossRef]
  39. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant map. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; IEEE: New York, NY, USA, 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  40. Settles, B. Active Learning Literature Survey; Department of Computer Sciences, University of Wisconsin-Madison: Madison, WI, USA, 2009. [Google Scholar]
  41. Bromley, J.; Guyon, I.; LeCun, Y.; Säckinger, E.; Shah, R. Signature verification using a “siamese” time delay neural network. In Advances in Neural Information Processing Systems; Morgan Kaufmann Pub: Burlington, MA, USA, 1993; Volume 6. [Google Scholar]
  42. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Similarity-Based Pattern Recognition, Proceedings of the Third International Workshop, SIMBAD 2015, Copenhagen, Denmark, 12–14 October 2015; Proceedings 3; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 84–92. [Google Scholar]
  43. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  44. He, K.; Fan, H.; Wu, Y.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9729–9738. [Google Scholar]
  45. Zhao, S.; Li, W.; Du, Q.; Ran, Q. Hyperspectral classification based on siamese neural network using spectral-spatial feature. In Proceedings of the IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; IEEE: New York, NY, USA, 2018; pp. 2567–2570. [Google Scholar]
  46. Cao, Z.; Li, X.; Jianfeng, J.; Zhao, L. 3D convolutional siamese network for few-shot hyperspectral classification. J. Appl. Remote Sens. 2020, 14, 048504. [Google Scholar] [CrossRef]
  47. Jia, S.; Jiang, S.; Lin, Z.; Xu, M.; Sun, W.; Huang, Q.; Zhu, J.; Jia, X. A semisupervised Siamese network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5516417. [Google Scholar] [CrossRef]
  48. Di, X.; Xue, Z.; Zhang, M. Active learning-driven siamese network for hyperspectral image classification. Remote Sens. 2023, 15, 752. [Google Scholar] [CrossRef]
  49. Yang, J.; Qin, J.; Qian, J.; Li, A.; Wang, L. AL-MRIS: An active learning-based multipath residual involution siamese network for few-shot hyperspectral image classification. Remote Sens. 2024, 16, 990. [Google Scholar] [CrossRef]
  50. Patel, U.; Patel, V. Active learning-based hyperspectral image classification: A reinforcement learning approach. J. Supercomput. 2024, 80, 2461–2486. [Google Scholar] [CrossRef]
  51. Zhuang, H.; Zhang, Y.; Wu, Q. Disconnection-based active learning for hyperspectral image classification. Remote Sens. 2020, 12, 1484. [Google Scholar]
  52. Ma, W.; Zhou, T.; Qin, J. Adaptive multi-feature fusion via cross-entropy normalization for effective image retrieval. Information Process. Management 2023, 60, 103119. [Google Scholar] [CrossRef]
  53. Raj, A.; Bach, F. Convergence of uncertainty sampling for active learning. arXiv 2021, arXiv:2110.15784. [Google Scholar]
  54. Chen, C.; Xu, J. A marginal sampling approach for active learning in hyperspectral image classification. Remote Sens. Lett. 2023, 14, 152–161. [Google Scholar]
  55. Li, J. Active learning for hyperspectral image classification with a stacked autoencoders based neural network. In Proceedings of the 2015 7th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Tokyo, Japan, 2–5 June 2015; IEEE: New York, NY, USA, 2015; pp. 1–4. [Google Scholar]
  56. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Li, J.; Plaza, A. Active learning with convolutional neural networks for hyperspectral image classification using a new Bayesian approach. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6440–6461. [Google Scholar] [CrossRef]
  57. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  58. Lei, Z.; Zeng, Y.; Liu, P.; Su, X. Active deep learning for hyperspectral image classification with uncertainty learning. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5502405. [Google Scholar] [CrossRef]
  59. Wang, Z.; Zhao, S.; Zhao, G.; Song, X. Dual-Branch Domain Adaptation Few-Shot Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 5506116. [Google Scholar] [CrossRef]
  60. Wang, H.; Wang, L. Collaborative active learning based on improved capsule networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5522126. [Google Scholar] [CrossRef]
  61. Parvaneh, A.; Abbasnejad, E.; Teney, D.; Haffari, R.; Van Den Hengel, A.; Shi, J.Q. Active learning by feature mixing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12237–12246. [Google Scholar]
  62. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  63. Liu, C.; Yi, Z.; Huang, B.; Zhou, Z.; Fang, S.; Li, X.; Zhang, Y.; Wu, X. A deep learning method based on triplet network using self-attention for tactile grasp outcomes prediction. IEEE Trans. Instrum. Meas. 2023, 72, 2518914. [Google Scholar] [CrossRef]
  64. Zhang, L.; Zhang, H.; Liu, T. An active learning framework for hyperspectral image classification. Remote Sens. 2020, 12, 40. [Google Scholar]
  65. Wang, J.; Li, Y. Improving classification accuracy through active learning: A case study of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4782–4796. [Google Scholar]
  66. Sener, O.; Savarese, S. Active learning for convolutional neural networks: A core-set approach. arXiv 2017, arXiv:1708.00489. [Google Scholar]
  67. Joshi, A.J.; Porikli, F.; Papanikolopoulos, N. Multi-class active learning for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 2372–2379. [Google Scholar]
  68. Cao, X.; Yao, J.; Xu, Z.; Meng, D. Hyperspectral image classification with convolutional neural network and active learning. IEEE Trans. Geosci. Remote Sens 2020, 58, 4604–4616. [Google Scholar] [CrossRef]
  69. Lesniak, D.; Sieradzki, I.; Podolak, T. Distribution-interpolation trade off in generative models. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  70. Parvaneh, A.; Abbasnejad, E.; Teney, D.; Shi, J.Q.; Van den Hengel, A. Counterfactual vision-and-language navigation: Unravelling the unseen. Adv. Neural Inf. Process. Syst. 2020, 33, 5296–5307. [Google Scholar]
  71. Zhang, H.; Cisse, M.; Dauphin, Y.; Lopez-Paz, D. Mixup: Beyond empirical risk minimization. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–13. [Google Scholar]
  72. Li, Z.; Zhang, L.; Liu, D. A small-sample hyperspectral image classification method based on spectral–spatial features. IEEE Trans. Geosci. Remote Sens. 2014, 52, 5045–5057. [Google Scholar]
  73. Pei, S.; Song, H. A small-sample hyperspectral image classification method based on dual-channel spectral enhancement network. Electronics 2022, 11, 2540. [Google Scholar] [CrossRef]
  74. Lee, D.H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Proceedings of the Workshop on Challenges in Representation Learning, ICML, Atlanta, GA, USA, 16–21 June 2013; Volume 3, pp. 896–901. [Google Scholar]
  75. Zhang, H.; Liu, L.; Long, Y.; Shao, L. Unsupervised deep hashing with pseudo labels for scalable image retrieval. IEEE Trans. Image Process. 2017, 27, 1626–1638. [Google Scholar] [CrossRef]
  76. Zhu, W.; Shi, B.; Feng, Z. A transfer learning method using high-quality pseudo labels for bearing fault diagnosis. IEEE Trans. Instrum. Meas. 2022, 72, 3502311. [Google Scholar] [CrossRef]
  77. Licciardi, G.; Marpu, P.R.; Chanussot, J.; Benediktsson, J.A. Linear versus nonlinear PCA for the classification of hyperspectral data based on the extended morphological profiles. IEEE Geosci. Remote Sens. Lett. 2011, 9, 447–451. [Google Scholar] [CrossRef]
  78. Bai, Y.; Xu, M.; Zhang, L.; Liu, Y. Pruning multi-scale multi-branch network for small-sample hyperspectral image classification. Electrics 2023, 12, 674. [Google Scholar] [CrossRef]
  79. Zhang, K.; Yan, J.; Zhang, F.; Ge, C.; Wan, W.; Sun, J.; Zhang, H. Spectral-spatial dual graph unfolding network for multispectral and hyperspectral image fusion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5508718. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the proposed MSTN for HIS classification. Sensors 24 06647 i001: encoding layer vector of the sample; Sensors 24 06647 i002: the classifier output vector of the sample; ⊕: add selected sample to the train-set; ⊖: remove selected sample from the test-set; filter 1: a filter used in active learning strategies to screen samples using the FMAX method; filter 2: a filter used in the pseudo-active learning strategy, which includes two filters that filter high-confidence samples and high-uncertainty samples, respectively.
Figure 1. The overall architecture of the proposed MSTN for HIS classification. Sensors 24 06647 i001: encoding layer vector of the sample; Sensors 24 06647 i002: the classifier output vector of the sample; ⊕: add selected sample to the train-set; ⊖: remove selected sample from the test-set; filter 1: a filter used in active learning strategies to screen samples using the FMAX method; filter 2: a filter used in the pseudo-active learning strategy, which includes two filters that filter high-confidence samples and high-uncertainty samples, respectively.
Sensors 24 06647 g001
Figure 2. TNC detailed model structure. k 0 i @ k 1 i k 2 i k 3 i : k 0 i denote the size of the convolution kernel; k 1 i k 2 i k 3 i represents the three dimensions of the input data; i represents the i-th layer convolution.
Figure 2. TNC detailed model structure. k 0 i @ k 1 i k 2 i k 3 i : k 0 i denote the size of the convolution kernel; k 1 i k 2 i k 3 i represents the three dimensions of the input data; i represents the i-th layer convolution.
Sensors 24 06647 g002
Figure 3. Feature mixing-based active learning (FMAL) illustration. ⊕: add selected sample to the train-set; ⊖: remove selected sample from the test-set; Repeat: number of iterations; Sensors 24 06647 i003: training TNC with train-set.
Figure 3. Feature mixing-based active learning (FMAL) illustration. ⊕: add selected sample to the train-set; ⊖: remove selected sample from the test-set; Repeat: number of iterations; Sensors 24 06647 i003: training TNC with train-set.
Sensors 24 06647 g003
Figure 4. The proposed DSPAL architecture. A filter is established to screen samples classified by the classifier. The samples are labeled using the predicted values of the classifier, and the qualified pseudo-labeled samples are then added to D t r a i n for iterative training.
Figure 4. The proposed DSPAL architecture. A filter is established to screen samples classified by the classifier. The samples are labeled using the predicted values of the classifier, and the qualified pseudo-labeled samples are then added to D t r a i n for iterative training.
Sensors 24 06647 g004
Figure 5. Indian Pines dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Figure 5. Indian Pines dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Sensors 24 06647 g005
Figure 6. Pavia University dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Figure 6. Pavia University dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Sensors 24 06647 g006
Figure 7. Wuhan Han Chuan dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Figure 7. Wuhan Han Chuan dataset: (a) false-color map; (b) ground-truth map. The numbers in parentheses represent the total number of samples in each class.
Sensors 24 06647 g007
Figure 8. Classification maps of the different models for the Indian Pines dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Figure 8. Classification maps of the different models for the Indian Pines dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Sensors 24 06647 g008
Figure 9. Classification maps of the different models for the Pavia University dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Figure 9. Classification maps of the different models for the Pavia University dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Sensors 24 06647 g009
Figure 10. Classification maps of the different models for the WHU-Hi-HanChuan dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Figure 10. Classification maps of the different models for the WHU-Hi-HanChuan dataset: (a) 3DCNN; (b) DFSL-NN; (c) DFSL-SVM; (d) Gia-CFSL; (e) DBDAFSL; (f) CapsGLOM; (g) MSTNC.
Sensors 24 06647 g010
Figure 11. Evolution of OA as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Figure 11. Evolution of OA as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Sensors 24 06647 g011
Figure 12. Evolution of AA as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Figure 12. Evolution of AA as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Sensors 24 06647 g012
Figure 13. Evolution of Kappa as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Figure 13. Evolution of Kappa as a function of the number of training samples per class: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Sensors 24 06647 g013
Figure 14. Training time and testing time of different methods.
Figure 14. Training time and testing time of different methods.
Sensors 24 06647 g014
Figure 15. Ablation experiments of the triplet contrastive learning.
Figure 15. Ablation experiments of the triplet contrastive learning.
Sensors 24 06647 g015
Figure 16. Ablation experiments of the proposed method on.
Figure 16. Ablation experiments of the proposed method on.
Sensors 24 06647 g016
Figure 17. Comparison of accuracy of three different sample selection methods: (a) OA; (b) AA; (c) Kappa.
Figure 17. Comparison of accuracy of three different sample selection methods: (a) OA; (b) AA; (c) Kappa.
Sensors 24 06647 g017
Figure 18. Classification maps of different methods for Indian Pines. Seven labeled samples were selected for each class. Light blue dots are randomly selected samples. The red dots is samples selected with a specific method. Dark blue dots are samples that were incorrectly predicted. (a) Random method; (b) BvSB method; (c) FMAL method.
Figure 18. Classification maps of different methods for Indian Pines. Seven labeled samples were selected for each class. Light blue dots are randomly selected samples. The red dots is samples selected with a specific method. Dark blue dots are samples that were incorrectly predicted. (a) Random method; (b) BvSB method; (c) FMAL method.
Sensors 24 06647 g018
Figure 19. Analysis of the role of iteration in active learning methods. The data in parentheses in the horizontal axis represent the number of iterations of ALn. (a) OA; (b) AA; (c) Kappa.
Figure 19. Analysis of the role of iteration in active learning methods. The data in parentheses in the horizontal axis represent the number of iterations of ALn. (a) OA; (b) AA; (c) Kappa.
Sensors 24 06647 g019
Figure 20. Effect of the number of DSPAL iterations on accuracy: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Figure 20. Effect of the number of DSPAL iterations on accuracy: (a) IP dataset; (b) PU dataset; (c) WHU dataset.
Sensors 24 06647 g020
Table 1. Classification results of each method in Indian Pines.
Table 1. Classification results of each method in Indian Pines.
Number/Class3DCNN [20]DFSL-NN [36]DFSL-SVM [36]Gia-CFSL [38]DBDAFSL [59]CapsGLOM [60]MSTNC (Ours)
1. Alfalfa90.1198.2397.56100100100100
2. Corn-notill58.5649.2452.9368.3745.241.296.76
3. Corn-mintill43.4262.9669.2166.3361.6265.6277.91
4. Corn44.9853.7757.3187.0183.2379.2388.05
5. Grass-pasture67.8884.8984.7171.7576.7286.7269.21
6. Grass-trees89.4292.3892.5684.7292.5593.5586.63
7. Grass-pasture-mowed91.9298.0198.6599.45100100100
8. Hay-windrowed81.2285.5985.5998.7399.2210099.78
9. Oats98.1190.990.999.7198.6684.21100
10. Soybean-notill66.9365.6630.2171.2162.5726.5773.34
11. Soybean-mintill19.6773.6495.7844.7870.7785.7773.71
12. Soybean-clean30.7321.0839.9654.9638.0289.0279.06
13. Wheat92.6775.2681.6389.6397.4410092.85
14. Woods89.3790.8189.8776.8784.5993.5984.92
15. Buildings-Grass-Trees-Drives55.158.4340.4280.4288.1878.18100
16. Stone-Steel-Towers95.0696.5397.899.0293.5281.52100
OA54.75 ± 3.5365.90 ± 4.6466.49 ± 3.3767.52 ± 3.7970.52 ± 3.6274.63 ± 3.6282.97 ± 2.44
AA63.91 ± 1.4571.25 ± 3.3470.34 ± 2.5180.94 ± 1.7880.97 ± 1.7781.58 ± 1.7788.89 ± 1.51
Kappa52.21 ± 3.7160.63 ± 4.1760.09 ± 3.2863.67 ± 3.9266.67 ± 2.8970.60 ± 2.8980.70 ± 2.48
Table 2. Classification results of each method in Pavia University.
Table 2. Classification results of each method in Pavia University.
Number/Class3DCNN [20]DFSL-NN [36]DFSL-SVM [36]Gia-CFSL [38]DBDAFSL [59]CapsGLOM [60]MSTNC (Ours)
1. Asphalt80.9475.6347.2284.0180.0179.5185.06
2. Meadows90.5683.7885.3184.8381.8395.2292.74
3. Gravel36.6151.4183.2554.9892.9835.3179.44
4. Trees61.2074.2266.9686.9965.9994.8771.06
5. Painted metal sheets99.6998.1599.83100.0099.77100.00100.00
6. Bitumen55.9285.4685.273.2299.2260.7983.85
7. Bare Soil90.7696.2895.398.1499.1489.2494.23
8. Self-Blocking Bricks37.2271.1270.4119.4154.4192.4788.69
9. Shadows81.4994.9834.1262.0290.0299.3686.62
OA75.75 ± 5.5680.47 ± 5.6876.33 ± 3.8676.79 ± 6.7482.07 ± 6.7485.60 ± 4.8587.94 ± 3.45
AA70.49 ± 3.3981.34 ± 3.8575.20 ± 2.5973.73 ± 5.6384.90 ± 5.6382.98 ± 2.6486.85 ± 2.37
Kappa67.47 ± 6.4574.64 ± 6.7169.55 ± 4.7769.73 ± 4.5977.02 ± 4.5980.76 ± 5.6984.11 ± 4.22
Table 3. Classification results of each method in WHU-Hi-HanChuan.
Table 3. Classification results of each method in WHU-Hi-HanChuan.
Number/Class3DCNN [20]DFSL-NN [36]DFSL-SVM [36]Gia-CFSL [38]DBDAFSL [59]CapsGLOM [60]MSTNC (Ours)
1. Strawberry85.9768.2655.1651.5257.3671.2291.25
2. Cowpea80.8945.9346.9667.8658.8576.4583.07
3. Soybean42.7086.2185.579.6278.0077.4586.90
4. Sorghum88.9296.3196.8595.7497.3597.3397.36
5. Water spinach30.8086.7199.8489.8196.99100.0098.58
6. Watermelon0.1159.9366.9784.7348.5074.5581.76
7. Greens2.5581.0298.7195.3882.4979.9483.94
8. Trees44.7561.2935.2959.1431.2165.9876.21
9. Grass11.3556.3226.9653.6162.2163.2083.70
10. Red roof39.7766.5373.7461.0391.3981.5484.47
11. Gray roof0.0487.7467.865.5279.5989.0696.75
12. Plastic34.3443.8659.2470.5674.7073.5751.94
13. Bare soil13.3439.1948.6336.2036.4441.2255.11
14. Road24.3765.2134.4630.5152.4557.5980.80
15. Bright object68.1053.0865.6857.2467.9958.2157.07
16. Water98.0672.6798.196.3998.9797.7296.48
OA62.91 ± 2.2167.85 ± 2.8168.04 ± 1.6570.15 ± 1.8572.22 ± 2.7179.50 ± 2.6786.57 ± 1.17
AA41.71 ± 1.7567.05 ± 2.6566.08 ± 2.3868.51 ± 1.7869.58 ± 2.1475.60 ± 1.6579.18 ± 1.05
Kappa55.30 ± 2.7163.53 ± 2.3163.00 ± 2.4265.67 ± 1.5667.97 ± 2.3176.22 ± 2.5184.36 ± 1.13
Table 4. Quantitative results for the three data sets.
Table 4. Quantitative results for the three data sets.
DATASETMeasureSSRN [31]3DAES [47]ConGcn [33]TensorSSA [22]SmaT-HOSVD [23]MSTNC (Ours)
IPOA (%)76.1678.9379.9487.5592.4596.36
AA (%)82.2578.2578.4288.9291.8996.29
κ × 10072.4975.2677.8586.6893.0195.85
PUOA (%)91.2792.7691.1894.4896.88100
AA (%)88.50 88.50 86.50 93.4594.45100
κ × 10088.3792.60 90.60 92.5295.79 100
WHUOA (%)96.2195.5495.0496.7897.88100
AA (%)97.2197.9196.0198.0299.01100
κ × 10092.5992.6190.5596.2497.95100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bai, Y.; Liu, D.; Zhang, L.; Wu, H. A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme. Sensors 2024, 24, 6647. https://doi.org/10.3390/s24206647

AMA Style

Bai Y, Liu D, Zhang L, Wu H. A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme. Sensors. 2024; 24(20):6647. https://doi.org/10.3390/s24206647

Chicago/Turabian Style

Bai, Yu, Dongmin Liu, Lili Zhang, and Haoqi Wu. 2024. "A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme" Sensors 24, no. 20: 6647. https://doi.org/10.3390/s24206647

APA Style

Bai, Y., Liu, D., Zhang, L., & Wu, H. (2024). A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme. Sensors, 24(20), 6647. https://doi.org/10.3390/s24206647

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop