Next Article in Journal
Distribution Modeling and Gap Analysis of Shorebird Conservation in Northern Brazil
Next Article in Special Issue
Semantic Segmentation Model for Wide-Area Coseismic Landslide Extraction Based on Embedded Multichannel Spectral–Topographic Feature Fusion: A Case Study of the Jiuzhaigou Ms7.0 Earthquake in Sichuan, China
Previous Article in Journal
What Are We Missing? Occlusion in Laser Scanning Point Clouds and Its Impact on the Detection of Single-Tree Morphologies and Stand Structural Variables
Previous Article in Special Issue
Feature Engineering of Geohazard Susceptibility Analysis Based on the Random Forest Algorithm: Taking Tianshui City, Gansu Province, as an Example
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Attention-Based 3D Convolutional Autoencoder for Few-Shot Hyperspectral Unmixing and Classification

1
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100094, China
2
School of Aeronautics and Astronautics, University of Chinese Academy of Sciences, Beijing 100049, China
3
Investigation College of People’s Public Security University of China, Beijing 100038, China
4
China Aero Geophysical Survey and Remote Sensing Center for Land and Resources, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(2), 451; https://doi.org/10.3390/rs15020451
Submission received: 2 December 2022 / Revised: 4 January 2023 / Accepted: 5 January 2023 / Published: 12 January 2023

Abstract

:
Few-shot hyperspectral classification is a challenging problem that involves obtaining effective spatial–spectral features in an unsupervised or semi-supervised manner. In recent years, as a result of the development of computer vision, deep learning techniques have demonstrated their superiority in tackling the problems of hyperspectral unmixing (HU) and classification. In this paper, we present a new semi-supervised pipeline for few-shot hyperspectral classification, where endmember abundance maps obtained by HU are treated as latent features for classification. A cube-based attention 3D convolutional autoencoder network (CACAE), is applied to extract spectral–spatial features. In addition, an attention approach is used to improve the accuracy of abundance estimation by extracting the diagnostic spectral features associated with the given endmember more effectively. The endmember abundance estimated by the proposed model outperforms other convolutional neural networks (CNNs) with respect to the root mean square error (RMSE) and abundance spectral angle distance (ASAD). Classification experiments are performed on real hyperspectral datasets and compared to several supervised and semi-supervised models. The experimental findings demonstrate that the proposed approach has promising potential for hyperspectral feature extraction and has better performance relative to CNN-based supervised classification under small-sample conditions.

1. Introduction

In earth observation, hyperspectral technology is one of the top trends in the remote sensing community, and it plays a significant role. The hundreds of spectral bands provided by hyperspectral images (HSIs) give them an inherent advantage in quantitative applications such as mineral mapping, environmental monitoring and classification [1].
Earlier work mainly focused on discriminative spectral feature extraction, such as the spectral angle mapper (SAM) [2] and spectral information divergence (SID) [3]. These methods utilize extensive prior knowledge and typically require no sample. They do not, however, fully use the spatial features of HSIs. Later on, some supervised spectral classifiers, such as support vector machines (SVM) [4], random forest [5] and neural networks [6], were proposed to achieve a more accurate classification and have since gained widespread acceptance. Nonetheless, the curse of dimensionality is still a bottleneck for supervised classifiers. To resolve this concern, a large number of spectral–spatial joint feature extraction methods have been developed for HSI classification [7,8,9,10,11]. In addition, numerous band selection methods [12,13] and hand-crafted feature extraction methods combined with a learning strategy are proposed [8,14]. However, these shallow features still have limitations regarding more precise hyperspectral classification in complex scenes.
Thanks to advances in computer vision, deep learning technology has made significant strides in remote sensing applications in recent years. CNNs’ potent learning capabilities have enabled end-to-end supervised deep learning models to achieve highly competitive hyperspectral classification results [15,16,17,18] when large amounts of labeled data are available. The various deep leaning classification modes for HSIs are summarized in [19]. However, since hyperspectral classification is a few-shot problem in most cases, it is challenging to collect the large number of hand-crafted training samples required for supervised classification models. In addition, larger models applied to scenarios with limited data tend to result in overfitting and reduce the robustness of the model. Currently, developing an unsupervised or semi-supervised method suitable for small-sample scenarios is a significant challenge in hyperspectral classification.
Various deep learning-based approaches with different learning paradigms have been proposed to address this issue [20,21], including transfer learning [22,23], few-shot learning [24,25] and self-learning [26,27]. The purpose of transfer learning is to initialize the network weights, thereby reducing training time and improving accuracy. Few-shot learning also employs the strategy of transfer learning but focuses more on mining meaningful representation information from labeled data [28]. However, this method has stringent requirements for the labeled quality and diversity of the data. Unlike few-shot learning methods, self-supervised learning can learn deep representations by reconstructing the input completely. However, this type of data representation is regarded as the compression of all information rather than the effective screening of meaningful information, and information with a small proportion is often disregarded [29].
Summarizing previous research, we believe that in order to solve the problem of few-shot hyperspectral classification, three conditions are necessary: 1. Strong feature extraction capabilities; 2. Effective representation information is obtained from data; 3. The model is robust and less dependent on data. Therefore, the question of how to combine self-supervised learning with traditional physical models, such as HU, to achieve effective expression of representational information has become an important area of study.
Let us take a look at the HSI unmixing methods first. The HU technique is applied in order to split the mixed pixel spectrum into materials in their purest forms (endmembers) as well as their proportions (abundances). Generally, the common HU model can be summarized in two categories: the linear mixing (LM) model and the nonlinear mixing (NLM) model [30]. The LM model operates under the presumption that every pixel of the HSIs is a linear combination of the pure endmembers. These methods can be further subdivided into pure pixel-based methods [31,32], and non-pure pixel-based methods [33,34], which concentrate on leveraging the data structure by making geometrical or sparseness assumptions. However, the LM model does not consider the multiple-scattering function and the interaction between objects, which makes it unsuitable to solve the HU problem in complex scenes [35]. In recent years, many neural network (NN) algorithms have been proposed to handle NLM challenges [36,37,38]. Deep learning methods have been proven to improve the accuracy of HU. Recently, the convolutional autoencoder (CAE) has emerged as a new trend in HU applications [39,40,41]. A denoising and sparseness autoencoder is introduced in [42,43] to estimate the abundance of endmembers. Additionally, stacked autoencoders are further employed for HU [44,45]. Most recently, 3D-CNN autoencoders [35,46] were employed to handle HU problems in a supervised setting. Existing research has acknowledged the significance of spatial–spectral joint characteristics for classification [47]. New advances indicate that HU theory can guide the network to learn more effective and regularized representational features; thus, the robustness of the model can be enhanced. The endmember abundance obtained through HU can provide useful spatial–spectral features for semi-supervised classification [48,49,50]. This also sheds new light on solving the few-shot classification problem.
In this research, we introduce a novel end-to-end convolutional autoencoder that we call CACAE. Furthermore, an attention mechanism is utilized to increase the accuracy of abundance estimation by extracting the diagnostic spectral characteristics associated with a given endmember more precisely. In addition, a semi-supervised classification pipeline based on CACAE that uses endmember abundance maps as classification features is introduced. Experiments are carried out on real hyperspectral datasets, and the outcomes are compared using a variety of supervised and semi-supervised models.
The remaining sections of this paper are structured as follows: In Section 2, the proposed method is described. The experimental dataset is described in Section 3. Experiments and analysis are discussed in Section 4, while the final section provides the conclusion.

2. Methods

2.1. Problem Definition

The LM model typically assumes that the reflectance spectra are linearly mixed by different endmembers. The HU formula can be expressed as:
M = E A + N
where M represents a mixed pixel, E stands for the endmembers, A stands for the abundances and N is the additive vector. However, considering the multiple-scattering function and the interaction between objects, the nonlinear model is more suitable to solve the HU problem in complex scenes. The nonlinear model is defined as:
M = g ( E A ) + N
where g refers to a nonlinear function.
Autoencoders are unsupervised training networks that are designed to force the learned representations to exhibit useful properties. During self-learning, input data are compressed, and data structures are subsequently learned and utilized. The decoder can be viewed as the reconstruction of HSIs by endmembers and abundances. The abundances can be estimated via an autoencoder by providing the endmembers as the weight of the rebuilt layer. Furthermore, the stacked convolutional and activation layers of CNNs can be interpreted as a nonlinear transformation of the linear function.

2.2. Attention-Based 3D Autoencoder

The spectral dimension is redundant for hyperspectral data. Consequently, we employ 3D convolution to make the extraction of effective features more feasible. The formula is expressed as follows:
p l f x y z = λ ( n h = 0 H k 1 w = 0 W k 1 d = 0 C k 1 w l f n h w d p ( l 1 ) n ( x + h ) ( y + w ) ( z + d ) ) + b l f
where p l f x y z is the value at position ( x ,   y ,   z ) on the f th feature map in the l th layer and n is an index of the sets on the previous ( l 1 ) layer before that; H k ,   W k and C k are the kernel’s height, width and channel, while w l f n h w d   is the weight at position ( h , w , d ) associated with the f th feature map; b represents the bias and λ stands for the activation function.
We use the hyperspectral cube as input to make sure that both the spatial and spectral information is considered when estimating the abundance (Figure 1). Assume that a hyperspectral cube has the dimensions S × S × C, where S represents the size of the spatial window and C represents the total number of spectral bands. The spatial and spectral information around the center pixel can be used to estimate the related abundance vector of the central pixel. Table 1 shows the architecture of the CACAE model. An encoder with an attention mechanism consisting of four 3D convolutional layers can further filter the effective channel information of HSIs. In order to implement the non-negativity and summation constraints [51,52] of HU, we add an NSC layer to the decoder part through the use of the softmax activation function.
HSIs contain hundreds of spectral bands, but not all spectral information is useful. The attention mechanism performs the extraction of the diagnostic spectral features that are associated with the given endmember by reassigning the weights to the spectral band. Figure 2 illustrates the suggested module for the channel attention mechanism, which is described in this section.
In the first step, the dimensions of the input feature map must be transformed into the channel-last form. Then, global average polling is utilized in the process of integrating and compressing the data of each channel into a one-dimensional vector. The weighted layer, which represents the importance of the feature map of each channel, is created using a 1 × 1 convolutional layer and sigmoid activation.
s = g ( z , W ) = λ ( W 2 δ ( W 1 z ) )
in which W 1 C r × C and W 2 C × C r .   λ and δ are activation functions.
  x ˜ c = s c F c
where X ˜ = [ x ˜ 1 , x ˜ 2 , , x ˜ C ] represents the channel-wise multiplication of the vector s c and the feature map F c H × W .

2.3. Loss Function

The spectral angle distance function (SADF) is employed for an abundance estimation process that deals with calculating the similarity between the rebuilt spectrum and the original spectrum. The SADF is defined as follows:
S A D F = a r c c o s ( < V ^ j , V j > | | V ^ j | | 2 | | V j | | 2 )
where V ^ j is the spectrum reconstructed by the model and the V j is the original input.
Cross-entropy is used in the classification process to measure the discrepancy between the predicted probability distribution and the actual probability distribution. It shows how far the actual output is from the expected output. This means that the two probability distributions are closer when the cross-entropy value is smaller. The equation can be written as:
L = 1 m i L i = 1 N i c = 1 k y i c log ( p i c )
where m is the total number of the sample size, k denotes the number of categories and y i c is an indicator variable that can take on the values 0 or 1 (it takes on the value 1 if the category c is the same as the category of sample i and it takes on the value 0 otherwise). Finally, p i c represents the predicted probability that the sample i is a member of category c .

3. Materials and Experimental Setup

3.1. Dataset

To test the suggested method, it was applied to two publicly available datasets with GT. Following is a description of the datasets’ specifics.

3.1.1. Jasper Ridge Dataset

The AVIRIS sensor is responsible for the collection of the Jasper Ridge dataset, which is now one of the most popular hyperspectral datasets utilized in hyperspectral research. It features 224 bands ranging from 0.38 to 205 μm and a resolution of 100 by 100 pixels. There are some bands associated with atmospheric effects that were eliminated. Experimentation was conducted on the remaining 198 bands. The Jasper Ridge data include four endmembers, namely “Tree”, “Water”, “Soil” and “Road” (Figure 3).

3.1.2. Urban Dataset

Additionally, the Urban dataset is a widely utilized HSI dataset for unmixing research. It consists of 307 by 307 pixels and 210 bands between 0.4 and 2.5 um. The 48 undesirable bands were eliminated, leaving 162 bands for experimentation. The data contain six endmembers, namely “Asphalt”, “Grass”, “Tree”, “Roof” and “Dirt” (Figure 4).

3.1.3. Shenyang Dataset

The Shenyang dataset was captured in Shenyang, Liaoning Province, China, by a next-generation Chinese airborne hyperspectral sensor, the airborne multi-modular imaging spectrometer (AMMIS), which was developed by the Shanghai Institute of Technical Physics (SITP). The spatial resolution of this dataset is 0.75 m/pixel. In experiments, 190 spectral bands were used, and the wavelength range was 0.4–0.9 µm [53,54,55]. The dataset consists of 651 × 837 pixels and five different ground objects, including “Tree”, “Grass”, “Rice”, “Corn” and “Bare land”. GT endmembers were manually labeled with reference to the ground investigations. A visualization image and the spectral curves of the endmembers in the scene depicted are shown in Figure 5.

3.2. Experiment Setting

In our experiments, all comparative CNN models used the same data and training parameter settings as the proposed methods. For Jasper and Urban HSIs datasets, only 10% of pixels were used for training, and all pixels of the datasets were used for evaluation. For the Shenyang dataset, we constructed 7 sets of training datasets through random sampling according to the sampling rates of 10%, 1%, 0.5%, 0.2%, 0.1% and 0.05%, which were used to test the robustness of the algorithm under small-sample conditions. During training, experiments were run in the TensorFlow (2.9.1) environment on an NVIDIA Tesla A100 GPU resource supplied by the HeyWhale data mining platform [56] and optimized using an adaptive moment estimate technique with an initial learning rate of 0.0005. In order to avoid overfitting (despite the low number of training data samples), another approach called dropout with a rate of 0.2 was used. Training consisted of a total of 100 epochs, and the batch size was 30.

4. Experiments and Analysis

4.1. Comparison Model

To demonstrate the efficacy of the input data and the attention mechanism, we developed three additional 3DCAE-based models for comparison: PCAE, CCAE and PACAE. They utilize the same infrastructure as CACAE but differ in specific details. The characteristic features of CACAE are that its input is a three-dimensional image cube and it utilizes the attention module. The difference between PACAE and CACAE is that PACAE’s input is a central pixel rather than an image cube. The main difference between PCAE, CCAE and CACAE is that the structures of PCAE and CCAE lack an attention module. Table 2 details the differences between the various models.
In ground object classification, in addition to PACAE, the classic spectral feature analysis methods of SAM and SID are applied as comparative approaches. A traditional machine learning method for abundance estimation, fully constrained least squares (FCLS) [57], was also employed to assess the effectiveness of the recommended approach for feature extraction. In order to further assess the advantages and disadvantages of the suggested semi-supervised and supervised models, a basic CNN model with a fundamental architecture similar to that of CACAE is presented (Table 3).

4.2. Evaluation

Due to the fact that the method proposed in this paper involves two main processes, endmember abundance map estimation and hyperspectral classification, a quantitative evaluation of the two processes is necessary. As mentioned earlier, the higher the accuracy of endmember abundance map estimation, the more meaningful the extracted spatial–spectral features are for classification. The RMSE judges the similarity between the predicted value and the estimated value by measuring their absolute distance, whereas the ASAD evaluates from the perspective of spectral similarity, with a smaller value indicating a smaller difference. Both of the aforementioned metrics are appropriate for evaluating the gap between the estimated endmember abundance and the ground truth (GT). RMSE and ASAD are defined as follows:
R M S E = i N ( A P i A T i ) 2 N
A S A D = c o s 1 ( T T P T P )    
where N denotes the dataset size, true abundance is represented by T , while predicted abundance is represented by P .
During the hyperspectral classification process, several regularly adopted classification accuracy evaluation criteria, including precision, recall, accuracy, F1-score and MIoU, are used to assess the comparative approaches. A comprehensive analysis of classification accuracy metrics is presented in [58]. These metrics have the following definitions:
P = T P T P + F P
R = T P T P + F N    
O A = T P + T N T P + T N + F P + F N    
F = 2 1 Recall + 1 Precision    
M I o U = 1 k + 1 i = 0 k T P F N + F P + T P    
where P , R , F , O A and M I o U represent precision, recall, F1-score, overall accuracy and mean intersection over union, respectively. TP, TN, FP and FN represent true positive, true negative, false positive and false negative, respectively, for each class k.

4.3. Abundance Map Estimation

In this section, the presented model is implemented to Jasper Ridge and Urban datasets for the purposes of quantitative and visual analysis. In this step, the estimated abundances of each endmember are compared to the corresponding GT. Several 3D-CNN autoencoders with similar structures are also employed to evaluate the abundance estimation performance of the proposed model. Then, classification applications based on abundance results are carried out, using multiple semi-supervised and supervised models for classification accuracy comparison.

4.3.1. Abundance Map Estimation for Jasper Ridge Dataset

The estimated abundance and the reference GT abundance for the Jasper Ridge dataset are shown in Figure 6. At first glance, all algorithms work quite well in this task, indicating that the 3D-CNN-based autoencoder has great potential for HU. However, through careful comparison of the estimation error (EE) map, it is not difficult to find that the proposed method has a superior visual performance when compared to the method that was used for comparison.
Table 4, with the best results highlighted in bold, indicates that CACAE has higher prediction accuracy for most endmembers, with smaller RMSE and ASAD. For water objects, both CACAE and CCAE achieved good and close estimation results for abundance. It is worth noting that the performance of the model with HSI cube input is superior to that with central pixels as the input. In addition, under the assumption of the same input (cube or pixel), the performance of the attention-based model performs better in most cases, as evidenced by and highlighted in the Urban dataset.

4.3.2. Abundance Map Estimation for Urban Dataset

The Urban dataset, with more endmembers and a larger image size, is considered to provide a more objective assessment compared to the Jasper Ridge dataset. Figure 7 and Figure 8 show the GT abundance and the estimated abundance, respectively, for the Urban dataset, for a straightforward visual comparison. When compared to the other 3D-CNN models, the experimental results show that CACAE achieves better performance, with lower EE and abundance estimation results that are most similar to the GT. As shown in Figure 8a,e,f, the performance enhancement achieved by the present work is clearly apparent on the Asphalt, road, Metal and Dirt EE maps.
Table 5 reports the RMSE and ASAD of each endmember for the Urban dataset. It shows that CACAE has the lowest estimation error according to the RMSE and ASAD values for all endmembers, with values of 0.83 × 10−1 and 2.50 × 10−1, respectively. Once again, the attention-based model still outperforms the others. Additionally, it is observed that the models that take the HIS cube as input lead to better results in most cases.

4.4. Classification

In this section, classification applications based on abundance results are carried out, using multiple semi-supervised and supervised models for classification accuracy comparison.

4.4.1. Ground Object Classification for Urban Dataset

Feature extraction is crucial for hyperspectral analysis. Over the past few years, as a result of the practical learning capabilities of CNNs, end-to-end supervised deep learning models have shown very competitive results in hyperspectral classification. However, since hyperspectral classification is a small-sample problem in most cases, the application of supervised classification is more labor-intensive and prone to overfitting. The endmember abundance obtained through HU has become an effective spatial–spectral feature source for semi-supervised classification. The semi-supervised classification process recommended in this paper is shown in Figure 9.
Based on the encoder of CACAE, trained through self-supervised learning, the Urban dataset is predicted to obtain its abundance estimation map. Then, we use a traditional classifier to calculate the maximum abundance in each abundance feature category as the label of the classification result. Comparative experiments are performed on the classic semi-supervised and supervised methods (Figure 10). The figure demonstrates that the CNN-based method has visual performance comparable to the conventional method but is improved greatly. There are obvious misclassifications in the SID results. FCLS has poor classification performance due to inaccurate abundance estimation. SAM is slightly better than the former two methods but misclassifies the Soil and Metal categories more seriously. It is clear from careful observation that the visual effects of PACAE and CACAE are more accurate than the basic CNN model.
Table 6 records the quantitative evaluation results of all six models. The results reveal that the proposed method CACAE achieves the best outcomes across all metrics, and is slightly higher than the PACAE model, indicating that using image cubes as input data can provide rich spectral–spatial information that is more conductive to abundance estimation and classification. It is noticed that, in the case of 1/10 sample data, the accuracy of the proposed semi-supervised classification methods outperforms that of the CNN-based supervised classification model. Furthermore, the MIoU metrics have improved by nearly 7%, indicating that the proposed method has better performance under the conditions of relatively sufficient samples.

4.4.2. Ground Object Classification for Shenyang Dataset

To further analyze the benefits of the proposed method for few-shot hyperspectral classification, we employ a real hyperspectral dataset from Shenyang, acquired by the airborne hyperspectral sensor AMMIS, for testing purposes. In the experiment, the classification performances of supervised classification methods (Basic-CNN and SVM) and semi-supervised classification methods (CACAE and PACAE) were compared at various sampling rates. In the supervised classification method, the performance of Basic-CNN is superior to that of the traditional machine learning method SVM, particularly under the condition of relatively sufficient samples, but both are controlled by the sample size, which is consistent with our prior knowledge (Figure 11). The superior performance of CACAE and PACAE proposed in this paper (Table 7) validates the earlier conclusion. In addition, their performance did not degrade significantly as the sampling rate decreased, suggesting that the proposed method is more robust. The MIoU value of the CACAE drops by only 1.7% when the sampling rate is decreased from 1/10 to 1/2000.
Figure 12 illustrates the classification outcomes predicted by each of the four methods at varying sampling rates. It is evident that the prediction results of CACAE and PACAE are essentially unaffected by the sampling rate, and that the overall results are relatively consistent with the GT. With a higher sampling rate, the supervised classification method is more competitive, but as the sample size decreases, errors and omissions gradually appear, resulting in a decline in classification performance. Since CACAE uses more spatial information as input, the classification results are smoother and less noisy than those of other methods.

5. Conclusions

In this contribution, we present a new semi-supervised pipeline for few-shot hyperspectral classification, where endmember abundance maps obtained by HU are treated as latent features for classification. It includes two main processes. First, it uses 3D-CNN to build a self-supervised learning model and realizes the non-negativity and summation constraints through the NSC layer and realizes the extraction of endmember abundance information with given endmembers. Secondly, the abundance map can be treated as a diagnostic spatial–spectral feature, enabling claffication with only a small number of samples.
The first experiment is designed to verify the effectiveness of the proposed model for abundance map extraction. In this experiment, the performance of CACAE and other 3D-CNN methods (including PCAE, CCAE and PACAE) for HU is assessed (Section 4.3). The results suggest that the proposed model is capable of accurately estimating the abundance map of a given endmember with the lowest RMSE and ASAD. Additionally, the two strategies of utilizing the attention mechanism and taking an image cube as input are useful for improving the estimation accuracy. The second experiment tests the performance of proposed methods and other algorithms (including SID, FCLS, SAM, Basic-CNN and PACAE) for HSI classification (Section 4.4.1). The experimental results demonstrated that the performance of the proposed model not only outperformed traditional unsupervised and semi-supervised classification methods but also exceeded that of CNN-based supervised classification models, even when samples were relatively sufficient. It also illustrates the effectiveness of the abundance maps for HSI classification. The last experiment is designed to assess the robustness of the proposed model with a small sample size. The traditional machine learning algorithms SVM and Basic-CNN are employed as comparison models (Section 4.4.2). The proposed model shows advantages over the supervised classification model at all levels of sampling rates and shows stable and robust characteristics such that the model’s performance is not restricted by the number of samples. These three experiments fully prove the potential of the proposed method in few-shot classification.
The advantage of the proposed method is that it combines a 3D-CNN with HU theory and uses the estimated endmember abundance as the input diagnostic feature for classification, which effectively improves the classification accuracy and robustness of the model. Because sufficient data are required to train a model with a large number of parameters, which tends to result in overfitting problems with small sample sizes, the suggested model employs a lightweight architecture design. The CNN model used for comparison also uses a similar backbone structure to make the comparison more fair and to avoid the difference in network structure affecting the reliability of the experiment.
The primary limitation of this method is that it is still dependent on the precision of the given endmembers. Although there are methods such as N-FINDR and VCA that can estimate endmembers, there is still room for improvement in accuracy. On the other hand, however, the current loss design makes it difficult to distinguish targets with very small spectral differences. Future research will focus on utilizing the autoencoder network for unsupervised endmember abundance estimation and developing a new loss function for more accurate classification. The hyperspectral datasets used in this paper and example code are available in the GitHub repository via https://github.com/lichunyu123/3DCAE_Hyper (accessed on 1 December 2022).

Author Contributions

Conceptualization, C.L.; methodology, C.L. and J.Y.; article writing and figure drawing, C.L.; writing—review and editing, C.L., R.C. and J.Y.; sample collection, C.L. and J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Key Research and Development Program of China (2021YFC3000400).

Acknowledgments

The authors would like to thank Shanghai HeyWhale information Technology Company for providing online GPU computing resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Dobigeon, N.; Parente, M.; Du, Q.; Gader, P.; Chanussot, J. Hyperspectral unmixing overview: Geometrical, statistical, and sparse regression-based approaches. IEEE J. Sel. Top. Appl. Earth Observat. Remote Sens. 2012, 5, 354–379. [Google Scholar] [CrossRef] [Green Version]
  2. Hecker, C.; Van der Meijde, M.; van der Werff, H.; van der Meer, F.D. Assessing the influence of reference spectra on synthetic SAM classification results. IEEE Trans. Geosci. Remote Sens. 2008, 46, 4162–4172. [Google Scholar] [CrossRef]
  3. Chang, C.I. An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis. IEEE Trans. Inf. Theory. 2000, 46, 1927–1932. [Google Scholar] [CrossRef] [Green Version]
  4. Melgani, F.; Bruzzon, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  5. Gislason, P.O.; Benediktsson, J.A.; Sveinsson, J.R. Random forests for land cover classification. Pattern Recognit. Lett. 2006, 27, 294–300. [Google Scholar] [CrossRef]
  6. Benediktsson, J.A.; Swain, P.H.; Ersoy, O.K. Neural network approaches versus statistical methods in classification of multisource remote sensing data. IEEE Trans. Geosci. Remote Sens. 1989, 2, 489–492. [Google Scholar]
  7. Plaza, A.; Martínez, P.; Plaza, J.; Pérez, R. Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations. IEEE Trans Geosci. Remote Sens. 2005, 43, 466–479. [Google Scholar] [CrossRef] [Green Version]
  8. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based K-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  9. Tarabalka, Y.; Fauvel, M.; Chanussot, J.; Benediktsson, J.A. SVM-and MRF-based method for accurate classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2010, 7, 736–740. [Google Scholar] [CrossRef] [Green Version]
  10. Li, W.; Chen, C.; Su, H.; Du, Q. Local binary patterns and extreme learning machine for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3681–3693. [Google Scholar] [CrossRef]
  11. Jia, S.; Hu, J.; Xie, Y.; Shen, L.; Jia, X.; Li, Q. Gabor cube selection based multitask joint sparse representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2016, 51, 3174–3187. [Google Scholar] [CrossRef]
  12. Shang, X.; Song, M.; Wang, Y.; Yu, C.; Yu, H.; Li, F.; Chang, C.I. Target-constrained interference-minimized band selection for hyperspectral target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6044–6064. [Google Scholar] [CrossRef]
  13. Sun, K.; Geng, X.; Chen, J.; Ji, L.; Tang, H.; Zhao, Y.; Xu, M. A robust and efficient band selection method using graph representation for hyperspectral imagery. Int. J. Remote Sens. 2016, 37, 4874–4889. [Google Scholar] [CrossRef]
  14. Wang, P.; Wang, L.; Leung, H.; Zhang, G. Super-resolution mapping based on spatial-spectral correlation for spectral imagery. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2256–2268. [Google Scholar] [CrossRef]
  15. Li, X.; Ding, M.; Pižurica, A. Deep feature fusion via two-stream convolutional neural network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2615–2629. [Google Scholar] [CrossRef] [Green Version]
  16. Zhang, B.; Zhao, L.; Zhang, X. Three-dimensional convolutional neural network model for tree species classification using airborne hyperspectral images. Remote Sens. Environ. 2020, 247, 111938. [Google Scholar] [CrossRef]
  17. Li, S.T.; Song, W.W.; Fang, L.Y.; Chen, Y.S.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  18. Yu, J.; Li, Y.; Zheng, X.; Zhong, Y.; He, P. An Effective Cloud Detection Method for Gaofen-5 Images via Deep Learning. Remote Sens. 2020, 12, 2106. [Google Scholar] [CrossRef]
  19. Audebert, N.; Le Saux, B.; Lefèvre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef] [Green Version]
  20. Wambugu, N.; Chen, Y.; Xiao, Z.; Tan, K.; Wei, M.; Liu, X.; Li, J. Hyperspectral image classification on insufficient-sample and feature learning using deep neural networks: A review. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102603. [Google Scholar] [CrossRef]
  21. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  22. Nowakowski, A.; Mrziglod, J.; Spiller, D.; Bonifacio, R.; Ferrari, I.; Mathieu, P.P.; Garcia-Herranz, M.; Kim, D.H. Crop type mapping by using transfer learning. Int. J. Appl. Earth Obs. Geoinf. 2021, 98, 102313. [Google Scholar] [CrossRef]
  23. Briechle, S.; Krzystek, P.; Vosselman, G. Classification of tree species and standing dead trees by fusing UAV-based lidar data and multispectral imagery in the 3D deep neural network PointNet++. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 2, 203–210. [Google Scholar] [CrossRef]
  24. Liu, B.; Yu, X.; Zhang, P.; Tan, X.; Yu, A.; Xue, Z. A semi-supervised convolutional neural network for hyperspectral image classification. Remote Sens. Lett. 2017, 8, 839–848. [Google Scholar] [CrossRef]
  25. Mei, S.; Ji, J.; Hou, J.; Li, X.; Du, Q. Learning sensor-specific spatial-spectral features of hyperspectral images via convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4520–4533. [Google Scholar] [CrossRef]
  26. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  27. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial–spectral feature learning by 3D convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  28. Zhu, Y.; Yeung, C.H.; Lam, E.Y. Digital holographic imaging and classification of microplastics using deep transfer learning. Appl. Opt. 2021, 60, A38–A47. [Google Scholar] [CrossRef]
  29. Guan, P.; Lam, E.Y. Cross-domain Contrastive Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  30. Dobigeon, N.; Tourneret, J.Y.; Richard, C.; Bermudez, J.C.M.; McLaughlin, S.; Hero, A.O. Nonlinear unmixing of hyperspectral images: Models and algorithms. IEEE Signal Process. Mag. 2014, 31, 82–94. [Google Scholar] [CrossRef] [Green Version]
  31. Winter, M.E. N-FINDR: An algorithm for fast autonomous spectral end-member determination in hyperspectral data. In Proc. SPIE. 1999, 3753, 266–276. [Google Scholar]
  32. Nascimento, J.M.; Dias, J.M. Vertex component analysis: A fast algorithm to unmix hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 898–910. [Google Scholar] [CrossRef]
  33. Miao, L.; Qi, H. Endmember extraction from highly mixed data using minimum volume constrained nonnegative matrix factorization. IEEE Trans. Geosci. Remote Sens. 2007, 45, 765–777. [Google Scholar] [CrossRef]
  34. Gao, G.; Gu, Y. Tensorized principal component alignment: A unified framework for multimodal high-resolution images classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 46–61. [Google Scholar] [CrossRef]
  35. Zhang, X.; Sun, Y.; Zhang, J.; Wu, P.; Jiao, L. Hyperspectral Unmixing via Deep Convolutional Neural Networks. IEEE Geoence Remote Sensing Lett. 2018, 15, 1755–1759. [Google Scholar] [CrossRef]
  36. Halimi, A.; Altmann, Y.; Dobigeon, N.; Tourneret, J.-Y. Nonlinear unmixing of hyperspectral images using a generalized bilinear model. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4153–4162. [Google Scholar] [CrossRef] [Green Version]
  37. Licciardi, G.A.; Frate, F.D. Pixel unmixing in hyperspectral data by means of neural networks. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4163–4172. [Google Scholar] [CrossRef]
  38. Zhu, F.; Honeine, P. Biobjective nonnegative matrix factorization: Linear versus kernel-based models. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4012–4022. [Google Scholar] [CrossRef] [Green Version]
  39. Palsson, B.; Ulfarsson, M.O.; Sveinsson, J.R. Convolutional Autoencoder for Spectral-Spatial Hyperspectral Unmixing. IEEE Trans. Geosci. Remote Sens. 2021, 59, 535–549. [Google Scholar] [CrossRef]
  40. Su, Y.; Li, J.; Plaza, A.; Marinoni, A.; Gamba, P.; Chakravortty, S. DAEN: Deep autoencoder networks for hyperspectral unmixing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4309–4321. [Google Scholar] [CrossRef]
  41. Palsson, B.; Sigurdsson, J.; Sveinsson, J.R.; Ulfarsson, M.O. Hyperspectral unmixing using a neural network autoencoder. IEEE Access. 2018, 6, 25646–25656. [Google Scholar] [CrossRef]
  42. Savas, O.; Berk, K.; Bozdagi, A.G. EndNet: Sparse AutoEncoder Network for Endmember Extraction and Hyperspectral Unmixing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 482–496. [Google Scholar]
  43. Qu, Y.; Qi, H. uDAS: An Untied Denoising Autoencoder with Sparsity for Spectral Unmixing. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1698–1712. [Google Scholar] [CrossRef]
  44. Su, Y.; Marinoni, A.; Li, J.; Plaza, J.; Gamba, P. Stacked Nonnegative Sparse Autoencoders for Robust Hyperspectral Unmixing. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1427–1431. [Google Scholar] [CrossRef]
  45. Févotte, C.; Dobigeon, N. Nonlinear hyperspectral unmixing with robust nonnegative matrix factorization. IEEE Trans. Image Process. 2015, 24, 4810–4819. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Khajehrayeni, F.; Ghassemian, H. Hyperspectral Unmixing Using Deep Convolutional Autoencoders in a Supervised Scenario. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 567–576. [Google Scholar] [CrossRef]
  47. Han, M.X.; Cong, R.M.; Li, X.Y.; Fu, H.Z.; Lei, J.J. Joint spatial-spectral hyperspectral image classification based on convolutional neural network. Pattern Recognit. Lett. 2020, 130, 38–45. [Google Scholar] [CrossRef]
  48. Ibarrola-Ulzurrun, E.; Drumetz, L.; Marcello, J.; Gonzalo-Martín, C.; Chanussot, J. Hyperspectral Classification Through Unmixing Abundance Maps Addressing Spectral Variability. IEEE Trans. Geosci. Remote Sens. 2019, 57, 4775–4788. [Google Scholar] [CrossRef]
  49. Xu, X.; Tong, X.; Plaza, A.; Zhong, Y.; Xie, H.; Zhang, L. Using Linear Spectral Unmixing for Subpixel Mapping of Hyperspectral Imagery: A Quantitative Assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1589–1600. [Google Scholar] [CrossRef]
  50. Dópido, I.; Li, J.; Gamba, P.; Plaza, A. A New Hybrid Strategy Combining Semisupervised Classification and Unmixing of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 3619–3629. [Google Scholar] [CrossRef]
  51. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 221–231. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Foody, G.M. Approaches for the production and evaluation of fuzzy land cover classifications from remotely-sensed data. Int. J. Remote Sens. 1996, 17, 1317–1340. [Google Scholar] [CrossRef]
  53. Xue, T.; Wang, Y.; Chen, Y.; Jia, J.; Wen, M.; Guo, R.; Wu, T.; Deng, X. Mixed noise estimation model for optimized kernel minimum noise fraction transformation in hyperspectral image dimensionality reduction. Remote Sens. 2021, 13, 2607. [Google Scholar] [CrossRef]
  54. Jia, J.; Chen, J.; Zheng, X.; Wang, Y.; Guo, S.; Sun, H.; Jiang, C.; Karjalainen, M.; Karila, K.; Duan, Z.; et al. Tradeoffs in the spatial and spectral resolution of airborne hyperspectral imaging systems: A crop identification case study. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  55. Jia, J.; Zheng, X.; Guo, S.; Wang, Y.; Chen, J. Removing stripe noise based on improved statistics for hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  56. HeyWhale Data Mining Platform. Available online: www.modelwhale.com.
  57. Heinz, D.C. Fully constrained least squares linear spectral mixture analysis method for material quantification in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 529–545. [Google Scholar] [CrossRef] [Green Version]
  58. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
Figure 1. Diagram of the proposed 3D convolutional neural network.
Figure 1. Diagram of the proposed 3D convolutional neural network.
Remotesensing 15 00451 g001
Figure 2. The architecture of the channel attention module.
Figure 2. The architecture of the channel attention module.
Remotesensing 15 00451 g002
Figure 3. The Jasper Ridge data shown in RGB form and ground truth endmembers.
Figure 3. The Jasper Ridge data shown in RGB form and ground truth endmembers.
Remotesensing 15 00451 g003
Figure 4. The Urban dataset shown in RGB format as well as GT endmembers.
Figure 4. The Urban dataset shown in RGB format as well as GT endmembers.
Remotesensing 15 00451 g004
Figure 5. The Shenyang dataset shown in RGB format as well as GT endmembers.
Figure 5. The Shenyang dataset shown in RGB format as well as GT endmembers.
Remotesensing 15 00451 g005
Figure 6. Abundance maps estimated by CACAE and the true abundance maps from the Jasper Ridge dataset. (a) Tree; (b) Water; (c) Soil; (d) Road.
Figure 6. Abundance maps estimated by CACAE and the true abundance maps from the Jasper Ridge dataset. (a) Tree; (b) Water; (c) Soil; (d) Road.
Remotesensing 15 00451 g006
Figure 7. Abundance maps estimated by CACAE and the true abundance maps from the Urban dataset. (a) Asphalt; (b) Grass; (c) Tree; (d) Roof; (e) Metal; (f) Dirt.
Figure 7. Abundance maps estimated by CACAE and the true abundance maps from the Urban dataset. (a) Asphalt; (b) Grass; (c) Tree; (d) Roof; (e) Metal; (f) Dirt.
Remotesensing 15 00451 g007
Figure 8. Absolute difference from the abundance of the Urban dataset estimated by different methods. (a) Asphalt; (b) Grass; (c) Tree; (d) Roof; (e) Metal; (f) Dirt.
Figure 8. Absolute difference from the abundance of the Urban dataset estimated by different methods. (a) Asphalt; (b) Grass; (c) Tree; (d) Roof; (e) Metal; (f) Dirt.
Remotesensing 15 00451 g008
Figure 9. An overview of the proposed 3D autoencoder-based semi-supervised classification framework.
Figure 9. An overview of the proposed 3D autoencoder-based semi-supervised classification framework.
Remotesensing 15 00451 g009
Figure 10. The classification results for the Urban dataset by different methods. (a) Original HSI in RGB format; (b) SID; (c) FCLS; (d) SAM; (e) CNN; (f) PACAE; (g) CACAE; (h) Ground Truth.
Figure 10. The classification results for the Urban dataset by different methods. (a) Original HSI in RGB format; (b) SID; (c) FCLS; (d) SAM; (e) CNN; (f) PACAE; (g) CACAE; (h) Ground Truth.
Remotesensing 15 00451 g010
Figure 11. The trend of MIoU at different sampling rates.
Figure 11. The trend of MIoU at different sampling rates.
Remotesensing 15 00451 g011
Figure 12. The classification prediction of all models at different sampling rates. (a) Sampling rate 1:10; (b) Sampling rate 1:100; (c) Sampling rate 1:200; (d) Sampling rate 1:1000; (e) Sampling rate 1:2000.
Figure 12. The classification prediction of all models at different sampling rates. (a) Sampling rate 1:10; (b) Sampling rate 1:100; (c) Sampling rate 1:200; (d) Sampling rate 1:1000; (e) Sampling rate 1:2000.
Remotesensing 15 00451 g012
Table 1. The structure of the CACAE model.
Table 1. The structure of the CACAE model.
LayersKernelFilterActivationFeature
3DConv(3, 3, 8)32LeakRelu(3, 3, C-7, 32)
3DConv(3, 3, 8)16LeakRelu(1, 1, C-14, 16)
3DConv(1, 1, 8)8LeakRelu(1, 1, C-21, 8)
3DConv(1, 1, 8)2LeakRelu(1, 1, C-28, 2)
Attention
Flatten---(C-28) × 2
Dense-32LeakRelu32
Dense-NLeakReluN
NSC--softmaxN
Dense-3-ClinearC
Table 2. Comparison table of the differences between all 3DCAE-based models.
Table 2. Comparison table of the differences between all 3DCAE-based models.
ModelInput DataAttentionBasic Model
CACAECubeYes3D-CAE
PACAEPixelYes3D-CAE
PCAEPixelNo3D-CAE
CCAECubeNo3D-CAE
Table 3. The structure of the basic CNN model.
Table 3. The structure of the basic CNN model.
LayersKernelFilterActivationFeature
Conv2D(3, 3)32Relu(3, 3, C-7)
Conv2D(3, 3)16Relu(1, 1, C-14)
Conv2D(1, 1)8Relu(1, 1, C-21)
Conv2D(1, 1)2Relu(1, 1, C-28)
Flatten---(C-28) × 2
Dense-32Relu32
Dense-NsoftmaxN
Table 4. RMSE and ASAD results for all endmembers of the Jasper Ridge dataset (×10−1).
Table 4. RMSE and ASAD results for all endmembers of the Jasper Ridge dataset (×10−1).
DatasetEndmemberAlgorithmsPCAECCAEPACAECACAE
JasperTreeRMSE0.4680.4220.5390.419
ASAD0.9200.8361.0080.830
WaterRMSE1.1460.8310.9750.835
ASAD1.9151.3651.5941.380
DirtRMSE1.1370.8741.0510.792
ASAD2.2692.0782.3321.910
RoadRMSE0.9080.8180.7630.741
ASAD3.3923.0463.3922.657
SumRMSE0.9550.7590.8560.716
ASAD2.2401.7732.0041.671
Table 5. RMSE and ASAD results for all endmembers of the Urban dataset (×10−1).
Table 5. RMSE and ASAD results for all endmembers of the Urban dataset (×10−1).
DatasetEndmemberAlgorithmsPCAECCAEPACAECACAE
UrbanRoadRMSE1.6941.2431.6671.017
ASAD4.1893.2564.1082.711
GrassRMSE1.4561.0711.1360.981
ASAD2.8052.1622.1831.980
TreeRMSE1.0380.8810.8720.833
ASAD2.4822.1232.0992.006
RoofRMSE0.5380.4690.4590.450
ASAD2.4632.1712.1432.100
MetalRMSE1.1660.7430.7530.536
ASAD9.4855.5756.0023.780
DirtRMSE1.3471.1171.5690.977
ASAD4.5383.3613.7712.760
SumRMSE1.2600.9561.1600.830
ASAD3.8502.8933.5322.503
Table 6. Quantitative analysis of prediction results on the Urban dataset.
Table 6. Quantitative analysis of prediction results on the Urban dataset.
MethodsPrecisionRecallOAMIoUF1-Score
SID0.51 0.67 0.63 0.40 0.54
FCLS0.62 0.80 0.75 0.51 0.66
SAM0.85 0.81 0.83 0.68 0.80
CNN0.84 0.86 0.90 0.75 0.85
PACAE0.89 0.88 0.90 0.80 0.88
CACAE0.91 0.89 0.92 0.82 0.90
Table 7. Comparison table of MIoU predicted by all models at different sampling rates.
Table 7. Comparison table of MIoU predicted by all models at different sampling rates.
MethodsMIoU at Different Sample RatesMean
1/20001/10001/5001/2001/1001/501/10
SVM0.6590.6830.7010.7260.7510.8010.8480.752
Basic-CNN0.6650.6870.7500.7390.7950.8430.8930.784
PACAE0.8740.8730.8800.8830.8870.8900.8900.884
CACAE0.8730.8850.8820.8890.8920.8920.8990.890
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, C.; Cai, R.; Yu, J. An Attention-Based 3D Convolutional Autoencoder for Few-Shot Hyperspectral Unmixing and Classification. Remote Sens. 2023, 15, 451. https://doi.org/10.3390/rs15020451

AMA Style

Li C, Cai R, Yu J. An Attention-Based 3D Convolutional Autoencoder for Few-Shot Hyperspectral Unmixing and Classification. Remote Sensing. 2023; 15(2):451. https://doi.org/10.3390/rs15020451

Chicago/Turabian Style

Li, Chunyu, Rong Cai, and Junchuan Yu. 2023. "An Attention-Based 3D Convolutional Autoencoder for Few-Shot Hyperspectral Unmixing and Classification" Remote Sensing 15, no. 2: 451. https://doi.org/10.3390/rs15020451

APA Style

Li, C., Cai, R., & Yu, J. (2023). An Attention-Based 3D Convolutional Autoencoder for Few-Shot Hyperspectral Unmixing and Classification. Remote Sensing, 15(2), 451. https://doi.org/10.3390/rs15020451

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop