Next Article in Journal
Autonomous Extraction Technology for Aquaculture Ponds in Complex Geological Environments Based on Multispectral Feature Fusion of Medium-Resolution Remote Sensing Imagery
Previous Article in Journal
Sea Surface Temperature Prediction Using ConvLSTM-Based Model with Deformable Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical Spectral–Spatial Transformer for Hyperspectral and Multispectral Image Fusion

1
School of Software Engineering, Tongji University, Shanghai 200070, China
2
School of Computer Science, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(22), 4127; https://doi.org/10.3390/rs16224127
Submission received: 18 September 2024 / Revised: 26 October 2024 / Accepted: 4 November 2024 / Published: 5 November 2024
(This article belongs to the Special Issue Remote Sensing Image Thorough Analysis by Advanced Machine Learning)

Abstract

:
This paper presents the Hierarchical Spectral–Spatial Transformer (HSST) network, a novel approach applicable to both drone-based and broader remote sensing platforms for integrating hyperspectral (HSI) and multispectral (MSI) imagery. The HSST network improves upon conventional multi-head self-attention transformers by integrating cross attention, effectively capturing spectral and spatial features across different modalities and scales. The network’s hierarchical design facilitates the extraction of multi-scale information and employs a progressive fusion strategy to incrementally refine spatial details through upsampling. Evaluations on three prominent hyperspectral datasets confirm the HSST’s superior efficacy over existing methods. The findings underscore the HSST’s utility for applications, including drone operations, where the high-fidelity fusion of HSI and MSI data is crucial.

1. Introduction

Hyperspectral imagery (HSI), embodying an array of slender spectral bands that span from visible to near-infrared wavelengths, enables the meticulous discernment of terrestrial object compositions. This proficiency proves to be especially beneficial in fields such as environmental surveillance and precision agriculture [1]. The utility of HSI spans across various domains, including land cover classification and anomaly detection, thanks to their detailed object attribute characterization [2,3,4,5]. Nevertheless, the high spectral resolution inherent in HSI often comes at the expense of spatial resolution, a trade-off imposed by the constraints of imaging platforms, including those mounted on drones. The scattering of electromagnetic waves with narrow bandwidths into the instantaneous field of view typically requires a compromise in spatial resolution to maintain an acceptable signal-to-noise ratio, thereby limiting the broader application of HSI [6]. Consequently, there is a pressing need for research aimed at algorithmically fusing low-resolution hyperspectral images (LR-HSI) with high-resolution multispectral images (HR-MSI) to generate high-resolution hyperspectral images (HR-HSI), particularly for use in drone-based platforms where enhanced spatial detail is critical [7]. This integration is essential for maximizing the potential of HSI in both aerial and terrestrial applications, ensuring that drones can effectively utilize the resulting HR-HSI for a wide range of tasks.
Several machine learning techniques have been proposed for the integration of LR-HSI (low-resolution hyperspectral imaging) and HR-MSI (high-resolution multispectral imaging), including methods based on matrix and tensor factorization [8,9,10,11,12]. However, many of these techniques rely on manually designed priors, which not only limit their representational capacity but also prove to be time-consuming. The swift proliferation of deep learning has illustrated its immense potential for the fusion of hyperspectral and multispectral imaging, owing to its sturdy capabilities in the extraction of features. This technique employs a multi-tiered, deep-seated neural network to discern the correlation among high-resolution multispectral imaging, low-resolution hyperspectral imaging, and the corresponding high-resolution hyperspectral imaging, thereby aiding the accomplishment of fusion tasks. Fusion methodologies predicated on deep learning not only extrapolate information from the input imagery but also utilize learned correlations as prior cognizance to reconstruct spectral and spatial details that are absent in multispectral and hyperspectral imaging. As a result, in contrast to conventional fusion techniques that rely on manually stipulated prior information, fusion methods underpinned by deep learning exhibit superior efficacy. In recent epochs, several fusion techniques predicated on convolutional neural networks have been conceived. To give an example, Yang et al. [13] proposed a fusion technique that amalgamates convolutional neural networks and spatial attention to extract intricate textures and enhance spatial structure. Cai et al. [14] incorporated a super-resolution module and progressive learning into their network, which enables the capture of spatial details at varying scales and their integration into upsampled multispectral images. Despite the enhancements in generalization capabilities by current convolutional neural network-based fusion algorithms [15,16,17], these algorithms fall short in effectively exploiting spatial location information and extracting long-range dependencies in images, thereby resulting in a deficiency of global context information. Inspired by the triumphant implementation of transformers in NLP, scholars have begun to suggest the use of vision transformers for fusion tasks. However, most extant transformer-based hyperspectral and multispectral imaging fusion methods extract features from a singular modality. This approach regrettably neglects the interplay between spatial and spectral modalities.
The observations previously discussed lay the groundwork for the introduction of a novel Hierarchical Spectral–Spatial Transformer (HSST) network. The HSST is a two-branch network that incorporates the self-attention mechanism of the transformer to extract and merge spectral features of HSI with spatial features of MSI. To effectively leverage the extensive spatial information in remote sensing images, a hierarchical structure is employed to facilitate multi-scale information extraction. For the reconstruction of HR-HSI, a hierarchical progressive fusion is employed to gradually restore spatial detail information through progressive upsampling, thereby harnessing the acquired multi-level feature representation. The main contributions of this paper are described in detail as follows:
  • We introduce a Hierarchical Spectral–Spatial Transformer network (HSST) for the fusion of HSI and MSI. The HSST is designed to extract and merge deep spectral and spatial features via hierarchical Spectral–Spatial Transformers and subsequently reconstruct HR-HSI through a process of hierarchical progressive fusion.
  • We also propose the use of a Hierarchical Spectral–Spatial Transformer to more effectively capture cross-modality spectral and spatial features at multiple scales. In addition to the traditional multi-head self-attention transformers, cross attention is incorporated to enhance the extraction of cross-modality features.
  • To optimize the spatial details of the reconstructed HR-HSI, hierarchical progressive fusion is proposed to gradually recover spatial detail information through progressive upsampling and fusion. This cumulative process facilitates the gradual reconstruction of the HR-HSI result.
The remainder of this article is as follows: Section 2 provides an overview of relevant HSI and MSI fusion methods and a detailed description of the proposed HSST. Section 3 presents and analyses the experimental results on three datasets. Section 4 consists of discussions and ablation studies. Finally, Section 5 concludes the paper.

2. Materials and Methods

2.1. Related Works

2.1.1. HSI and MSI Fusion

The current fusion strategies for hyperspectral images (HSI) and multispectral images (MSI) primarily fall into four categories: matrix factorization, tensor factorization, pan-sharpening, and deep learning.
Matrix factorization-based techniques transform the 3D HSI into a 2D matrix. Thereafter, an HR-HSI is concocted utilizing the endmember matrix and abundance matrix extricated from the LR-HSI and HR-MSI. For example, the CNMF [18] method employs non-negative matrix factorization to disintegrate HR-MSI and LR-HSI into mixed pixels, thereby engendering a superior HR-HSI through the utilization of the abundance matrix of LR-HSI and the endmember matrix of HR-MSI.
Tensor factorization-based methods are utilized to maintain the spatial and spectral structure of images, as opposed to reshaping them into matrices. Two frequently used decompositions in the fusion of HSI and MSI are the Tucker decomposition and the Canonical Polyadic decomposition. Dian et al. [19] introduced a nonlocal sparse tensor factorization method for semi-blind fusion of HSI and MSI. To mitigate computational strain, Kanatsoulis et al. [20] implemented CP decomposition on the HR-HSI, obtaining each factor matrix by resolving the least squares equation.
Pan-sharpening methods integrate the fusion of HSI and MSI images. Grohnfeldt et al. [21] introduce a sparse representation (SR)-based pan-sharpening method for HSI and MSI fusion. Although it achieves satisfactory performance with a limited number of MSI bands, it fails to yield desirable outcomes when the number of bands increases due to the diminished correlation between the missing bands of MSI and the high-resolution images.
With the emergence of deep learning, it is now feasible to learn all parameters from training data using deep learning networks, eliminating the necessity for assumptions about the images [22]. Bearing this in mind, Dian et al. [23] propose a novel HSI sharpening method called DHSI for the fusion of HSI and MSI data. The method employs a deep residual network to learn image priors, thereby obviating the need for manually crafted priors. Palsson et al. [24] have proposed a method that utilizes a trained three-dimensional convolutional neural network to acquire filters for effectively merging MSI and HSI. To address the computational complexity associated with the 3D CNN, they have employed principal component analysis (PCA) as a means of reducing the dimensionality prior to fusion [25]. In a similar vein, Zheng et al. [26] have developed EC-FTN that aims to preserve low-level structural details, including sharp edges. However, it is worth noting that many of the existing learning-based approaches are supervised in nature, requiring a substantial amount of training data, which can pose challenges in real-world applications.

2.1.2. Hyperspectral Image Transformer

The proliferation of convolutional neural network (CNN)-based fusion methodologies for amalgamating hyperspectral and multispectral images is noteworthy. Nonetheless, the limiting receptive field of CNNs poses a challenge in extracting comprehensive information from images. This is particularly crucial in hyperspectral imaging (HSI) where strong correlations in the spectral dimension necessitate the extraction of global characteristics for enhanced fusion performance. The transformer model, renowned for its self-attention mechanism that accommodates long-range information, has gained momentum in various applications. In the domain of HSI processing, the transformer model has demonstrated its prowess in handling sequential data. For example, Hong et al. [27] proposed SpectralFormer, a primary network for HSI classification, adept at learning locally sequential spectral information from adjacent HSI bands and generating group-wise spectral embeddings. He et al. [28] introduced a spectral–spatial transformer classification network, which combines a well-structured CNN for spatial feature extraction with an improved transformer to capture sequential spectral relationships. Selen et al. [29] developed a spectral-swin transformer (SpectralSWIN) classification network, employing a spectral-swin module to simultaneously process spatial and spectral features. Transformers have gained significant traction in the realm of HSI reconstruction. Cai et al. [30] were the pioneers in proposing a transformer-based method for HSI reconstruction. They utilized the feature map of each spectral channel to compute self-attention. Similarly, Bandara et al. [31] employed the self-attention mechanism of the transformer to transfer high-resolution textual features to low-resolution features for pan-sharpening.

2.2. Proposed Method

Our network utilizes a bifurcated structure, as illustrated in Figure 1. The input HSI and MSI experience a corresponding upsampling and downsampling, respectively, facilitated by linear and convolutive procedures. This process extracts spatial information from low-level detailed features alongside semantic information from high-level semantic features. The Spectral–Spatial Transformer is harnessed to execute comprehensive feature extraction, foster meaningful interplay of information, and facilitate the amalgamation of feature maps on a homogeneous scale. The image reconstruction incorporates a hierarchical progressive fusion procedure, which is designed to incrementally combine extracted features across varying scales.
Let Y 0 R r × c × S represent the input LR-HSI, where r, c, and S denote the number of rows, columns, and spectral bands respectively. Similarly, let Z 0 R R × C × s denote the input HR-MSI, where R, C, and s correspond to the number of rows, columns, and spectral bands respectively. Initially, we perform upsampling on the LR-HSI and downsampling on the HR-MSI to generate Y u p s a m p l e R R × C × S and Z d o w n s a m p l e R r × c × s , respectively, employing the bilinear interpolation method, which can be expressed as follows:
Y u p s a m p l e U P S Y 0
Z d o w n s a m p l e D O W N S Z 0
where UPS and DOWNS denote the functions of upsampling and downsampling. Subsequently, we perform concatenation of Y 0 and Z d o w n s a m p l e to get Y c o n c a t R h × w × S + s , and concatenate Z and Y u p s a m p l e to get Z c o n c a t R H × W × s + S , which can be written as
Y c o n c a t = c o n c a t Y 0 , Z d o w n s a m p l e
Z c o n c a t = c o n c a t Z 0 , Y u p s a m p l e
where concat refers to the concatenation operation within the channel dimension. This cross-modality concatenation enables the interaction of cross-modality information between the two branches.
The intricate computational convolutions of the transformer are commensurate with the sequence length, rendering it impractical to condense the input image into a sequence for transformer assimilation. To mitigate this predicament, the Vision Transformer (ViT) [32] advocates dividing the image into static-size segments. Initially, LR-HSI Y c o n c a t is upsampled to generate Y 1 R H × W × C , which is of the same scale as the HR-MSI Z 1 . The same downsampling operation is then performed on Y 1 and Z 1 to produce a pair of feature maps of identical scale, Y 2 R H 4 × W 4 × 2 C and Z 2 R H 4 × W 4 × 2 C . The second downsampling can be represented as Y 3 R H 8 × W 8 × 4 C and Z 3 R H 8 × W 8 × 4 C . Throughout the gradual downscaling procedure, numerous local spatial features are harnessed. Subsequently, Y n and Z n are inputted into Spectral–Spatial Transformers to further apprehend the long-range correlation in a global context.

2.2.1. Spectral–Spatial Transformer

In order to integrate data from the spectral and spatial modalities and construct a comprehensive image representation, a Spectral–Spatial Transformer (SST) was developed. This concept is visually depicted in Figure 2. In the preliminary stage, the LR-HSI and HR-MSI feature maps are subjected to a linear mapping procedure. Following this, the overall dependency of features on both modalities is modeled to facilitate the fusion of cross-modality data, utilizing a fusion attention block. This block serves to coalesce the interactive data interwoven between the two modalities. Subsequently, the Multilayer Perceptron (MLP) is structured as a bi-layer perceptron, supplemented with a hidden layer expansion ratio. To conclude, Layer Normalization (LN) is executed.
We began with two feature maps of identical dimensions, denoted as F R h × w × C . Subsequently, an LN was carried out on these features:
x 1 = L N L P F 1 , x 2 = L N L P F 2
where LP denotes the linear projection operation, and x 1 , x 2 R d × h × w signify the feature embeddings. LN signifies the layer normalization.
The Spectral–Spatial Fusion Self-Attention block is crafted with the intention of capturing the intricate interdependence of spectral and spatial feature maps within an image, wherein each map contains distinct semantic information at diverse scales. This design is specifically customized by calculating self-attention across spatial and spectral dimensions, thereby revealing the correlations present. Initially, we feed the feature embedding Yn into the system to obtain the query matrix Q, the key matrix K, and the value matrix V through a trainable linear projection:
Q = x n W Q , K = x n W K , V = x n W V
where W q , W k , W v are learnable projection matrices. Then, given two feature vectors x 1 and x 2 , the fusion attention block can be expressed as:
y 1 = a t t e n t i o n Q 1 , K 1 , V 2 = s o f t m a x Q 1 K 1 T d k V 2
y 2 = a t t e n t i o n Q 2 , K 2 , V 1 = s o f t m a x Q 2 K 2 T d k V 1
where Q, K, and V denote the query, the key, and the value respectively, while y 1   and y 2 represent the output feature maps. This cross-attention mechanism encourages a more effective interaction between the two feature maps in the fusion task, resulting in an improved fusion outcome. Once the feature map is processed through another layer of normalization and MLP, its dimensions cannot align with the subsequent network structure. To rectify this, we employ a feature mapping module to reconfigure the output sequences into a standard 3D feature map of dimensions H × W × d. Subsequently, the channel count of the feature maps is reduced by the convolution operation, resulting in feature maps z 1 and z 2 that share the same dimensions as feature maps F 1 and F 2 . The final step involves combining the two feature maps to generate the feature map F o u t , thereby enhancing information fusion. The above process can be described as:
z 1 = F M L N y 1 + M L P L N y 1 , z 2 = F M L N y 2 + M L P L N y 2
F o u t = z 1 + z 2
where LN represents the layer normalization, FM represents the feature mapping, and MLP refers to the multilayer perceptron network.

2.2.2. Hierarchical Progressive Fusion

The quality of reconstructed images is profoundly influenced by the precise restoration of spatial intricacies. Furthermore, the efficacy of feature fusion utilization is pivotal for acquiring a comprehensive multi-level feature representation. Consequently, we have engineered a feature amalgamation reconstruction module to synchronously merge these three distinctive feature maps. This method incrementally recovers spatial detail intelligence via sequential elevation in sampling. Our methodology merges the spatial detail data from the reduction layer with the introductory information from the amplified layer. This initiative enhances the image’s semantic features while concurrently maintaining the spatial particulars of each band, thereby aiding the systematic reconstruction of the HR-HSI result.
f u s i o n n = S t r i p C o n v C o n c a t Y n , Z n , S S T Y n , Z n
where StripConv is the Strip Convolution Block, the Concat is the concatenation operation, and SST is the Spectral–Spatial Transformer.
The strip convolution block captures long-range context information from four different directions: horizontal, vertical, left diagonal, and right diagonal. In the strip convolution block, F is input to four different shapes of strip convolution paths after a 1 × 1 convolution. The output feature maps of the four paths are concatenated. Then, the upsampling operation and a 1 × 1 convolution are performed to obtain the final output of the strip convolution block. Let w R 2 k + 1 be a strip convolution filter of size 2 k + 1 , and let D = ( D h , D w ) represent the direction of filter w. Let H be the result of strip convolution. The strip convolution can be defined as follows:
H D i , j = ( F w ) D [ i , j ]

3. Results

3.1. Experimental Settings

The performance of the HSST network was evaluated on three openly accessible remote sensing datasets. Specifically, we employed the Pavia Center dataset, the Botswana dataset, and the Urban dataset for conducting experiments. We conducted comparative experiments between HSST and five SOTA fusion models. The five fusion algorithms were CNMF [18], MSD-CNN [33], TFNET [34], SSF-CNN [35], and MCT-NET [36]. CNMF is a matrix factorization-based method, while the other models are deep learning methods. MSD-CNN proposes a multi-scale and multi-depth CNN for remote sensing image fusion, which is based on residual learning and multiscale feature extraction. TFNet is a two-stream network that encodes spatial and spectral features independently, and then decodes the HR-HSI using the fusion of spatial and spectral features. SSF-CNN utilizes the direct concatenation of LR-HSI and HR-MSI to predict the HR-HSI, with the HR-MSI being concatenated in each convolutional layer. Lastly, MCT-NET employs a cross transformer to fuse spatial and spectral features.
The experimental procedure was carried out using LR-HSI, obtained from HR-HSI through the application of Gaussian blur and a subsequent four times downsampling. HR-MSI were generated by extracting red–green–blue bands at regular intervals that corresponded to their physical wavelengths from using specific dataset or sensors. For testing purposes, subsets with dimensions of 128 × 128 were extracted from the data center while the remaining components served as the training set. In each iterative cycle, the training area was randomly cropped to the size of 128 × 128 as part of the training process. The procedure was performed on a computational system equipped with an Intel i7-11700K 3.60 GHz CPU and an Nvidia RTX 3080Ti 12 GB GPU, with the assistance of PyTorch 2.2.2. The Adam optimization algorithm was selected for this particular experiment.

3.2. Evaluation Metrics

This paper utilizes four prominent indexes to thoroughly assess the quality of the reconstructed HR-HSI at a reduced resolution, as subsequently detailed.
To appraise the spatial quality, the Peak Signal-to-Noise Ratio (PSNR) is employed. PSNR serves as an objective evaluation metric, gauging the noise level or image distortion. A higher PSNR value indicates less distortion and superior image quality. The definition of PSNR is as follows:
P S N R X , X = 10 l o g 10 m a x X k 2 1 H W X k X k 2 2
where X represents the estimated HR-HSI, X represents the ground truth HR-HSI, and X k and X k denote the k t h band of the reference HR-HSI and the estimated HR-HSI, respectively.
We evaluated the spectral quality utilizing the Spectral Angle Mapper (SAM). SAM serves as a metric to estimate an image’s spectral quality, achieved through the computation of the average spectral angle across the entire spatial domain, as defined below:
S A M X , X = 1 H W i = 1 H j = 1 W a r c c o s X T i , j X i , j X i , j 2 X i , j 2
where H and W represent the number of rows and columns in the HR-HSI. The pixel vector of the reference HR-HSI is represented by X ( i , j ) , while the estimated HR-HSI at the same position ( i , j ) is represented by X ( i , j ) . Spectral Angle Mapper (SAM) is a measure of spectral distortion, with a lower value indicating less distortion.
Erreur relative globale adimensionnelle de synthèse (ERGAS): The ERGAS index is specifically crafted to assess the comprehensive quality of fused images, defined as follows:
E R G A S X , X = 100 r 1 S k = 1 S X k X k 2 2 μ 2 X k
where r signifies the downsampling ratio, and μ represents the mean value. A lower ERGAS value indicates a more favorable fusion outcome.
The Root Mean Squared Error (RMSE) is a statistical measure that represents the discrepancy between the values of X and X′. A lower value of RMSE indicates fewer reconstruction errors, thus implying superior quality of reconstruction. It is defined as follows:
R M S E = k = 1 S i = 1 H j = 1 W X k i , j X k i , j 2 H W S
where X k i , j and X k i , j represent the values at the position (i, j) in the kth band of the reference HR-HSI and the estimated HR-HSI.

3.3. Experimental Results on the Pavia Center Dataset

The Pavia Center dataset was obtained using drone-based ROSIS sensors. The sensor initially comprised 115 bands, which, post-processing, were reduced to 102. The dimensions of the Pavia Center dataset are 1096 × 715 pixels.
The fusion performances and quantitative evaluations are illustrated in Figure 3 and Table 1. Overall, the deep learning methods outperformed the traditional methods. The proposed HSST achieved three optimal results for PSNR, SAM, and RMSE, and a sub-optimal result for ERGAS. The best values of PSNR and RMSE demonstrated the superior elementwise reconstruction quality of HSST, while the best value of SAM showed the spectral reconstruction quality of HSST. First place for ERGAS, which indicates the spatial reconstruction quality, was achieved by TFNet. TFNET has the deepest network, which gives it an advantage in extracting nonlinear deep features that are beneficial for spatial reconstruction.

3.4. Experimental Results on the Botswana Dataset

The Botswana dataset comprises a sequence of datasets procured by NASA satellites during the period from 2001 to 2004. The dataset encompasses a total of 145 spectral bands, subsequent to the exclusion of the uncalibrated band and the noise band, which encompasses the water absorption characteristic. The dimensions of the dataset amount to 1476 × 256 pixels.
The fusion performances and quantitative evaluations are depicted in Figure 4 and Table 2. Our HSST experimental results are in the second tier, trailing slightly behind MCT-NET. In the Botswana dataset, the pixel’s spatial resolution reaches up to 30 m, making the spatial information more intricate compared to other datasets that have a higher requirement for feature extraction. The smaller size of HSST, in contrast to MCT-NET, constrains the performance of our model in extracting spatial features from a large receptive field.

3.5. Experimental Results on the Urban Dataset

The Urban dataset, acquired in 1995 using a drone-mounted HYDICE sensor, was centered on Copper Tree Bay, located in the state of Texas, USA. This dataset, characterized by image dimensions of 307 × 307, originally consisted of 210 spectral bands. However, after eliminating the bands affected by noise and water absorption, only 162 bands were preserved for further processing and analysis.
The fusion performances and quantitative evaluations are presented in Figure 5 and Table 3. It is evident that the proposed HSST outperforms all other comparative methods in terms of all four evaluation metrics. The superiority of the proposed HSST can be observed from the perspective of elementwise reconstruction quality (RMSE and PSNR), spectral reconstruction quality (SAM), and spatial reconstruction quality (ERGAS).

4. Discussion

The Pavia Center dataset, obtained via drone-based ROSIS sensors, presents a challenging platform for evaluating the effectiveness of various fusion algorithms. This dataset is particularly beneficial for urban analysis, wherein detailed spatial information is crucial for applications like infrastructure monitoring, land use classification, and urban planning. Our proposed HSST model showcased optimal results in three out of the four evaluation metrics, namely PSNR, RMSE, and SAM, against this dataset. A close inspection of the fusion outcomes in Figure 3 suggests that deep learning-based approaches, including HSST, are superior at preserving spectral and spatial details compared to traditional methods like CNMF. The high PSNR and low RMSE values achieved by HSST indicate a high degree of fidelity in the reconstructed images, which is essential for accurately identifying and monitoring minor urban features like road networks, building footprints, and vegetation patches. The best SAM score of HSST underlines its capability in spectral reconstruction, preserving the spectral signatures necessary for distinguishing between different land cover types and materials in remote sensing applications. TFNet, another deep learning-based method, achieved the best ERGAS score, highlighting the advantage of deeper networks in extracting spatial features that maintain spatial resolution and contextual information in the fused image. However, despite TFNet’s deeper architecture, HSST outperformed it in other metrics, suggesting that HSST strikes a balance between network depth and efficiency in feature extraction and reconstruction.
The Botswana dataset, taken from NASA satellite imagery, posed a unique challenge due to its high spatial resolution, which increases the complexity of spatial information. This complexity requires a model capable of effectively extracting features from a larger receptive field. In this context, MCT-NET outperformed HSST, which may be due to its larger size and enhanced capacity for spatial feature extraction. Despite having a smaller model size, HSST’s performance was commendable, achieving sub-optimal results across all metrics. This suggests that HSST offers computational efficiency without significant compromises in fusion quality. However, the results imply that for datasets with higher spatial resolution, like the Botswana dataset, the model’s complexity might need to be increased to improve feature extraction capabilities. The Botswana dataset is rich in spectral information, which is pivotal for environmental monitoring and ecosystem health assessment. HSST’s strength in spectral reconstruction, as evidenced by its SAM score, is particularly relevant for these applications. The accurate spectral information provided by HSST can facilitate more precise monitoring of vegetation health, which is essential for understanding the effects of climate change, identifying drought stress, and assessing the proliferation of diseases among plant populations. The slightly sub-optimal spatial reconstruction performance of HSST, likely due to its smaller size, suggests that the model could potentially be enhanced for satellite-based datasets with high spatial resolution.
The Urban dataset presented a scenario that played to the strengths of HSST, where it excelled over all other methods across all evaluation metrics, including RMSE, PSNR, SAM, and ERGAS. The dataset’s characteristics, featuring smaller dimensions and a significant number of spectral bands, appear to be ideally matched to HSST’s capabilities. HSST’s consistently optimal performance on the Urban dataset highlights its comprehensive skill in addressing the multifaceted challenges of image fusion, including elementwise, spectral, and spatial reconstruction. This robust and versatile performance indicates that HSST is well-suited for a broad spectrum of hyperspectral image fusion tasks. With a focus on industrial and urban land cover, the Urban dataset greatly benefits from HSST’s proficiency in capturing both spatial and spectral details. In applications such as pollution monitoring, HSST’s superior reconstruction quality is instrumental in pinpointing pollution hotspots and tracking the spread of pollutants over time. For urban planning purposes, the detailed spatial and spectral information provided by HSST is invaluable for informed decision-making regarding land use, infrastructure development, and environmental impact assessments. The outstanding performance of HSST on the Urban dataset underscores its potential role in managing the intricate and dynamic interactions within urban environments, where the relationship between human activities and the environment is particularly complex.
The experimental outcomes across the Pavia Center, Botswana, and Urban datasets strongly support the efficacy and robustness of the proposed HSST model for hyperspectral image fusion. HSST‘s capability to maintain a balance between elementwise, spectral, and spatial reconstruction qualities positions it as a promising tool for applications in remote sensing and other hyperspectral imagery-dependent fields. The depth of networks like TFNet may enhance spatial reconstruction, but HSST‘s overall balance and efficiency suggest that increased network depth does not always equate to optimal performance. The results also underscore the significance of model complexity in relation to dataset characteristics such as spatial resolution and the number of spectral bands.

4.1. Ablation Studies

In order to explore the role of the components of HSST in HSI and MSI fusion, some ablation experiments were conducted on the Urban dataset based on the three model variations of a model with only a spectral transformer, a model with only a spatial transformer, and a model without progressive fusion.
The fusion performances and quantitative evaluations of the ablation experiments are presented in Figure 6 and Table 4. When comparing the model with only the spectral transformer and the model with only the spatial transformer, it can be observed that the first model, which reconstructs the spectral information, performed significantly better than the second model, which reconstructs the spatial information. These results suggest that the reconstruction of spectral information is easier compared to spatial information, indicating that spatial information is more complex. Ablation studies on progressive fusion demonstrate that the use of progressive fusion significantly improves the accurate reconstruction of spatial details, thereby affecting the quality of the final reconstructed images.

4.2. Classification Performance Studies

To further evaluate the advantages of HSI-MSI image fusion, we conducted additional classification experiments on the Pavia Center dataset using unfused LR-HSI and fused HR-HSI. The Pavia Center dataset is a high-resolution city scene located in Pavia, northern Italy. For this dataset, we utilized the Self-Adaptive 3D ASPP Multi-Scale Feature Fusion Network (SAAFN) [37] for HSI classification. For all experiments, only 10% of labeled samples were randomly selected for training, with the rest used for testing. Both qualitative maps and quantitative evaluations comprehensively analyzed performance using four common metrics: producer‘s accuracy, overall accuracy, average accuracy, and Kappa.
The classification maps and the corresponding quantitative evaluations for the classification experiments using SAAFN conducted at the Pavia Center are presented in Figure 7 and Table 5, respectively. The optimal value for each line is highlighted in bold, while the sub-optimal value is underlined.
The HSST method achieved the highest overall accuracy (OA) with a score of 96.74%, closely followed by the TFNET method at 96.68%. It is worth noting that the fused HR-HSI methods outperformed the unfused LR-HSI, which only achieved an OA of 81.93%. Both the TFNET and HSST methods achieved the highest Kappa score of 0.95. On the other hand, the CNMF method performed the worst among the fused HR-HSI methods, with an OA of 82.54%. Overall, the fused HR-HSI methods, particularly TFNET and HSST, demonstrated higher classification accuracy compared to the unfused LR-HSI, as evidenced by their higher OA, AA, and Kappa scores.

5. Conclusions

In this article, we introduced the innovative Hierarchical Spectral–Spatial Transformer (HSST) network, a technique particularly well-suited for enhancing the capabilities of drone-based imaging systems. The HSST network comprises two branches that harness the self-attention mechanism inherent in transformers, enabling the extraction and integration of spectral details from hyperspectral images (HSI) with the spatial details from multispectral images (MSI). To leverage the rich spatial information inherent in remote sensing imagery, including that captured by drones, we incorporated a hierarchical structure that captures multi-scale information. Furthermore, our hierarchical progressive fusion strategy is designed for the reconstruction of high-resolution hyperspectral images (HR-HSI), progressively restoring spatial detail through upsampling and effectively utilizing the multi-level feature representation. Comparative experiments of the proposed HSST and five state-of-the-art methods were conducted on three widely used remote sensing hyperspectral datasets, including the Pavia Center, the Botswana, and the Urban. The superior experimental results of HSST demonstrate the effectiveness of the proposed method, and can be a competitive method for practical applications.

Author Contributions

Conceptualization, T.Z.; methodology, T.Z. and L.Z.; software, T.Z.; writing—original draft preparation, T.Z.; writing—review and editing, Q.L. and L.Z.; supervision, Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 10 December 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhuang, L.; Ng, M.K.; Fu, X.; Bioucas-Dias, J.M. Hy-Demosaicing: Hyperspectral Blind Reconstruction from Spectral Sub-Sampling. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  2. Bian, J.; Li, A.; Zhang, Z.; Zhao, W.; Lei, G.; Yin, G.; Jin, H.; Tan, J.; Huang, C. Monitoring Fractional Green Vegetation Cover Dynamics over a Seasonally Inundated Alpine Wetland Using Dense Time Series HJ-1A/B Constellation Images and an Adaptive Endmember Selection LSMM Model. Remote Sens. Environ. 2017, 197, 98–114. [Google Scholar] [CrossRef]
  3. Jia, S.; Shen, L.; Zhu, J.; Li, Q. A 3-D Gabor Phase-Based Coding and Matching Framework for Hyperspectral Imagery Classification. IEEE Trans. Cybern. 2018, 48, 1176–1188. [Google Scholar] [CrossRef]
  4. Zhao, J.; Zhong, Y.; Hu, X.; Wei, L.; Zhang, L. A Robust Spectral-Spatial Approach to Identifying Heterogeneous Crops Using Remote Sensing Imagery with High Spectral and Spatial Resolutions. Remote Sens. Environ. 2020, 239, 111605. [Google Scholar] [CrossRef]
  5. Fu, X.; Jia, S.; Zhuang, L.; Xu, M.; Zhou, J.; Li, Q. Hyperspectral Anomaly Detection via Deep Plug-and-Play Denoising CNN Regularization. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9553–9568. [Google Scholar] [CrossRef]
  6. Zhuang, L.; Fu, X.; Ng, M.K.; Bioucas-Dias, J.M. Hyperspectral Image Denoising Based on Global and Nonlocal Low-Rank Factorizations. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10438–10454. [Google Scholar] [CrossRef]
  7. Ghassemian, H. A Review of Remote Sensing Image Fusion Methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  8. Wei, Q.; Dobigeon, N.; Tourneret, J.-Y. Fast Fusion of Multi-Band Images Based on Solving a Sylvester Equation. IEEE Trans. Image Process. 2015, 24, 4109–4121. [Google Scholar] [CrossRef] [PubMed]
  9. Dian, R.; Li, S.; Fang, L.; Wei, Q. Multispectral and Hyperspectral Image Fusion with Spectral-Spatial Sparse Representation. Inf. Fusion 2019, 49, 262–270. [Google Scholar] [CrossRef]
  10. Fu, X.; Jia, S.; Xu, M.; Zhou, J.; Li, Q. Fusion of Hyperspectral and Multispectral Images Accounting for Localized Inter-Image Changes. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–18. [Google Scholar] [CrossRef]
  11. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J.M. Fusing Hyperspectral and Multispectral Images via Coupled Sparse Tensor Factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef] [PubMed]
  12. Dian, R.; Fang, L.; Li, S. Hyperspectral Image Super-Resolution via Non-Local Sparse Tensor Factorization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5344–5353. [Google Scholar]
  13. Yang, Q.; Xu, Y.; Wu, Z.; Wei, Z. Hyperspectral and Multispectral Image Fusion Based on Deep Attention Network. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  14. Cai, J.; Huang, B. Super-Resolution-Guided Progressive Pansharpening Based on a Deep Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5206–5220. [Google Scholar] [CrossRef]
  15. Wang, X.; Wang, X.; Zhao, K.; Zhao, X.; Song, C. Fsl-Unet: Full-Scale Linked Unet with Spatial–Spectral Joint Perceptual Attention for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  16. Dong, M.; Li, W.; Liang, X.; Zhang, X. MDCNN: Multispectral Pansharpening Based on a Multiscale Dilated Convolutional Neural Network. J. Appl. Remote Sens. 2021, 15, 036516. [Google Scholar] [CrossRef]
  17. Benzenati, T.; Kessentini, Y.; Kallel, A. Pansharpening Approach via Two-Stream Detail Injection Based on Relativistic Generative Adversarial Networks. Expert Syst. Appl. 2022, 188, 115996. [Google Scholar] [CrossRef]
  18. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion. IEEE Trans. Geosci. Remote Sens. 2011, 50, 528–537. [Google Scholar] [CrossRef]
  19. Dian, R.; Li, S.; Fang, L.; Lu, T.; Bioucas-Dias, J.M. Nonlocal Sparse Tensor Factorization for Semiblind Hyperspectral and Multispectral Image Fusion. IEEE Trans. Cybern. 2020, 50, 4469–4480. [Google Scholar] [CrossRef]
  20. Kanatsoulis, C.I.; Fu, X.; Sidiropoulos, N.D.; Ma, W.-K. Hyperspectral Superresolution: A Coupled Tensor Factorization Approach. IEEE Trans. Signal Process. 2018, 66, 6503–6517. [Google Scholar] [CrossRef]
  21. Grohnfeldt, C.; Zhu, X.X.; Bamler, R. Jointly Sparse Fusion of Hyperspectral and Multispectral Imagery. In Proceedings of the 2013 IEEE International Geoscience and Remote Sensing Symposium-IGARSS, Melbourne, Australia, 21–26 July 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 4090–4093. [Google Scholar]
  22. Xie, Q.; Zhou, M.; Zhao, Q.; Meng, D.; Zuo, W.; Xu, Z. Multispectral and Hyperspectral Image Fusion by MS/HS Fusion Net. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 1585–1594. [Google Scholar]
  23. Dian, R.; Li, S.; Guo, A.; Fang, L. Deep Hyperspectral Image Sharpening. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5345–5355. [Google Scholar] [CrossRef]
  24. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. Multispectral and Hyperspectral Image Fusion Using a 3-D-Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 639–643. [Google Scholar] [CrossRef]
  25. Wold, S.; Esbensen, K.; Geladi, P. Principal Component Analysis. Chemom. Intell. Lab. Syst. 1987, 2, 37–52. [Google Scholar] [CrossRef]
  26. Zheng, Y.; Li, J.; Li, Y.; Guo, J.; Wu, X.; Shi, Y.; Chanussot, J. Edge-Conditioned Feature Transform Network for Hyperspectral and Multispectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  27. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  28. He, X.; Chen, Y.; Lin, Z. Spectral-Spatial Transformer for Hyperspectral Image Classification. Remote Sens. 2021, 13, 498. [Google Scholar] [CrossRef]
  29. Selen, A.; Esra, T.-G. SpectralSWIN: A Spectral-Swin Transformer Network for Hyperspectral Image Classification. Int. J. Remote Sens. 2022, 43, 4025–4044. [Google Scholar]
  30. Cai, Y.; Lin, J.; Hu, X.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Van Gool, L. Maskguided Spectral-Wise Transformer for Efficient Hyperspectral Image Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17502–17511. [Google Scholar]
  31. Bandara, W.G.C.; Patel, V.M. HyperTransformer: A Textural and Spectral Feature Fusion Transformer for Pansharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1767–1777. [Google Scholar]
  32. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  33. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef]
  34. Liu, X.; Liu, Q.; Wang, Y. Remote Sensing Image Fusion Based on Two-Stream Fusion Network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef]
  35. Han, X.H.; Shi, B.; Zheng, Y. SSF-CNN: Spatial and Spectral Fusion with CNN for Hyperspectral Image Super-Resolution. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 2506–2510. [Google Scholar]
  36. Wang, X.; Wang, X.; Song, R.; Zhao, X.; Zhao, K. MCT-Net: Multi-Hierarchical Cross Transformer for Hyperspectral and Multispectral Image Fusion. Knowl. Based Syst. 2023, 264, 108630. [Google Scholar] [CrossRef]
  37. Zhu, T.; Liu, Q.; Zhang, L. An Adaptive Atrous Spatial Pyramid Pooling Network for Hyperspectral Classification. Electronics 2023, 12, 5013. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed network.
Figure 1. Framework of the proposed network.
Remotesensing 16 04127 g001
Figure 2. The schematic diagram of a multi-head Spectral–Spatial Transformer feature fusion block.
Figure 2. The schematic diagram of a multi-head Spectral–Spatial Transformer feature fusion block.
Remotesensing 16 04127 g002
Figure 3. The fusion outcomes of various models on the Pavia Center dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Figure 3. The fusion outcomes of various models on the Pavia Center dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Remotesensing 16 04127 g003
Figure 4. The fusion outcomes of various models on the Botswana dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Figure 4. The fusion outcomes of various models on the Botswana dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Remotesensing 16 04127 g004
Figure 5. The fusion outcomes of various models on the Urban dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Figure 5. The fusion outcomes of various models on the Urban dataset. The first row shows the R-G-B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) CNMF; (c) MSD_CNN; (d) TFNET; (e) SSF-CNN; (f) MCT-NET; (g) HSST.
Remotesensing 16 04127 g005
Figure 6. The fusion outcomes of ablation studies on the Urban dataset. The first row shows the R–G–B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) Spectral Transformer only; (c) Spatial Transformer only; (d) Without progressive fusion; (e) HSST.
Figure 6. The fusion outcomes of ablation studies on the Urban dataset. The first row shows the R–G–B images after fusion, and the second row shows the pseudo-color processed differential images between the fused and reference images. (a) Original image; (b) Spectral Transformer only; (c) Spatial Transformer only; (d) Without progressive fusion; (e) HSST.
Remotesensing 16 04127 g006
Figure 7. The results for the Pavia Center classification experiment: (a) Original image; (b) Ground truth; (c) LR-HSI; (d) CNMF; (e) MSD-CNN; (f)TFNET; (g) SSF-CNN; (h) MCT-NET; (i) HSST.
Figure 7. The results for the Pavia Center classification experiment: (a) Original image; (b) Ground truth; (c) LR-HSI; (d) CNMF; (e) MSD-CNN; (f)TFNET; (g) SSF-CNN; (h) MCT-NET; (i) HSST.
Remotesensing 16 04127 g007
Table 1. Quantitative evaluations for the Pavia Center dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
Table 1. Quantitative evaluations for the Pavia Center dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
MethodsPSNRSAMERGASRMSE
CNMF25.22214.363511.436113.9777
MSD-CNN35.75665.25404.71254.1563
TFNET35.65754.89313.76174.2040
SSF-CNN34.98984.73084.98074.5401
MCT-NET36.98094.15044.13253.6099
HSST37.51834.13383.84513.3933
Table 2. Quantitative evaluations for the Botswana dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
Table 2. Quantitative evaluations for the Botswana dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
MethodsPSNRSAMERGASRMSE
CNMF26.34572.48669.484926.3457
MSD-CNN35.71602.79773.22490.5964
TFNET36.54352.44792.96300.5422
SSF-CNN30.06265.164116.27641.1434
MCT-NET37.89552.18032.63030.4640
HSST37.18242.22742.48980.5037
Table 3. Quantitative evaluations for the Urban dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
Table 3. Quantitative evaluations for the Urban dataset. The optimal and sub-optimal values are bolded and underlined, respectively.
MethodsPSNRSAMERGASRMSE
CNMF36.04682.49711.56124.0198
MSD-CNN35.74403.21171.84283.2133
TFNET35.95843.02551.78853.1350
SSF-CNN37.39122.59041.46382.6582
MCT-NET37.32942.70761.43122.6772
HSST37.62492.46441.39542.5877
Table 4. Quantitative evaluations for the ablation studies on the Urban dataset. The optimal values are bolded.
Table 4. Quantitative evaluations for the ablation studies on the Urban dataset. The optimal values are bolded.
MethodsPSNRSAMERGASRMSE
Spectral Transformer only34.31253.49712.93803.7891
Spatial Transformer only21.28988.03709.451316.9695
Without progressive fusion37.32942.70761.43122.6772
HSST37.62492.46441.39542.5877
Table 5. Quantitative evaluations for Pavia Center classification experiment (%). The optimal and sub-optimal values are bolded and underlined, respectively.
Table 5. Quantitative evaluations for Pavia Center classification experiment (%). The optimal and sub-optimal values are bolded and underlined, respectively.
Unfused
LR-HSI
Fused HR-HSI
CNMFMSD-CNNTFNETSSF-CNNMCT-NETHSST
OA81.9382.5494.3796.6894.2596.1196.74
AA48.2671.1568.9583.2578.0791.4086.72
Kappa0.730.760.920.950.920.940.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, T.; Liu, Q.; Zhang, L. Hierarchical Spectral–Spatial Transformer for Hyperspectral and Multispectral Image Fusion. Remote Sens. 2024, 16, 4127. https://doi.org/10.3390/rs16224127

AMA Style

Zhu T, Liu Q, Zhang L. Hierarchical Spectral–Spatial Transformer for Hyperspectral and Multispectral Image Fusion. Remote Sensing. 2024; 16(22):4127. https://doi.org/10.3390/rs16224127

Chicago/Turabian Style

Zhu, Tianxing, Qin Liu, and Lixiang Zhang. 2024. "Hierarchical Spectral–Spatial Transformer for Hyperspectral and Multispectral Image Fusion" Remote Sensing 16, no. 22: 4127. https://doi.org/10.3390/rs16224127

APA Style

Zhu, T., Liu, Q., & Zhang, L. (2024). Hierarchical Spectral–Spatial Transformer for Hyperspectral and Multispectral Image Fusion. Remote Sensing, 16(22), 4127. https://doi.org/10.3390/rs16224127

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop