Next Article in Journal
Coastal High-Temporal Sea-Surface Altimetry Using the Posterior Error Estimations of Ionosphere-Free PPP and Information Fusion for Multi-GNSS Retrievals
Next Article in Special Issue
CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration
Previous Article in Journal
IRA-MRSNet: A Network Model for Change Detection in High-Resolution Remote Sensing Images
Previous Article in Special Issue
Supervised Contrastive Learning-Based Classification for Hyperspectral Image
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion

1
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
2
Peng Cheng Laboratory, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(21), 5600; https://doi.org/10.3390/rs14215600
Submission received: 27 September 2022 / Revised: 28 October 2022 / Accepted: 29 October 2022 / Published: 6 November 2022
(This article belongs to the Special Issue Deep Learning for the Analysis of Multi-/Hyperspectral Images)

Abstract

:
The process of fusing the rich spectral information of a low spatial resolution hyperspectral image (LR-HSI) with the spatial information of a high spatial resolution multispectral image (HR-MSI) to obtain an HSI with the spatial resolution of an MSI image is called hyperspectral image fusion (HIF). To reconstruct hyperspectral images at video frame rate, we propose a lightweight multi-level information network (MINet) for multispectral and hyperspectral image fusion. Specifically, we develop a novel lightweight feature fusion model, namely residual constraint block based on global variance fine-tuning (GVF-RCB), to complete the feature extraction and fusion of hyperspectral images. Further, we define a residual activity factor to judge the learning ability of the residual module, thereby verifying the effectiveness of GVF-RCB. In addition, we use cascade cross-level fusion to embed the different spectral bands of the upsampled LR-HSI in a progressive manner to compensate for lost spectral information at different levels and to maintain spatial high frequency information at all times. Experiments on different datasets show that our MINet outperforms the state-of-the-art methods in terms of objective metrics, in particular by requiring only 30% of the running time and 20% of the number of parameters.

Graphical Abstract

1. Introduction

Hyperspectral image (HSI) can be obtained by acquiring the rich spectral information of the scene target as well as the detailed spatial information [1,2], and the spectral and spatial information in its three-dimensional data cube enables the differentiation of different substances in the scene. Accordingly, HSI has outstanding performance for uses in many fields, ranging from remote sensing, change detection [3,4], to land-use classification [5], and other fields such as Digital Soil Mapping (DSM) and Geologic mapping [6,7]. However, it is difficult to obtain 3D hyperspectral data from a single shot using a 2D planar detector [8]. In this condition, a flexible alternative is to acquire both a high spatial resolution low spectral resolution multispectral image (HR-MSI) and a low spatial resolution high spectral resolution hyperspectral image (LR-HSI) of the same static scene, and then combine the advantages of both images to obtain a high resolution (HR) image in both the spatial and spectral domains. The pansharpening technique in remote sensing is closely related to the hyperspectral image fusion problem. This task aims at generating a high resolution multi-spectral (MS) image from inputs of a high spatial resolution single band panchromatic (PAN) image and a low spatial resolution multi-spectral image. The two different sets of inputs crreate the two forms of pansharpening concepts—multispectral and panchromatic images, and low-resolution hyperspectral and high-resolution multispectral images different. The PAN image has only one spectral channel, which means it cannot express RGB colors and, on the contrary, the MS image carries a high expression ability of color.
There are many hyperspectral image fusion techniques that have been mentioned in the last two decades. These methods can be roughly divided into two categories depending on whether deep learning is used or not. The first category is the classical traditional methods, which mainly include the component substitution (CS)-based methods [9,10,11], multiresolution analysis (MRA)-based methods [12,13,14,15,16,17] and variational optimization (VO)-based methods [18,19,20]. The second category is the deep learning methods [21,22,23,24] that have become popular over the years.

1.1. Traditional Methods

CS-based pansharpening methods have achieved wide application in the field of remote sensing images pansharpening due to their simplicity and efficiency. The general steps of this method are: firstly, the LR-HSI is upsampled and then the result is mapped to other new representation spaces, separating the spectral and structural information of the hyperspectral image in the new transform domain and then replacing the structural information component with a panchromatic image, improving the spatial resolution of the fused image in this way. Representative works of CS-based methods mainly include the principal component analysis (PCA) [25], the Brovey [26], the intensity-hue-saturation (IHS) method [27], and the Gram–Schmidt (GS) method [28]. However, CS-based methods lose part of the spectral information while replacing the spectral image structure information, so the fusion results usually lead to severe spectral distortion and oversharpening.
MRA-based pansharpening methods deal with remote sensing image fusion problems from a spatial perspective. Such methods use pyramid or wavelet transform to decompose LR-HSI and HR-MSI into multiple scales, and then extract the spatial information of HR-MSI of a certain scale and inject it into the corresponding scale of LR-HSI, and finally use the inverse transform to generate the fused image. The decomposition methods used usually include high pass filter (HPF) fusion [29], generalized Laplacian pyramid (GLP) transform [17], wavelet-transform-based [30], and smoothing filter-based intensity modulation (SFIM) [12]. MRA-based methods can solve the spectral distortion problem of the CS-based method to a certain extent. However, in the process of injecting structural information into the panchromatic image at the multi-resolution scale, the fusion result will degrade the spatial information.
CS-based and MRA-based methods both focus on retaining a certain aspect of detailed information, and the fusion results cannot balance structural information and spectral information. VO-based methods regard pansharpening as an ill-posed problem and are build on variational theory. The major process mainly includes two steps: constructing the energy function and designing an optimisation algorithm for the model function. The sparse-based methods [31,32] and the model-based methods [33,34] that evolved from these two steps are two representative VO-related approaches. Since VO-methods obtain optimal solutions with the help of different valid a priori information and effective optimisation algorithms, they can improve the spatial structure information while maintaining spectral fidelity, but it is easy to fall into local optimisation solutions during the optimisation process, resulting in less than optimal final fused image quality results.

1.2. DNN-Based Methods

In recent decades, due to the increasing amount of hyperspectral data obtained, the development of theoretical research in deep learning and the increase in computing power have made it possible to process these data [35,36]. In spectral image pansharpening, DL-based methods are gaining widespread attention and recognition for their excellent fusion performance and speed advantages. To be specific, the inputs to the network are specified as LR-HSI and HR-MSI, and the designed network module performs feature extraction and fusion to obtain the desired spectral data cube for comparison with the ground-truth. By making the reconstruction result infinitely close to the real image, a stable network structure can be determined.
Since Huang et al. first applied DL-based methods [37] to remote sensing image fusion, DL-based methods have been promoted and improved in various ways, resulting in many fusion methods based on various types of network architectures. For instance, Masi et al. applied a super-resolution convolutional neural network (SRCNN) [35] to remote sensing image fusion, and proposed a simple and effective pansharpening neural network (PNN) with a three-layer structure [38]. An improved version of advanced PNN (A-PNN) [39] then optimises network performance and achieves better results. Dong et al. proposed a non-negative structured sparse representation framework based on clustering to mine spatial and spectral correlations [21], and made good progress. To address the problem of inadequate extraction of HR-MSI spatial information with existing deep learning methods, Jiang et al. proposed a convolutional neural network using a differential information mapping strategy to process residual information between HR-MSI and LR-HSI [40]. In addition, Zhao et al. proposed a fusion network based on spatial attention and channel attention mechanisms to improve the network’s ability to extract features from input images [41]. On the other hand, Liu et al. used a generative adversarial network (GAN) to synthesize high-quality panchromatic sharpened images for the first time, using a two-stream fusion architecture to generate the desired HR-HSI [42]. To alleviate the dependence on real datasets, unsupervised spectral fusion methods have also received attention [43,44]. Ma et al. proposed a new unsupervised fusion framework [45], where the generator builds an adversarial game with spectral discriminator and spatial discriminator, respectively, to preserve the rich spectrum of hyperspectral images information and spatial information for panchromatic images. A network that learns high-level features by constraining the receptive from increasing in the deep layers was proposed to handle the fusion problem and achieved good results [46].

1.3. Motivation and Contribution

The ideal fusion image should have spatial information close to HR-MSI and retain the spectral information of LR-HSI. The DL-based methods improve the traditional fusion methods (CS-based, MRA-based and VO-based) and provide various experiences to deal with the pansharpening problem. Although deep learning has been successful in pansharpening, DL-based methods only focus on fused HR-HSI visual effects and objective evaluation metrics, ignoring the running time issue. To the best of our knowledge, Deep blind HSI fusion network (DBIN+) [47] used a recurrent neural network that cannot be parallelized, with an average forward inference time of 2.17 s. Multispectral and hyperspectral image fusion network (MHF) [48,49] took advantage of the large-scale parallelization of neural networks and relieves the pressure of memory congestion to a certain extent, but it still does not break through the constraints of its network structure, weakening the forward reasoning process, and it still takes 1.67 s to reconstruct a HR-HSI image. A hyperspectral image fusion method based on zero-center residual learning (SpeNet) [50] proposed by Zhu et al. is currently the fastest pansharpening method, requiring only 0.49 s for one forward inference. Although Cheng et al. proposed a multiscale information fusion network for semantic segmentation [51], it uses deeply separable convolution to increase the perceptual field to extract local and global features. This is very different from extracting different levels of spectral image features in a hierarchical manner in the spectral image fusion process. Over the past few decades, many computational spectral imaging systems [52,53,54] have been proposed that can image in real time, computational spectral imaging does not require scanning operations, it projects the spatially and spectrally encoded 3D data cube on a certain plane, the detector superimposes the aliased images of spatial and spectral information, restores the original data through calculation, and obtains the spatial dimension and spectral dimension at the same time. However, neither traditional fusion algorithms nor DL-based methods can meet the requirements of video frame rate reconstruction, it is still a challenge to build a lightweight feature extraction and fusion network for real-time hyperspectral image synthesis.
This motivates us to design a lightweight multi-level information network (MINet) for multispectral and hyperspectral image fusion to meet the real-time reconstruction requirements of hyperspectral imaging. In particular, our starting point is not to prune the best current pansharpening network to reduce network parameters and running time. We have analysed the working mechanism of the residual module, which plays a key role in existing multispectral and hyperspectral image fusion networks, and defined a residual activity factor η to determine the capability of the residual module used (described in detail in Section 2.1). A lightweight fusion network was then designed based on this principle, in order to increase the correlation between spectral channels, we use multi-level spectral information and feature maps share a hidden layer state in batches, mutually perceive the mean information of each channel map, and complete feature extraction and fusion at the same time. While performing feature extraction, LR-HSI and HR-MSI with different depths are selected for joint-guided detail extraction, which is progressive layer by layer to improve the correlation between spectral images and RGB images and reduce spectral distortion. The main contributions of this paper are summarized as follows.
(1)
We developed a lightweight multi-level information network (MINet) for multispectral and hyperspectral image fusion to meet the real-time reconstruction requirements of hyperspectral imaging. The residual information between the different levels of LR-HSI and HR-MSI provides additional spectral detail features that contribute to the fusion results;
(2)
We defined the activity factor η of residual learning for quantifying the capability of the residual learning module, and compared the proposed network module with the hyperspectral image super-resolution reconstruction method that includes the residual learning module for experiments;
(3)
A lightweight residual constraint block based on global variance fine-tuning (GVF-RCB) was designed to extract and fuse the spectral feature information of LR-HSI and the spatial structure information of HR-MSI, which strengthen spatial details and reduce parameters.
The remainder of the paper is organized as follows. Section 2 describes our proposed framework in detail. Section 3 introduces comparative experiments on three datasets and discusses the experimental results. Section 4 draws conclusions.

2. Proposed Method

In this section, we will first introduce the proposed residual activity factor, then explain the overall network framework in detail, and finally analyse the working principle of the fully connected beta network.

2.1. Residual Activity Factor

Residual learning has proven its effectiveness in image feature extraction and fusion [47,49]. Although fusion networks that include residual blocks have achieved good results, their workings and operational speed still deserve more attention. In fact, using the cross-layer connection strategy used by Dense Block in SpeNet with good performance leads to the increase of network data buffer, which not only increases the computation, but also consumes a large amount of memory in the GPU. Therefore, we first consider the ResNet [55] structure instead of Dense Block to improve the synthesis speed.
Residual learning is one of the most used network structures in deep learning networks, relying on ‘shortcut connections’ to achieve a very good fit. Formally, denoting the desired underlying mapping as H ( x ) (an underlying mapping to be fit by a few stacked layers), with x denoting the inputs to the first of these layers. Multiple nonlinear layers can asymptotically approximate a residual function F ( x ) . So the desired underlying mapping H ( x ) can be expressed as Equation (1).
H ( x ) = x + F ( x )
We suppose that the residual mapping F to be learned is independent of the number of layers rather considering it as a function F ( x , W i ) with W i is the weight matrix layers for a given block. In general, the residual connection directly outputs the input feature x to H ( x ) , and F ( x ) learns the tiny residual feature between x and H ( x ) . Compared with the network structure of the VGG [56] tiled convolutional layer, residual learning can better transfer the features extracted by the previous network layer to the subsequent network layers. Unlike extracting features from a single natural image, the core process of hyperspectral image fusion is the spatial-spectral fusion of a three-dimensional data cube stitched together from upsampled multichannel hyperspectral images and multispectral images (RGB images). Therefore, the problem we are dealing with can be transformed into: extracting the features of the data block formed by the connection of the LR-HSI upsampled with the HR-MSI. The traditional CS-based and MRA-based methods accomplish the task by extracting the spatial information of the multispectral image (HR-MSI) and adding it to the blurred hyperspectral image (LR-HSI). Inspired by this, we analogize the form of residual network, assuming that the upsampled LR-HSI represents x and HR-MSI represents F ( x ) . It is worth noting that HR-MSI is no longer the meaning of residuals. This is a fusion problem of fusing data from two different distributions. Further, we hope that the two kinds of data are in the same order of magnitude as possible as shown Equation (2).
x 1 F ( x ) 1 ,
where · 1 is 1 norm of a vector, to quantitatively analyze the performance of the residual learning block, we define η to denote the activity of residual learning as Equation (3).
η = x 1 F ( x ) 1 .
This can be understood as a restriction on LR-HSI and HR-MSI data in the process of feature extraction. Indirectly, it is also possible to determine whether the input and output of the residual block are of one magnitude. As shown in Figure 1, when x 1 F ( x ) 1 , the structure of the network is similar to the residual module, which extracts hyperspectral information from x and spatial information of the same magnitude from F(x), respectively. When x 1 F ( x ) 1 , the learning ability of the residual decreases sharply. For the residual module, the network will degenerate into the VGG-Net structure, which is equivalent to considering more spatial information in HR-MSI while ignoring the hyperspectral information in LR-HSI. When x 1 F ( x ) 1 , the residual learning block degenerates to a constant mapping with infinite learning activity, and the residual fraction does not learn any knowledge and has a learning capacity of 0. At this point, the fusion process did not learn any spatial information from HR-MSI.
According to our definition of the residual activity factor, the residual learning block has an activity value of approximately 1.0 when the magnitudes of the extracted spatial information and inter-spectral information are not particularly significant. In order to verify the effectiveness of the residual activity factor, we use the HARVARD dataset to test three networks with residual modules, DBIN+, MHF-Net and SpeNet, and record the η and the corresponding PSNR at different stages in the training process. Figure 2 shows the corresponding PSNR under different η by adjusting the distribution of LR-HSI and HR-MSI. At the same time, we also compare the SAM and ERGAS of the three methods as shown in Table 1. An intuitive conclusion is that the closer the residual activity factor is to 1, the better the performance of the network. Therefore, when evaluating the performance of the residual block in the fusion network, the proposed η can be used to evaluate whether the currently used residual block can achieve the desired effect. It is worth noting that η is not suitable for evaluating the residual activity in the extraction process of a single natural image, because this is different from the principle that the spectral information and spatial information need to be extracted in the fusion process as close as possible in order of magnitude.
According to the definition of Equation (3), the value of η varies as the input to the neural network changes, but the residual activity factor should be a characterising quantity in a uniform sense. Therefore, for each data i with a sample set of N, we count the η i corresponding to each input and calculate the average residual factor η for sample N together reflecting the learning activity of the residuals throughout the network as shown in Equation (4). For the sake of concise representation, η appearing in subsequent comparison trials all denote the average result for the overall dataset.
η = 1 N i = 1 N η i ,
where N represents the number of samples. We illustrate the proposed method experimentally in comparison with current hyperspectral super-resolution reconstruction methods in Section 3.4.

2.2. Lightweight Multi-Level Information Fusion Network

The proposed lightweight multi-level information network (MINet) for multispectral and hyperspectral image fusion framework is displayed in Figure 3. The network is composed of several residual constraint blocks (orange boxes) based on a global variance fine-tuning (GVF-RCB) module, and a hyperspectral image reconstruction module (green boxes). The input to the network is the LR-HSI L R h × w × B and the HR-MSI P R H × W × b , where ( h , w , b ) are the reduced height, width, and number of spectral bands, respectively, and ( H , W , B ) are corresponding high-resolution version ( h H , w W , b B ). First, L is upsampled to L ˜ R H × W × B of the same size as P by bilinear interpolation. In order to reduce the amount of computation and improve the running speed, we choose a bilinear interpolation algorithm that is more suitable for large-scale parallel computation and stable for upsampling operations. Then, the GVF-RCB module estimates a residual image from both the input P and L ˜ progressively along the spectral dimension. Inspired by the success of the progressive reconstruction strategy in image super-resolution, we progressively embed the spectral bands of upsampled LR-HSI instead of feeding them all into the network at the beginning. The resulting residual image is further superimposed on L ˜ to generated a blurred HR-HSI, which is finally fed into the reconstruction module. After further fine-tuning and correction of the reconstruction module, the network outputs the final reconstructed high-resolution hyperspectral image.
The GVF-RCB module is the core module of our proposed multi-level information network, and the flowchart of the network architecture is shown in Figure 4. Its main function is to fuse the spatial information extracted from HR-MSI and the spectral information extracted from LR-HSI to reconstruct the corresponding HR-HSI. The GVF-RCB module is composed of a global residual connection branch and multiple local residual fusion unit cascade branches. The global residual connection branch can extract and transfer low-frequency information of hyperspectral images quickly and efficiently. The goal of residual branch learning is to fuse the information extracted from LR-HSI and MSI into a map of the detail level of the desired HR-HSI to be reconstructed. Define the output of the fusion network as Z R H × W × B , which can be expressed as Equation (5).
Z = L ˜ + f ( L ˜ + P ) ,
where f ( · , · ) represents the mapping of the detailed level of information extracted from LR-HSI and MSI fused into the desired HR-HSI to be reconstructed. The network structure of the proposed GVF-RCB module mainly uses convolution kernels with receptive field sizes of 3 × 3 and 1 × 1 as the basic configuration. The 3 × 3 convolution kernel is used to extract the spatial feature information and build a fusion model for local spatial information features of hyperspectral images, and the 1 × 1 convolution kernel is used to fuse the hyperspectral information for channel compression or dimension enhancement. Inspired by the success of the zero-mean normalization strategy in SpeNet, we also add this constraint to transform the data distribution of feature maps to zero when extracting residual feature information.
The overall network consists of four key steps: (1) concatenate the output of the upper layer with HR-MSI (UP-HSI sampling numbers are 8, 16, 31) and increase the number of channels to be consistent with the up-sampled hyperspectral image (UP-HSI); (2) fuse two kinds of data with different distributions. A fully connected beta network is used to perform mutual induction global spectral information fusion on the output feature map and UP-HSI, adaptively adjust the data distribution of the two inputs, and accelerate the fusion of local spatial information and spectral information in the next stage; (3) use a 3 × 3 convolutional layer to fuse the local spatial structure feature information of feature map0 and feature map1, and use a 1 × 1 convolutional layer to perform spectral fusion of the channel dimension and reduce the number of channels; (4) use convolution for further fusion, and add it to the feature map and output it.
The GVF-RCB module mainly completes two tasks. First, it extracts the spectral feature information of the hyperspectral image and the spatial structure information of the multispectral image, and fuses the two kinds of information into the detailed information of the hyperspectral image to be output. It is responsible for MINet extraction and fusion of complementary information from each input. In addition, as the basic unit of residual branch information fusion, the spatial structure information and spectral information of different levels are gradually fused through the cascade of multiple GVF-RCB modules. The two fused parts have the same spatial resolution, but the number of channels and spectral images are different for each input to the GVF-RCB module, and the gradually increasing number of channels enhances the feature extraction of spectral information. The two different feature vectors rely on the core unit beta network for fusion.
The input of the hyperspectral image reconstruction module is the coarse HR-HSI output by the fusion network, and the reconstructed images in the first stage of the network are fine-tuned and distilled. The network structure of the reconstruction module is relatively simple, the overall structure is a residual block, and the residual branch is mainly composed of 3 × 3 and 1 × 1 convolutional layers alternately.

2.3. Fully Connected Beta Network

The fully connected beta network is the core unit of the GVF-RCB module, which can be used to adjust the mean and variance of hyperspectral images. Although hyperspectral images and feature maps generated by convolutional neural networks have the same spatial dimension, their data distributions are quite different. We visualized the input and output of the beta network in the fifth GVF-RCB module, and selected three images from the CAVE [57] and HARVARD [58] datasets, respectively, as shown in Figure 5, where blue represents the mean value of each channel of the hyperspectral image and orange represents the mean value of each channel of the feature map. The average value of each channel is calculated as in the average pooling operation. From the mean value of each channel of the hyperspectral image and feature map, and the beta coefficient of each channel (output of the network), it can be seen that the mean value of the hyperspectral image channel is small and densely distributed, while the mean value of each channel of the feature map is large and irregularly distributed. Due to the obvious data gap, directly cascading hyperspectral images and feature maps will result in the ineffective fusion of the spatial feature information of the feature maps and the spectral information of the hyperspectral images. Therefore, it is usually necessary to pass the extracted feature information to convolutional layers using Dense Connection.
Adjusting the internal data distribution of the neural network can often make the network converge quick and stable, and improve the generalization ability of the network. Batch normalization is one of the most commonly used methods to quickly adjust the distribution of feature map data. During training, the local mean and variance of each mini-batch data are used for normalization and to estimate the global value. Finally, in the testing process adjust the data distribution with the global mean and variance instead of the local mean and variance. However, super-resolution reconstruction methods [59] for natural images point out that batch normalization will introduce huge randomness and cause the network to underfit, so it is not suitable for image reconstruction problems.
To reduce the difficulty of training neural networks, we introduce a coefficient β ( 0 β 1 ) on the residual learning branch, and the refined residual information is multiplied by β to achieve convergence. Assuming X R H × W × B is a feature map, let μ and σ 2 be the mean and variance of X , respectively, calculated according to the following equation:
μ = 1 H × W × B i = 1 H × W × B X i
σ 2 = 1 H × W × B i = 1 H × W × B ( X i μ ) 2 ,
when the output feature map X of the residual branch is multiplied by a small coefficient β , the mean μ and variance σ 2 of the feature map are calculated according to the following equation:
μ = 1 H × W × B i = 1 H × W × B ( β · X i ) = β μ μ
σ 2 = 1 H × W × B i = 1 H × W × B ( β · X i μ ) 2 = β 2 σ 2 σ 2 .
Therefore, multiplying the residual branch by a small coefficient β changes the mean and variance of the feature map data, which essentially adjusts the data distribution of the feature map in the direction where the mean tends to zero, reducing the absolute value of the feature map. Further, the magnitude of the learning parameters in the process of training the network is reduced, so that the variable parameters can better fit the data map. The reduction of variance makes the data distribution more compact, suppresses the fluctuation range of data changes, and at the same time enhances the stability of the data, making the neural network converge towards a more stable direction.
The proposed fully connected beta network is shown in Figure 6. We used the beta network to make the hyperspectral image and the feature map perceive each other’s feature information, and then adaptively changed the data distribution to an appropriate degree. The selection of the beta value was obtained by adaptive learning of the beta network, for each channel will learn a different beta value, and then the small coefficient beta was multiplied by each feature map. The detailed calculation steps of the beta network algorithm are shown in Algorithm 1.
Algorithm 1: Data distribution adjustment in fusion network
   Input: X and Y
   Output:  X and Y
      STEP1: compute mean vector
             μ X 1 H × W i = 1 H × W X i j , j = 1 , 2 , , C
             μ Y 1 H × W i = 1 H × W Y i j , j = 1 , 2 , , C
      STEP2: concatenate μ X and μ Y to obtain neural input μ
      STEP3: feed μ into fully connected network and output the coefficient vector β
      STEP4: split β to β X and β Y
      STEP5: compute outputs X and Y
             X β X ( i ) X ( i ) , i = 1 , 2 , C
             Y β Y ( i ) Y ( i ) , i = 1 , 2 , C

3. Experiments and Analysis

3.1. Experimental Setup

This section is for experimental evaluation. All experiments were run on a server with a Inter(R) Core(TM) i7-6850K CPU 3.60GHZ processor, 32GB of RAM, 2TB disk space, Ubuntu 18.04 and NVIDIA TITAN XP GPUs. We completed the construction of the network model using the PyTorch framework, using a batch size range of [ 10 , 64 ] during training. We adopted ADAM optimizer [60] to train the network. The initial learning rate of the MINet network was set to 1 × 10 3 and the number of iterations was fixed at 3000. For comparison, we selected nine state-of-the-art fusion based HSI SR methods, including coupled sparse tensor factorization (CSTF) [61], non-negative structured sparse representation based method (NSSR) [21], coupled non-negative matrix factorization (CNMF) [62], the low tensor-train rank-based method (LTTR) [63], MHF [48], DBIN+ [47], DMDNet [64], ResTFNet [65], and SpeNet [50].
Seven quantitative metrics that are widely used in multispectral and hyperspectral image fusion tasks were selected to analyse and evaluate the quality of the fused images. First, we used the mean square error (MSE), which is most commonly used in image quality evaluation. It averages the difference between the true and predicted values after a squaring operation, reflecting the difference between the two sets of images. For each spectral band, the specific formula is shown below.
M S E ( X ˜ , X ) = 1 H × W × B H × W × B ( X ˜ i X i ) 2 ,
where H, W and B denote the numbers of rows, columns and channels of the image, respectively. X ˜ i and X i represent the predicted image of the i-th spectral band with ground-truth, respectively. The smaller the value of MSE, the closer the predicted image is to the real image.
Peak Signal-to-Noise Ratio (PSNR) is another index for assessing the fusion quality of each band. The definition is based on the MSE and the difference in MSE can be seen by comparing the PSNR. The average PSNR of each band is given by:
P S N R ( X ˜ , X ) = 10 B k = 1 B l o g ( M S E ( X ˜ k X k ) ) ,
where X ˜ k and X k represent the image of the k-th channel. Unlike MSE, a higher PSNR value means that the spatial information of the two sets of images is more similar and better between reconstructions.
The relative average spectral error (RASE) estimates the global spectral quality of the pan-sharpened image. It is defined as:
R A S E ( X ˜ , X ) = 100 M 1 N i = 1 N R M S E ( B i ) 2 ,
where R M S E ( B i ) is the root mean square error between the i-th band of the pan-sharpened image and the i-th band of the reference image. M is the mean value of the N spectral bands ( B 1 , B 2 , B N ) .
Erreur Relative Global Adimensionnelle Synthese (ERGAS) [28] is an error indicator, which represents the effect of pansharpened images from a global perspective. It mainly reflects the spectral distortion of the enhanced image compared to the reference image. The ERGAS index is a normalized dissimilarity index:
E R G A S ( X ˜ , X ) = 100 h l 1 B k = 1 B M S E ( k ) μ ( k ) 2 ,
where h l is the ratio of the spatial resolution size of HR-MSI and LR-HSI, and μ ( k ) represents the average value in spectral band B. The lower ERGAS indicates that the spectral distribution of the enhanced images is similar to that of the ground-truth.
Spectral Angle Mapper (SAM) measures the spectral similarity between the reference image and the fused image by treating the spectrum of each image element as a high-dimensional vector, and by calculating the angle between the two vectors to reflect the likelihood of belonging to the same feature. The SAM index is defined as:
S A M ( X ˜ , X ) = 1 H × W j = 1 H × W a r c c o s x ˜ j x j x ˜ j 2 x j 2 ,
where x ˜ j and x j are the spectral signatures of the j-th ( 1 j H W ) pixels of X ˜ and X , respectively, and · 2 is 2 norm of a vector.
Average Structural Similarity Index (ASSIM) represents the proximity of structural information between the ground truth and pansharpened images. The ASSIM index is defined as:
A S S I M ( X ˜ , X ) = 1 B k = 1 B S S I M ( X k , X ˜ k ) ,
where S S I M ( · , · ) [66] computes the SSIM value of a typical spectral band. When S S I M = 1 , the fusion result is the best.
The Correlation Coefficient (CC) is another widely used indicator measuring the spectral quality of the pan-sharpened images. It calculates the correlation coefficient between a pan-sharpened image X ˜ and the corresponding reference image X as:
C C ( X ˜ , X ) = i = 1 w j = 1 h ( X ˜ i , j μ X ˜ ) ( X i , j μ X ) i = 1 w j = 1 h ( X ˜ i , j μ X ˜ ) 2 i = 1 w j = 1 h ( X i , j μ X ) 2 ,
where w and h are the width and height of the images, μ indicates mean value of an image. CC ranges from −1 to +1, and the ideal value is +1.
We also use QNR [67], which is composed of spectral distortion index D λ and spatial distortion index D s , as the reference-free measure.

3.2. Datasets

To fully evaluate the performance of our proposed MINet network for multispectral and hyperspectral image fusion, we chose two simulated datasets and two real-world hyperspectral datasets for our experiments, including the CAVE dataset, the HARVARD dataset, the National Center for Airborne Laser Mapping (NCALM) and WorldView-2. The following information is specific to the datasets.
(1)
The HARVARD [58] dataset consists of 50 indoor and outdoor images captured under daylight illumination. Each image contains 1024 × 1392 pixels and 31 spectral bands in a wavelength range of 420 nm to 720 nm. Following [47], in terms of data quantity allocation, we randomly selected 30 images as the training set and the remaining 20 images as the test set.
(2)
The CAVE [57] dataset contains 32 images of scenes taken indoors, each hyperspectral image has a spatial resolution of 512 × 512 and a spectral range from 400 nm to 700 nm at 10 nm intervals, containing a total of 31 channels. Following [48], 20 images were randomly selected for training and the remaining 12 images were used as the test set in the data allocation process. In order to objectively evaluate the reconstruction performance of each spectral image fusion method, we combine the spectral bands of HSI with the widely used response function of the Nikon D700 camera to generate HR-MSI images of the same scene. Following [47], to obtain LR-HSI, we first filter the image with a 7 × 7 Gaussian blur kernel with a mean of 0 and a standard deviation of 2, and then take pixels with step r in both the row and column directions of the spatial dimension of the hyperspectral image.
(3)
We selected the remote sensing image dataset released by the Hyperspectral Image Analysis Laboratory of the University of Houston and the National Center for Airborne Laser Mapping (NCALM), which was part of the dataset of the 2018 IEEE GRSS Data Fusion Competition [68] as a pair of real remote sensing image datasets. It contains an RGB image of spatial dimensions 24,040 × 83,440, which was captured by with the iMAC ULTRALIGHT+ with the focal length of 70 mm, and a 48-band HSI of spatial dimensions 1202 × 4172 , which was captured with the ITRES CASI 1500 with an interval wavelength of 10nm in the range of 380–1050 nm. These sensors are installed on a Piper PA-31-350 Navajo Chieftain aircraft. To obtain ground-truth data, we first intercept HSI images with dimension 1202 × 4172 and RGB images spatially downsampled to the same dimension as real data. Then we divided the hyperspectral images and RGB images into 2 × 7 networks, extracted hyperspectral images and RGB images with dimensions of 576 × 576 from each grid, and used the 7 image patches extracted from the upper half of the hyperspectral images as the training set, and the 7 image patches extracted from the lower half are used as the test set. In order to obtain LR-HSI, we adopted the preprocessing method of the CAVE dataset, and added random additive Gaussian noise with mean 0 and variance 0.0001 to LR-HSI to simulate the low-noise environment of the real environment.
(4)
The WV-2 dataset was captured on a commercial satellite and contains an LR-HSI image of size 419 × 658 × 8 and an LR-MSI image (RGB images) of size 1676 × 2632 × 3 , while the LR-MSI image is not available. As the ground-truth data are not avaiable, following [48,49], we generated the training data in the following way. To be specific, we select the top half part of LR-HSI ( 836 × 2632 × 3 ) and HR-MSI ( 208 × 656 × 8 ) image to train the MINet and utilize the remaining parts of the dataset as testing data. We first extract 144 × 144 × 3 patches from HR-MSI as network input color images, and 36 × 36 × 8 overlapping patches from LR-HSI as network input low-resolution hyperspectral images, and finally generate training samples.

3.3. Experiments and Analysis

We selected two simulated datasets and two real-world hyperspectral datasets for our experiments, using different metrics for comparison. The experimental results of our proposed MINet compared with different methods on the Harvard dataset are shown in Table 2 and bold fonts represent more favorable results. For the fairness principle, the ten methods adopted the same spectral response curve and downsampling process. The significant superiority of our method over state-of-the-art methods is validated. To be specific, the proposed method is only second to DBIN+ on the ASSIM metric, while achieving the best results on PSNR, SAM, ERGAS, Params, FLOPs and GPU memory. It only takes 197.4 ms for MINet to perform a forward inference, which is 12.5% of the time required by MHF-Net, 10% of the time required by DBIN+, and 33% of the time required by SpeNet. In terms of GPU memory consumption, MHF-Net and SpeNet need up to 8 G graphics card memory for inference, DBIN+ needs 4.4 G memory, and MINet only needs 2.5 G to complete the calculation.
The results of the experiments comparing our proposed MINet with different methods on the CAVE dataset are shown in Table 3. All methods were performed using 32× upsampling and in the same test environment. MINet uses 5 GVF-RCB blocks to reduce the number of parameters and run time without reducing the reconstruction effect. The various evaluation metrics demonstrate the superiority of the proposed method, specifically, except that the ASSIM metric is slightly lower than DBIN+, the proposed MINet achieves the best results on PSNR, SAM and ERGAS. Not only have state-of-the-art results been achieved in the various image evaluation indicators, our proposed lightweight network has fewer parameters and faster running time, which can meet the needs of real-time reconstruction.
To more clearly represent the differences in the resulting images, we visualized the residual images on the CAVE data as shown in Figure 7, which takes the ground-truth image as a reference. The images reconstructed by NSSR have local artifacts and noise. The lack of reconstruction of high-frequency details in images by CNMF results in bright foreground contours. Obviously, the spectral fusion method based on deep learning is better than the traditional method based on optimization. The image reconstructed by MINet has less error and the residual map is closer to the ground-truth. In addition, we selected three sets of reconstructed images from the CAVE and HARVARD datasets respectively and compared their PSNR in different spectral bands quantitatively with the comparison method, and the results are shown in Figure 8. It can be seen that the PSNR of MINet is better than that of each comparison method.
Table 4 shows the results of different methods on the NCALM dataset. The three compared methods decreases significantly and fail to reconstruct high-frequency spatial details, resulting in blurring boundaries. We randomly select a reconstructed image from the dataset, and select the images of the 15th, 33rd, and 44th channels of the reconstructed images of each hyperspectral fusion method for visual comparison. The results are shown in Figure 9. Although DBIN+ achieves high PSNR on the test set, the detail features of the reconstructed image are blurred and have obvious noise. This method cannot effectively suppress the noise contained in the image. The contours of MHF-Net reconstructed images are smoother, but there are still problems of dark spots and blurred details.
Figure 10 shows the comparison results of the proposed method and four methods on the WV-2 dataset. We selected three small areas from the lower right block area of the dataset for detailed magnification and comparison. Our reconstructed high-frequency detail features and smooth parts are better than DBIN+ and MHF, and the Spe method has blurred boundaries. The visualization results show that the proposed method has good visual effect.
On the whole, the MINet method outperforms existing methods, and the results on the four datasets show that the objective evaluation of the MINet method is well consistent with the subjective evaluations. In particular, our proposed method operates with the fewest parameters and the fastest speed, highlighting the advantages of light weight.

3.4. Ablation Study

In this section, we evaluated the performance of the residual activity factor proposed in Section 2.1 as shown in Table 5. The residual activity value of our proposed MINet on HARVARD data is 0.9873, which is closer to 1 than the state-of-the-art hyperspectral fusion network. Not only is it demonstrated that residual factors can be used to evaluate network activity, but also the superiority of the proposed network is demonstrated. At the same time, we also compared the MINet without the beta network structure to illustrate the performance of the beta network.
Then, to verify the performance of the proposed beta network, we compare the performance of MINet and the network after removing the beta network as shown in Table 6. The results of the four evaluation metrics show that the beta network has a positive gain effect on the hyperspectral reconstruction performance. Although the added beta network increases the number of parameters by about 7k, the PSNR is improved by 0.94 dB, and ASSIM, SAM and ERGAS were all significantly improved.

4. Conclusions

In this paper, we proposed a lightweight multi-level information network (MINet) for multispectral and hyperspectral image fusion to meet the real-time reconstruction requirements of hyperspectral imaging. MINet not only has efficient calculation speed, but also maintains satisfactory accuracy, which is suitable for resource-constrained equipment.
In contrast to the traditional pruning of generalised sharpening networks to obtain lightweight networks, we first analysed the working mechanism of the residual module, which plays a key role in existing hyperspectral fusion networks, and then defined a residual activity factor to quantify the residual learning capability of the module and assess the effectiveness of the proposed module. This inspired us to design a constraint block based on global variance fine-tuning (GVF-RCB) to improve the feature extraction and fusion module. To further improve the efficiency of the fusion network, the proposed network uses multi-level spectral information and feature maps to share a hidden layer state in batches, mutually perceive the mean information of each channel map, and complete feature extraction and fusion simultaneously, simplifying the network complexity while improving the efficiency of feature extraction and information fusion. The ablation experiments also further validate the effectiveness of this module. In addition, we cascade multiple GVF-RCBs in MINet to progressively inherit different levels of feature content into the fused image.
To fully evaluate the performance of our proposed MINet network for pansharpen-ng, we chose two simulated datasets and two real-world hyperspectral datasets for our experiments. The experimental results demonstrate that the proposed method achieves an average improvement of 0.3 dB in PSNR and achieves state-of-the-art results in most evaluation metrics. More importantly, the network forward inference is three times faster than the best current methods, while reducing the number of parameters by a factor of five. This lightweight design idea and the results validate the effectiveness of the network.
In future work, we will focus on the hardware deployment of the network model, applying it to existing spectral imaging systems and testing it, modifying and debugging the network for specific problems in the engineered system. In addition, we will also test the performance of the residual activity factor on the dataset used in ILSVRC and COCO 2015 competitions.

Author Contributions

Conceptualization, M.M., Y.N. and F.L.; methodology, M.M.; software, Y.N.; validation, F.L., C.L. and G.S.; formal analysis, M.M.; investigation, M.M. and Y.N.; resources, G.S.; data curation, C.L. and F.L.; writing—original draft preparation, M.M. and Y.N.; writing—review and editing, C.L. and G.S.; visualization, M.M. and F.L.; supervision, Y.N. and G.S.; project administration, Y.N.; funding acquisition, Y.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural National Science Foundation of China under grants 61875157, 61672404 and 61751310, in part by the National Defense Basic Scientific Research Program of China under grants JCKY2017204B102, in part by Science and Technology Plan of Xi’an under grant 20191122015KYPT011JC013, in part by the Fundamental Research Funds for the Central Universities of China under grants No. RW200141, JC1904 and JX18001, in part by the National Key Research and Development Project of China under grant 2018YFB2202400.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  2. Khan, M.J.; Khan, H.S.; Yousaf, A.; Khurshid, K.; Abbas, A. Modern Trends in Hyperspectral Image Analysis: A Review. IEEE Access 2018, 6, 14118–14129. [Google Scholar] [CrossRef]
  3. Taskesen, B.; Koz, A.; Alatan, A.; Weatherbee, O. Change Detection for Hyperspectral Images Using Extended Mutual Information and Oversegmentation. In Proceedings of the 2018 9th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 23–26 September 2018. [Google Scholar]
  4. Hs, A.; Zw, A.; Axzbc, D.; Qian, D.E. Low rank and collaborative representation for hyperspectral anomaly detection via robust dictionary construction. ISPRS J. Photogramm. Remote Sens. 2020, 169, 195–211. [Google Scholar]
  5. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised Spatial–Spectral Feature Learning by 3D Convolutional Autoencoder for Hyperspectral Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  6. Gomez, C.; Viscarra Rossel, R.A.; McBratney, A.B. Soil organic carbon prediction by hyperspectral remote sensing and field vis-NIR spectroscopy: An Australian case study. Geoderma 2008, 146, 403–411. [Google Scholar] [CrossRef]
  7. Gasmi, A.; Gomez, C.; Chehbouni, A.; Dhiba, D.; Elfil, H. Satellite Multi-Sensor Data Fusion for Soil Clay Mapping Based on the Spectral Index and Spectral Bands Approaches. Remote Sens. 2022, 14, 1103. [Google Scholar] [CrossRef]
  8. Dian, R.; Li, S.; Guo, A.; Fang, L. Deep Hyperspectral Image Sharpening. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5345–5355. [Google Scholar] [CrossRef]
  9. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE Pan Sharpening of Very High Resolution Multispectral Images. IEEE Trans. Geosci. Remote Sens. 2007, 46, 228–236. [Google Scholar] [CrossRef]
  10. Vivone, G. Robust Band-Dependent Spatial-Detail Approaches for Panchromatic Sharpening. IEEE Trans. Geosci. Remote Sens. Publ. IEEE Geosci. Remote Sens. Soc. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  11. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2010, 49, 295–309. [Google Scholar] [CrossRef]
  12. Liu, J.G. Smoothing Filter-based Intensity Modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  13. Otazu, X.; Gonzalez-Audicana, M.; Fors, O.; Nunez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  14. Shensa, M. The discrete wavelet transform: Wedding the a trous and Mallat algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef] [Green Version]
  15. Burt, P.; Adelson, E. The Laplacian Pyramid as a Compact Image Code. IEEE Trans. Commun. 1983, 31, 532–540. [Google Scholar] [CrossRef]
  16. Vivone, G.; Restaino, R.; Mura, M.D.; Licciardi, G.; Chanussot, J. Contrast and Error-Based Fusion Schemes for Multispectral Image Pansharpening. IEEE Geosci. Remote Sens. Lett. 2013, 11, 930–934. [Google Scholar] [CrossRef] [Green Version]
  17. Vivone, G.; Restaino, R.; Chanussot, J. Full Scale Regression-Based Injection Coefficients for Panchromatic Sharpening. IEEE Trans. Image Process. 2018, 27, 3418–3431. [Google Scholar] [CrossRef]
  18. Fu, X.; Lin, Z.; Huang, Y.; Ding, X. A Variational Pan-Sharpening with Local Gradient Constraints. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  19. Chen, C.; Li, Y.; Wei, L.; Huang, J. Image Fusion with Local Spectral Consistency and Dynamic Gradient Sparsity. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  20. He, X.; Condat, L.; Bioucas-Dias, J.M.; Chanussot, J.; Xia, J. A New Pansharpening Method Based on Spatial and Spectral Sparsity Priors. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2014, 23, 4160–4174. [Google Scholar] [CrossRef]
  21. Dong, W.; Fu, F.; Shi, G.; Cao, X.; Wu, J.; Li, G.; Li, X. Hyperspectral Image Super-Resolution via Non-Negative Structured Sparse Representation. IEEE Trans. Image Process. 2016, 25, 2337–2352. [Google Scholar] [CrossRef]
  22. Qin, Z.; Zeng, Q.; Zong, Y.; Xu, F. Image inpainting based on deep learning: A review. Displays 2021, 69, 102028. [Google Scholar] [CrossRef]
  23. Ma, M.; Niu, Y.; Li, R. Bayesian network based general correspondence retrieval method for depth sensing with single-shot structured light. Displays 2021, 67, 102001. [Google Scholar] [CrossRef]
  24. Wang, Z.; Li, J.; Zhang, T. Discriminative graph convolution networks for hyperspectral image classification. Displays 2021, 70, 102114. [Google Scholar] [CrossRef]
  25. Shah, V.P.; Younan, N.H.; King, R.L. An Efficient Pan-Sharpening Method via a Combined Adaptive PCA Approach and Contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  26. Meng, X.; Shen, H.; Li, H.; Zhang, L.; Fu, R. Review of the Pansharpening Methods for Remote Sensing Images Based on the Idea of Meta-analysis: Practical Discussion and Challenges. Inf. Fusion 2018, 46, 102–113. [Google Scholar] [CrossRef]
  27. Yang, C.; Zhan, Q.; Liu, H.; Ma, R. An IHS-Based Pan-Sharpening Method for Spectral Fidelity Improvement Using Ripplet Transform and Compressed Sensing. Sensors 2018, 18, 3624. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Meng, X.; Xiong, Y.; Shao, F.; Shen, H.; Zhang, H. A Large-Scale Benchmark Data Set for Evaluating Pansharpening Performance: Overview and Implementation. IEEE Geosci. Remote Sens. Mag. 2021, 9, 18–52. [Google Scholar] [CrossRef]
  29. Wang, Z.; Ziou, D.; Armenakis, C.; Li, D.; Li, Q. A comparative analysis of image fusion methods. IEEE Trans. Geoence Remote Sens. 2005, 43, 1391–1402. [Google Scholar] [CrossRef]
  30. Alparone, L.; Baronti, S.; Aiazzi, B.; Garzelli, A. Spatial Methods for Multispectral Pansharpening: Multiresolution Analysis Demystified. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2563–2576. [Google Scholar] [CrossRef]
  31. Li, S.; Yang, B. A New Pan-Sharpening Method Using a Compressed Sensing Technique. IEEE Trans. Geosci. Remote Sens. 2011, 49, 738–746. [Google Scholar] [CrossRef]
  32. Zhu, X.X.; Bamler, R. A Sparse Image Fusion Algorithm With Application to Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
  33. Ballester, C.; Caselles, V.; Igual, L.; Verdera, J.; Rougé, B. A Variational Model for P+XS Image Fusion. Int. J. Comput. Vis. 2006, 69, 43–58. [Google Scholar] [CrossRef]
  34. Meng, X.; Shen, H.; Yuan, Q.; Li, H.; Zhang, L.; Sun, W. Pansharpening for Cloud-Contaminated Very High-Resolution Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2840–2854. [Google Scholar] [CrossRef]
  35. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object Detection With Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Huang, W.; Xiao, L.; Wei, Z.; Liu, H.; Tang, S. A New Pan-Sharpening Method With Deep Neural Networks. IEEE Geosci. Remote Sens. Lett. 2017, 12, 1037–1041. [Google Scholar] [CrossRef]
  38. Giuseppe, M.; Davide, C.; Luisa, V.; Giuseppe, S. Pansharpening by Convolutional Neural Networks. Remote Sens. 2016, 8, 594. [Google Scholar]
  39. Scarpa, G.; Vitale, S.; Cozzolino, D. Target-Adaptive CNN-Based Pansharpening. IEEE Trans. Geosci. Remote Sens. 2017, 56, 5443–5457. [Google Scholar] [CrossRef] [Green Version]
  40. Jiang, M.; Shen, H.; Li, J.; Yuan, Q.; Zhang, L. A differential information residual convolutional neural network for pansharpening. ISPRS J. Photogramm. Remote Sens. 2020, 163, 257–271. [Google Scholar] [CrossRef]
  41. Zhao, Z.; Zhang, J.; Xu, S.; Sun, K.; Huang, L.; Liu, J.; Zhang, C. FGF-GAN: A Lightweight Generative Adversarial Network for Pansharpening via Fast Guided Filter. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021. [Google Scholar] [CrossRef]
  42. Liu, X.; Wang, Y.; Liu, Q. PSGAN: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018. [Google Scholar]
  43. Diao, W.; Zhang, F.; Sun, J.; Xing, Y.; Zhang, K.; Bruzzone, L. ZeRGAN: Zero-Reference GAN for Fusion of Multispectral and Panchromatic Images. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–15. [Google Scholar] [CrossRef]
  44. Tba, B.; Yka, C.; Aka, B. Pansharpening approach via two-stream detail injection based on relativistic generative adversarial networks. Expert Syst. Appl. 2022, 188, 115996. [Google Scholar]
  45. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  46. Bandara, W.; Valanarasu, J.; Patel, V.M. Hyperspectral Pansharpening Based on Improved Deep Image Prior and Residual Reconstruction. IEEE Trans. Geosci. Remote Sens. 2021. [Google Scholar] [CrossRef]
  47. Wang, W.; Zeng, W.; Huang, Y.; Ding, X.; Paisley, J. Deep Blind Hyperspectral Image Fusion. Int. Conf. Comput. Vis. 2019. [Google Scholar]
  48. Xie, Q.; Zhou, M.; Zhao, Q.; Meng, D.; Zuo, W.; Xu, Z. Multispectral and Hyperspectral Image Fusion by MS/HS Fusion Net. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  49. Xie, Q.; Zhou, M.; Zhao, Q.; Xu, Z.; Meng, D. MHF-Net: An Interpretable Deep Network for Multispectral and Hyperspectral Image Fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1457–1473. [Google Scholar] [CrossRef] [PubMed]
  50. Zhu, Z.; Hou, J.; Chen, J.; Zeng, H.; Zhou, J. Hyperspectral Image Super-resolution via Deep Progressive Zero-centric Residual Learning. IEEE Trans. Image Process. 2020, 1423–1438. [Google Scholar] [CrossRef]
  51. Cheng, J.; Peng, X.; Tang, X.; Tu, W.; Xu, W. MIFNet: A lightweight multiscale information fusion network. Int. J. Intell. Syst. 2022, 37, 5617–5642. [Google Scholar] [CrossRef]
  52. Hellman, B.; Takashima, Y. Angular and spatial light modulation by single digital micromirror device for multi-image output and nearly-doubled étendue. Opt. Express 2019, 27, 21477–21496. [Google Scholar] [CrossRef] [Green Version]
  53. Dong, X.; Xiao, X.; Pan, Y.; Wang, G.; Yu, Y. DMD-based hyperspectral imaging system with tunable spatial and spectral resolution. Opt. Express 2019, 27, 16995. [Google Scholar] [CrossRef]
  54. Qiao, M.; Liu, X.; Yuan, X. Snapshot spatial-temporal compressive imaging. Opt. Lett. 2020, 45, 1659–1662. [Google Scholar] [CrossRef]
  55. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  56. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. International Conference on Learning Representations. 2015. Available online: Https://arxiv.org/abs/1409.1556 (accessed on 26 September 2022).
  57. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized Assorted Pixel Camera: Postcapture Control of Resolution, Dynamic Range, and Spectrum. IEEE Trans. Image Process. Publ. IEEE Signal Process. Soc. 2010, 19, 2241. [Google Scholar] [CrossRef] [Green Version]
  58. Chakrabarti, A.; Zickler, T. Statistics of real-world hyperspectral images. In Proceedings of the IEEE Conference on Computer Vision Pattern Recognition, Springs, CO, USA, 20–25 June 2011. [Google Scholar]
  59. Santurkar, S.; Tsipras, D.; Ilyas, A.; Mądry, A. How Does Batch Normalization Help Optimization? In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; NIPS’18. Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 2488–2498. [Google Scholar]
  60. Simoes, M.; Bioucas-Dias, J.; Almeida, L.B.; Chanussot, J. A Convex Formulation for Hyperspectral Image Superresolution via Subspace-Based Regularization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3373–3388. [Google Scholar] [CrossRef] [Green Version]
  61. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J.M. Fusing Hyperspectral and Multispectral Images via Coupled Sparse Tensor Factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef] [PubMed]
  62. Yokoya, N.; Iwasaki, A.; Yairi, T. Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion. IEEE Trans. Geosci. Remote Sens. 2012, 50, 528–537. [Google Scholar] [CrossRef]
  63. Dian, R.; Li, S.; Fang, L. Learning a Low Tensor-Train Rank Representation for Hyperspectral Image Super-Resolution. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2672–2683. [Google Scholar] [CrossRef] [PubMed]
  64. Fu, X.; Wang, W.; Huang, Y.; Ding, X.; Paisley, J. Deep Multiscale Detail Networks for Multiband Spectral Image Sharpening. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 2090–2104. [Google Scholar] [CrossRef]
  65. Liu, X.; Liu, Q.; Wang, Y. Remote Sensing Image Fusion Based on Two-stream. Fusion Netw. 2017, 55, 1–15. [Google Scholar] [CrossRef] [Green Version]
  66. Zhou, W.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar]
  67. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and Panchromatic Data Fusion Assessment Without Reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef] [Green Version]
  68. Xu, Y.; Du, B.; Zhang, L.; Cerra, D.; Saux, B.L. Advanced Multi-Sensor Optical Remote Sensing for Urban Land Use and Land Cover Classification: Outcome of the 2018 IEEE GRSS Data Fusion Contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1709–1724. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of residual learning degradation. Dashed lines indicate non-existent paths. (a) represents x 1 F ( x ) 1 ; (b) represents x 1 F ( x ) 1 ; (c) represents x 1 F ( x ) 1 .
Figure 1. Schematic diagram of residual learning degradation. Dashed lines indicate non-existent paths. (a) represents x 1 F ( x ) 1 ; (b) represents x 1 F ( x ) 1 ; (c) represents x 1 F ( x ) 1 .
Remotesensing 14 05600 g001
Figure 2. The relationship between the normalized PSNR and the active residual activity factor η at different stages of the residual learning module.
Figure 2. The relationship between the normalized PSNR and the active residual activity factor η at different stages of the residual learning module.
Remotesensing 14 05600 g002
Figure 3. The overall flowchart of the proposed MINet for multispectral and hyperspectral image fusion.
Figure 3. The overall flowchart of the proposed MINet for multispectral and hyperspectral image fusion.
Remotesensing 14 05600 g003
Figure 4. The architecture of the lightweight residual constraint block based on global variance fine-tuning (GVF-RCB).
Figure 4. The architecture of the lightweight residual constraint block based on global variance fine-tuning (GVF-RCB).
Remotesensing 14 05600 g004
Figure 5. Mean and corresponding beta value of hyperspectral image and feature map. (a,c,e) represent the data of the three groups of images in the CAVE dataset; (b,d,f) represent the data of the three groups of images in the HARVARD dataset.
Figure 5. Mean and corresponding beta value of hyperspectral image and feature map. (a,c,e) represent the data of the three groups of images in the CAVE dataset; (b,d,f) represent the data of the three groups of images in the HARVARD dataset.
Remotesensing 14 05600 g005
Figure 6. The schematic diagram of beta network structure.
Figure 6. The schematic diagram of beta network structure.
Remotesensing 14 05600 g006
Figure 7. Comparisons of the error maps between the spectral bands of reconstructed HR-HSIs.
Figure 7. Comparisons of the error maps between the spectral bands of reconstructed HR-HSIs.
Remotesensing 14 05600 g007
Figure 8. Compare PSNR values for each spectral band of the reconstructed HR-HSI. (a,c,e) represent the data of the three groups of images in the CAVE dataset; (b,d,f) represent the data of the three groups of images in the HARVARD dataset.
Figure 8. Compare PSNR values for each spectral band of the reconstructed HR-HSI. (a,c,e) represent the data of the three groups of images in the CAVE dataset; (b,d,f) represent the data of the three groups of images in the HARVARD dataset.
Remotesensing 14 05600 g008
Figure 9. The visual comparisons of fusion results obtained by different methods on NCALM dataset.
Figure 9. The visual comparisons of fusion results obtained by different methods on NCALM dataset.
Remotesensing 14 05600 g009
Figure 10. The visual comparisons of fusion results obtained by different methods on World View-2(WV-2) dataset.
Figure 10. The visual comparisons of fusion results obtained by different methods on World View-2(WV-2) dataset.
Remotesensing 14 05600 g010
Table 1. Comparison of residual activity factors with PSNR, SAM and ERGAS in several super-resolution networks based on residual modules, where η is the defined residual activity factor. The best results are shown in bold.
Table 1. Comparison of residual activity factors with PSNR, SAM and ERGAS in several super-resolution networks based on residual modules, where η is the defined residual activity factor. The best results are shown in bold.
MethodDBIN+MHF-NetSpeNet
η 0.74070.82150.9229
PSNR46.2046.5946.69
SAM3.81713.54823.4380
ERGAS0.33980.32060.3171
Table 2. Experiment with ten PS methods on the HARVARD dataset. The best performance is shown in bold.
Table 2. Experiment with ten PS methods on the HARVARD dataset. The best performance is shown in bold.
MethodPSNR (dB)ASSIMSAMERGASParams (M)FLOPs (T)TimeGPU Memory (MB)
NSSR34.860.93127.18100.9435--492.5 s-
CNMF36.140.93275.94710.9492--200.2 s-
LTTR44.320.98094.45280.5223--2316.7 s-
CSTF43.640.97175.96560.5943--97.3 s-
MHF-Net46.590.99363.54820.32063.0072.1361647.2 ms8087
DBIN+46.200.99383.81710.33983.0834.1252176.8 ms4433
DMDNet46.440.99323.75620.33152.2512.4871456.6 ms6847
ResTFNet45.860.99163.82540.33961.5252.3651263.5 ms6954
SpeNet46.690.99343.43800.31710.7131.086597.1 ms7885
Ours46.910.99363.39270.31110.1280.255197.4 ms2511
Table 3. Experiment with ten PS methods on the CAVE dataset. The best performance is shown in bold.
Table 3. Experiment with ten PS methods on the CAVE dataset. The best performance is shown in bold.
MethodPSNR (dB)ASSIMSAMERGASParams (M)FLOPs (G)TimeGPU Memory (MB)
NSSR30.450.915910.35951.5184--117.5 s-
CNMF32.440.887114.65561.3654--50.5 s-
LTTR41.620.970610.81580.4541--506.5 s-
CSTF41.730.958912.25540.4753--19.6 s-
MHF-Net45.240.99326.96550.29543.007534974.6 ms5783
DBIN+43.340.99367.06300.36503.0832063531.6 ms1619
DMDNet45.750.99297.12060.32452.2511768654.3 ms3517
ResTFNet45.120.99157.04590.33691.5252635362.2 ms3268
SpeNet46.490.99316.77350.25580.713271146.0 ms6857
Ours46.610.99336.68530.24820.12864 G48.1 ms1169
Table 4. Quantitative performance comparison with the investigated methods on the NCALM dataset. The best results are shown in bold.
Table 4. Quantitative performance comparison with the investigated methods on the NCALM dataset. The best results are shown in bold.
MethodPSNR (dB)ASSIMSAMERGASCCRASEQNR
MHF-Net39.780.97893.83171.22910.85615.48480.9549
DBIN+39.530.97563.85111.25820.84625.98480.9436
DMDNet39.260.97353.86211.26450.82195.68470.9477
ResTFNet39.760.97413.84561.24560.84985.16780.9587
SpeNet39.960.97933.82881.22220.88544.86440.9650
MINet40.220.98153.80761.21870.89014.85010.9654
Table 5. Residual activity comparison with hyperspectral image fusion network on the HARVARD dataset. The best results are shown in bold.
Table 5. Residual activity comparison with hyperspectral image fusion network on the HARVARD dataset. The best results are shown in bold.
MethodDBIN+MHF-NetSpeNetMINet-No-BetaMINet
η 0.74040.82150.92290.89570.9873
Table 6. Contrast effect of beta network in MINet. The best results are shown in bold.
Table 6. Contrast effect of beta network in MINet. The best results are shown in bold.
NetworkPSNR (dB)ASSIMSAMERGASParams (K)FLOPs (G)
MINet-no-beta45.670.98588.65440.3278121.704128
MINet46.610.99336.68530.2482128.345128
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, M.; Niu, Y.; Liu, C.; Li, F.; Shi, G. A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion. Remote Sens. 2022, 14, 5600. https://doi.org/10.3390/rs14215600

AMA Style

Ma M, Niu Y, Liu C, Li F, Shi G. A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion. Remote Sensing. 2022; 14(21):5600. https://doi.org/10.3390/rs14215600

Chicago/Turabian Style

Ma, Mingming, Yi Niu, Chang Liu, Fu Li, and Guangming Shi. 2022. "A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion" Remote Sensing 14, no. 21: 5600. https://doi.org/10.3390/rs14215600

APA Style

Ma, M., Niu, Y., Liu, C., Li, F., & Shi, G. (2022). A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion. Remote Sensing, 14(21), 5600. https://doi.org/10.3390/rs14215600

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop