Next Article in Journal
Ionospheric Perturbations Due to Large Thunderstorms and the Resulting Mechanical and Acoustic Signatures
Next Article in Special Issue
From Model-Based Optimization Algorithms to Deep Learning Models for Clustering Hyperspectral Images
Previous Article in Journal
Mapping Waterlogging Damage to Winter Wheat Yield Using Downscaling–Merging Satellite Daily Precipitation in the Middle and Lower Reaches of the Yangtze River
Previous Article in Special Issue
High-Resolution Mapping of Soil Organic Matter at the Field Scale Using UAV Hyperspectral Images with a Small Calibration Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rethinking 3D-CNN in Hyperspectral Image Super-Resolution

School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(10), 2574; https://doi.org/10.3390/rs15102574
Submission received: 23 April 2023 / Revised: 8 May 2023 / Accepted: 12 May 2023 / Published: 15 May 2023
(This article belongs to the Special Issue Advances in Hyperspectral Remote Sensing Image Processing)

Abstract

:
Recently, CNN-based methods for hyperspectral image super-resolution (HSISR) have achieved outstanding performance. Due to the multi-band property of hyperspectral images, 3D convolutions are natural candidates for extracting spatial–spectral correlations. However, pure 3D CNN models are rare to see, since they are generally considered to be too complex, require large amounts of data to train, and run the risk of overfitting on relatively small-scale hyperspectral datasets. In this paper, we question this common notion and propose Full 3D U-Net (F3DUN), a full 3D CNN model combined with the U-Net architecture. By introducing skip connections, the model becomes deeper and utilizes multi-scale features. Extensive experiments show that F3DUN can achieve state-of-the-art performance on HSISR tasks, indicating the effectiveness of the full 3D CNN on HSISR tasks, thanks to the carefully designed architecture. To further explore the properties of the full 3D CNN model, we develop a 3D/2D mixed model, a popular kind of model prior, called Mixed U-Net (MUN) which shares a similar architecture with F3DUN. Through analysis on F3DUN and MUN, we find that 3D convolutions give the model a larger capacity; that is, the full 3D CNN model can obtain better results than the 3D/2D mixed model with the same number of parameters when it is sufficiently trained. Moreover, experimental results show that the full 3D CNN model could achieve competitive results with the 3D/2D mixed model on a small-scale dataset, suggesting that 3D CNN is less sensitive to data scaling than what people used to believe. Extensive experiments on two benchmark datasets, CAVE and Harvard, demonstrate that our proposed F3DUN exceeds state-of-the-art HSISR methods both quantitatively and qualitatively.

1. Introduction

Hyperspectral imaging systems receive reflections of multiple electromagnetic spectrum intervals from objects and acquire multi-band images. Thus, hyperspectral images (HSI) preserve the complete spectrum curve for each pixel, and this information is widely used in numerous areas, such as mineral exploration [1], medical diagnosis [2], and plant detection [3], etc. However, due to hardware limitations, most hyperspectral cameras must compromise between spatial resolution and spectral resolution. As a result, the spatial resolution of HSIs is relatively low compared to natural images, which becomes an obstacle to downstream high-level tasks such as image classification [4] and anomaly detection [5].
To solve this challenge, hyperspectral image super-resolution (HSISR) is in high demand, with the aim of reconstructing high-resolution (HR) MSIs given one or more inputs. Since the loss of spatial resolution comes from the multi-band property of an HSI, a natural idea is to fuse high-resolution multispectral images (HR-MSIs) and low-resolution hyperspectral images (LR-HSIs) to generate high-resolution hyperspectral outputs. Some works use corresponding RGB images as supplementary information [6,7,8], while others use panchromatic (PAN) images [9,10]. The main reason for the shortage of fusion-based methods is because co-registered HSI-MSI pairs are hard to collect, which limits their application scenarios. Therefore, we focus on an HSISR that is executed solely using LR-HSIs, i.e., single hyperspectral image super-resolution (SHISR), in this article.
Single hyperspectral image super-resolution is an ill-posed problem, as multiple high-resolution equivalents may correspond to a single low-resolution image. In the early period, this kind of method is highly dependent on handcrafted priors, such as low-rank [11] and sparsity [12]. However, these manually designed priors only reflect some aspects of hyperspectral data. Recently, convolutional neural networks (CNNs) have shown a strong representation ability, and they have been widely introduced to improve natural image super-resolution (SR) [13,14]. However, the mainstream of natural image SR methods is 2D CNN, which can lead to severe spectral distortion when applied to multi-band HSIs. Recently, some methods have emerged that combine hand-craft priors and neural networks and introduce mathematical modeling considering the properties of HSI, such as [15,16,17].
The problem caused by the high dimensionality in HSIs is severe. In [18], Cai et al. tried to solve this issue with a 3D attention mechanism and information bottleneck for classification tasks. However, in super-resolution, the model needs to preserve sufficient information of the input; thus, a natural way to solve this problem is to introduce 3D convolutions, which are good at extracting spatial–spectral correlations. The key difference between 3D and 2D convolutions is that 3D convolutions share parameters not only on spatial sides, but also on spectral dimension (see Figure 1). Thus, 3D convolutions can utilize the similarities between adjacent bands. The pioneering work of 3D CNN on HSISR is 3DFCNN [19], which consists of five cascaded 3D convolutional layers. However, the wave of full 3D CNN models did not last long, and 3D CNNs have received two main critiques. On the one hand, it is believed that 3D models are too complex to have good performance on HSI datasets with limited scale and are at risk of overfitting. On the other hand, some argue that the HSISR model should focus more on spatial enhancement, and 3D models cannot accomplish this. The latter comment leads to 3D/2D mixed models, which introduce extra 2D convolutions to boost spatial details. Representatives of such methods are MCNet [20] and ERCSR [21]. The former introduces multiple 2D branches to extract multi-scale spatial information from the features generated by 3D convolutions in each block (see Figure 2d), and the latter adds one 2D unit after every 3D unit to concentrate on spatial features (Figure 2c). Recently, graph-based methods have been applied to many tasks in hyperspectral image processing, such as classification [22,23,24], clustering [25,26,27], target detection [28], and anomaly detection [29], and have achieved good performance.
However, existing 3D HSISR models are very simple and plain, and they are not combined with some advanced inventions of deep learning. Therefore, it is worth questioning whether the critical views on 3D models still hold. We believe that the potential of 3D models has not been fully explored yet. This drives us to revisit those structures that have been proven effective in the super-resolution field and combine them with 3D CNN. Specifically, we propose a simple yet effective full 3D CNN model with a U-Net architecture, called Full 3D U-Net (F3DUN). F3DUN replaces the 2D blocks in normal U-Net with 3D convolutions and removes the downsampling–upsampling process in the middle. Experiments show that F3DUN outperforms state-of-the-art methods, which proves the effectiveness of full 3D models in HSISR tasks. Moreover, we design a 3D/2D mixed model that shares the same U-Net structure with F3DUN called Mixed U-Net (MUN), where cascaded 3D convolutions are used for feature extraction in the shallow half while 2D convolutions are gathered in the deep half for spatial enhancement. F3DUN and MUN provide an excellent pair to compare the two kinds of model priors, 3D/2D mixed models and full 3D CNN models. In Figure 2, we summarize the main model priors in HSISR. The experimental results show that full 3D CNN models have larger modeling capacities than the 3D/2D mixed model with the same number of parameters; that is, full 3D CNN models have better performance with sufficient training data. Another common concern regarding full 3D CNN models is that they cannot perform well on small-scale datasets, since they demand a lot of training samples. A comparison on a small-scale dataset disproves this point, since F3DUN still obtains competitive results with MUN. With little harm on performance, F3DUN eliminates severe overfitting. Thus, we conclude that full 3D CNN models are less sensitive to the scale of training sets than we used to think and that overfitting can be prevented with carefully designed architectures.
In summary, the contributions we made in this paper are listed here:
  • We rethink the role that 3D CNN plays in the HSISR field and design a novel full 3D CNN model based on thd U-Net architecture called Full 3D U-Net (F3DUN). Experimentally, it outperforms existing state-of-the-art, single-image SR methods, which proves the effectiveness of full 3D CNN models in this field.
  • We develop a mixed 3D/2D model that shares the same structure with F3DUN, termed Mixed U-Net (MUN), for comparison. Extensive analysis on the two models shows that the full 3D CNN model has a larger modeling capacity than the 3D/2D mixed model with the same number of parameters; thus, it performs better with large-scale datasets.
  • We explore the relationship between the scale of training samples and the prior of the model. We argue that the full 3D CNN model can obtain competitive results on small-scale training sets compared with the 3D/2D mixed model, which concludes that the full 3D CNN model is more robust with respect to the amount of training samples than commonly thought.
The rest of this paper is organized as follows: In Section 2, we briefly introduce related works on hyperspectral and natural image super-resolution. Then, in Section 3, we describe the model designs of F3DUN and MUN in detail. Experiments on public datasets are shown and discussed in Section 4, and the conclusion is in Section 5. For ease of reading, we have provided a list of abbreviations in Table 1 for reference.

2. Related Work

In this section, we briefly review some research fields related to our work. First, we introduce 3D convolution and its applications in computer vision tasks. Then, we give a big picture of hyperspectral image super-resolution, including single hyperspectral image super-resolution and fusion-based hyperspectral image super-resolution. At the end of this section, the development of single natural image super-resolution is summarized.

2.1. Three-Dimensional Convolution

Three-dimensional convolution shares parameters not only on spatial dimensions but also on channel dimension, which is suitable to deal with cube data such as HSIs or video. A typical application of 3D convolution is hyperspectral image classification [30,31,32,33], since 3D convolution can tackle the rich spatial–spectral information contained in HSIs. Furthermore, Jiang et al. [34] used a 3D convolutional neural network for dehazing in HSIs. The application of 3D convolution expands to some other remote sensing data, such as SLAM [35] and LiDAR [36,37]. Three-dimensional convolution also plays an important role in video processing due to its ability to capture inter-frame relationships. Li et al. [38] proposed a 3D CNN for video person re-identification, and Ying et al. [39] employed deformable 3D convolution in a video super-resolution task.

2.2. Single Hyperspectral Image Super-Resolution

Single hyperspectral image super-resolution (SHISR) aims to reconstruct an HR-MSI solely from its LR opponent. Without the need of an auxiliary image, SHISR can be widely applied and thus attracts considerable research attention. The pioneering work is [40], in which Akgun et al. proposed an acquisition model and applied the projection onto convex sets (POCS) algorithm [41] on SHISR. Subsequently, Huang et al. [12] introduced low-rank and sparsity constraints on dictionary learning to reconstruct HR-MSIs. The aforementioned works can be categorized as model-driven methods, since they rely on manually designed priors. With the coming of the deep learning era, data-driven frameworks are also being applied in the SHISR field. Yuan et al. [42] tried to transfer knowledge from natural image SR to SHISR and combined CNN and non-negative matrix factorization (NMF) to super-resolve MSIs. Xie et al. [43] utilized CNN and NMF to extract spatial and spectral features separately, super-resolving HSIs in a non-end-to-end manner. Due to the strong representative ability, Mei et al. [19] first proposed a five-layer full 3D CNN (3D-FCNN) for SHISR. Subsequently, Li et al. [20] designed a mixed 2D/3D CNN (MCNet), introducing multiple 2D branches to extract multi-scale spatial information. Furthermore, in [21], the authors replaced the redundant multi-branch structure with interleaved 2D/3D blocks. On the other hand, Li et al. [44] proposed a full 2D CNN model (GDRRN) with grouped convolutions and recursive architecture. Then, Jiang et al. [45] divided HSIs into overlapped band groups based on the similarity of adjacent bands, and designed a multi-branch model (SSPSR) in a divide-and-merge manner. Li et al. [46] presented intra-group and inter-group fusion to gather features from different band groups. Another category of SHISR methods is sequence models, motivated by video SR. Wang et al. [47] proposed a dual-channel model, fusing spatial and spectral features from neighboring bands and super-resolving LR-MSIs band by band. Recently, Liu et al. [48] proposed a parallel network called Interactformer for SHISR, which contains a Transformer branch and a 3D-CNN branch; the Transformer branch is used to capture long-range dependencies to obtain global features, and the 3-D CNN branch is used to extract local features while preserving the spectral correlation of HSIs.

2.3. Fusion-Based Hyperspectral Image Super-Resolution

Fusion-based hyperspectral image super-resolution is a classical method of HSISR with a long history. Its basic idea is to fuse the information of LR-HSIs and HR-MSIs to improve the spatial resolution of the LR-HSIs. Yokoya et al. [49] proposed an approach based on coupled non-negative matrix factorization (CNMF). In addition, more researchers have taken the abundant spectral information encoded in HSIs into consideration and have exploited various priors, such as sparsity [50,51], non-local similarity [52], tensor and low-rank constraints [53]. Recently, deep learning methods have shown their superiority in feature representation and they rely less on assumptions; thus, they have gradually become popular [54]. Some deep unfolding networks have been proposed, which turn a mathematical model to a neural network [6,7,55,56]. However, all the above methods require co-registered HSI-MSI pairs, which severely limits the scope of their application, since the collection of paired data can be extremely hard.

2.4. Single Natural Image Super-Resolution

Single-image super-resolution (SISR) has been a research topic for a long time. In the early stages, researchers focused on designing good priors to super-resolve RGB or gray images [57,58,59]. With the rise of deep learning, CNN-based SISR methods have played an important role. Dong et al. [60] proposed the first end-to-end model consisting of three convolutional layers. Later, Kim et al. [13] presented a very deep convolutional model of 20 layers for SISR, and Lim et al. [14] introduced residual connection and removed unnecessary modules to make the model more effective and robust. Afterwards, enormous CNN-based works emerged, such as RCAN [61], SAN [62], HAN [63], NLSA [64], IPT [65], and SwinIR [66], which achieved considerable success.

3. Method

In this section, we describe the details of the proposed Full 3D U-Net (F3DUN). Our main interest in F3DUN is whether the performance of previous full 3D CNN models is limited by their inherent nature or by flaws in the designs of their architectures. We also introduce F3DUN’s 3D/2D mixed counterpart, Mixed U-Net (MUN), in order to analyze the effect of 3D convolutions in HSISR models.

3.1. Full 3D U-Net

F3DUN follows the idea of a full 3D CNN that can be traced back to 3DFCNN and, at the same time, integrates advanced experience in model design. As is shown in Figure 3, our full 3D U-Net consists of three parts: shallow feature extraction, nonlinear mapping, and an upsampler. To prevent memory loss and generate multi-scale features, multiple skip connections are introduced.
For clarity, let I L R R L × H × W and I H R R L × s H × s W denote the LR-MSI and HR-MSI, respectively, where H and W represent the height and weight of each band, L is the number of spectrum bands, and s indicates the scale factor. The super-resolved HSI has the same size of I H R and is remarked as I S R . Given I L R as input, we first use a 3D convolution of 3 × 3 × 3 to preliminarily extract information:
F 3 D 0 = C o n v 3 D 0 ( u n s q u e e z e ( I L R ) ) .
The aim of the unsqueeze operation is to transfer I L R to size of 1 × L × H × W so that it can be handled by 3D convolution.
Then, the shallow feature F 3 D 0 is sent to the nonlinear mapping part, which consists of cascaded 3D blocks. Each 3D block in the shallow half of this part has two 3 × 3 convolutions and a ReLU activation function in between. In addition, all 3D blocks use a residual connection for stability, and its formula can be written as:
F 3 D k = B l o c k 3 D k ( F 3 D k 1 ) + F 3 D k 1
= C o n v 3 × 3 × 3 ( R e L U ( C o n v 3 × 3 × 3 ( F 3 D k 1 ) ) ) + F 3 D k 1 .
We use skip connections to link 3D blocks at symmetrical locations in order to prevent memory loss and provide cross-scale information for deep 3D blocks:
F 3 D K k = B l o c k 3 D K k ( [ F 3 D K k , F 3 D k ] ) f o r k = 1 , 2 , . . . , K / 2 ,
where K is the total number of 3D blocks for non-linear mapping, and [ · , · ] denotes concatenation operation. In addition, 3D blocks in the deep half add an extra 1 × 1 × 1 convolution to fuse the information from the previous block and the skip connection. Beyond skip connections within the U-Net architecture, we further introduce three long-distance connections. Two of them are located in the first and second halves of the non-linear mapping part:
F 3 D K / 2 = F 3 D K / 2 + F 3 D 0 ,
F 3 D K = F 3 D K / 2 + F 3 D K ,
and the other one is a global skip connection:
I S R = U p s a m p l e r ( F 3 D K ) + n e a r e s t u p s a m p l e ( I L R ) .
As for the upsampler, the input is F 3 D K , and we use a transposed 3D convolutions to scale it up to b × C × L × s H × s W . At last, a pseudo-3D convolution, which is made of a 1 × 3 × 3 and a 3 × 1 × 1 filter, squeezes the channel number to 1. For 4 × super-resolution, we directly use one transposed 3D convolution layer for upsampling. When the scale factor is 8, we first use one transposed 3D convolution layer to upsample to 4 × , and then use another transposed 3D convolution layer to scale up to the desired size. This progressive upsampling strategy prevents the upsampling factor from being too large beyond the capabilities of a single deconvolution layer.

3.2. Mixed U-Net

The basic idea of Mixed U-Net (MUN) is to divide the non-linear mapping of the network into two parts: a group of 3D convolutions for feature extraction and cascaded 2D convolutions for spatial enhancement. The specialized structure of MUN can effectively avoid the feature confusion problem caused by interleaved structures of 3D and 2D modules in previous 3D/2D mixed models, and skip connections can also effectively suppress memory forgetting. MUN provides a useful tool to compare 3D/2D mixed models and full 3D CNN models, since it shares similar architecture with F3DUN.
As is shown in Figure 4, MUN shares the same architecture with F3DUN, only replacing 3D Blocks with 2D blocks in the deep half of the non-linear part. The 2D blocks follow the structure of a resblock and use a 1 × 1 convolution to fuse the two inputs from the previous block and the skip connection, respectively. The pipeline of the 2D block in MUN can be depicted as:
F 2 D k = B l o c k 2 D ( [ F 2 D k 1 , r e s h a p e ( F 3 D K k ] ) ,
where K is the total number of 3D and 2D blocks. Here, 3D features F 3 D K k are reshaped from b × C × L × H × W to ( b × L ) × C × H × W before being fed into the corresponding 2D block, where b and C denote batch size and feature channels, respectively. At the junction of the 3D group and the 2D group, i.e., after the Nth 3D block, a particular 2D block is added without the fusion convolution.

3.3. Loss Function

Following the main stream of HSISR methods, we employ L1 loss to lead the model to generate sharp results:
L r e c = 1 M i = 1 M | | I S R i I H R i | | 1 ,
where M denotes the number of samples. However, in practice, we find that although reconstruction loss can effectively constrain spatial similarity, spectral distortion is still a serious problem. Here, we additionally introduce a loss based on the SAM [67] criterion for HSI quality assessment to constrain the spectral similarity:
L S A M = 1 s H × s W ( x , y ) arccos ( < I S R ( x , y ) , I H R ( x , y ) > | | I S R ( x , y ) | | 2 | | I H R ( x , y ) | | 2 ) ,
where < · , · > means dot production and ( x , y ) denotes the spatial coordinate of the pixel. Since the derivative of arccos ( x ) is 1 1 x 2 , there exists a gradient explosion problem when x 2 1 . Therefore, we filter the pair of pixels whose cosine similarity lies in [ 1 + ϵ , 1 ϵ ] and only apply SAM loss on them. In practice, we choose ϵ = 10 8 . In general, the total loss function is:
L t o t a l = L r e c + λ L S A M ,
in which λ is a hyperparameter to balance the two losses. In practice, we set λ = 0.05 . This loss function is applied to both F3DUN and MUN.

3.4. Relationship with Other Methods

(a)
3DFCNN: 3DFCNN is the pioneering work of full 3D CNN models for HSISR. It consists of 5 3D convolution layers and uses MSE loss to train. However, there are noticeable drawbacks to it. Most importantly, the model architecture of 3DFCNN is too shallow and it does not combine with advanced ideas of deep learning, such as residual learning and long skip connections. Therefore, in this paper, we rethink the prior of full 3D CNN in HSISR and argue that full 3D CNN models’ bad results are not caused by 3D CNN, but rather a lack of model design. It is obvious that F3DUN is much deeper than 3DFCNN and successfully prevents overfitting and achieves SOTA results.
(b)
Two- and three-dimensional mixed models: The main idea of 3D/2D mixed models is to introduce 2D convolutions into 3D CNN in order to boost the ability of spatial enhancement. They have become popular recently and can be seen as a modification to full 3D CNN models. However, there are two potential problems. On the one hand, the 2D convolutions in 3D/2D models are shared in the spectral dimension, which leads to a risk of spectral distortion. On the other hand, 3D/2D mixed models always face the dilemma of balancing the two kinds of convolutions, where features enhanced by 2D convolutions can be polluted by cascading 3D modules. In MUN, we partially solve the two above problems, but we make a step forward: could a full 3D CNN model outperform 3D/2D mixed ones? Model analysis on MUN and F3DUN supports our idea and proves the advantages of the full 3D CNN model compared with 3D/2D mixed models.

4. Experiments and Analysis

In this section, we conduct comprehensive experiments to evaluate the effectiveness of the proposed F3DUN. Two benchmark datasets, i.e., CAVE [68] and Harvard [69], are used for the comparisons. We present the quantitative and perceptual results of our F3DUN together with seven existing HSISR methods, including Bicubic, EDSR [14], 3DFCNN [19], SSPSR [45], MCNet [20], ERCSR [21], and SFCSR [47]. Furthermore, analysis between F3DUN and MUN using different data scales reveals the data efficiency of our proposed method.

4.1. Datasets

(a)
CAVE dataset: The CAVE dataset [68] is taken by a cooled CCD camera, and the range of the wavelength is from 400 nm to 700 nm, at a step of 10 nm (31 bands). The 32 hyperspectral images are divided into five sections: real and fake, skin and hair, paints, food and drinks, and stuff. Each image has a size of 512 × 512 × 31 , and every band is stored as a grayscale picture separately in the form of a PNG.
(b)
Harvard dataset: The Harvard dataset [69] contains 77 indoor and outdoor hyperspectral images under daylight illumination collected by the Nuance FX, CRI Inc., camera. Each image covers the wavelength range of 400 nm to 700 nm, evenly divided into 31 bands. The spatial resolution is 1040 × 1392 , and all images are stored as .mat files.

4.2. Implementation Details

For the CAVE dataset, we select 12 images as the test set, and the remaining 20 images are used for training. When the scale factor is 4, each training image is cropped into 96 × 96 patches with a stride of 48 pixels. For an 8 × super-resolution, the patch size and stride are 192 × 192 and 36. After that, we use bicubic downsampling to generate LR patches and LR images for test. 10% of the cropped patches are randomly sampled to form a validation set. Several data argumentation methods are applied, including rotations of 90 , 180 and 270 , vertical and horizontal flips, and their combinations.
The training samples of the Harvard dataset are cropped randomly on the spatial dimensions, and each training image generates 72 patches for both 4 × and 8 × super-resolution. The patch sizes of the training samples are 96 × 96 and 128 × 128 for scale factors of 4 and 8, respectively. Additionally, we select 5 and 3 images from the outdoor and indoor categories as test set. Due to the image size in the Harvard dataset being too large, we crop the top-left 1024 × 1024 part of each test image and divide it into 4 sub-images of 512 × 512 . Then the LR images are generated by bicubic downsampling.
We trained F3DUN for scale factors of 4 and 8, with random initialization. The model contains 10 and 12 3D blocks for 4 × and 8 × super-resolutions, respectively, and the number of filters in each convolutional layer is set to 64. We adopt the Adam optimizer [70] and set β 1 = 0.9 and β 2 = 0.999 . The learning rate is initialized as 10 4 and halved every 35 epochs. We employ the batch size to be 16 and train our model for 200 epochs. Our algorithm is performed using PyTorch library with NVIDIA GeForce GTX 2080ti GPUs.

4.3. Evaluation Metrics

We adopt six common metrics to evaluate the quality of super-resolved images: peak signal-to-noise ratio (PSNR), structure similarity (SSIM) [71], cross correlation (CC) [72], spectral angle mapper (SAM) [67], root mean squared error (RMSE), and erreur relative globale adimensionnelle de synthese (ERGAS) [73]. We compute the mean of all bands for PSNR and SSIM, which are generally used to measure image restoration quality quantitatively:
P S N R = 1 L l = 1 L 10 log 10 ( M A X l 2 M S E l ) ,
S S I M = 1 L l = 1 L ( 2 μ I S R l μ I H R l + c 1 ) ( 2 σ I S R l σ I H R l + c 2 ) m n ,
where
m = ( μ I S R l ) 2 + ( μ I H R l ) 2 + c 1 ,
n = ( σ I S R l ) 2 + ( σ I H R l ) 2 + c 2 .
Here, L denotes the total number of bands of the hyperspectral image, and will refer to the same hereinafter. SAM is widely used to compare the spectral difference between two hyperspectral images, and its definition can be found in Equation (10). RMSE measures the absolute error between the reconstructed image and the ground truth. CC and ERGAS are often employed in HS fusion, and they reflect the overall quality of the reconstructed images. Their mathematical definitions are as follows:
C C = 1 L l = 1 L c o v I S R l , I H R l σ I S R l σ I H R l ,
E R G A S = 100 s 1 L l = 1 L R M S E l m e a n I H R l 2 .
Here, s represents the scale factor, and m e a n denotes the average operation. The limit values of the six above indicators are + , 1, 0, 0, 0, and 1, respectively.

4.4. Model Analysis

4.4.1. Comparison with 3D-2D Model Prior

In this section, we perform a comparison on F3DUN and MUN to analyze the advantages and disadvantages of the 3D/2D mixed model and the full 3D CNN model for the HSISR task. Without loss of generality, this model analysis is performed on the CAVE dataset, and the scale factor is set to 4. And we made comparisons on both normal and reduced datasets, in order to explore the relationship between data scale and model priors. The results can be found in Table 2 and Table 3 respectively. The reason why MUN is chosen as a representative of 3D/2D mixed model is manifold. Firstly, MUN has been proven to outperform other HSISR models, especially all previous 3D/2D mixed models (see Table 2 and Table 4). Secondly, F3DUN and MUN share similar architectures, making the comparison fairer and more significant. Thirdly, it is easy to control the number of parameters of the two models so that they are on the same level.
In Table 2, we summarize the quantitative results of the two models on the original CAVE dataset. We control the two models to have roughly the same amount of parameters, thereby ruling out performance gains for one model due to having more parameters. It can be seen that F3DUN beats MUN on all indicators, excluding SSIM, which indicates that the full 3D CNN model has a larger modeling capacity than the 3D/2D model. Considering that the main difference between F3DUN and MUN is the use of 3D convolution and 2D convolution in the deep half of the nonlinear mapping, we can reason that the advantage of F3DUN is basically brought about by 3D convolutions. In Figure 5, we present a visual comparison between MUN and F3DUN, and F3DUN shows some superiority in edge and complex context recovery tasks.
A long-standing concern about full 3D CNN models is that they are computationally more complex and therefore prone to poor results due to overfitting on hyperspectral image datasets, the scale of which is limited by various sensors. This concern follows common sense in the field of deep learning. However, we want to more precisely analyze the conditions in which the full 3D CNN model could work. To further investigate this issue, we adjust the step size of patch cropping in the CAVE dataset to 72 pixels and thus construct a smaller training set. In the original training set, each image generates 81 patches, while only 36 patches are collected per image in the reduced training set.
The performances of MUN and F3DUN on the reduce dataset are shown in Table 3. With fewer training samples, both models have a certain decline in most indicators, which is consistent with the common sense of deep learning. However, the full 3D CNN model still achieves acceptable results on this smaller dataset, even outperforming the 3D/2D mixed model on several metrics such as PSNR, RMSE, and ERGAS. This demonstrates that the full 3D CNN model is less sensitive to the scale of training samples, contrary what people commonly thought, and it could achieve competitive performance compared with the 3D/2D mixed model on the small-scale dataset. In addition, if we put the two tables together, it suggests that the full 3D CNN model is more effective than the 3D/2D mixed model when sufficiently trained.

4.4.2. Results on Noisy Images

Hyperspectral images are prone to suffering from various distortions. However, the robustness to degradation is not our main concern. On the one hand, we argue that benchmarks are collected by real-world sensors, so a model’s performance on them is meaningful for practical cases. On the other hand, super-resolution is a subset of hyperspectral image reconstructions, and there are many techniques that deal with degradation, such as denoising [74,75] and deblurring [76,77].
In Figure 6 and Figure 7, we present reconstruction results with noisy inputs on the CAVE dataset. It can be seen that our model is somewhat robust to Gaussian noise, but can be severely affected by impulse noise. The main reason for this is that we do not introduce noise manually in training, and the model has no chance to learn to remove it.

4.5. Comparison with State-of-the-Art Methods

In this section, we compare the proposed F3DUN with several existing single-image hyperspectral super-resolution methods. The selected methods widely cover popular model priors in this field. Among them, bicubic upsampling is a traditional interpolation-based super-resolution approach, and EDSR [14] is a full 2D convolutional neural network that achieves remarkable success in natural image super-resolution. The 3DFCNN [19] model is the pioneer work that applies the full 3D CNN model to the HSISR task. Moreover, SSPSR [45] is a full 2D model designed for HSISR specifically using the similarity of neighboring bands. Another kind of method is 3D/2D mixed models, including MCNet [20] and ERCSR [21]. Finally, SFCSR [47] is a sequence model that treats bands in an HSI as frames in a video.
A quantitative comparison on the CAVE dataset is shown in Table 4. It can be seen that our method outperforms other models on all six metrics for both 4 × and 8 × super-resolutions. Bicubic upsampling obtains the worst results in both cases, since it cannot recover missing high-frequency components in LR-MSIs. The comparison between 3DFCNN and F3DUN supports our hypothesis: the full 3D CNN model can achieve SOTA performance by carefully designing the network structure. A 3D/2D mixed model, MCN, achieves the second highest score on PSNR for a 4 × scale factor, but it fails to generate good SR images when the scale factor increases to 8. In contrast, ERCSR performs better on 8 × upsampling. The sequence model SFCSR obtained results second only to F3DUN in most of the indicators of the 8 × super-resolution experiment, showing good reconstruction ability and balance in several indicators. However, the advantage of our model is obvious, proving the effectiveness of full 3D CNN in HSISR. In the 4 × super-resolution experiment, F3DUN outperforms the advanced 3D/2D mixed model MCNet by 0.22dB on the PSNR metric. Furthermore, in the case of 8 × , F3DUN is 0.27dB higher than the second best model on the PSNR index and significantly outperforms other methods in the SSIM index, which illustrates that F3DUN is more robust and capable for larger scale factors. Additionally, there are clear gains on the SAM score, namely, −0.28 rad for 4 × and −0.45 rad for 8 × , showing the remarkable effect of introducing SAM loss on suppressing spectral distortion.
Table 4. Quantitative evaluation of different HSISR methods on the CAVE dataset. The best and second best results are in bold and underlined, respectively.
Table 4. Quantitative evaluation of different HSISR methods on the CAVE dataset. The best and second best results are in bold and underlined, respectively.
MethodScalePSNR↑SSIM↑RMSE↓SAM↓ERGAS↓  CC↑
Bicubic434.61630.92870.02255.18696.80470.9728
EDSR [14]436.15690.93930.02024.41495.97930.9758
3DFCNN [19]435.34790.93490.02094.04156.07500.9758
SSPSR [45]436.59340.94310.01953.96915.76080.9766
MCNet [20]436.71820.94340.01913.95905.62810.9769
ERCSR [21]436.63180.94360.01933.91405.66100.9772
SFCSR [47]436.67180.94380.01913.92185.64090.9774
F3DUN436.93680.94580.01863.63405.44290.9786
Bicubic830.07190.85110.03677.094810.86640.9349
EDSR [14]830.95240.87030.03476.027010.15620.9393
3DFCNN [19]830.36220.86690.03545.671410.24980.9391
SSPSR [45]830.45240.86700.03675.609310.74480.9322
MCNet [20]831.21930.87900.03395.60569.76720.9403
ERCSR [21]831.34910.88080.03355.57219.72040.9421
SFCSR [47]831.40080.88140.03305.60699.62300.9432
F3DUN831.67300.88490.03255.12309.43270.9441
The results of the Harvard dataset are summarized in Table 5. The salient feature of the Harvard dataset is that the training samples are sparser than the CAVE dataset, since they are randomly cropped and possibly unable to cover the entire training images. Under these conditions, our F3DUN still generates competitive SR results. It can be seen that SSPSR achieves sub-optimal scores for most indicators on the Harvard dataset with respect to 4 × SR. SSPSR shows slight advantage on SAM, and this margin possibly comes from its divide-and-merge strategy. Since SSPSR first processes each band separately and then merge all bands together, this architecture can preserve spectral information. Moreover, F3DUN is optimal on all six indices. When the scale factor is 8, our method achieves the highest scores, except in RMSE and SAM. Specifically, F3DUN performs best on the SSIM and ERGAS metrics, which indicates that the reconstructed images preserve more complete structures and have higher overall quality. The lightweight 3D/2D mixed model ERCSR obtains the second-best scores on most indices, and the sequence model SFCSR achieves pretty good results, which suggests the advantages of models with less complexity. However, with the help of a deliberately designed architecture, our model still obtains an outstanding output.
In Table 6, we compare the model scale and computational complexity of our F3DUN and other state-of-the-art methods. It is shown that our model achieves superior performance with little extra computational cost. In specific, the number of parameters and FLOPs of F3DUN are at the same level of MCNet, while F3DUN outperforms MCNet on all six quantitative metrics. In addition, the main cost of our model comes from its pure 3D CNN structure, since 3D convolutions are very dense in terms of parameter and calculations. Our model is slightly heavier than the mixed 3D/2D models (i.e., MCNet and ERCSR), and 3DFCNN is lightweight due to its shallow structure. Though sequence models such as SFCSR have fewer parameters, they can run slowly due to their bad parallelism.
To compare the reconstruction effect more straightforwardly, we also present some visual results. Figure 8 shows the reconstruction results and absolute error maps with the × 4 upsampling of some bands in the CAVE dataset. The results shown in the figure are a reconstructed single spectral band (represented in grayscale), and all images are normalized. Generally, the images generated by our model have a lower error, and spatial details are recovered better. Specifically, our F3DUN could better recover sharp edges such as those in the red rectangles in Figure 8. As is seen in the third line in Figure 8, F3DUN handles images with higher contrast successfully, while the other methods have the problem that the overall brightness is too dark or too light. Some super-resolution results and error maps with × 4 upsampling in the Harvard dataset are shown in Figure 9. Obviously, our method is better at recovering complex contexts, such as the lattice of the chair (line 3). In Figure 10, we show the reconstructed results and absolute error maps with × 8 upsampling of some bands in the CAVE dataset and Harvard dataset. From the seventh row, we can see that different degrees of deformation and artifacts appear in the reconstruction results of several other deep learning methods. In Figure 11, we visualize the spectrum curves at the pixel positions of (65, 125), (250, 250), and (340, 340) for chart_and_stuffed_toy in the CAVE dataset. Among all the methods, the reconstructed spectrum vectors are quite similar at each location. Though the difference is tiny, the curve of EDSR has an observable deviation from the ground truth, while the spectral distortion of our model is minimal. Similar phenomena also occur in the Harvard dataset, the results of which are shown in Figure 12.

5. Conclusions

In this paper, we rethink an important model prior for HSISR, namely, the full 3D CNN model, in view of model design. We argue that this model prior is not inevitable, and it could achieve good performance with combination of advanced architectures. Thus, we combine the full 3D CNN model with the U-Net architecture, introducing skip connections to save memory and using a multi-scale structure. Extensive experiments show that the full 3D CNN model can achieve a performance that surpasses the current single-image hyperspectral image super-resolution methods and rediscovers the effectiveness of full 3D models for this task. Through comparison with MUN, the 3D/2D mixed counterpart of F3DUN, we find that 3D convolutions can bring higher performance improvements than 2D ones using the same number of parameters; that is, the modeling capacity of the full 3D model exceeds that of the 3D/2D mixed model. Furthermore, contrary to common belief, the full 3D CNN model can work in a setting with fewer training samples, and the requirements of the full 3D model on training data are not as demanding as one might imagine. All these illustrate that, combined with more advanced network structures, the full 3D model can be a simple and effective method to play a greater role in the field of hyperspectral image super-resolution.
In the future, we plan to explore a more lightweight full 3D CNN for hyperspectral image super-resolution in order to reduce the computational workload. There are some successful works for 2D convolution, such as separable convolution [78] and dilated convolution [79], but lightweight 3D convolution still needs more effort. In addition, some useful designs, such as an attention mechanism, can be combined with full 3D models. Our experiments prove the advancement of full 3D models in hyperspectral image super-resolution, and we believe F3DUN opens a new door for more 3D CNN-based methods to come.

Author Contributions

Conceptualization, J.J., Methodology, Z.L.; Writing and original draft preparation, Z.L., W.W. and Q.M.; Writing, review, and editing, X.L. and J.J.; Supervision: J.J. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the National Natural Science Foundation of China (61971165, 92270116).

Data Availability Statement

The CAVE dataset used in this study is accessible from https://www.cs.columbia.edu/CAVE/databases/multispectral, and the Harvard dataset used in this study is accessible from http://vision.seas.harvard.edu/hyperspec/index.html.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sabins, F.F. Remote sensing for mineral exploration. Ore Geol. Rev. 1999, 14, 157–183. [Google Scholar] [CrossRef]
  2. Lu, G.; Fei, B. Medical hyperspectral imaging: A review. J. Biomed. Opt. 2014, 19, 010901. [Google Scholar] [CrossRef] [PubMed]
  3. Lowe, A.; Harrison, N.; French, A.P. Hyperspectral image analysis techniques for the detection and classification of the early onset of plant disease and stress. Plant Methods 2017, 13, 80. [Google Scholar] [CrossRef] [PubMed]
  4. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plaza, A.; Chanussot, J. Graph convolutional networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5966–5978. [Google Scholar] [CrossRef]
  5. Yang, Y.; Song, S.; Liu, D.; Chan, J.C.W.; Li, J.; Zhang, J. Hyperspectral anomaly detection through sparse representation with tensor decomposition-based dictionary construction and adaptive weighting. IEEE Access 2020, 8, 72121–72137. [Google Scholar] [CrossRef]
  6. Dong, W.; Zhou, C.; Wu, F.; Wu, J.; Shi, G.; Li, X. Model-guided deep hyperspectral image super-resolution. IEEE Trans. Image Process. 2021, 30, 5754–5768. [Google Scholar] [CrossRef]
  7. Xie, Q.; Zhou, M.; Zhao, Q.; Meng, D.; Zuo, W.; Xu, Z. Multispectral and hyperspectral image fusion by MS/HS fusion net. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1585–1594. [Google Scholar]
  8. Dian, R.; Li, S.; Guo, A.; Fang, L. Deep hyperspectral image sharpening. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 5345–5355. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, H.; Ma, J. GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening. ISPRS J. Photogramm. Remote Sens. 2021, 172, 223–239. [Google Scholar] [CrossRef]
  10. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  11. Wang, Y.; Chen, X.; Han, Z.; He, S. Hyperspectral image super-resolution via nonlocal low-rank tensor approximation and total variation regularization. Remote Sens. 2017, 9, 1286. [Google Scholar] [CrossRef]
  12. Huang, H.; Yu, J.; Sun, W. Super-resolution mapping via multi-dictionary based sparse representation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 3523–3527. [Google Scholar]
  13. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  14. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  15. Doicu, A.; Doicu, A.; Efremenko, D.S.; Loyola, D.; Trautmann, T. An overview of neural network methods for predicting uncertainty in atmospheric remote sensing. Remote Sens. 2021, 13, 5061. [Google Scholar] [CrossRef]
  16. Todorov, V. Advanced Monte Carlo Methods to Neural Networks. In Proceedings of the Recent Advances in Computational Optimization: Results of the Workshop on Computational Optimization WCO 2021, Online, 17 September 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 303–314. [Google Scholar]
  17. Zhu, L.; Huang, L.; Fan, L.; Huang, J.; Huang, F.; Chen, J.; Zhang, Z.; Wang, Y. Landslide susceptibility prediction modeling based on remote sensing and a novel deep learning algorithm of a cascade-parallel recurrent neural network. Sensors 2020, 20, 1576. [Google Scholar] [CrossRef] [PubMed]
  18. Cai, W.; Ning, X.; Zhou, G.; Bai, X.; Jiang, Y.; Li, W.; Qian, P. A Novel Hyperspectral Image Classification Model Using Bole Convolution with Three-Directions Attention Mechanism: Small sample and Unbalanced Learning. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–17. [Google Scholar] [CrossRef]
  19. Mei, S.; Yuan, X.; Ji, J.; Zhang, Y.; Wan, S.; Du, Q. Hyperspectral image spatial super-resolution via 3D full convolutional neural network. Remote Sens. 2017, 9, 1139. [Google Scholar] [CrossRef]
  20. Li, Q.; Wang, Q.; Li, X. Mixed 2D/3D convolutional network for hyperspectral image super-resolution. Remote Sens. 2020, 12, 1660. [Google Scholar] [CrossRef]
  21. Li, Q.; Wang, Q.; Li, X. Exploring the relationship between 2D/3D convolution for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8693–8703. [Google Scholar] [CrossRef]
  22. Ding, Y.; Zhao, X.; Zhang, Z.; Cai, W.; Yang, N.; Zhan, Y. Semi-supervised locality preserving dense graph neural network with ARMA filters and context-aware learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  23. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Li, W.; Cai, W.; Zhan, Y. AF2GNN: Graph convolution with adaptive filters and aggregator fusion for hyperspectral image classification. Inf. Sci. 2022, 602, 201–219. [Google Scholar] [CrossRef]
  24. Ding, Y.; Zhang, Z.; Zhao, X.; Hong, D.; Cai, W.; Yang, N.; Wang, B. Multi-scale receptive fields: Graph attention neural network for hyperspectral image classification. Expert Syst. Appl. 2023, 223, 119858. [Google Scholar] [CrossRef]
  25. Ding, Y.; Zhang, Z.; Zhao, X.; Cai, W.; Yang, N.; Hu, H.; Huang, X.; Cao, Y.; Cai, W. Unsupervised self-correlated learning smoothy enhanced locality preserving graph convolution embedding clustering for hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  26. Ding, Y.; Zhang, Z.; Zhao, X.; Cai, Y.; Li, S.; Deng, B.; Cai, W. Self-supervised locality preserving low-pass graph convolutional embedding for large-scale hyperspectral image clustering. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Ding, Y.; Zhao, X.; Siye, L.; Yang, N.; Cai, Y.; Zhan, Y. Multireceptive field: An adaptive path aggregation graph neural framework for hyperspectral image classification. Expert Syst. Appl. 2023, 217, 119508. [Google Scholar] [CrossRef]
  28. Chen, Z.; Lu, Z.; Gao, H.; Zhang, Y.; Zhao, J.; Hong, D.; Zhang, B. Global to Local: A Hierarchical Detection Algorithm for Hyperspectral Image Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  29. Fan, G.; Ma, Y.; Mei, X.; Fan, F.; Huang, J.; Ma, J. Hyperspectral anomaly detection with robust graph autoencoders. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar]
  30. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  31. Ji, S.; Zhang, C.; Xu, A.; Shi, Y.; Duan, Y. 3D convolutional neural networks for crop classification with multi-temporal remote sensing images. Remote Sens. 2018, 10, 75. [Google Scholar] [CrossRef]
  32. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial–spectral feature learning by 3D convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  33. Yang, X.; Zhang, X.; Ye, Y.; Lau, R.Y.; Lu, S.; Li, X.; Huang, X. Synergistic 2D/3D convolutional neural network for hyperspectral image classification. Remote Sens. 2020, 12, 2033. [Google Scholar] [CrossRef]
  34. Jiang, H.; Lu, N. Multi-scale residual convolutional neural network for haze removal of remote sensing images. Remote Sens. 2018, 10, 945. [Google Scholar] [CrossRef]
  35. Quan, Y.; Fu, D.; Chang, Y.; Wang, C. 3D Convolutional Neural Network for Low-Light Image Sequence Enhancement in SLAM. Remote Sens. 2022, 14, 3985. [Google Scholar] [CrossRef]
  36. Xu, Z.; Guan, K.; Casler, N.; Peng, B.; Wang, S. A 3D convolutional neural network method for land cover classification using LiDAR and multi-temporal Landsat imagery. ISPRS J. Photogramm. Remote Sens. 2018, 144, 423–434. [Google Scholar] [CrossRef]
  37. Dong, H.; Zhang, L.; Zou, B. PolSAR image classification with lightweight 3D convolutional networks. Remote Sens. 2020, 12, 396. [Google Scholar] [CrossRef]
  38. Li, J.; Zhang, S.; Huang, T. Multi-scale 3d convolution network for video based person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, Hawaii, USA, 27 January–1 February 2019; Volume 33, pp. 8618–8625. [Google Scholar]
  39. Ying, X.; Wang, L.; Wang, Y.; Sheng, W.; An, W.; Guo, Y. Deformable 3d convolution for video super-resolution. IEEE Signal Process. Lett. 2020, 27, 1500–1504. [Google Scholar] [CrossRef]
  40. Akgun, T.; Altunbasak, Y.; Mersereau, R.M. Super-resolution reconstruction of hyperspectral images. IEEE Trans. Image Process. 2005, 14, 1860–1875. [Google Scholar] [CrossRef]
  41. Bauschke, H.H.; Borwein, J.M. On projection algorithms for solving convex feasibility problems. SIAM Rev. 1996, 38, 367–426. [Google Scholar] [CrossRef]
  42. Yuan, Y.; Zheng, X.; Lu, X. Hyperspectral image superresolution by transfer learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1963–1974. [Google Scholar] [CrossRef]
  43. Xie, W.; Jia, X.; Li, Y.; Lei, J. Hyperspectral image super-resolution using deep feature matrix factorization. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6055–6067. [Google Scholar] [CrossRef]
  44. Li, Y.; Zhang, L.; Dingl, C.; Wei, W.; Zhang, Y. Single hyperspectral image super-resolution with grouped deep recursive residual network. In Proceedings of the 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), Xi’an, China, 13–16 September 2018; pp. 1–4. [Google Scholar]
  45. Jiang, J.; Sun, H.; Liu, X.; Ma, J. Learning spatial-spectral prior for super-resolution of hyperspectral imagery. IEEE Trans. Comput. Imaging 2020, 6, 1082–1096. [Google Scholar] [CrossRef]
  46. Li, Q.; Wang, Q.; Li, X. Hyperspectral image super-resolution via adjacent spectral fusion strategy. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1645–1649. [Google Scholar]
  47. Wang, Q.; Li, Q.; Li, X. Hyperspectral image super-resolution using spectrum and feature context. IEEE Trans. Ind. Electron. 2021, 68, 11276–11285. [Google Scholar] [CrossRef]
  48. Liu, Y.; Hu, J.; Kang, X.; Luo, J.; Fan, S. Interactformer: Interactive transformer and CNN for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  49. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled nonnegative matrix factorization unmixing for hyperspectral and multispectral data fusion. IEEE Trans. Geosci. Remote Sens. 2011, 50, 528–537. [Google Scholar] [CrossRef]
  50. Akhtar, N.; Shafait, F.; Mian, A. Bayesian sparse representation for hyperspectral image super resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  51. Tian, X.; Zhang, W.; Yu, D.; Ma, J. Sparse tensor prior for hyperspectral, multispectral, and panchromatic image fusion. IEEE/CAA J. Autom. Sin. 2022, 10, 284–286. [Google Scholar] [CrossRef]
  52. Dong, W.; Fu, F.; Shi, G.; Cao, X.; Wu, J.; Li, G.; Li, X. Hyperspectral image super-resolution via non-negative structured sparse representation. IEEE Trans. Image Process. 2016, 25, 2337–2352. [Google Scholar] [CrossRef] [PubMed]
  53. Veganzones, M.A.; Simoes, M.; Licciardi, G.; Yokoya, N.; Bioucas-Dias, J.M.; Chanussot, J. Hyperspectral super-resolution of locally low rank images from complementary multisource data. IEEE Trans. Image Process. 2015, 25, 274–288. [Google Scholar] [CrossRef] [PubMed]
  54. Dian, R.; Li, S.; Kang, X. Regularizing hyperspectral and multispectral image fusion by CNN denoiser. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1124–1135. [Google Scholar] [CrossRef] [PubMed]
  55. Wen, B.; Kamilov, U.S.; Liu, D.; Mansour, H.; Boufounos, P.T. DeepCASD: An end-to-end approach for multi-spectral image super-resolution. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 6503–6507. [Google Scholar]
  56. Wang, W.; Zeng, W.; Huang, Y.; Ding, X.; Paisley, J. Deep blind hyperspectral image fusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 4150–4159. [Google Scholar]
  57. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef]
  58. Kim, K.I.; Kwon, Y. Single-image super-resolution using sparse regression and natural image prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1127–1133. [Google Scholar]
  59. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization. IEEE Trans. Image Process. 2011, 20, 1838–1857. [Google Scholar] [CrossRef]
  60. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
  61. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  62. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  63. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 191–207. [Google Scholar]
  64. Mei, Y.; Fan, Y.; Zhou, Y. Image super-resolution with non-local sparse attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  65. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 12299–12310. [Google Scholar]
  66. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  67. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In Proceedings of the JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop, Volume 1: AVIRIS Workshop, Pasadena, CA, USA, 1–5 June 1992; Volume 1, pp. 147–149. [Google Scholar]
  68. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef]
  69. Chakrabarti, A.; Zickler, T. Statistics of real-world hyperspectral images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2011; pp. 193–200. [Google Scholar]
  70. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference for Learning Representations, Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  71. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  72. Loncan, L.; De Almeida, L.B.; Bioucas-Dias, J.M.; Briottet, X.; Chanussot, J.; Dobigeon, N.; Fabre, S.; Liao, W.; Licciardi, G.A.; Simoes, M.; et al. Hyperspectral pansharpening: A review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 27–46. [Google Scholar] [CrossRef]
  73. Wald, L. Data Fusion: Definitions and Architectures: Fusion of Images of Different Spatial Resolutions; Presses des MINES: Paris, France, 2002. [Google Scholar]
  74. Zhang, J.; Cai, Z.; Chen, F.; Zeng, D. Hyperspectral image denoising via adversarial learning. Remote Sens. 2022, 14, 1790. [Google Scholar] [CrossRef]
  75. Pang, L.; Gu, W.; Cao, X. TRQ3DNet: A 3D quasi-recurrent and transformer based network for hyperspectral image denoising. Remote Sens. 2022, 14, 4598. [Google Scholar] [CrossRef]
  76. Hao, J.; Xue, J.; Zhao, Y.; Chan, J.C.W. Transformed Structured Sparsity with Smoothness for Hyperspectral Image Deblurring. IEEE Geosci. Remote Sens. Lett. 2022, 20, 1–5. [Google Scholar] [CrossRef]
  77. Liao, W.; Goossens, B.; Aelterman, J.; Luong, H.Q.; Pižurica, A.; Wouters, N.; Saeys, W.; Philips, W. Hyperspectral image deblurring with PCA and total variation. In Proceedings of the 2013 5th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Gainesville, FL, USA, 26–28 June 2013; pp. 1–4. [Google Scholar]
  78. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  79. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
Figure 1. Comparison of 2D (a) and 3D convolution (b) on multi-band data.
Figure 1. Comparison of 2D (a) and 3D convolution (b) on multi-band data.
Remotesensing 15 02574 g001
Figure 2. Simplified flowchart of several models: (a) full 2D CNNs, (b) full 3D CNNs, (c) ERCSR, (d) MCNet, (e) MUN, (f) F3DUN.
Figure 2. Simplified flowchart of several models: (a) full 2D CNNs, (b) full 3D CNNs, (c) ERCSR, (d) MCNet, (e) MUN, (f) F3DUN.
Remotesensing 15 02574 g002
Figure 3. Overview of F3DUN.
Figure 3. Overview of F3DUN.
Remotesensing 15 02574 g003
Figure 4. Overview of MUN.
Figure 4. Overview of MUN.
Remotesensing 15 02574 g004
Figure 5. Some visual results and error maps of MUN (the upper line) and F3DUN (the bottom line).
Figure 5. Some visual results and error maps of MUN (the upper line) and F3DUN (the bottom line).
Remotesensing 15 02574 g005
Figure 6. Some reconstruction results with Gaussian noise ( σ = 5 ) on the CAVE dataset ( 4 × ) .
Figure 6. Some reconstruction results with Gaussian noise ( σ = 5 ) on the CAVE dataset ( 4 × ) .
Remotesensing 15 02574 g006
Figure 7. Some reconstruction results with impulse noise ( 3 % ) on the CAVE dataset ( 4 × ) .
Figure 7. Some reconstruction results with impulse noise ( 3 % ) on the CAVE dataset ( 4 × ) .
Remotesensing 15 02574 g007
Figure 8. Reconstruction results and error maps in the CAVE dataset. From top to bottom: cd_ms at 530 nm ( 4 × ), fake_and_real_food_ms at 690 nm ( 4 × ).
Figure 8. Reconstruction results and error maps in the CAVE dataset. From top to bottom: cd_ms at 530 nm ( 4 × ), fake_and_real_food_ms at 690 nm ( 4 × ).
Remotesensing 15 02574 g008
Figure 9. Reconstruction results and error maps in the Harvard dataset. From top to bottom: img4 at 690 nm ( 4 × ), imga6 at 560 nm ( 4 × ).
Figure 9. Reconstruction results and error maps in the Harvard dataset. From top to bottom: img4 at 690 nm ( 4 × ), imga6 at 560 nm ( 4 × ).
Remotesensing 15 02574 g009
Figure 10. Reconstruction results and error maps in CAVE dataset and Harvard dataset. From top to bottom: clay_ms at 670 nm ( 8 × ) from CAVE dataset and img4 at 640 nm ( 8 × ) from Harvard dataset.
Figure 10. Reconstruction results and error maps in CAVE dataset and Harvard dataset. From top to bottom: clay_ms at 670 nm ( 8 × ) from CAVE dataset and img4 at 640 nm ( 8 × ) from Harvard dataset.
Remotesensing 15 02574 g010
Figure 11. Visual comparison of spectral curves for image chart_and_stuffed_toy at three pixel positions in the CAVE dataset ( 4 × ).
Figure 11. Visual comparison of spectral curves for image chart_and_stuffed_toy at three pixel positions in the CAVE dataset ( 4 × ).
Remotesensing 15 02574 g011
Figure 12. Visual comparison of spectral curves for image img4 at three pixel positions in the Harvard dataset ( 4 × ).
Figure 12. Visual comparison of spectral curves for image img4 at three pixel positions in the Harvard dataset ( 4 × ).
Remotesensing 15 02574 g012
Table 1. Abbreviations used in the paper.
Table 1. Abbreviations used in the paper.
AbbreviationFull Name
HSIhyperspectral image
LRlow-resolution
HRhigh-resolution
SRsuper-resolution
HSISRhyperspectral image super-resolution
SHISRsingle-image hyperspectral super-resolution
MUNMixed U-Net
F3DUNFull 3D U-Net
Table 2. Comparison of the full 3D CNN model and the 3D/2D mixed model on the original CAVE dataset.
Table 2. Comparison of the full 3D CNN model and the 3D/2D mixed model on the original CAVE dataset.
MethodParam.PSNR↑SSIM↑RMSE↓SAM↓ERGAS↓CC↑
MUN2.4 M36.88600.94520.01883.65005.49750.9779
F3DUN2.5 M36.93680.94580.01863.63405.44290.9786
Table 3. Comparison of the full 3D CNN model and the 3D/2D mixed model on the reduced CAVE dataset.
Table 3. Comparison of the full 3D CNN model and the 3D/2D mixed model on the reduced CAVE dataset.
MethodParam.PSNR↑SSIM↑RMSE↓SAM↓ERGAS↓CC↑
MUN2.4M36.70150.94540.01893.64725.55220.9783
F3DUN2.5M36.72490.94530.01883.64965.53300.9783
Table 5. Quantitative evaluation of different HSISR methods on the Harvard dataset. The best and second best results are in bold and underlined, respectively.
Table 5. Quantitative evaluation of different HSISR methods on the Harvard dataset. The best and second best results are in bold and underlined, respectively.
MethodScalePSNR↑SSIM↑RMSE↓SAM↓ERGAS↓  CC↑
Bicubic442.41280.92650.01373.05203.48970.9519
EDSR [14]443.05120.93590.01262.78793.25910.9578
3DFCNN [19]442.80210.93100.01292.85933.34950.9554
SSPSR [45]443.45220.93870.01202.67883.09800.9615
MCNet [20]443.31260.93750.01242.74283.15900.9602
ERCSR [21]443.29830.93760.01242.74163.18390.9600
SFCSR [47]443.38170.93830.01232.75713.14620.9605
F3DUN443.48550.93890.01202.65553.09230.9618
Bicubic838.39090.86670.02253.79805.42310.8967
EDSR [14]838.67190.93590.02163.42335.15300.9025
3DFCNN [19]838.86760.87370.02103.46195.23090.9047
SSPSR [45]839.18890.88290.02043.26374.93240.9127
MCNet [20]839.17020.88190.02073.51174.94750.9123
ERCSR [21]839.31850.88390.02023.46494.87990.9144
SFCSR [47]839.26070.88290.02033.44264.90070.9137
F3DUN839.32650.88490.02043.27394.85520.9149
Table 6. Comparison of computational complexity among the proposed F3DUN and state-of-the-art methods. All metrics are measured for × 4 upscaling on the CAVE dataset.
Table 6. Comparison of computational complexity among the proposed F3DUN and state-of-the-art methods. All metrics are measured for × 4 upscaling on the CAVE dataset.
MethodEDSR [14]3DFCNN [19]SSPSR [45]MCNet [20]ERCSR [21]SFCSR [47]F3DUN
Params × 10 6 1.540.0412.892.171.591.222.40
FLOPs × 10 9 36.98328.99977.504483.564460.102011.94932.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Wang, W.; Ma, Q.; Liu, X.; Jiang, J. Rethinking 3D-CNN in Hyperspectral Image Super-Resolution. Remote Sens. 2023, 15, 2574. https://doi.org/10.3390/rs15102574

AMA Style

Liu Z, Wang W, Ma Q, Liu X, Jiang J. Rethinking 3D-CNN in Hyperspectral Image Super-Resolution. Remote Sensing. 2023; 15(10):2574. https://doi.org/10.3390/rs15102574

Chicago/Turabian Style

Liu, Ziqian, Wenbing Wang, Qing Ma, Xianming Liu, and Junjun Jiang. 2023. "Rethinking 3D-CNN in Hyperspectral Image Super-Resolution" Remote Sensing 15, no. 10: 2574. https://doi.org/10.3390/rs15102574

APA Style

Liu, Z., Wang, W., Ma, Q., Liu, X., & Jiang, J. (2023). Rethinking 3D-CNN in Hyperspectral Image Super-Resolution. Remote Sensing, 15(10), 2574. https://doi.org/10.3390/rs15102574

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop