Next Article in Journal
The Application of Augmented Reality Technology in Perioperative Visual Guidance: Technological Advances and Innovation Challenges
Previous Article in Journal
Improving the Rapidity of Magnitude Estimation for Earthquake Early Warning Systems for Railways
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mask-Guided Spatial–Spectral MLP Network for High-Resolution Hyperspectral Image Reconstruction

1
Graduate School of Artificial Intelligence and Science, Rikkyo University, Tokyo 171-8501, Japan
2
School of Information Science and Engineering, Shandong Normal University, Jinan 250358, China
3
College of Information Science and Engineering, Ritsumeikan University, Osaka 603-8577, Japan
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7362; https://doi.org/10.3390/s24227362
Submission received: 23 September 2024 / Revised: 7 November 2024 / Accepted: 13 November 2024 / Published: 18 November 2024

Abstract

:
Hyperspectral image (HSI) reconstruction is a critical and indispensable step in spectral compressive imaging (CASSI) systems and directly affects our ability to capture high-quality images in dynamic environments. Recent research has increasingly focused on deep unfolding frameworks for HSI reconstruction, showing notable progress. However, these approaches have to break the optimization task into two sub-problems, solving them iteratively over multiple stages, which leads to large models and high computational overheads. This study presents a simple yet effective method that passes the degradation information (sensing mask) through a deep learning network to disentangle the degradation and the latent target’s representations. Specifically, we design a lightweight MLP block to capture non-local similarities and long-range dependencies across both spatial and spectral domains, and investigate an attention-based mask modelling module to achieve the spatial–spectral-adaptive degradation representationthat is fed to the MLP-based network. To enhance the information flow between MLP blocks, we introduce a multi-level fusion module and apply reconstruction heads to different MLP features for deeper supervision. Additionally, we combine the projection loss from compressive measurements with reconstruction loss to create a dual-domain loss, ensuring consistent optical detection during HS reconstruction. Experiments on benchmark HS datasets show that our method outperforms state-of-the-art approaches in terms of both reconstruction accuracy and efficiency, reducing computational and memory costs.

1. Introduction

Hyperspectral images (HSIs) contain many more spectral bands than conventional RGB images, allowing them to capture richer details. Because of their abundant spectral information, HSIs are extensively used in various applications, such as image recognition [1,2,3], remote sensing [4,5,6,7], medical image analysis [8,9], and environmental monitoring. Traditionally, HSIs are acquired using spectrometers that scan scenes across their spectral or spatial dimensions. The scan process is usually time-consuming and unsuitable for capturing dynamic objects. To address this limitation, snapshot compressive imaging (SCI) systems [10,11,12,13] have been developed to capture HSIs at video rates. Among these, coded aperture snapshot spectral imaging (CASSI) [14,15,16] stands out as an promising imaging modality for capturing high-quality spatial and spectral details in real time. In general, CASSI uses a coded aperture and a disperser to modulate HSI signals at different wavelengths, compressing them into a 2D measurement. THe reconstruction of the full HSI from the compressed data, known as the CASSI inverse problem, is a challenging task due to its ill-posed nature. To address this challenge, various approaches, primarily ranging from traditional model-based methods [17,18,19,20,21] and end-to-end (E2E) techniques [22,23,24,25,26,27,28,29] to deep unrolling pipelines [25,30,31,32,33], have been proposed, resulting in substantial advancements.
Traditional model-based approaches typically formulate the compressed detection process as a mathematical model and utilize various handcrafted priors, such as total variation [34,35], low ranks [17], sparsity [36,37], and many others [38,39], to enhance reconstruction accuracy. Although these techniques enable practical spectral reconstruction, they come with the drawback of high computational costs and require the manual design of priors for each specific scene, limiting their broader applicability in real-world systems. Inspired by the remarkable success of deep learning in vision tasks, researchers have extensively explored the use of convolutional neural networks (CNNs) [23,24,26,40] and transformers [28,41,42] for hyperspectral image reconstruction. These deep learning methods are generally implemented in an end-to-end (E2E) learning way to establish brute-force mapping between compressed measurements and the target hyperspectral images [23,24,29,40,42], and they are limited in their interpretability as black-box models. Recent research has increasingly turned to deep unfolding networks (DUNs) as a promising hybrid approach, aiming to combine the strengths of both model-based and deep learning techniques [25,30,31,32]. DUNs decompose the whole optimization process into two distinct sub-problems and are usually realised in a multi-stage iterative way, which unavoidably causes a large computational cost and memory footprint.
In addition, the backbone architectures of the E2E and DUN method for HSI reconstruction have progressed from CNNs [22,23,24] to transformers [31,32,42]. Benefiting from the powerful modelling capability of their long-range dependencies, transformers have demonstrated promising performances. To reduce computational complexity and effectively model spectral correlations, transformer-based HSI models typically focus on self-attention (SA) mechanisms along the spectral dimension [31,42], rather than in the spatial domain. This spectral SA modelling method generally uses a few conventional convolution layers to exploit local spatial dependencies. However, while leveraging non-local similarities and long-range relationships in the spatial domain can be highly beneficial for recovering missing information from HSIs, it usually leads to a high computational cost and large memory overhead. Despite the efficiency of improvements made with window-based transformers [43], the computational load for real imaging sensors remains a challenging aspect that needs to be reduced further. In contrast, recent developments have shown that replacing the self-attention mechanism in transformers with MLP-like operations offers a promising alternative. These operations excel at capturing non-local similarities and modelling long-range dependencies, with considerable potential applicability in the vision community. Notably, they have delivered comparable performance in various vision tasks while reducing memory usage and computational costs. Building on the above insights, we explore the use of the MLP block as the core element for constructing a baseline network for hyperspectral (HS) image reconstruction and aim to offer a compelling solution to overcome the locality constraints of CNNs and the high computational demands of Vision Transformers (ViTs).
To this end, the study introduces a novel spatial and spectral MLP (S2MLP) block, designed to effectively capture long-distance representations from both non-local spatial positions and spectral bands. This proposed S2MLP block serves as the core component used to construct a simple yet effective deep model (S2MLPNet) for HSI reconstruction. In particular, the spatial MLP in the S2MLP block replaces the conventional SpatialFC with a CycleFC layer [44], which maintains a comparable linear computational cost to that of the ChannelFC while enhancing non-local spatial modelling. Additionally, the S2MLPNet integrates a multi-level information fusion module, which promotes the flow of valuable information between different MLP blocks. Moreover, we equip the reconstruction heads with multiple S2MLP features for deep supervision at various stages, facilitating the effective representation learning of the overall model. Then, an attention-based mask modelling module has been investigated to achieve the spatial–spectral-adaptive degradation representation fed to the S2MLPNet, disentangling the representation learning of the latent HSI and the degradation information. Finally, a dual-domain loss function, combining losses from both the compressive and reconstruction domains, is introduced to ensure consistency in optical detection. Extensive experiments on benchmark datasets demonstrate that our proposed method significantly outperforms existing state-of-the-art (SOTA) approaches.
In summary, our main contributions are as follows:
(1)
We propose a simple yet effective S2MLPNet for hyperspectral image reconstruction. As far as we know, this study marks the first effort to investigate the potential of MLPs in HSI reconstruction. We believe it will offer new perspectives for future research on deep modelling architectures for HSIs.
(2)
We design a lightweight spatial and spectral MLP (S2MLP) module to efficiently capture non-local similarities and long-range dependencies across both spatial and spectral domains with linear complexity to the image size.
(3)
We investigate the use of an attention-based mask modelling module to capture the degradation representation and incorporate it into the S2MLP for learning the disentangled representations of latent HSI data.
(4)
We exploit the multi-level information fusion mechanism seen between different levels and the deep supervision strategy used to enhance robust representation learning and further employ a dual-domain loss function for ensuring optical measurement consistency.

2. Related Work

This section firstly presents a brief review of the incorporation strategies of the degradation mask used for the HSI reconstruction and then presents the network architectures employed in recent learning-based methods, including E2E and deep unrolling approaches.

2.1. Incorporation of Degradation into HSI Reconstruction

A spectral compressive imaging system such as CASSI captures 3D spatial and spectral data as a 2D measurement, where the detection process can be expressed as a mathematical model by applying the degradation mask to the original HSIs. Traditional model-based methods [17,18,19,20,21] minimize the regularised objective function formulated based on the degradation process and the handcrafted priors of the latent HSIs. Then, optimization of the model-based methods iteratively achieves the recovery HSI by an inverse projection of the previous reconstruction error according to the degradation mask. Although this kind of method makes full use of the physical degradation process and its clear interpretability, it usually suffer from an insufficient representation ability due to the challenge of creating prior designs for each specific scene. Deep neural networks have been extensively applied for HSI reconstruction to achieve a brute-force mapping (E2E methods) between the measurement and the target HSI while usually ignoring the potential degradation information found in the detection process [22,23,24,40]. A few works have attempted to integrate the degradation mask into a representation learning network to enhance modelling capabilities [29,42] and have demonstrated some performance enhancements in HSI reconstruction. Recently, the deep unrolling method (DUM) [25,30,31,32] has been proposed to integrate the degradation process-formulated model method and the deep learning approach. Benefiting from the advantages of two pipelines, DUM methods usually have great potential for HSI reconstruction and demonstrate significant performance gains [31,32,33,45]. For example, DGSMP [31] employed an iterative approach that combines Maximum A Posteriori (MAP) estimation with a learnable Gaussian Scale Mixture prior. DAUHST [32] introduced a degradation-aware framework and half-shuffle attention that incorporated a prior learning network to tackle issues of degradation and ill-posedness. RDLUF [45] proposed a residual degradation learning unfolding framework to bridge the gap between the sensing matrix and the degradation process. Despite the great progress made in HSI reconstruction, DUM is generally implemented in a multi-stage, iterative way and thus has a high computational cost and heavy memory footprint, restricting its wide applicability in real imaging systems.

2.2. Network Architectures for HSI Representation Learning

The network architectures used for HSI representation learning have evolved from convolution-based to transformer-based backbones [23,24,26,28,31,32,40,42]. Initially, CNN-based approaches were exploited to capture local spatial similarities in HSI data. For instance, λ -net [24] proposed a dual-branch convolution network to reconstruct HSIs, while TSA-Net [22] is designed to model spatial–spectral correlations via elaborating various combinations of convolution operations. Further, HypermixNet [26] explored improved variants of the convolution to enhance its representation ability as well as reduce computational complexity. Although CNNs are effective to some extent, their inherent inductive biases usually limit their ability to capture non-local similarities. To overcome this, transformer-based methods have emerged [41,42], which are used to employ multi-head self-attention mechanisms to capture long-range dependencies. For example, MST [42] investigated a the use of a spectral transformer via capturing dependencies across spectral dimensions, generating attention maps that implicitly encode global context. CST [41] incorporated spatially sparse nature into a transformer learning model to capture the interactions of closely related patterns and reduce computational complexity. PADUT [33] investigated the use of a non-local spectral transformer to model the representative 3D features of HSIs. More recently, several studies [46,47] have integrated the window-based spectral transformer with the spatial modelling module to capture both spatial and spectral correlations and demonstrated promising reconstruction performances. All of these transformer-based networks generally realize long-region dependencies by calculating the pair-wise correlations of all tokens as a self-attention map, which would unavoidably increase the computational complexity with the growth of the token number.
In recent years, a novel type of deep learning architecture built entirely on multi-layer perceptrons (MLPs) has gained significant interest in the computer vision domain. For example, MLPMixer [48] pioneered this shift by introducing the concept of MLP-based prototypes, while GMLP [49] advanced the idea further by incorporating gating mechanisms to boost performance. Additionally, ResMLP [50] made key improvements by adding residual connections, which enhanced its training stability. Contrarily, CycleMLP [44] tackled the persistent challenge of the computational reliance on the input sizes from an earlier transformer era. Although MLP-based models have proven effective in various high-level tasks, their application in low-level tasks remains underexplored. MLP-based networks have promising prospects in capturing long-range dependencies as they separately mix tokens and channels across layers, which is significantly different from the CNNs used for local representation learning and ViTs used for global dependency modelling with self-attention (SA) computation. Therefore, compared to CNNs and transformers with an SA mechanism, MLP-based models generally have lower computational complexity and are particularly suitable for low-level HSI reconstruction tasks that use high-dimensional information, where non-local information is essential for restoring missing or corrupted data. The investigation of MLP models will offer new perspectives on deep modelling architectures, opening up more opportunities for diverse and adaptable network configurations specifically in high-dimensional data representations.

2.3. The Coded Mask in Spectral Snapshot Imaging Systems

The coded mask in spectral snapshot imaging systems allows for the efficient capture of spectral data by encoding both spatial and spectral information into a single image, which can be later reconstructed to provide a full hyperspectral cube. One of the most popular systems in snapshot compressive imaging (SCI) is the Coded Aperture Snapshot Spectral Compressive Imaging (CASSI) system [14,15,16], which usually uses the same coded mask to encode different spectral bands and spatially shifted the coded spectral images with a disperser to integrate them into a single snapshot. In the early era of this research, most works usually randomly generated a binary matrix according to a Bernoulli distribution, with p = 0.5 as the coded masks [25,26,28], and then simulated the snapshot measurements for HS image reconstruction. Meng et al. [22] have released a coded mask obtained with optical hardware, and then subsequent studies have widely employed it to conduct simulation experiments for fair comparisons [23,24,31,32,42]. More recently, Motoki et al. [51] proposed a novel optical filter-based HS imaging system using a spatial–spectral coded 3D mask to capture compressive measurements, and released a coded mask. This study uses a 3D coded mask in its simulation experiments.

3. Spectral Snapshot Imaging Model

According to compressive sensing theory [52], a spectral snapshot imaging system captures a compressed measurement by encoding information from all spectral bands. Figure 1 provides an overview of the basic coding process. Given a three-dimensional (3D) spectral image X H × W × Λ of a scene, where H and W represent the height and width, while Λ refers to the wavelength number, the spectral snapshot imaging system employs a coded aperture (sensing mask) M λ H × W to modulate the image of each spectral band, and the resulted temporary data Y λ H × W can be expressed as
Y λ = = X ( : , : , λ ) M λ ,
where ⊙ represents the element-wise multiplication and Y H × W × Λ refers to the modulated cube. Then, Y is shifted in the horizontal direction using a dispersive function d (representing the spatial shifting step), resulting in the dispersed cube Y H × ( W + d ( Λ 1 ) ) × Λ , which can be described as follows:
Y ( w , h , λ ) = Y ( w , h + d ( λ ) , λ ) ,
where h and w refer to the spatial coordinates. Ultimately, the compressive measurement Y is achieved by performing spectral integration on the dispersed cube Y , which is expressed as follows:
Y = λ = 1 Λ Y ( : , : , λ ) + N ,
where N denotes the random noise generated during the measurement process. To simplify, we rewrite Equation (3) in a matrix–vector format as follows:
y = Φ x + n ,
where Φ refers to the sensing matrix, characterised by its large and highly structured form. In particular, the CASSI system employs the same sensing mask for different spectral bands and uses a disperser by spatially shifting the modulated cube used for integration. While the recent work [51] adopted an optimised spatial and spectral coded mask, which can be easily achieved by placing the encoding (Fabry–P e ´ rot) filter, along with an array CMOS devices, onto a monochromatic image sensor to encode the original HS cube and integrating the spectral information into snapshot measurements without spatial shift on the spectral bands. This novel spectral compressive imaging is usually referred to as an optical filter-based HS imaging system. This study obtains snapshot images using the optical filter-based HS imaging technique. After the compressive spectral measurement, it is an indispensable step to recover the 3D HSI x from the compressed 2D measurement y , which is a challenging task due to its ill-posed nature. Typically, the sensing mask varies spatially and causes different information losses in different spatial positions and spectral bands. Therefore, the incorporation of the sensing matrix into inverse HSI reconstruction is crucial to obtain a disentangled representation of the latent HSI while weakening the effect of the degradation.

4. Proposed Method

With the measured snapshot Y and the corresponding sensing mask Φ , the goal of HSI reconstruction is to accurately recover the underlying 3D cube X . The challenge lies in reversing the effects of the sensing process to retrieve high-dimensional spectral information from the compressed measurement. This study introduces a simple yet highly effective deep learning model for HSI reconstruction, as depicted in Figure 2. The architecture is composed of two distinct branches. The first branch is designed to capture both spatial and spectral information, utilising a series of mask-guided spatial and spectral MLP blocks (MG-S2MLP), to learn the complicated and abundant structures in high-dimensional HS data. The second branch focuses on mask modelling and incorporates several specialized modules that are integrated with each MG-S2MLP block. Furthermore, an interactive mechanism is established between the two branches, allowing for cross-branch communication, as well as interactions between the multi-level MG-S2MLP blocks, to enhance the reconstruction performance. This dual-branch scheme ensures that both spatial–spectral features and mask properties are jointly optimized throughout the reconstruction process.
Specifically, the snapshot measurement Y and the sensing mask Φ are separately passed through a convolutional layer, producing the initial representations X 0 for the measurement and M 0 for the mask, both with a channel dimension of C. To enhance the mask representation, an attention-based mask modelling module (AMMM) is applied, generating a mask context, X 0 M = f M A M M 0 ( M 0 ) , which interacts with the measurement representation X 0 of the S2MLP branch through a multiplication operation. Then, after the linear mapping of the aggregated measurement and mask context, we add X 0 to enhance its representation capability. The resulting representation is then fed into a S2MLP module, which is designed to model both spatial and spectral dependencies simultaneously. The above process, referred to as the MG-S2MLP module, is formally represented as follows:
X 0 , 1 = f S 2 M L P ( f l p ( X 0 M ) + X 0 ) .
where f l p ( · ) denotes a linear mapping function. Next, an another MG-S2MLP module is applied, constructing a full S2MLP block at the first level, which generates spatial–spectral feature representation, denoted as X 0 . This process is repeated over several levels, where the MG-S2MLP block at the l t h level produces the corresponding feature representation X l for that level, with l = 1 , 2 , , L , where L is the total number of levels. To ensure that valuable information from previous levels is effectively transmitted and utilised at each subsequent stage, we introduce a multi-level information fusion module (MIFM). This module aggregates important features from all preceding levels (i.e., from levels 1 , 2 , , l 1 ) to form the input for the current l t h level. The MIFM strategically fuses information across multiple levels to enhance the learning of spatial and spectral dependencies, and this process is mathematically formulated as
X l i n = f P W ( [ X 1 , X 2 , , X l 1 ] )
where f P W ( · ) refers to a point-wise convolution for multi-level feature fusion. Once the spatial–spectral features from all L levels have been learned, we apply a prediction head with simple convolutional layers to the output features of each level. These prediction heads are responsible for generating the HS reconstructions at all L levels, producing L different HS predictions. These predictions are then used to formulate multiple loss functions, enabling the effective optimization of the reconstruction model. The overall losses are obtained as follows:
L = L L ( f P H L ( X L ) ) + α l = 1 L 1 L l ( f P H l ( X l ) )
where f P H l refers to the prediction head applied at the l t h level, which is responsible for generating a reconstructed HSI from the learned representations X l . The loss function L * at each level consists of two components: a reconstruction loss and a projection loss. The reconstruction loss measures the difference between the ground-truth HSI X and the prediction at each level, while the projection loss evaluates the difference between the snapshot measurement X and the degraded prediction using the simulated compressive sensing process f C S . This dual-loss function ensures that our model not only reconstructs the HSI but also aligns it with the observed measurement. The loss at the l t h level can be expressed as
L l = X f P H l ( X l ) 2 + Y f C S ( f P H l ( X l ) ) 2
Next, we will present a detailed description of our proposed spatial and spectral representation leaning component, S2MLP, and the attention-based mask modelling module, AMMM.
Spatial and spectral MLP module (S2MLP): The S2MLP module is composed of two key components: a spatial MLP that facilitates the communication of information across the spatial domain and a spectral MLP that manages the interactions among spectral bands. To process the input feature, layer normalization (LN) is employed to standardize the input for the following exploitation of spatial correlations. Unlike the conventional Spatial Fully-Connected layer (SpatialFC) [48], which has a quadratic computational complexity related to the size of the feature map, S2MLP employs Cycle Fully-Connected layers (CycleFC) [44], offering a more efficient approach to capturing long-range dependencies with the linear complexity of number of feature channels. In detail, CycleFC achieves spatial interaction by sampling points cyclically along the channel dimension, maintaining the same computational complexity as Channel Fully-Connected layers (ChannelFC) [48], but with a significantly larger receptive field. This cyclical sampling allows CycleFC to more effectively capture spatial dependencies across the entire feature map without the large computational cost of traditional methods. In essence, S2MLP uses three parallel CycleFC layers, each with different step sizes: S H × 1 , 1 × 1 , and 1 × S W . The integration of these three layers captures potential spatial interactions across a receptive field of size S H × S W , offering a broad coverage of spatial relationships within the feature map. In our experiments, S W and S H are both set to 11, as illustrated in the bottom–center panel of Figure 2. In general, the operation of CycleFC with a step size of S H × S W can be mathematically expressed as follows:
F S H × S W = C y c l e F C ( F i , j , : ) = c = 0 C i n F i + σ i ( c ) , j + σ j ( c ) , c W c , : C F C + b σ i ( c ) = ( c m o d S H ) 1 , σ j ( c ) = ( c S H m o d S W ) 1 ,
where W C F C R C i n × C o u t and b R C o u t denote the weighting matrix and bias of the CycleFC operator, respectively, and the terms σ i ( c ) and σ j ( c ) refer to the spatial offset in the horizontal and vertical directions for the c t h channel, determining how the feature map is sampled during cyclical operation. These offsets enable the CycleFC layers to efficiently capture long-range spatial dependencies by shifting across the feature map in both spatial dimensions. Subsequently, we sum the outputs from the three parallel CycleFC layers, each operating with different step sizes to capture spatial interactions at varying scales. This summation aggregates spatial information from all three receptive fields, resulting in a richer spatial representation. To further refine this representation, we apply a linear mapping function, denoted as f l m ( · ) , which transforms the aggregated output into the final spatial feature F S p a . This process is formally expressed as:
F S p a = f l m ( F 11 × 1 + F 1 × 1 + F 1 × 11 )
Next, the spectral MLP is designed to capture interactions across the spectral domain. Similar to the spatial MLP, layer normalization is firstly employed to standardize the input features, and then a ChannelFC layer is applied to aggregate information across the channels with a step size of 1, effectively increasing the number of channels in the feature map by a factor of four. By expanding the channel dimension, the model can capture a broader range of spectral dependencies. After the channel expansion, a GeLU activation function is applied, introducing non-linearity to enhance the feature learning process. Finally, another ChannelFC layer is used to restore the channel number back to its original size, ensuring that the output retains the same dimensions as the input while preserving the spectral relationships learned during the process. This entire sequence of operations enables the spectral MLP to model complex spectral interactions efficiently, balancing both expansion and compression in the channel domain. The overall process of the spectral MLP can be mathematically expressed as follows:
F S p e = f F C + G e L U 2 ( f F C + G e L U 1 ( F S p a ) )
Attention-based mask modelling module: The mask modelling branch is composed of several independent attention-based mask modelling modules (AMMMs), each aligned with a corresponding S2MLP module. These AMMMs are designed to capture the 3D spatial–spectral mask context, which is then used to modulate the input features of the S2MLP. This integrated process, referred to as MG-S2MLP, is crucial for disentangling the latent HS data from the degradation mask, thereby improving the network’s ability to generalize across different conditions and datasets. Specifically, the AMMM begins by transforming the sensing mask M through a convolutional layer, generating an initial mask M 0 with the same dimensions as the input feature of the corresponding S2MLP. This ensures that both the mask and feature map are aligned for subsequent operations. Next, a depth-wise convolution is applied to the mask feature, followed by a Sigmoid activation function, producing a mask attention map A M . By element-wise multiplying the mask attention map A M and feature M 0 , we obtain a modulated mask feature, and this is then passed through another convolution layer to generate the final mask context, which is subsequently used to interact with the S2MLP module. By incorporating mask-specific information, the network can better understand the relationship between the mask and the underlying HS data, leading to more accurate reconstructions. This entire process can be mathematically described as follows:
F m a s k = M 0 f S i g m o i d ( f D w ( M 0 ) ) + f C o n v ( M 0 )
where f D w refers to the depth-wise convolution operation and ⊙ represents the element-wise multiplication.

5. Experiments

This section firstly introduces the experimental settings including the datasets, the network training process, and the evaluation metrics used. Next, we assess the performance of our proposed MG-S2MLPNet using both simulated and real-world HSI datasets and then conduct an ablation study to highlight the impact and effectiveness of different proposed components.

5.1. Experimental Settings

Datasets: We utilised two distinct HSI datasets—CAVE [53] and KAIST [54]—for our simulation experiments. The CAVE dataset comprises 32 high-quality HSIs, each with a spatial resolution of 512 × 512 pixels. These images capture a variety of objects and scenes and have good variety for model training. In contrast, the KAIST dataset offers 30 HSIs with significantly larger spatial dimensions of 2704 × 3376 pixels, providing more detailed visual information. In our experiments, the CAVE dataset was employed exclusively for model training due to its compact image size and diverse content, and we synthesised the snapshot measurements using the released sensing mask in the recent work [51], without using spatial shift. To assess the model’s performance, we selected a subset of 10 cropped scenes from the KAIST dataset for evaluation. These cropped scenes were chosen following same experimental settings as in [22,23,32,33,42,45], ensuring a fair and consistent benchmarking comparison.
In the real-world data experiment, we used five HSIs collected from the TSA-Net work [22] to evaluate the performance of our method. These real scenes were collected using the general CASSI setup with a shift step d = 2 in the disperser. Each of these HSIs consists of 660 × 660 pixels and 28 spectral bands, capturing a rich and detailed representation of the scene’s spectral properties. For training, we followed the strategy used in [42], using samples extracted from both the CAVE and KAIST datasets. These training samples were selected to match the dimensions of the test data, with patches sized at 660 × 660 × 28 .
Implementation Details: We implemented our proposed MG-S2MLPNet model using the PyTorch framework. The model was trained using the Adam optimizer, configured with hyperparameters β 1 = 0.9 and β 2 = 0.999 , to control the momentum and smooth the optimization process. We trained the model for a total of 300 epochs and employed a cosine annealing scheduler to adjust the learning rate during training. The initial learning rate was set to 4 × 10 4 , and the batch size was set as 2. To enhance the diversity of the training samples, data augmentations such as random flipping and rotation were applied during training.
Evaluation Metrics: The effectiveness of our HSI restoration method will be evaluated using two quantitative metrics: the Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM). These metrics provide a quantitative assessment of the quality of the restored image, with PSNR measuring the fidelity of the reconstruction by comparing the restored image to the ground truth, while SSIM evaluates the structural consistency and perceptual quality between the two images.

5.2. Quantitative Results

We conducted a thorough comparative analysis of our proposed MG-S2MLPNet and several state-of-the-art (SOTA) HSI restoration methods. The SOTA techniques included three model-based approaches: TwIST [35], GAP-TV [20], and DeSCI [17]; four E2E methods: TSA-Net [22], HDNet [23], MST [42], and CST [41]; and eight DUMs: DGSMP [31], GAP-Net [55], ADMM-Net [40], DAUHST [32], PADUT [33], MAUN [56], and RDLUF-L [45]. To ensure a fair comparison, all methods were trained on the same datasets and evaluated under identical conditions. The results, compared using the evaluation metrics PSNR and SSIM and based on 10 simulated scenes, are presented in Table 1. From Table 1, it can be observed both that E2E models significantly outperform traditional model-based methods and that the DUMs usually obtain further performance improvements compared to the E2E methods. Our proposed MG-S2MLPMet surpasses all competing techniques. Specifically, when compared to HDNet [23], MST-L [42], CST-L [41], DAUHST-L [32], PADUT-L [33], MAUN-L [56], and RDLUF-L [45], our proposed model delivers notable improvements in its average PSNR, with gains of 5.15 dB, 4.96 dB, 4.00 dB, 1.76 dB, 1.23 dB, 1.07 dB, and 0.55 dB, respectively. In addition to the superior performance it achieves, our model also boasts lower computational costs and memory overhead, as demonstrated, making it both highly efficient and effective for HSI restoration tasks.

5.3. Qualitative Results

Simulation Results: We also provide a visualization of the comparison of the proposed MG-S2MLPNet for HSI reconstruction by evaluating 3 out of the 28 spectral channels from Scene 1 and juxtaposing the results against eight SOTA methods. As illustrated in Figure 3, our approach yields reconstructions with visually smoother and cleaner textures, while also maintaining the spatial coherence of homogeneous regions. Moreover, the difference between the ground-truth images and the reconstruction made by our proposed model is much smaller than that of other methods at almost spatial positions. These findings demonstrate the efficacy of our method in producing high-quality HSIs characterised by enhanced texture representations and the preservation of spatial information. Additionally, we evaluated the spectral fidelity of the reconstructed images by comparing the spectral density curves of the restored regions with the ground truth. As presented in Figure 4, our method attained the highest correlation coefficient, underscoring the effectiveness of our mask-guided S2MLP learning strategy in preserving spectral consistency and accuracy.
Real Results: To evaluate the reconstruction capabilities of our MG-S2MLPNet on real-world data, we retrained the model using both the CAVE and KAIST datasets, applying a real mask with dimensions of 660 × 660 . For testing, we utilised five real-world scenes from the collected dataset [22]. The reconstruction outcomes for two of these real scenes, as depicted in Figure 5, enable a comparative analysis of our proposed MG-S2MLPNet and seven SOTA methods. Upon closer examination of the visual reconstructions, our model consistently delivers a superior performance, particularly in terms of texture clarity and spatial coherence. Fine details and textures appear more pronounced and distinct in the MG-S2MLPNet results compared to the competing approaches. Additionally, our method demonstrates a better preservation of spatial integrity, avoiding the distortions or blurring that often diminish the quality of the reconstructions produced by the other SOTA techniques. Overall, MG-S2MLPNet shows strong competitiveness, offering more faithful and visually accurate reconstructions.

5.4. Ablation Study

To thoroughly analyse the contribution of each component to the performance improvements of our model, we conducted a detailed ablation study. Initially, we used the S2MLP model without the mask modelling branch as the baseline. In this baseline, we varied the number of feature channels (C) within the S2MLP module and the S2MLP block number (B) to examine their impact on performance. As shown in Table 2, the increased channel and block numbers are capable of improving the model’s reconstruction performance, while higher numbers cause high computational costs. It can be observed from Table 2 that three blocks of S2MLP with 28 spectral and 56 spatial channels achieve an impressive performance with a much lower computational cost, as seen in Table 1. Then, we conducted ablation study using the network configuration with three S2MLP blocks. We incrementally introduced the key components of our framework: the attention-based mask modelling branch (AMMM), the projection loss, the multi-level information fusion module (MIFM), and the multiple prediction heads for deep supervision (DS). Each of these components was added one by one to isolate their individual contributions to the overall enhancement of the model. The comparative results are presented in Table 2, clearly demonstrating that each of the proposed modules contributes positively to performance. The integration of the AMMM refines feature extraction according to the degradation degrees hidden in the sensing mask, while the projection loss enforces consistency in the optical measurement process, leading to more accurate reconstructions. The MIFM further enhances the model by fusing information at various levels of abstraction, improving the depth and richness of the features. Finally, deep supervision (DS) with multiple prediction heads ( α = 0.2 ) provides finer control over intermediate representations, allowing the model to refine predictions at different stages of processing. Together, these components work synergistically to significantly elevate the model’s performance beyond its baseline.
To evaluate the effectiveness of S2MLP in comparison to CNN-based architectures and conventional transformer models, we substituted the S2MLP blocks with ResBlock, the Swin Transformer (Swin) [43], and the Spectral Transformer (Spe-Trans) [42]. In this experimental setup, the step size of the CycleFC in S2MLP was set to 11. As indicated in Table 3, S2MLP achieved the highest performance, exceeding SW-MSA and S-MSA by 4.23, 0.65, and 0.77 dB, respectively, with lower computational complexity. These findings underscore that the S2MLP-based model provides a streamlined and efficient approach for achieving superior performance in hyperspectral image reconstruction. Further, we assess the impact of the step-size parameter within the CycleFC layer by testing step sizes of 7, 9, 11, and 13. As shown in Table 3, a step size of 11 results in the highest performance among the configurations examined. Finally, we adjust the hyperparameter α in Equation (7) to assess its influence on reconstruction performance. The results indicate that variations in α do not significantly impact the overall performance.
To provide insights into the recovered spatial and spectral data both with (w/) and without (w/o) degradation information incorporated, we present difference images that show the discrepancies between the ground-truth data and the reconstructed data across 9 out of 28 spectral bands. Figure 6 illustrates comparative visualizations of Scenes 2 and 5, confirming that reconstruction errors with the AMMM method are significantly reduced across most spatial regions. Additionally, we calculate the PSNR values for each spectral band w/ and w/o the use of ADMM; the resulting comparison curves are illustrated on the right side of Figure 6. These curves also demonstrate that incorporating degradation information (using AMMM) improves PSNR values across all spectral bands. For Scenes 2 and 5, while the PSNR improvements for spectral bands within the 480–500 nm range show a slight degradation, they still exceed an increase of 2 dB.

6. Conclusions

In this study, we introduced a novel yet simple S2MLP model designed to effectively tackle HSI reconstruction. Our approach concentrated on leveraging a powerful yet computationally efficient spatial and spectral modelling mechanism by using the CycleFC framework. This integration ensures that our model can capture complex spatial and spectral patterns without imposing a heavy computational burden. Furthermore, we enriched the model’s learning capacity by incorporating a attention-based mask modelling branch, which facilitates the disentangling of representations, allowing the network to better separate and handle different features during the learning process. To further refine the model’s capabilities, we explored multi-level information fusion, where information from different levels of the network is combined. This, coupled with supervision applied to multiple feature layers (deep supervision), strengthens the model’s ability to learn more robust and detailed representations. Extensive experimental evaluations underscored the effectiveness of our proposed approach. The results consistently demonstrate that our model excelled in producing high-quality reconstructions, surpassing SOTA methods in terms of its accuracy and preservation of detail.

Author Contributions

Conceptualization, methodology, software, validation, formal analysis, writing—original draft preparation, and visualization, X.-H.H.; writing—review and editing, X.-H.H., J.W. and Y.-W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a research grant from the Okawa Foundation for Information and Telecommunications Foundation.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available in a publicly accessible repository. The datasets used are the CAVE and KAIST datasets. The CAVE dataset is available at https://cave.cs.columbia.edu/repository/Multispectral 4 April 2024, and the Harvard dataset is available at http://dx.doi.org/10.1145/3130800.3130810 5 April 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Maggiori, E.; Charpiat, G.; Tarabalka, Y.; Alliez, P. Recurrent neural networks to correct satellite image classification maps. IEEE Trans. Geosci. Remote Sens. 2016, 55, 4962–4971. [Google Scholar] [CrossRef]
  2. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  3. Hanachi, R.; Sellami, A.; Farah, I.R.; Mura, M.D. Multi-view graph representation learning for hyperspectral image classification with spectral–spatial graph neural networks. Neural Comput. Appl. 2024, 36, 3737–3759. [Google Scholar] [CrossRef]
  4. Borengasser, M.; Hungate, W.S.; Watkins, R. Hyperspectral Remote Sensing: Principles and Applications; CRC Press: Boca Raton, FL, USA, 2007. [Google Scholar]
  5. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sensin 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  6. Solomon, J.; Rock, B. Imaging spectrometry for earth remote sensing. Science 1985, 42, 1778–1790. [Google Scholar]
  7. Ding, C.; Zheng, M.; Zheng, S.; Xu, Y.; Zhang, L.; Wei, W. Integrating prototype learning with graph convolution network for effective active hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5504816. [Google Scholar] [CrossRef]
  8. Zhi, L.; Zhang, D.; Yan, J.Q.; Li, Q.L.; Tang, Q.L. Classification of hyperspectral medical tongue images for tongue diagnosis. Comput. Med. Imaging Graph. 2007, 31, 672–678. [Google Scholar] [CrossRef]
  9. Fei, B. Hyperspectral imaging in medical applications. In Data Handling in Science and Technology; Elsevier: Amsterdam, The Netherlands, 2019; Volume 32, pp. 523–565. [Google Scholar]
  10. Llull, P.; Liao, X.; Yuan, X.; Yang, J.; Kittle, D.; Carin, L.; Sapiro, G.; Brady, D.J. Coded aperture compressive temporal imaging. Opt. Express 2013, 21, 10526–10545. [Google Scholar] [CrossRef]
  11. Wagadarikar, A.; John, R.; Willett, R.; Brady, D. Single disperser design for coded aperture snapshot spectral imaging. Appl. Opt. 2008, 47, B44–B51. [Google Scholar] [CrossRef]
  12. Wagadarikar, A.A.; Pitsianis, N.P.; Sun, X.; Brady, D.J. Video rate spectral imaging using a coded aperture snapshot spectral imager. Opt. Express 2009, 17, 6368–6388. [Google Scholar] [CrossRef]
  13. Cao, X.; Yue, T.; Lin, X.; Lin, S.; Yuan, X.; Dai, Q.; Carin, L.; Brady, D.J. Computational snapshot multispectral cameras: Toward dynamic capture of the spectral world. IEEE Signal Process. Mag. 2016, 33, 95–108. [Google Scholar] [CrossRef]
  14. Gehm, M.E.; John, R.; Brady, D.J.; Willett, R.M.; Schulz, T.J. Single-shot compressive spectral imaging with a dual-disperser architecture. Opt. Express 2007, 15, 14013–14027. [Google Scholar] [CrossRef] [PubMed]
  15. Arce, G.R.; Brady, D.J.; Carin, L.; Arguello, H.; Kittle, D.S. Compressive coded aperture spectral imaging: An introduction. IEEE Signal Process. Mag. 2013, 31, 105–115. [Google Scholar] [CrossRef]
  16. Wang, L.; Xiong, Z.; Gao, D.; Shi, G.; Zeng, W.; Wu, F. High-speed hyperspectral video acquisition with a dual-camera architectur. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4942–4950. [Google Scholar]
  17. Liu, Y.; Yuan, X.; Suo, J.; Brady, D.J.; Dai, Q. Rank minimization for snapshot compressive imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2990–3006. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, L.; Xiong, Z.; Shi, G.; Wu, F.; Zeng, W. Adaptive nonlocal sparse representation for dual-camera compressive hyperspectral imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 2104–2111. [Google Scholar] [CrossRef]
  19. He, W.; Yao, Q.; Li, C.; Yokoya, N.; Zhao, Q.; Zhang, H.; Zhang, L. Non-local meets global: An iterative paradigm for hyperspectral image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2089–2107. [Google Scholar] [CrossRef]
  20. Yuan, X. Generalized alternating projection based total variation minimization for compressive sensing. In Proceedings of the 2016 IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 2539–2543. [Google Scholar]
  21. Figueiredo, M.A.; Nowak, R.D.; Wright, S.J. Gradient projection for sparse reconstruction: Application to compressed sensing and other inverse problems. IEEE J. Sel. Top. Signal Process. 2017, 1, 586–597. [Google Scholar] [CrossRef]
  22. Meng, Z.; Ma, J.; Yuan, X. End-to-end low cost compressive spectral imaging with spatial-spectral self attention. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 187–204. [Google Scholar]
  23. Hu, X.; Cai, Y.; Lin, J.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Gool, L.V. Hdnet: High-resolution dual-domain learning for spectral compressive imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  24. Miao, X.; Yuan, X.; Pu, Y.; Athitsos, V. l-net: Reconstruct hyperspectral images from a snapshot measurement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  25. Wang, L.; Sun, C.; Fu, Y.; Kim, M.H.; Huang, H. Hyperspectral image reconstruction using a deep spatial-spectral prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  26. Yorimoto, K.; Han, X.H. HyperMixNet: Hyperspectral Image Reconstruction with Deep Mixed Network from a Snapshot Measurement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1184–1193. [Google Scholar]
  27. Zhang, X.; Zhang, Y.; Xiong, R.; Sun, Q.; Zhang, J. Herosnet: Hyperspectral explicable reconstruction and optimal sampling deep network for snapshot compressive imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 17532–17541. [Google Scholar]
  28. Takabe, T.; Han, X.; Chen, Y. Deep Versatile Hyperspectral Reconstruction Model from A Snapshot Measurement with Arbitrary Masks. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 2390–2394. [Google Scholar]
  29. Han, X.; Wang, J.; Chen, Y. Hyperspectral Image Reconstruction Using Hierarchical Neural Architecture Search from A Snapshot Image. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 2500–2504. [Google Scholar]
  30. Wang, L.; Sun, C.; Zhang, M.; Fu, Y.; Huang, H. Dnu: Deep non-local unrolling for computational spectral imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1661–1671. [Google Scholar]
  31. Huang, T.; Dong, W.; Yuan, X.; Wu, J.; Shi, G. Deep gaussian scale mixture prior for spectral compressive imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2021; pp. 16216–16225. [Google Scholar]
  32. Cai, Y.; Lin, J.; Wang, H.; Yuan, X.; Ding, H.; Zhang, Y.; Timofte, R.; Gool, L.V. Degradation-aware unfolding half-shuffle transformer for spectral compressive imaging. Adv. Neural Inf. Process. Syst. 2022, 35, 37749–37761. [Google Scholar]
  33. Li, M.; Fu, Y.; Liu, J.; Zhang, Y. Pixel adaptive deep unfolding transformer for hyperspectral image reconstruction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023. [Google Scholar]
  34. Eason, D.T.; Andrews, M. Total variation regularization via continuation to recover compressed hyperspectral images. IEEE Trans. Image Process. 2014, 24, 284–293. [Google Scholar] [CrossRef]
  35. Dias, J.M.B.; Figueiredo, M.A. A new twist: Two step iterative shrinkage-thresholding algorithms for image restoration. IEEE Trans. Image Process. 2007, 16, 2992–3004. [Google Scholar] [CrossRef]
  36. Zhang, S.; Dong, Y.; Fu, H.; Huang, S.L.; Zhan, L. A spectral reconstruction algorithm of miniature spectrometer based on sparse optimization and dictionary learning. Sensors 2018, 18, 644. [Google Scholar] [CrossRef] [PubMed]
  37. Lin, X.; Liu, Y.; Wu, J.; Dai, Q. Spatialspectral encoded compressive hyperspectral imaging. ACM Trans. Graph. (TOG) 2014, 33, 1–11. [Google Scholar] [CrossRef]
  38. Fu, Y.; Zheng, Y.; Sato, I.; Sato, Y. Exploiting spectral-spatial correlation for coded hyperspectral image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3727–3766. [Google Scholar]
  39. Zhang, S.; Wang, L.; Fu, Y.; Zhong, X.; Huang, H. Computational hyperspectral imaging based on dimension-discriminative low-rank tensor recovery. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 10183–10192. [Google Scholar]
  40. Ma, J.; Liu, X.; Shou, Z.; Yuan, X. Deep tensor admm-net for snapshot compressive imaging. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  41. Cai, Y.; Lin, J.; Hu, X.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Gool, L.V. Coarse-to-fine sparse transformer for hyperspectral image reconstruction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  42. Cai, Y.; Lin, J.; Hu, X.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Gool, L.V. Mask-guided spectral-wise transformer for efficient hyperspectral image reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 17502–17511. [Google Scholar]
  43. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  44. Chen, S.; Xie, E.; GE, C.; Chen, R.; Liang, D.; Luo, P. CycleMLP: A MLP-like Architecture for Dense Prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 14284–14300. [Google Scholar] [CrossRef] [PubMed]
  45. Dong, Y.; Gao, D.; Qiu, T.; Li, Y.; Yang, M.; Shi, G. Residual Degradation Learning Unfolding Framework with Mixing Priors across Spectral and Spatial for Compressive Spectral Imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  46. Zhang, J.; Zeng, H.; Chen, Y.; Yu, D.; Zhao, Y.P. Improving Spectral Snapshot Reconstruction with Spectral-Spatial Rectification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 25817–25826. [Google Scholar]
  47. Yao, Z.; Liu, S.; Yuan, X.; Fang, L. SPECAT: SPatial-spEctral Cumulative-Attention Transformer for High-Resolution Hyperspectral Image Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  48. Tolstikhin, I.O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J. Mlp-mixer: An all-mlp architecture for vision. Adv. Neural Inf. Process. Syst. 2021, 34, 24261–24272. [Google Scholar]
  49. Liu, H.; Dai, Z.; So, D.; Le, Q.V. Pay attention to mlps. Adv. Neural Inf. Process. Syst. 2021, 34, 9204–9215. [Google Scholar]
  50. Touvron, H.; Bojanowski, P.; Caron, M.; Cord, M.; ElNouby, A.; Grave, E.; Izacard, G.; Joulin, A.; Synnaeve, G.; Verbeek, J. Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5314–5321. [Google Scholar] [CrossRef]
  51. Motoki, Y.; Yoshikazu, Y.; Takayuki, K. Video-rate hyperspectral camera based on a cmoscompatible random array of fabry–pérot filters. Nat. Photonics 2023, 17, 218–223. [Google Scholar]
  52. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  53. Yasuma, F.; Mitsunaga, T.; Iso, D.; Nayar, S.K. Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum. IEEE Trans. Image Process. 2010, 19, 2241–2253. [Google Scholar] [CrossRef]
  54. Choi, I.; Jeon, D.S.; Nam, G.; Gutierrez, D.; Kim, M.H. High-quality hyperspectral reconstruction using a spectral prior. ACM Trans. Graph. 2017, 36, 1–13. [Google Scholar] [CrossRef]
  55. Meng, Z.; Jalali, S.; Yuan, X. Gap-net for snapshot compressive imaging. arXiv 2020, arXiv:2012.08364. [Google Scholar]
  56. Hu, Q.; Ma, J.; Gao, Y.; Jiang, J.; Yuan, Y. MAUN: Memory-Augmented Deep Unfolding Network for Hyperspectral Image Reconstruction. IEEE/CAA J. Autom. Sin. 2024, 11, 1139–1150. [Google Scholar] [CrossRef]
Figure 1. The conceptual scheme of the coding process in a spectral compressive imaging system.
Figure 1. The conceptual scheme of the coding process in a spectral compressive imaging system.
Sensors 24 07362 g001
Figure 2. An overview of the proposed MGS2MLP model.
Figure 2. An overview of the proposed MGS2MLP model.
Sensors 24 07362 g002
Figure 3. Comparison of visual quality of Scene 1, using 3 (out of 28) spectral channels, and the difference between the ground truth images and those reconstructed with 8 SOTA methods.
Figure 3. Comparison of visual quality of Scene 1, using 3 (out of 28) spectral channels, and the difference between the ground truth images and those reconstructed with 8 SOTA methods.
Sensors 24 07362 g003
Figure 4. Spectral density curves in ‘a’ and ‘b’ corresponding to two small regions: a and b in Scene 6. The correlation refers to the coincidence degree between the reconstruction and ground-truth curves.
Figure 4. Spectral density curves in ‘a’ and ‘b’ corresponding to two small regions: a and b in Scene 6. The correlation refers to the coincidence degree between the reconstruction and ground-truth curves.
Sensors 24 07362 g004
Figure 5. HSI reconstruction results of the proposed MG-S2MLP and seven SOTA methods on the real Scene 1.
Figure 5. HSI reconstruction results of the proposed MG-S2MLP and seven SOTA methods on the real Scene 1.
Sensors 24 07362 g005
Figure 6. The comparative visualisation of the difference images between the ground-truth data and reconstructed data across 9 out of 28 spectral bands, and the PSNR values of all spectral bands with and without degradation incorporation (AMMM) for Scenes 2 and 5.
Figure 6. The comparative visualisation of the difference images between the ground-truth data and reconstructed data across 9 out of 28 spectral bands, and the PSNR values of all spectral bands with and without degradation incorporation (AMMM) for Scenes 2 and 5.
Sensors 24 07362 g006
Table 1. Comparisons with SoTA methods over 10 simulation scenes.
Table 1. Comparisons with SoTA methods over 10 simulation scenes.
MethodsParamsGFLOPss1s2s3s4s5s6s7s8s9s10Avg
TwIST [35]--25.1623.0221.4030.1921.4120.9522.2021.8222.4222.6723.12
0.7000.6040.7110.8510.6350.6440.6430.6500.6900.5690.669
GAP-TV [20]--26.8222.8926.3130.6523.6421.8523.7621.9822.6323.124.36
0.7540.6100.8020.8520.7030.6630.6880.6550.6820.5840.669
DeSCI [17]--27.1323.0426.6234.9623.9422.3824.4522.0324.5623.5925.27
0.7480.6200.8180.8970.7060.6830.7430.6730.7320.5870.721
λ -Net [24]62.64M117.9830.1028.4927.7337.0126.1928.6426.4726.0927.5027.1328.53
0.8490.8050.8700.9340.8170.8530.8060.8310.8260.8160.841
TSA-Net [22]44.25M110.0632.0331.0032.2539.1929.3931.4430.3229.3530.0129.5931.46
0.8920.8580.9150.9530.8840.9080.8780.8880.8900.8740.894
DGSMP [31]3.76M646.6533.2632.0933.0640.5428.8633.0830.7431.5531.6631.4432.63
0.9150.8980.9250.9640.8820.9370.8860.9230.9110.9250.917
GAP-Net [55]4.27M78.5833.7433.2634.2841.0331.4432.4032.2730.4633.5130.2433.26
0.9110.9000.9290.9670.9190.9250.9020.9050.9150.8950.917
ADMM-Net [40]4.27M78.5834.1233.6235.0441.1531.8232.5432.4230.7433.7530.6833.58
0.9180.9020.9310.9660.9220.9240.8960.9070.9150.8950.918
HDNet [23]2.37M154.7635.1435.6736.0342.3032.6934.4633.6732.4834.8932.3834.97
0.9350.9400.9430.9690.9460.9520.9260.9410.9420.9370.943
MST-L [42]2.03M28.1535.4035.8736.5142.2732.7734.8033.6632.6735.3932.5035.18
0.9410.9440.9530.9730.9470.9550.9250.9480.9490.9410.948
CST-L [41]3.00M40.0135.9636.8438.1642.4433.2535.7234.8634.3436.5133.0936.12
0.9490.9550.9620.9750.9550.9630.9440.9610.9570.9450.957
DAUHST-L [32]6.15M79.5037.2539.0241.0546.1535.8037.0837.5735.1040.0234.5938.36
0.9580.9670.9710.9830.9690.9700.9630.9660.9700.9560.967
PADUT-L [33]5.38M90.4637.3640.4342.3846.6236.2637.2737.8335.3340.8634.5538.89
0.9620.9780.9790.9900.9740.9740.9660.9740.9780.9630.974
MAUN-L [56]3.77M143.8337.7840.5341.8846.8536.7437.7837.4436.0540.5434.9039.05
0.9630.9760.9730.9860.9730.9740.9610.9710.9730.9620.971
RDLUF [45]1.81M115.1637.9440.9543.2547.8337.1137.4738.5835.5041.8335.2339.57
0.9660.9770.9790.9900.9760.9750.9690.9700.9780.9620.974
MG-S2MLPNet0.31M15.1239.4742.2641.3945.0839.1539.8638.9737.0540.9337.0540.12
0.9820.9890.9820.9900.9880.9880.9760.9800.9870.9880.985
Table 2. Break-down of ablation study.
Table 2. Break-down of ablation study.
CS2MLP (C = 28, B = 3)
CS2MLP (C = 56, B = 2)
CS2MLP (C = 56, B = 4)
CS2MLP (C = 56, B = 3)
MAMM
P-loss
MIFM
DS
PSNR34.8939.5339.6639.2739.9739.1040.1240.25
SSIM0.9620.9830.9840.9810.9850.9840.9850.985
Table 3. Comparisons with other architectures and different step sizes in the CycleFC layer.
Table 3. Comparisons with other architectures and different step sizes in the CycleFC layer.
ArchitecturesStep Sizes in CycleFC α Value in Equation (7)
ResBlockSWinSpe-TransS2MLP7911130.10.20.5
PSNR35.8939.4739.3540.1239.9740.1140.1239.9640.1340.1240.10
SSIM0.9680.9810.9800.9850.9840.9850.9850.9840.9850.9850.985
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, X.-H.; Wang, J.; Chen, Y.-W. Mask-Guided Spatial–Spectral MLP Network for High-Resolution Hyperspectral Image Reconstruction. Sensors 2024, 24, 7362. https://doi.org/10.3390/s24227362

AMA Style

Han X-H, Wang J, Chen Y-W. Mask-Guided Spatial–Spectral MLP Network for High-Resolution Hyperspectral Image Reconstruction. Sensors. 2024; 24(22):7362. https://doi.org/10.3390/s24227362

Chicago/Turabian Style

Han, Xian-Hua, Jian Wang, and Yen-Wei Chen. 2024. "Mask-Guided Spatial–Spectral MLP Network for High-Resolution Hyperspectral Image Reconstruction" Sensors 24, no. 22: 7362. https://doi.org/10.3390/s24227362

APA Style

Han, X. -H., Wang, J., & Chen, Y. -W. (2024). Mask-Guided Spatial–Spectral MLP Network for High-Resolution Hyperspectral Image Reconstruction. Sensors, 24(22), 7362. https://doi.org/10.3390/s24227362

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop