Next Article in Journal
Analysis of Deformation Dynamics in Guatemala City Metropolitan Area Using Persistent Scatterer Interferometry
Next Article in Special Issue
Vehicle Detection in Multisource Remote Sensing Images Based on Edge-Preserving Super-Resolution Reconstruction
Previous Article in Journal
Analysis of Regional Ambient Seismic Noise in the Chukchi Sea Area in the Arctic Based on OBS Data from the Ninth Chinese National Arctic Scientific Survey
Previous Article in Special Issue
Guided Local Feature Matching with Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Adaptively Optimized PCNN Model for Hyperspectral Image Sharpening

1
Faculty of Geomatics, Lanzhou Jiaotong University, Lanzhou 730070, China
2
National-Local Joint Engineering Research Center of Technologies and Applications for National Geographic State Monitoring, Lanzhou 730070, China
3
Gansu Provincial Engineering Laboratory for National Geographic State Monitoring, Lanzhou 730070, China
4
State Key Laboratory of Resources and Environmental Information System, Institute of Geographical Sciences and Natural Resources Research, CAS, Beijing 100101, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(17), 4205; https://doi.org/10.3390/rs15174205
Submission received: 26 June 2023 / Revised: 12 August 2023 / Accepted: 24 August 2023 / Published: 26 August 2023

Abstract

:
Hyperspectral satellite imagery has developed rapidly over the last decade because of its high spectral resolution and strong material recognition capability. Nonetheless, the spatial resolution of available hyperspectral imagery is inferior, severely affecting the accuracy of ground object identification. In the paper, we propose an adaptively optimized pulse-coupled neural network (PCNN) model to sharpen the spatial resolution of the hyperspectral imagery to the scale of the multispectral imagery. Firstly, a SAM-CC strategy is designed to assign hyperspectral bands to the multispectral bands. Subsequently, an improved PCNN (IPCNN) is proposed, which considers the differences of the neighboring neurons. Furthermore, the Chameleon Swarm Optimization (CSA) optimization is adopted to generate the optimum fusion parameters for IPCNN. Hence, the injected spatial details are acquired in the irregular regions generated by the IPCNN. Extensive experiments are carried out to validate the superiority of the proposed model, which confirms that our method can realize hyperspectral imagery with high spatial resolution, yielding the best spatial details and spectral information among the state-of-the-art approaches. Several ablation studies further corroborate the efficiency of our method.

1. Introduction

Hyperspectral (HS) images are acquired by sampling the spectrum range into a large number of spectral channels, rendering them as enhanced multispectral (MS) images with multiple bands, a narrow spectral range, and abundant spectral information. Since the abundance of spectral bands enables HS imagery to identify ground cover types accurately, HS imagery is broadly applied in environmental monitoring [1], agricultural product assessment [2], geology [3], and mineralogical mapping [4]. However, because of the signal-to-noise ratio (SNR) constraints of satellite sensors, the spatial resolution and the spectral resolution of HS imagery would have to be inevitably compromised in a unique acquisition. Hence, HS images are characterized by high spectral resolution but low spatial resolution, which limits their applications in higher precision remote sensing (RS) interpretation. RS image sharpening is a cost-effective way to generate imagery with simultaneously rich spectral information and high spatial resolution by injecting spatial details into either HS or MS imagery, which could enhance the spatial resolution of the RS images. RS image sharpening has been extensively studied over the last four decades. Nonetheless, most of these methods aim to fuse the MS imagery and the panchromatic (PAN) imagery, commonly known as pansharpening. As more HS satellites have been launched more recently (such as Earth Observing-1, ZY-1-02D, AVIRIS, and GF-5), HS sharpening has become increasingly prominent.
To solve the HS sharpening problem, many traditional pansharpening methods can still be applied. Pansharpening methods fall into two broad categories, namely the component substitution (CS) method and the multiresolution analysis (MRA) method [5]. The CS method involves projecting the original MS image onto the transform domain and substituting the spatial component with the PAN image [6,7,8,9,10,11,12,13]. Since the relatively large differences in spectral ranges between the PAN component and the replaced MS spatial component, the fusion result suffers from significant spectral distortion in the CS method. On the contrary, the MRA method preserves spectral information by injecting spatial details from the PAN imagery into the upsampled MS image, which is achieved by multiscale spatial filtering [14,15,16,17,18,19,20]. The MRA method is able to conserve the spectral information well, whereas it cannot get impressive spatial details.
HS sharpening aims to produce HS imagery with high spatial resolution by fusing low-resolution HS imagery and high-resolution MS imagery. In order to tackle the HS sharpening issue, traditional CS and MRA pansharpening methods are typically utilized by simply replacing both the PAN imagery and the MS imagery with the MS imagery and the HS imagery simultaneously [21]. Gomez et al. are the pioneers to apply the pansharpening approach to HS sharpening [22] and used the 2D-wavelet transform to render the MS and HS images fused in the same wavelength range. Chen et al. [23] present a generic HS sharpening framework based on the pansharpening method, which is the primary inspiration for later transmigration methods from pansharpening to HS sharpening. More recently, some methods for HS sharpening have also been proposed. Picone et al. addressed the band-assignment problem of the HS sharpening [24]. Lu et al. propose a spectral modulation hyper-sharpening framework [25], which mitigates the problem of large spectral distortion after fusion. Yokoya et al. [26] propose a coupled nonnegative matrix factorization method (CNMF), which generates the end members and abundance matrices by alternately unmixing HS and MS images with the NMF algorithm. However, the three-dimensional structure of the HS image is hardly reserved by the matrix factorization. Thus, tensor factorization is utilized in place of the matrix factorization [21]. Dian et al. present a nonlocal sparse tensor factorization method (NLSTF_SMBF) for the semi-blind fusion of HS and MS images [27], which primarily constructs full-band patches (FBPs), and similar FBPs can share the same dictionary.
In recent years, deep learning (DL) methods have been extensively applied in the field of HS sharpening [28,29]. Zhang et al. [30] proposed an unsupervised deep learning network architecture for the simultaneous optimization of the HS super-resolution and the degradation estimation. Qu et al. [31] proposed an unsupervised, unaligned Mutual HS super-resolution Dirichlet-Net, which effectively improves the robustness in the face of alignment errors.
Although promising fusion results are realized by various HS sharpening methods, it is expected to further improve the HS fusion accuracy by neural network algorithms. Nonetheless, due to the independent spectral features of the unique RS image, it is difficult to establish a universal training database for the RS fusion applications. Besides, it is time-consuming to train the network. A pulse-coupled neural network (PCNN) is a kind of biologically inspired neural network without training, which exhibits the characteristics of the pulse-synchronous phenomenon. The property of pulse-synchronous allows the synchronously stimulated pixels to generate the segmentation results, which turns out to be in accordance with the human visual mechanism [32]. In order to improve the accuracy of image fusion, it is useful to adopt different fusion strategies for different segmentation regions that are consistent with human perception. Traditional PCNN fusion models include simplified PCNN (SPCNN) [33], dual-channel PCNN (DCPCNN) [34], and shuffled frog-leaping PCNN [35]. Particularly, Panigrahy et al. propose adaptive DCPCNN for multi-aggregation and medical image fusion [36,37]. However, traditional PCNN fusion approaches are applicable only to either the medical image fusion or the multi-focus image fusion, which cannot be applied to the RS image fusion. Recently, Li et al. presented a modified PCNN model for RS image fusion [32], but it only works on MS imagery.
In this paper, a novel adaptively optimized PCNN model for HS sharpening is proposed. Concretely, we summarize the main contributions of the paper as follows: (1) a SAM-CC band assignment method is proposed to group the HS bands with the MS band; (2) an improved PCNN (IPCNN) model is proposed to obtain irregular injection region of spatial details. (3) A Chameleon Swarm Optimization (CSA)-based IPCNN parameter optimization method is designed to achieve the optimal sharpening imagery. The comparative experiments were carried out on three datasets captured by the ZY-1-02D and GF-2 satellites, and the results substantiate the effectiveness of the proposed method.
The rest of the article is organized as follows. The description of the PCNN and CSA principles is given in Section 2. Section 3 presents the details of our proposed fusion approach. Experimental results and discussions are provided in Section 4. Section 5 contains the conclusion of the paper.

2. Related Work

2.1. Standard PCNN Principle

PCNN belongs to the third generation of artificial neural networks, which is a kind of neural network model proposed by Johnson [38] on the basis of the observation of pulse delivery experiments in the cerebral cortex of cats and monkeys. The individual neuron in the PCNN model is partitioned into the receptive field, the modulation field, and the pulse delivery field according to its functions. Mathematical formulae of the receptive field are given in Equations (1) and (2). The mathematical description of the modulation field can be found in Equation (3), and the mathematical formulae of the pulse delivery field can be defined in Equations (4) and (5).
F i j n = e α F F i j n 1 + V F k l M i j k l Y i j n 1 + I i j
L i j n = e α L L i j n 1 + V L k l W i j k l Y k l n 1
U i j n = F i j n 1 + β L i j n
Y i j n = 1 U i j n > E i j n 0 o t h e r w i s e
E i j n + 1 = e α E E i j n + V E Y i j n
where the subscripts ij and kl refer to the positions of the current neuron and the neighboring neurons, respectively. n denotes the nth iteration. I is the input image, whereas Iij corresponds to the external input of the neuron ij. F and L indicate the feeding input and the linking input, respectively, while the difference between them is that the linking input L can only receive the local pulse stimulation from the neighboring neuron via the synaptic matrix W. On the contrary, F receives not only the local stimulation but also the external stimulation from I. β denotes the linking strength, which ranges from 0 to 1. A larger value of β indicates that the linking relationships are greater, i.e., the current neuron is more susceptible to neighborhood neurons. M and W represent the connection coefficient matrixes, which are typically calculated by the Euclidean distance between the current neuron and the neighborhood. U refers to the internal activity, and E represents the dynamic threshold. U and E jointly determine whether or not the current neuron is stimulated in the current iteration, i.e., the boolean variable Yij becomes equal to one whenever Uij is greater than Eij. αF, αL, and αE are time decay constants, which are utilized to adjust the decay rate of F, L and E. VF, VL and VE indicate the normalization constants. The matrix Y[n] denotes the binary output of PCNN in each iteration n, whereas Yij is 0 (when the neuron ij is unstimulated) and 1 (when stimulated), respectively.
In addition, as illustrated in Figure 1, each PCNN neuron corresponds to a pixel in the image when applying PCNN to image processing applications. Therefore, the current neuron ij will release a pulse as long as Uij is greater than Eij. If the current pixel is stimulated in the current iteration, E will automatically spike the value VE; otherwise, E will decrease gradually as the iteration number n increases. After the nth iteration of PCNN, the neurons with either similar grayscale values or the adjacent positions will be synchronously stimulated, while the output Y will eventually form a binary ignition map.

2.2. Chameleon Swarm Optimization Algorithm

The CSA algorithm is a meta-heuristic method proposed by Malik [39]. The model imitates the socially intelligent synergistic behavior of chameleons when foraging and capturing food near woods, swamps, and deserts. It is a bio-optimization algorithm for finding the global optimum of nonlinear, nonconvex, and other complex problems, which may prevent entrapment in the local optimum.
The principle of CSA is shown in Figure 2. The algorithm mathematizes the behavioral stages of a chameleon in looking for food, which include initializing the starting position, tracking the prey from a distance, locating the prey by eye rotation, and catching the prey with a high-velocity sticky tongue.

3. Proposed Method

In order to tackle the problem of spectral distortion and fuzzy spatial details in HS sharpening, we propose a novel adaptively optimized PCNN fusion algorithm. The flowchart of the proposed HS sharpening algorithm is shown in Figure 3, which contains the following modules: (1) the SAM-CC band assignment part; (2) the improved PCNN (IPCNN) model part; (3) the automatic parameter optimization of IPCNN by CSA part; (4) the extracting MS detail part; (5) the adaptive injected gains part; (6) the fusion output part.

3.1. SAM-CC Band Assignment Block

In contrast to the multi-to-single band assignment in the pansharpening problem, HS sharpening is a multiband-to-multiband fusion. Thus, as shown in Figure 4, correctly assigning each HS band to the MS band is an essential procedure before the actual fusion operation, i.e., the choice of which band in the MS imagery can be used to sharpen the corresponding HS bands plays a crucial role. Classical band selection methods include the minimum spectral distance (MSD) assignment algorithm, the maximum cross-correlation (CC) assignment algorithm, and the minimum spectral distortion (SAM) assignment algorithm [24]. Since the optimal band selection is challenging for any single criterion, we propose a joint band selection algorithm using both the SAM and CC indices (SAM-CC) in order to group the HS bands more accurately.
Let H = {H}h = 1,…, Nh, stands for the HS images and M = {M}m = 1,…, Nm, represents the MS images, where h and m are the band numbers of HS and MS, respectively. Nh and Nm refer to the total number of bands of HS and MS. The proposed SAM-CC criterion is represented as Equation (6).
S A M - C C H h , M m r = arg min m E C C H h , M m r × S A M H h , M m r
C C H h , M m r = H h , M m r H h , H h M m r , M m r
S A M H h , M m r = 1 N h h = 1 N h arccos ( H h , M m r H h . M m r )
where E is a unit matrix with size NM × NH. The symbol < · > indicates the inner product operation. The symbol · stands for the l2 norm. r refers to the spatial resolution ratio of HS to MS. M j r represents the low-pass filtered down-sampled image of Mj.

3.2. Improved PCNN Model

To make the PCNN model more suitable for HS sharpening, we propose an IPCNN model in the paper. In the standard PCNN model, the feeding input F stands for the accumulative influence of the external stimuli. Nonetheless, the human eyes are more sensitive to the edges and orientations rather than individual pixels within an image. Therefore, a new feeding input SF is designed in the IPCNN model, which considers the local neighborhood differences in both the horizontal and vertical directions. The SF is calculated in Equations (9)–(11).
S F i j = R F i j 2 + C F i j 2
R F i j = 1 M 1 × M 2 i = 1 M 1 j = 2 M 2 [ F i , j F i , j 1 ] 2
C F i j = 1 M 1 × M 2 i = 2 M 1 j = 1 M 2 [ F i , j F i 1 , j ] 2
where RFij denotes the row difference of the two neighboring feeding inputs F via the local rectangular window M1 × M2, and CFij indicates the column difference. After performing a number of experiments, we take the window size to be 5 × 5. Furthermore, Iij′ is designed to strengthen the impact of the local neighborhood, which is given by Equation (12).
I i j = 1 2 H i j + 1 M 1 × M 2 i = 1 M 1 j = 1 M 2 H i j 2
where Hij refers to the pixel value of the upsampled HS in row i and column j.
After modeling the new SF and I′, the proposed IPCNN neuron model is illustrated in Figure 5. The IPCNN model has the following advantages compared to the standard PCNN: (1) the IPCNN model utilizes the spatial difference of the neighborhoods to stimulate the feeding input Fij of the neuron, which can describe the local detail features; (2) the external stimuli Iij considers the influence of the surrounding pixels towards the central pixels.

3.3. Automatic Parameter Optimization of IPCNN by CSA

Similar to the standard PCNN, the setting of appropriate parameters (i.e., αF, αL, αE, β) of the IPCNN model is most critical. As shown in Figure 6, different combinations of IPCNN parameters will lead to different segmentation results for the same input image. If there are too many segmentation pieces, the segmentation regions will be too small and will further affect the accuracy and complexity of the statistical computation of subsequent injection weights. Conversely, if the number of image segmentation pieces is too few, it is impossible to distinguish each region or to take advantage of the characteristics of each region. So, different parameters can lead to different segmentation results, and different segmentations will have corresponding impacts on the final fusion results. Most researchers choose to simplify the PCNN model or to use the manually set uniform parameters for the different images. However, fixed parameters for different input images do not lead to the optimal fusion results for all images. Thus, we propose a CSA-based optimization approach for setting IPCNN parameters αF, αL, αE, β, and W, which can adaptively generate their own optimal parameters for various input images.
The IPCNN parameters need to be jointly optimized together, so the CSA is employed to automatically optimize all five IPCNN parameters. For convenience, the connection weight W is denoted as W = [w, 1, w; 1, 0, 1; w, 1, w], where w is optimized rather than the entire matrix W. The flowchart of the automatic IPCNN parameter optimization algorithm based on CSA is shown in Figure 7. Firstly, the chameleon position is initialized with the classical parameter values. Secondly, the fitness function is set as the weighted summation of spectral fidelity SAM [5] and spatial detail representation ERGAS [40]. The fitness function for the proposed optimization method is given as follows.
f = S A M r E R G A S r + S A M r E R G A S + E R G A S r E R G A S r + S A M r S A M
E R G A S F H , F = 100 h l 1 N h h = 1 N h R M S E h μ h 2
S A M F H , F = arccos f h j , f j f h j 2 f j 2
where FH denotes the reference imagery, and F refers to the fusion imagery. SAMr and ERGASr stand for the ranges of SAM and ERGAS. h and l indicate the spatial resolutions of the MS and HS imagery, respectively. uh indicates the mean value of the hth band of the reference imagery. <·> means the inner product. ‖·‖2 refers to the second-order norm operation.
Since there are too many bands in the HS image, directly optimizing the parameters for all bands will be time-consuming. Now that the adjacent HS bands acquired by the SAM-CC band assignment method are usually strongly correlated, it would be more efficient to obtain optimized parameters by only using the representative bands of each group. Wang et al. proposed a HS band selection method with optimal domain reconstruction [41], which can adaptively achieve the optimal neighborhood reconstruction (ONR) to find the subsets of bands that can best represent the HS image. Thus, the ONR band selection method is used before IPCNN parameter optimization. In addition, the method can also reduce the impacts of noisy bands by taking advantage of the characteristics of neighborhood bands.
To be more efficient, the optimal IPCNN parameters are generated with the representative bands of HS images. Other HS bands could utilize these parameters as long as both bands are very similar to each other. Representative bands from each group are chosen by the ONR method. If there are more than five HS bands in each group, five of these are then selected using ONR to compose the optimal subset of bands. Otherwise, all of the bands from the original HS within the group are retained. Equation (16) is defined to express the procedure.
H φ i = H j j φ i ONR band selection H ϕ i = H j j j j ϕ i     j > 5 , j j = 5 H j j φ i       j 5

3.4. Extracting MS Details

The multiscale analysis has demonstrated excellent capability in RS image fusion. The “atrous” wavelet transform is an undecimated multiscale analysis method, which is widely used to address image fusion issues due to its fast decomposition and reconstruction as well as the richer extracted details. For each histogram-matched MS band Mhm (matched by HS image), the “atrous” wavelet decomposition is performed to extract the spatial details from the multispectral imagery. After wavelet decomposition, the low-frequency component stands for the approximate image, and the high-frequency component indicates the noise and the local features. After setting the high-frequency components to zero, the low-frequency MS imagery ML is then realized via wavelet reconstruction. Finally, the spatial detail information Md is obtained through Equation (17).
M d = M h m M L

3.5. Adaptive Injected Gains

The HS imagery is divided into different irregular segmentation regions by the proposed CSA-based IPCNN segmentation algorithm. The injected gains are, therefore, calculated in each irregular region. The overall steps of the adaptive gain approach are described as follows:
(1)
Initialize IPCNN. Let VF = 0.5, VL = 0.2, VE = 20, Y [0] = L [0] = U [0] = 0, E [0] = VE.
(2)
Optimize IPCNN parameters αF, αL, αE, β, and W using the CSA-based IPCNN optimization algorithm.
(3)
Obtain the irregular segmentation region of the IPCNN model in the current iteration n. And calculate the injected gain Gk[n] according to Equation (20).
R i j [ n ] = cov H k u i , j , M L i , j cov M L i , j , M L i , j i f   Y i j n 0 0   o t h e r w i s e
g i j k = c o r r H k u i , j , M L i , j s t d ( H h ) s t d ( M m )
G i j k [ n ] = g i j k s t d H k u i , j s t d M L i , j i f   R i j [ n ] > 0   a n d   Y i j [ n ] 0 1 o t h e r w i s e
where (i, j) stands for the pixel coordinates of the imagery. Hku denotes the upsampled image of HS, and ML indicates the low-resolution version of the MS imagery. Y[n] refers to the activated neurons in the current iteration. R is the correlation ratio between the HS and MS imagery. corr(A, B) represents the correlation coefficient between matrix A and matrix B. cov(A, B) denotes the covariance between matrix A and matrix B. std(A) refers to the standard deviation of A.

3.6. Fusion Output

The final fusion result can be calculated by Equation (21).
F H k = H k u + G k M d
where FHk denotes the high-resolution HS fusion image.
The fusion pseudo-code of the proposed HS sharpening algorithm (Algorithm 1) is described as follows.
Algorithm 1 AT-AIPCNN (“atrous” transform-adaptive IPCNN) method
Remotesensing 15 04205 i001

4. Experimental Results

4.1. Datasets

Three real datasets captured from the ZY1-02D and GF-2 satellites are utilized to test the validity of the proposed method. The HS sensor AHSI of the ZY1-02D satellite, which works in the spectral range from the visible to the short-wave infrared (with the wavelength from 395 nm to 2500 nm), provides 166 HS bands with a spatial resolution of 30 m. The GF-2 satellite captures four multispectral bands with a spatial resolution of 3.4 m, spanning the spectral range from the visible to the near-infrared (450–890 nm). For HS sharpening application, HS images from the ZY1-02D sensor would need to be sharpened to the spatial resolution of MS imagery of the GF-2 sensor. Thus, dataset1 consists of imagery taken over the Liujiaxia reservoir in China from the ZY1-02D and GF-2 sensors, which mainly include the lake and the village. Dataset2 stands for the suburban area of Linxia City in China, mainly containing mountainous terrain. The dataset3 is the farmland area of Yongchang City in China, which primarily includes farmland, mountains, and buildings. The original images of the three datasets are shown in Figure 8.

4.2. Experimental Setup

In the experiments, most program code is executed with matlab2020a on the Intel CPU Core i7-13700K and NVIDIA GPU GeForce GTX 3090. The number of chameleon populations is set to 20 in the CSA-based IPCNN adaptive optimization, and the maximum number of iterations is set to 30.
Nine classical and competitive methods of different sharpening categories are compared with the proposed approach, i.e., the Gram-Schmidt adaptive (GSA) algorithm [12], the smoothing filtered-based intensity modulation (SFIM) algorithm [42], the generalized Laplacian pyramid (GLP) algorithm [43], the CNMF algorithm [26], the nonlocal sparse tensor factorization (NLSTF) algorithm [27], the NLSTF_SMBF algorithm [27], the HYSURE algorithm [44], the fast fusion based on Sylvester equation (FUSE) algorithm [45], and the UDALN algorithm [46]. Among which, GSA, SFIM, and GLP are owned by the classical pansharpening methods, GSA belongs to the CS methods, SFIM and GLP are the MRA method, CNMF belongs to the matrix decomposition methods, NLSTF and NLSTF_SMBF are owned by tensor decomposition categories, while the HYSURE and FUSE belong to Bayesian-based methods. In addition, GSA, SFIM, and GLP do not require any parameter setting. The main parameters for the NLSTF and NLSTF_SMBF methods are set according to Table 1, and the parameters of other comparative approaches are set to be the same as in the literature [47]. The UDALN is implemented in PyTorch, and the parameter settings are the same as in the literature [46].
The performance of the proposed HS sharpening approach is evaluated by eight complementary quantitative indices to verify the spectral and spatial qualities of the fusion imagery, such as the peak signal-to-noise ratio (PSNR) [47], the root mean square error (RMSE) [48], the error relative global adimensionnelle de synthèse (ERGAS) [40], the spectral angle mapper (SAM) [5], the structural similarity index measurement (SSIM) [49,50], the universal image quality index (UIQI) [48], the inter-correlation (CC) [51], and the degree of distortion (DD) [52]. The ideal values for RMSE, ERGAS, SAM, and DD are 0, whereas 1 for SSIM, UIQI, and CC. Moreover, the ideal value for PSNR is positive infinity.
PSNR measures the spatial similarity between the fusion imagery and the reference imagery. In general, the larger the PSNR, the more spatially similar the fusion imagery is to the reference imagery. The PSNR is defined as:
P S N R F H , F = 1 N h h = 1 N h log 10 max F H h 2 1 L j = 1 L F H h j F h j 2
where FH denotes the reference imagery, and F refers to the fusion imagery. h stands for the band number, while Nh indicates the total number of all bands. log10(⸳) represents the logarithm function with base 10. FHh denotes the hth band of the reference imagery. max(⸳) denotes the maximum value. L denotes the total number of pixels per band. FHhj denotes the jth pixel of the hth band of the reference imagery, and Fhj stands for the jth pixel of the hth band of the fusion image.
RMSE refers to the measure of deviation between two images. The RMSE is defined as:
R M S E F H , F = F H , F F 2 T
where T stands for the total pixel number of the reference imagery.
ERGAS is a comprehensive index that reflects both spectral distortions and spatial detail differences. The ERGAS can be computed through Equation (14). SAM calculates the similarity of the spectral vectors. The smaller one indicates that the spectrum is better maintained. The SAM index of two images FH and F is computed as Equation (15). UIQI measures the similarity in terms of the brightness and contrast. For the window, a × b, Q is defined as:
Q a , b = 4 μ a μ b μ a 2 + μ b 2 σ a , b 2 σ a 2 + σ b 2
where μx and σ x 2 denote the expectation and variance of x, respectively, and σ x , y 2 represents the covariance of x and y.
Then, the UIQI can be computed through the following formula:
U I Q I F H , F = 1 N h h = 1 N h Q ( F H h , F h )
SSIM compares the structural similarities between two images, which indicate the luminance distortion, the contrast distortion, and the structural distortion. Its formula is described in Equations (26) and (27).
S S I M F H , F = 1 N h h = 1 N h S S I M S F H h , F h
S S I M S a , b = 2 μ a μ b + C 1 σ a σ b + C 2 μ a 2 + μ b 2 + C 1 σ a 2 + σ b 2 + C 2
where C1 and C2 are infinitesimal small constants to ensure stability.
DD is an indicator to verify the quality of the fusion spectra. The smaller the value, the better the spectral retention. If the DD is equal to 0, there is no spectral distortion. Its calculation formula is shown in Equation (28).
D D F H , F = 1 T v e c F H v e c F 1
where T denotes the total pixel number of the reference imagery. vec(x) represents matrix x as a vector, and ‖ ‖1 is the one-order norm operation.
CC represents the spatial correlation between two images, which measures the geometric distortion of the fusion imagery. It is calculated as follows:
C C F H , F = 1 N h h = 1 N h C C S F H h , F h
C C S a , b = i = 1 H j = 1 W a i j μ a b i j μ b i = 1 H j = 1 W a i j μ a 2 b i j μ b 2
where Xh refers to the hth band of X, and μx indicates the mean value of x.

4.3. Experimental Results

The first experiment is conducted on the Liujiaxia dataset. The fusion results are shown in Figure 9, in which the local area is enlarged for convenient observation. As can be seen from Figure 9, the color of the GSA, SFIM, GLP, CNMF, HYSURE, UDALN, and FUSE methods appear to be dark. Our proposed method, NLSTF and NLSTF_SMBF methods, perform best in spectral preservation; nevertheless, the details of both the NLSTF and NLSTF_SMBF methods are blurred into chunks. Figure 10 presents the SAM error maps of the fusion images in all 166 bands with the Liujiaxia dataset. It is noted that the spectral distortion for all methods primarily occurs at the reservoir-mountain boundary, of which NLSTF and NLSTF_SMBF exhibit the largest aberration. Besides, UDALN has larger distortions in the reservoir area, and the proposed method, SFIM, and GLP perform better in the SAM error maps. Furthermore, quantitative assessments with the Liujiaxia dataset are shown in Table 2. It is clear from Table 2 that the proposed method obtains the best performance across all quantitative metrics, which demonstrates the superiority of the proposed method with respect to spectral preservation and spatial details.
The fusion results with the Linxia dataset are shown in Figure 11. As can be seen, the large color difference and block blur are presented in the NLSTF and NLSTF_SMBF fusion methods. Moreover, the SFIM fusion image appears dark. GSA algorithm, as a kind of CS method, has greater spectral distortion, while MRA methods, such as SFIM and GLP, have less spatial detail. Furthermore, the proposed method, GSA, and UDALN, have fine spatial detail information. In addition, the spectra of the proposed method are the closest to the ground-truth image in subjective visualization. Figure 12 shows the SAM error maps of all the fusion results with the Linxia dataset. As can be seen from Figure 12, spectral distortion is more likely to occur in the ridge area and in the land-lake boundary. Although the SFIM method has less spectral distortion in most regions, it exhibits larger errors in some local areas (top left corner and bottom left of Figure 12d). NLSTF and NLSTF_SMBF appear to have more spectral distortion, whereas the proposed method performs well. Table 3 gives the quantitative evaluation results of the Linxia dataset, which demonstrates that our method achieves the best indicator results.
Figure 13 presents the fusion results of the Yongchang dataset. From Figure 13, we can see that the color error occurred in the fusion result of the CNMF, NLSTF, and NLSTF_SMBF methods. In addition, The SAM error maps of the Yongchang dataset are shown in Figure 14. We note from Figure 14 that the proposed method has the least amount of spectral distortion at the edges, which is also verified by Table 4.

4.4. Ablation Experiments

Three kinds of ablation experiments are carried out to test our three primary fusion modules, i.e., SAM-CC band assignment module, ONR band selection module, and CSA adaptive PCNN parameter module. In addition, different automatic parameter optimization strategies are also investigated.
The fusion results obtained on three different datasets with different band assignment strategies (SAM-CC, SAM, and CC) are presented in Table 5. For the Liujiaxia dataset, our proposed SAM-CC assignment indicator performs best. Furthermore, the SAM-CC achieves better fusion results in most cases with the Linxia dataset and the Yongchang dataset, which illustrates the effectiveness of the SAM-CC band assignment.
To validate the efficiency of the ONR band selection module, we compared the fusion results of using ONR and without ONR. In order to verify the effectiveness of ONR band selection, we fused the data sets with and without ONR band selection. Comparisons of the fusion accuracy and the running time are shown in Table 6. As can be seen from Table 6, the running time has been reduced by more than 84% with the ONR band selection module, while the fusion accuracy remains almost unchanged. Overall, this indicates significant time cost savings for the ONR band selection module.
In order to examine the rationality of the CSA-based parameter optimization strategy in IPCNN, Table 7 shows the comparison of the CSA fusion metrics with other three parameter optimization strategies, i.e., the sparrow search strategy (SSA) [53], the improved grey wolf optimizer (IGWO) strategy [54] and the enhanced whale optimization strategy (EWOA) [55]. As shown in Table 7, the CSA-based fusion algorithm achieves a better fusion performance as well as less time-consuming.
Table 8 exhibits the quantitative analysis of the fusion results with and without adaptive CSA optimization to verify the impact of the CSA algorithm. Besides, the traditional IPCNN parameters are set as classical values, i.e., αF = 0.1, αL = 1, αE =0.62, β = 0.1, w = 0.5. In all three datasets, the adaptive IPCNN approach is significantly superior to the traditional parameters in most cases, which indicates that the adaptive parameter optimization effectively improves the quality of the fusion images.

5. Conclusions

In the paper, a novel algorithm for HS sharpening is proposed. Firstly, the band groups are acquired by performing joint SAM-CC band matching of HS and MS assignment method, the band grouping results are obtained, and then ONR is applied to the HS images of each group to improve the efficiency for subsequent adaption. Besides, according to the characteristics of remote sensing image fusion application, an IPCNN model is proposed, which can obtain irregular injection regions of spatial details. In addition, a CSA-based IPCNN parameter optimization method is designed to achieve the optimal sharpening imagery. In summary, the proposed method is simple and easy to implement, which presents good fusion results on a variety of datasets, including reservoir, mountain, town, and river landscapes. Furthermore, several ablation experiments are also conducted to corroborate the efficiency of the proposed method.

Author Contributions

All the authors have contributed substantially to the manuscript. X.X. and X.L. proposed the methodology. X.X., X.L. and J.G. performed the experiments and software. X.X., X.L. and Y.L. wrote the paper. Y.L. and L.K. analyzed the data. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Natural Science Foundation of China (No. 41861055), China Postdoctoral Science Foundation (No. 2019M653795), and LZJTU EP Program (No. 201806).

Data Availability Statement

Not applicable.

Acknowledgments

The authors are grateful to the editor and anonymous reviewers for their helpful and valuable suggestions. We also want to express our sincere gratitude to Jie Li for her support and help.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kokila, S.; Jayachandran, A. Hybrid Behrens-Fisher- and Gray Contrast–Based Feature Point Selection for Building Detection from Satellite Images. J. Geovisualization Spat. Anal. 2023, 7, 8. [Google Scholar] [CrossRef]
  2. Zheng, Z.; Hu, Y.; Yang, H.; Qiao, Y.; He, Y.; Zhang, Y.; Huang, Y. AFFU-Net: Attention feature fusion U-Net with hybrid loss for winter jujube crack detection. Comput. Electron. Agric. 2022, 198, 107049. [Google Scholar] [CrossRef]
  3. Gadea, O.; Khan, S. Detection of Bastnäsite-Rich Veins in Rare Earth Element Ores through Hyperspectral Imaging. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–4. [Google Scholar] [CrossRef]
  4. Boubanga-Tombet, S.; Huot, A.; Vitins, I.; Heuberger, S.; Veuve, C.; Eisele, A.; Hewson, R.; Guyot, E.; Marcotte, F.; Chamberland, M. Thermal Infrared Hyperspectral Imaging for Mineralogy Mapping of a Mine Face. Remote Sens. 2018, 10, 1518. [Google Scholar] [CrossRef]
  5. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Restaino, R.; Scarpa, G.; Ulfarsson, M.O.; Alparone, L.; Chanussot, J. A New Benchmark Based on Recent Advances in Multispectral Pansharpening: Revisiting Pansharpening with Classical and Emerging Pansharpening Methods. IEEE Geosci. Remote Sens. Mag. 2020, 9, 53–81. [Google Scholar] [CrossRef]
  6. Jiayuan, L.; Ai, M.; Wang, S.; Hu, Q. GRF: Guided Residual Fusion for Pansharpening. Int. J. Remote Sens. 2022, 43, 3609–3627. [Google Scholar]
  7. Wu, Z.; Huang, Y.; Zhang, K. Remote Sensing Image Fusion Method Based on PCA and Curvelet Transform. J. Indian Soc. Remote Sens. 2018, 46, 687–695. [Google Scholar] [CrossRef]
  8. Kong, Y.; Hong, F.; Leung, H.; Peng, X. A Fusion Method of Optical Image and SAR Image Based on Dense-UGAN and Gram–Schmidt Transformation. Remote Sens. 2021, 13, 4274. [Google Scholar] [CrossRef]
  9. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 228–236. [Google Scholar] [CrossRef]
  10. Vivone, G. Robust Band-Dependent Spatial-Detail Approaches for Panchromatic Sharpening. IEEE Trans. Geosci. Remote Sens. Publ. IEEE Geosci. Remote Sens. Soc. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  11. Yang, Y.; Wan, C.; Huang, S.; Lu, H.; Wan, W. Pansharpening Based on Adaptive High-Frequency Fusion and Injection Coefficients Optimization. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 799–811. [Google Scholar] [CrossRef]
  12. Dian, R.; Li, S.; Sun, B.; Guo, A. Recent advances and new guidelines on hyperspectral and multispectral image fusion. Inf. Fusion 2021, 69, 40–51. [Google Scholar] [CrossRef]
  13. Bouslihim, Y.; Kharrou, M.H.; Miftah, A.; Attou, T.; Bouchaou, L.; Chehbouni, A. Comparing Pan-sharpened Landsat-9 and Sentinel-2 for Land-Use Classification Using Machine Learning Classifiers. J. Geovisualization Spat. Anal. 2022, 6, 35. [Google Scholar] [CrossRef]
  14. Burt, P.; Adelson, E. The Laplacian Pyramid as a Compact Image Code. Read. Comput. Vis. 1987, 31, 671–679. [Google Scholar]
  15. Gao, W.; Xiao, Z.; Bao, T. Detection and Identification of Potato-Typical Diseases Based on Multidimensional Fusion Atrous-CNN and Hyperspectral Data. Appl. Sci. 2023, 13, 5023. [Google Scholar] [CrossRef]
  16. Jindal, H.; Bharti, M.; Kasana, S.; Saxena, S. An ensemble mosaicing and ridgelet based fusion technique for underwater panoramic image reconstruction and its refinement. Multimed. Tools Appl. Available online: https://link.springer.com/article/10.1007/s11042-023-14594-9 (accessed on 8 March 2023). [CrossRef]
  17. Du, C.; Gao, S. Remote sensing image fusion based on nonlinear IHS and fast nonsubsampled contourlet transform. J. Indian Soc. Remote Sens. 2018, 46, 2023–2032. [Google Scholar] [CrossRef]
  18. Cheng, S.; Qiguang, M.; Pengfei, X. A novel algorithm of remote sensing image fusion based on Shearlets and PCNN. Neurocomputing 2013, 117, 47–53. [Google Scholar] [CrossRef]
  19. Restaino, R.; Vivone, G.; Dalla Mura, M.; Chanussot, J. Fusion of multispectral and panchromatic images based on morphological operators. IEEE Trans. Image Process. 2016, 25, 2882–2895. [Google Scholar] [CrossRef] [PubMed]
  20. Ren, C.; Liang, Y.; Lu, X.; Yan, H. Research on the soil moisture sliding estimation method using the LS-SVM based on multi-satellite fusion. Int. J. Remote Sens. 2019, 40, 2104–2119. [Google Scholar] [CrossRef]
  21. Vivone, G. Multispectral and hyperspectral image fusion in remote sensing: A survey. Inf. Fusion 2023, 89, 405–417. [Google Scholar] [CrossRef]
  22. Gomez, R.; Jazaeri, A.; Kafatos, M. Wavelet-based hyperspectral and multispectral image fusion. Proc. SPIE-Int. Soc. Opt. Eng. 2001, 4383, 36–42. [Google Scholar]
  23. Chen, Z.; Pu, H.; Wang, B.; Jiang, G.-M. Fusion of Hyperspectral and Multispectral Images: A Novel Framework Based on Generalization of Pan-Sharpening Methods. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1418–1422. [Google Scholar] [CrossRef]
  24. Picone, D.; Restaino, R.; Vivone, G.; Addesso, P.; Dalla Mura, M.; Chanussot, J. Band assignment approaches for hyperspectral sharpening. IEEE Geosci. Remote Sens. Lett. 2017, 14, 739–743. [Google Scholar] [CrossRef]
  25. Lu, X.; Zhang, J.; Yu, X.; Tang, W.; Li, T.; Zhang, Y. Hyper-sharpening based on spectral modulation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1534–1548. [Google Scholar] [CrossRef]
  26. Yokoya, N.; Yairi, T.; Iwasaki, A. Coupled Nonnegative Matrix Factorization Unmixing for Hyperspectral and Multispectral Data Fusion. IEEE Trans. Geosci. Remote Sens. 2012, 50, 528–537. [Google Scholar] [CrossRef]
  27. Dian, R.; Li, S.; Fang, L.; Lu, T.; Bioucas-Dias, J.M. Nonlocal Sparse Tensor Factorization for Semiblind Hyperspectral and Multispectral Images Fusion. IEEE Trans. Cybern. 2020, 50, 4469–4480. [Google Scholar] [CrossRef]
  28. Li, J.; Hong, D.; Gao, L.; Yao, J.; Zheng, K.; Zhang, B.; Chanussot, J. Deep learning in multimodal remote sensing data fusion: A comprehensive review. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102926. [Google Scholar] [CrossRef]
  29. Liu, J.; Yuan, Z.; Pan, Z.; Fu, Y.; Liu, L.; Lu, B. Diffusion Model with Detail Complement for Super-Resolution of Remote Sensing. Remote Sens. 2022, 14, 4834. [Google Scholar] [CrossRef]
  30. Zhang, L.; Nie, J.; Wei, W.; Li, Y.; Zhang, Y. Deep blind hyperspectral image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 2388–2400. [Google Scholar] [CrossRef]
  31. Qu, Y.; Qi, H.; Kwan, C.; Yokoya, N.; Chanussot, J. Unsupervised and unregistered hyperspectral image super-resolution with mutual Dirichlet-Net. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  32. Li, X.; Yan, H.; Xie, W.; Kang, L.; Tian, Y. An Improved Pulse-Coupled Neural Network Model for Pansharpening. Sensors 2020, 20, 2764. [Google Scholar] [CrossRef] [PubMed]
  33. Zhang, L.; Zeng, G.; Wei, J.; Xuan, Z. Multi-modality image fusion in adaptive-parameters SPCNN based on inherent characteristics of image. IEEE Sens. J. 2020, 20, 11820–11827. [Google Scholar] [CrossRef]
  34. Bhagyashree, V.; Manisha, D.; Mohammad, F.; Avinash, G.; Deep, G. Saliency Detection Using a Bio-inspired Spiking Neural Network Driven by Local and Global Saliency. Appl. Artif. Intell. 2022, 36, 2900–2928. [Google Scholar]
  35. Huang, C.; Tian, G.; Lan, Y.; Peng, Y.; Ng, E.; Hao, Y.; Che, W. A new pulse coupled neural network (PCNN) for brain medical image fusion empowered by shuffled frog leaping algorithm. Front. Neurosci. 2019, 13, 210. [Google Scholar] [CrossRef]
  36. Panigrahy, C.; Seal, A.; Mahato, N. Fractal dimension based parameter adaptive dual channel PCNN for multi-focus image fusion. Opt. Lasers Eng. 2020, 133, 106141. [Google Scholar] [CrossRef]
  37. Panigrahy, C.; Seal, A.; Gonzalo-Martín, C.; Pathak, P.; Jalal, A. Parameter adaptive unit-linking pulse coupled neural network based MRI-PET/SPECT image fusion. Biomed. Signal Process. Control 2023, 83, 104659. [Google Scholar] [CrossRef]
  38. Johnson, J.; Padgett, M. PCNN models and applications. IEEE Trans. Neural Netw. 1999, 10, 480–498. [Google Scholar] [CrossRef]
  39. Braik, M. Chameleon Swarm Algorithm: A Bio-inspired Optimizer for Solving Engineering Design Problems. Expert Syst. Appl. 2021, 174, 114685. [Google Scholar] [CrossRef]
  40. Ren, K.; Sun, W.; Meng, X.; Yang, G.; Du, Q. Fusing China GF-5 Hyperspectral Data with GF-1, GF-2 and Sentinel-2A Multispectral Data: Which Methods Should Be Used? Remote Sens. 2020, 12, 882. [Google Scholar] [CrossRef]
  41. Wang, Q.; Zhang, F.; Li, X. Hyperspectral band selection via optimal neighborhood reconstruction. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8465–8476. [Google Scholar] [CrossRef]
  42. Liu, J. Smoothing Filter-based Intensity Modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  43. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored Multiscale Fusion of High-resolution MS and Pan Imagery. Photogramm. Eng. Remote Sens. 2015, 72, 591–596. [Google Scholar] [CrossRef]
  44. Simões, M.; Bioucas-Dias, J.; Almeida, L.B.; Chanussot, J. A Convex Formulation for Hyperspectral Image Superresolution via Subspace-Based Regularization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3373–3388. [Google Scholar] [CrossRef]
  45. Wei, Q.; Bioucas-Dias, J.; Dobigeon, N.; Tourneret, J.-Y. Hyperspectral and Multispectral Image Fusion Based on a Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 3658–3668. [Google Scholar] [CrossRef]
  46. Li, J.; Zheng, K.; Yao, J.; Gao, L.; Hong, D. Deep Unsupervised Blind Hyperspectral and Multispectral Data Fusion. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6007305. [Google Scholar] [CrossRef]
  47. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and Multispectral Data Fusion: A comparative review of the recent literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  48. Sara, D.; Mandava, A.K.; Kumar, A.; Duela, S.; Jude, A. Hyperspectral and multispectral image fusion techniques for high resolution applications: A review. Earth Sci. Inform. 2021, 14, 1685–1705. [Google Scholar] [CrossRef]
  49. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  50. Brunet, D.; Vrscay, E.R.; Wang, Z. On the mathematical properties of the structural similarity index. IEEE Trans. Image Process. 2011, 21, 1488–1499. [Google Scholar] [CrossRef]
  51. Tian, X.; Li, K.; Zhang, W.; Wang, Z.; Ma, J. Interpretable Model-Driven Deep Network for Hyperspectral, Multispectral, and Panchromatic Image Fusion. IEEE Trans. Neural Netw. Learn. Syst. Available online: https://ieeexplore.ieee.org/document/10138912 (accessed on 31 May 2023). [CrossRef]
  52. Li, S.; Dian, R.; Fang, L.; Bioucas-Dias, J. Fusing Hyperspectral and Multispectral Images via Coupled Sparse Tensor Factorization. IEEE Trans. Image Process. 2018, 27, 4118–4130. [Google Scholar] [CrossRef] [PubMed]
  53. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  54. Nadimi-Shahraki, M.; Taghian, S.; Mirjalili, S. An improved grey wolf optimizer for solving engineering problems. Expert Syst. Appl. 2020, 166, 113917. [Google Scholar] [CrossRef]
  55. Nadimi-Shahraki, M.; Zamani, H.; Mirjalili, S. Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study. Comput. Biol. Med. 2022, 148, 105858. [Google Scholar] [CrossRef]
Figure 1. Connection relationship diagram of PCNN.
Figure 1. Connection relationship diagram of PCNN.
Remotesensing 15 04205 g001
Figure 2. Flowchart of CSA. See reference [39] for all variable definitions.
Figure 2. Flowchart of CSA. See reference [39] for all variable definitions.
Remotesensing 15 04205 g002
Figure 3. Flowchart of the proposed fusion algorithm.
Figure 3. Flowchart of the proposed fusion algorithm.
Remotesensing 15 04205 g003
Figure 4. Schematic diagram of HS and MS bands grouping.
Figure 4. Schematic diagram of HS and MS bands grouping.
Remotesensing 15 04205 g004
Figure 5. IPCNN neuron model.
Figure 5. IPCNN neuron model.
Remotesensing 15 04205 g005
Figure 6. IPCNN segmentation results under pseudo-color maps with different parameters of IPCNN. (a) Original image of GF-2 satellite sensor. (b) Classical parameters combination, [αF, αL, αE, β, w] = [0.1, 1, 0.62, 0.1, 0.707]. (c) Change the parameters αE and w, [αF, αL, αE, β, w] = [0.1, 1, 1, 0.1, 0.9]. (d) Change all parameters, [αF, αL, αE, β, w] = [0.5, 2.5, 2, 0.3, 0.9].
Figure 6. IPCNN segmentation results under pseudo-color maps with different parameters of IPCNN. (a) Original image of GF-2 satellite sensor. (b) Classical parameters combination, [αF, αL, αE, β, w] = [0.1, 1, 0.62, 0.1, 0.707]. (c) Change the parameters αE and w, [αF, αL, αE, β, w] = [0.1, 1, 1, 0.1, 0.9]. (d) Change all parameters, [αF, αL, αE, β, w] = [0.5, 2.5, 2, 0.3, 0.9].
Remotesensing 15 04205 g006
Figure 7. Flowchart of CSA-based automatic IPCNN parameter optimization algorithm.
Figure 7. Flowchart of CSA-based automatic IPCNN parameter optimization algorithm.
Remotesensing 15 04205 g007
Figure 8. The original images of three datasets. (a) HS image of Liujiaxia dataset. (b) HS image of Linxia dataset. (c) HS image of Yongchang dataset. (d) MS image of Liujiaxia dataset. (e) MS image of Linxia dataset. (f) MS image of Yongchang dataset.
Figure 8. The original images of three datasets. (a) HS image of Liujiaxia dataset. (b) HS image of Linxia dataset. (c) HS image of Yongchang dataset. (d) MS image of Liujiaxia dataset. (e) MS image of Linxia dataset. (f) MS image of Yongchang dataset.
Remotesensing 15 04205 g008
Figure 9. Color composite image of fusion results with Liujiaxia dataset (bands 35,14,7 are input to RGB channels respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 9. Color composite image of fusion results with Liujiaxia dataset (bands 35,14,7 are input to RGB channels respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g009
Figure 10. SMA error map of Liujiaxia dataset. (a) Proposed. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 10. SMA error map of Liujiaxia dataset. (a) Proposed. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g010
Figure 11. Color composite image of fusion results with Linxia dataset (bands 35,14,7 are input to RGB channel respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 11. Color composite image of fusion results with Linxia dataset (bands 35,14,7 are input to RGB channel respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g011
Figure 12. SMA error map of Linxia dataset. (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 12. SMA error map of Linxia dataset. (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g012
Figure 13. Color composite image of fusion results of with Yongchang dataset (bands 29,19,9 are input to RGB channel respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 13. Color composite image of fusion results of with Yongchang dataset (bands 29,19,9 are input to RGB channel respectively). (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g013
Figure 14. SMA error map of Yongchang dataset. (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Figure 14. SMA error map of Yongchang dataset. (a) Proposed method. (b) GSA. (c) SFIM. (d) GLP. (e) CNMF. (f) NLSTF. (g) NLSTF_SMBF. (h) HYSURE. (i) FUSE. (j) UDALN.
Remotesensing 15 04205 g014
Table 1. Main parameter settings of NLSTF/NLSTF_SMBF methods.
Table 1. Main parameter settings of NLSTF/NLSTF_SMBF methods.
MethodMain Parameters
NLSTF/NLSTF_SMBFThe atomic numbers for three different dictionaries:
lW = 10, lH = 10, lS = 14.
Parameters of sparse regularization:
λ = 10−6, λ1 = 10−5, λ2 = 10−5, λ3 = 10−6.
Cluster scaling parameter: K = 151.
The spectral response matrix R was estimated by HYSURE [44].
Table 2. Quantitative evaluation results of the Liujiaxia dataset.
Table 2. Quantitative evaluation results of the Liujiaxia dataset.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCC
Proposed37.66775.05601.29711.59610.72520.92962.37480.9863
GSA27.624817.23375.11952.13570.55350.887511.60320.9767
SFIM25.872421.23436.88821.78520.59020.861314.60260.9768
GLP25.614621.88127.20561.75730.59200.858815.05080.9777
CNMF26.756418.91975.87962.17200.46190.870712.86210.9653
NLSTF26.725318.39084.48325.32020.35530.739311.59770.8935
NLSTF_SMBF25.452419.59917.44069.69970.26210.683312.13360.8094
HYSURE27.951216.59094.84752.27590.47250.886211.20040.9685
FUSE23.145829.170011.55443.76710.44660.774620.14690.9606
UDALN30.194813.09023.13106.75270.46110.89178.44850.9555
The best result is in bold, and the second best is underlined.
Table 3. Quantitative evaluation results of the Linxia dataset.
Table 3. Quantitative evaluation results of the Linxia dataset.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCC
Proposed36.23315.72540.99601.96340.84980.90673.46080.9645
GSA27.716416.00833.03272.65270.81580.871612.47410.9393
SFIM17.720152.065922.73152.56780.36560.538742.04730.9452
GLP29.109113.57212.49552.37620.82560.882610.58570.9526
CNMF29.445312.48782.33892.82210.78540.87238.94180.8916
NLSTF26.787316.46223.53014.55790.67110.782612.64800.8645
NLSTF_SMBF22.445427.29269.938313.55310.48170.676320.54410.7075
HYSURE29.113913.48092.47102.54690.80840.866510.37410.9299
FUSE27.710316.05563.04282.81510.79590.865712.58060.9403
UDALN28.325013.56252.55552.78750.75350.870910.33360.8834
The best result is in bold, and the second best is underlined.
Table 4. Quantitative evaluation results of the Yongchang farmland area data set.
Table 4. Quantitative evaluation results of the Yongchang farmland area data set.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCC
Proposed31.881611.08281.87164.45850.49790.64177.16030.7057
GSA30.694812.95072.45695.14390.47440.63858.66610.6708
SFIM30.454013.08302.52905.46650.47850.62438.58270.6538
GLP31.469711.75462.21685.32640.48850.62847.24070.6606
CNMF26.215321.56734.45036.32500.28440.558215.79050.4523
NLSTF19.022651.342423.863512.80570.14180.385041.46620.3664
NLSTF_SMBF19.561548.562121.165210.66300.14340.445138.92300.3652
HYSURE26.124722.04034.68975.43680.40610.595316.45660.5652
FUSE23.800729.03027.09057.25280.33470.534422.48660.5202
UDALN27.9588 17.2212 2.9768 6.1002 0.3413 0.6091 12.1402 0.4833
The best result is in bold, and the second best is underlined.
Table 5. SAM-CC ablation.
Table 5. SAM-CC ablation.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCC
SAM-CC (dataset1)37.66775.05601.29711.59610.72520.92962.37480.9863
SAM (dataset1)37.17265.29191.41421.72930.71390.92152.51950.9847
CC (dataset1)37.26045.31981.33801.65970.71130.92372.53110.9852
SAM-CC (dataset2)36.23315.72540.99601.96340.84980.90673.46080.9645
SAM (dataset2)36.30035.69170.99881.97210.84930.90593.43700.9645
CC (dataset2)35.09486.47391.15042.00090.83830.90174.22710.9634
SAM-CC (dataset3)31.881611.08281.87164.45850.49790.64177.16030.7057
SAM (dataset3)28.432616.82802.27484.10730.56380.716812.67650.7755
CC (dataset3)31.205612.20601.91224.86040.50260.69678.17910.7334
The best result is in bold, and the second best is underlined.
Table 6. ONR band selection for ablation.
Table 6. ONR band selection for ablation.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCCTime
(s)
Using ONR (dataset1)37.66775.05601.29711.59610.72520.92962.37480.9863221.1
Without ONR (dataset1)37.37465.20961.31661.61990.71430.92722.52020.98591937.6
Using ONR (dataset2)36.23315.72540.99601.96340.84980.90673.46080.9645132.1
Without ONR (dataset2)36.34675.63660.99181.90850.85330.90823.39770.96461794.2
Using ONR (dataset3)31.881611.08281.87164.45850.49790.64177.16030.7057179.2
Without ONR (dataset3)30.986012.28641.87644.12960.55320.70818.63890.76651219.2
The best result is in bold, and the second best is underlined.
Table 7. Automatic parameter optimization ablation.
Table 7. Automatic parameter optimization ablation.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCCTime
(s)
CSA (dataset1)37.66775.05601.29711.59610.72520.92962.37480.9863221.1
SSA (dataset1)37.01125.42241.42001.63150.69990.92892.81510.9860267.9
IGWO (dataset1)37.17605.32241.39471.62530.70510.92912.71330.9861394.6
EWOA (dataset1)37.30935.25041.37301.63610.70990.92912.62810.9861162.6
CSA (dataset2)36.23315.72540.99601.96340.84980.90673.46080.9645132.1
SSA (dataset2)36.27385.71561.00861.99550.84660.90563.45950.9643214.3
IGWO (dataset2)36.23485.71391.00241.97430.84780.90593.45370.9640246.4
EWOA (dataset2)36.23045.74581.01282.00580.84730.90563.49520.9644122.1
CSA (dataset3)31.881611.08281.87164.45850.49790.64177.16030.7057179.2
SSA (dataset3)30.193213.56011.98494.13810.55540.70709.75180.7679199.2
IGWO (dataset3)30.759912.62621.90814.12420.55390.70808.94010.7672280.4
EWOA (dataset3)30.399713.21741.96434.16520.55250.70439.44560.7671137.6
The best result is in bold, and the second best is underlined.
Table 8. CSA adaptive IPCNN parameter ablation.
Table 8. CSA adaptive IPCNN parameter ablation.
MethodPSNR
(dB)
RMSEERGASSAM
(°)
UIQISSIMDDCC
Adaptive (dataset1)37.66775.05601.29711.59610.72520.92962.37480.9863
Traditional (dataset1)29.471013.80643.84691.69110.64800.90209.11790.9818
Adaptive (dataset2)36.23315.72540.99601.96340.84980.90673.46080.9645
Traditional (dataset2)34.10577.26301.29451.99780.83460.90435.00360.9629
Adaptive (dataset3)31.881611.08281.87164.45850.49790.64177.16030.7057
Traditional (dataset3)29.748514.32522.07094.21380.54810.691610.34730.7591
The best result is in bold, and the second best is underlined.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, X.; Li, X.; Li, Y.; Kang, L.; Ge, J. A Novel Adaptively Optimized PCNN Model for Hyperspectral Image Sharpening. Remote Sens. 2023, 15, 4205. https://doi.org/10.3390/rs15174205

AMA Style

Xu X, Li X, Li Y, Kang L, Ge J. A Novel Adaptively Optimized PCNN Model for Hyperspectral Image Sharpening. Remote Sensing. 2023; 15(17):4205. https://doi.org/10.3390/rs15174205

Chicago/Turabian Style

Xu, Xinyu, Xiaojun Li, Yikun Li, Lu Kang, and Junfei Ge. 2023. "A Novel Adaptively Optimized PCNN Model for Hyperspectral Image Sharpening" Remote Sensing 15, no. 17: 4205. https://doi.org/10.3390/rs15174205

APA Style

Xu, X., Li, X., Li, Y., Kang, L., & Ge, J. (2023). A Novel Adaptively Optimized PCNN Model for Hyperspectral Image Sharpening. Remote Sensing, 15(17), 4205. https://doi.org/10.3390/rs15174205

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop