Next Article in Journal
An Extraction Method for Large Gradient Three-Dimensional Displacements of Mining Areas Using Single-Track InSAR, Boltzmann Function, and Subsidence Characteristics
Next Article in Special Issue
ADD-UNet: An Adjacent Dual-Decoder UNet for SAR-to-Optical Translation
Previous Article in Journal
Tree Species Classification in UAV Remote Sensing Images Based on Super-Resolution Reconstruction and Deep Learning
Previous Article in Special Issue
Hyperspectral Denoising Using Asymmetric Noise Modeling Deep Image Prior
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Generalized IHS Based on Total Variation for Pansharpening

1
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430072, China
2
Institute of Aerospace Science and Technology, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(11), 2945; https://doi.org/10.3390/rs15112945
Submission received: 27 April 2023 / Revised: 25 May 2023 / Accepted: 30 May 2023 / Published: 5 June 2023
(This article belongs to the Special Issue Machine Vision and Advanced Image Processing in Remote Sensing II)

Abstract

:
Pansharpening refers to the fusion of a panchromatic (PAN) and a multispectral (MS) image aimed at generating a high-quality outcome over the same area. This particular image fusion problem has been widely studied, but until recently, it has been challenging to balance the spatial and spectral fidelity in fused images. The spectral distortion is widespread in the component substitution-based approaches due to the variation in the intensity distribution of spatial components. We lightened the idea using the total variation optimization to improve upon a novel GIHS-TV framework for pansharpening. The framework drew the high spatial fidelity from the GIHS scheme and implemented it with a simpler variational expression. An improved L1-TV constraint to the new spatial–spectral information was introduced to the GIHS-TV framework, along with its fast implementation. The objective function was solved by the Iteratively Reweighted Norm (IRN) method. The experimental results on the “PAirMax” dataset clearly indicated that GIHS-TV could effectively reduce the spectral distortion in the process of component substitution. Our method has achieved excellent results in visual effects and evaluation metrics.

Graphical Abstract

1. Introduction

Some satellites can obtain multispectral (MS) and panchromatic (PAN) images simultaneously, such as WorldView-2, 3, 4 and GeoEye-1, benefitting the alignment of the two images. MS images contain a wealth of spectral information whose wavelength is from the visible to the near-infrared. MS images usually perform with a relatively poor spatial resolution, limiting their application scenarios. A PAN image can provide a powerful complement to spatial detail. Pansharpening refers to the fusion of MS images and PAN images in the same region, aiming to produce an image with rich spatial and spectral information. Pansharpening has been widely used in many fields, such as environmental monitoring, agriculture, forestry and geological survey.
The pansharpening problem has been widely studied for over three decades. There are still some challenges in the filed of pansharpening. The fusion task requires consideration of alignment, noise amplification due to upsampling, information loss by downsampling and how information is chosen between source images. The inconsistent spectral response between the panchromatic and multispectral sensors is prone to spatial and spectral structure distortion. Suboptimal image registration between the MS and PAN images will lead to problems such as false edges or artifacts. In addition, increasing the practicality of algorithms is a demanding issue in remote sensing image spatial–spectral fusion. Thus, it is a difficult task for pansharpening to balance spatial and spectral fidelity due to these factors. Figure 1 illustrates that spatial fidelity and spectral fidelity usually cannot be achieved at the same time. The increase in spatial detail tends to destroy the spectral fidelity, as in the IHS [1] and the GS [2] methods. Conversely, the fused images of the EXP [3] and BDSD [4] methods are too blurred. From a mathematical point of view, MS and PAN images are incomplete complementary observations of ideal images, and image fusion is the reconstruction of ideal images. Therefore, the focus of pansharpening is to extract geometric information from the PAN image to add this into the fused images and to leave the spectral information unchanged as much as possible.
Approaches of pansharpening can be classified into four categories: component substitution-based (CS-based), multi-resolution analysis-based (MRA-based), variational optimization-based and machine learning-based [5]. The main idea of the CS-based methods is that projection transforms the MS image into a separated spatial component and then enhances it with the PAN image. It is worth mentioning that Tu [6] made a mathematical derivation to simplify the process of component substitution and established a detail injection scheme called generalized IHS (GIHS), where the forward and backward transformations are omitted. The CS approach has high fidelity for spatial details and robustness for spatial misregistration [7]. Based on multiscale decomposition methods, the MRA approaches directly act on the spatial domain [8,9,10]. The basic idea of the MRA-based class methods is to keep the whole content of the LRMS images and add further information from the PAN image. The difference between the PAN image and its low-pass version is regarded as the complement of the spatial details to the fused images [8]. Multiscale decomposition methods such as pyramid transform and wavelet transform are widely used in pansharpening [10]. The process of the MRA-based methods is generally divided into three steps: decomposing the source image into sub-images at different levels with pyramid or wavelet transform, then fusing the sub-images of MS and PAN images at each level and finally performing inverse transform and reconstructing to obtain the fused image [11]. The iterative decomposition scheme is the distinguishing feature from other classes of methods [12]. The MRA-based methods can maintain high spectral quality, and the research hotspots are utilizing different decomposition schemes and optimizing the injection coefficients [5]. The variational optimization-based methods hold the common idea of considering PAN and MS images as coarse measurements of high-spatial MS images, which can be estimated based on the regularized solution of this ill-inverse problem [13]. The variational optimization-based techniques perform well in both spectral and spatial fidelity, with the drawback that they take more optimization iterations than the CS and MRA methods [14]. The last category is machine learning-based methods, where compressed perception-based [15] and dictionary-based [16] versions are the early algorithms, and CNN [17,18,19,20,21,22,23,24,25,26] and GAN [27,28,29,30,31] in deep learning are moving into the field of pansharpening with promising achievements. The pansharpening network (PNN) [32] has achieved encouraging results and a following from researchers. Most super-resolution algorithms cannot be used directly for pansharpening because they cannot fully utilize the spatial information of PAN images. The requirement of CNN for the number of training samples is also an essential limitation of deep learning for pansharpening [5,17,22].
Among these methods, CS-based methods have the best spatial fidelity and worst spectral fidelity. The spatial component synthesis in the earlier CS methods is simpler and uses less spectral information, laying the groundwork for their spectral distortion. BDSD [4] and PRACS [33] use downsampled PAN for spatial component synthesis, which explains the poor spatial fidelity in these improved CS methods. These methods cannot maintain spectral and spatial fidelity simultaneously. The important advantages of CS methods are their high spatial fidelity and good tolerance for misregistration. In the component substitution framework, histogram matching and fitting is the most common way to reduce spectral distortion, but these weak constraints lead to worse results. The new spatial component ( I new ) could be regarded as the function of the original spatial component ( I 0 ) and PAN image. Some proper constraints could be imposed to construct the spatial component ( I new ) by minimizing the functional of the objective function. In this paper, we proposed an optimization framework that takes an optimization perspective to build the new spatial component, which was called the GIHS-TV framework. The goal is to improve the spectral distortion in the CS-based approaches and maintain the spatial details. Furthermore, L1 norm total variation as a constraint for the spatial components generated by the IHS transform was proposed as a simple implementation of the GIHS-TV framework. Experimental results on the “PAirMax” dataset indicated that the GIHS-TV performs well, especially in spectral fidelity.

2. Related Works

CS-based methods assume that this transformation can separate the spatial component from the spectral information of different bands. However, the separated spatial component is the weighted sums of the MS image, which usually contains spectral information and does not precisely match the spectral response of panchromatic sensors. Some spectral information is lost after the component substitution, which accounts for the poor performance on spectral fidelity [5,6]. IHS [1,34], PCA [35], BT [36] and GS [2] are the early classical algorithms that are easy to implement and fast to compute. These methods focused on the construction of the spatial component. The spatial component of the IHS transform [1,34] and Brovey transform (BT) [36] are the means of all bands. PCA [35] considers the first principal component as the spatial component. The spatial component of these transforms contains certain spectral information. The GS [2] method conducts Gram–Schimidt orthogonalization with the original multi-band data and the means of each band and then performs the subsequent component substitution and its backward transformation. The mean intensity information is incorporated before the transformation in the GS. In the improved algorithm of GS, GSA [37], using multispectral data to fit the low-passed filtered PAN image, generates the spatial component by a linear combination of MS images whose weights are solved by minimizing the RMSE.
The focus of subsequent research on most CS approaches after GIHS [6] was proposed shifts from constructing spatial variables to designing injection forms. BDSD [4] jointly estimates the weights and gain coefficients by minimizing MSE. An improved method based on physical constraint optimization, BDSD-PC [38], was proposed to raise the fusion quality. The PRACS [33] approach proposed the concept of the partial substitution of the spatial components. The weighted sum of MS and PAN images was used to construct the new spatial component. Scale factors and correlation coefficients are adopted for optimization to remove the local spectral instability errors.
Poor performance in terms of spatial fidelity usually exists with the MRA-based and variational methods. The typical MRA-based methods include ATWT [8], ATWT-M2 [9], ATWT-M3 [9], MTF-GLP [10], MTF-GLP-CBD [10], MTF-GLP-HPM-PP [33], MTF-GLP-HPM [39], HPF [40], AWLP [12], Indusion [41], SFIM [42,43], etc. Early MRA methods often took the strictly sampled discrete wavelet transform (DWT) approach, which does not have translation invariance, so the undecimated discrete wavelet transform gradually replaced it [8,9,12,39]. The Laplacian pyramid extracts spatial details from PAN images and adds these details to MS images. The generalized Laplacian pyramid (GLP) generalizes the LP to arbitrary fraction ratios [3]. The difference of the modulation transfer functions (MTF) between the MS and PAN sensors tends to cause spatial and spectral distortion in the fused images. Introducing the spectral response information of sensors into the multiscale decomposition framework is a milestone of the MRA-based approaches [10,33,39,44]. The MTF-GLP [10] method conducts operations such as MTF filtering and interpolation to obtain detailed images, which are injected into the multispectral images to obtain pansharpened images. The MTF-GLP achieved similar fused results to the ATWT method. MTF-GLP-HPM [39], using high pass modulation, improved the spectral quality on the basis of MTF-GLP. MTF-GLP-CBD [10] adopted injection coefficients estimated by multivariate linear regression, which achieved good performance on spatial and spectral fidelity. Some critical approaches employ a combination of CS and MRA in pansharpening [7]. The variational optimization-based methods rely on the models describing how the low-resolution MS and PAN images degrade from the high-resolution MS images. The complicated spectral and spatial constraints were constructed with large amounts of basic assumptions and priors. These models usually consist of at least three items that depict the subsampled and degradation process where the level sets [45], the sparse representation [46,47,48] and Bayesian [15,49] were introduced to regulate the ill-posed problem. P+XS [45] began the pioneering of variational pansharpening in the form of the regularization of the total variation. The subsequent variational fusion model adopted the TV or its derivatives [15,50,51], the nonlocal extension form [52] or fractional order model [13], as the regularization terms. Even with the noise amplification considered in these models depicting the degradation process, there is still a large upside to the variational optimization-based methods because the framework proposed by Ballester [45] too deeply binds them, making it difficult to incorporate approaches from other frameworks. Otherwise, sparse representation-based approaches form an essential class of pansharpening methods due to their effectiveness in local structure extraction. They are also considered variational optimization-based algorithms because these methods use variational models directly or indirectly [15]. It is worth noting that the performance of variational optimization-based methods is very sensitive to the value of hyperparameters, both in terms of computation time and performance. Therefore, determining hyperparameters usually requires a precise optimization phase, which may limit its performance and application scenarios. The variational optimization-based method may become more practical if suitable hyperparameters can be determined quickly. Machine learning works effectively in image processing, including image fusion [7,14]. Masi [32] introduced conventional neural networks with a simple three-layer architecture to the pansharpening problem, which achieved a competitive result. Yang [53] designed a deep network architecture called PanNet with a strong generalization ability incorporating domain-specific priors for spectral and spatial preservation. Recently, the basic idea of model-driven networks has become popular. The GPPNN [54], the first model-driven deep network for pansharpening, formulated two optimization problems for the generative models for PAN and LRMS images, which perform well visually and quantitatively.

3. GIHS-TV Fusion Framework

Tu [6] simplified the component substitution framework into the details injection scheme (GIHS) by mathematical derivation. In this paper, the module for constructing new spatial components using optimization methods was integrated into the GIHS scheme, which is called the GIHS-TV framework. The general steps of the GIHS-TV framework in the component substitution method are as follows: up-sampling low-resolution multispectral images (LRMS) to PAN image size to obtain M S ^ , calculating the spatial component (mean intensity, I 0 ) with the weights ( ω k ) for LRMS bands and constructing the objective function F ( I new , I 0 , P A N ) based on custom constraints. Then, the iterative method is selected to solve I new by minimizing the function; according to the selected method of constructing the spatial component, the detail gain coefficients ( g k ) are determined and the detail residuals calculated ( δ = I new I 0 ). The detail injection, whose expression is as (1), is calculated to obtain the fused image.
M S k ^ = M S k ˜ + g k δ , k ,
In addition, we proposed a new L1-TV optimized method under the GIHS-TV framework. Figure 2 clearly illustrates the construction process of the objective function and the vital role of total variation in the GIHS-TV framework.

3.1. Generalized IHS Transform

To visualize the construction process of the injection expression, we first performed the derivation of the three-channel form of the IHS method. The IHS transform extracts the spatial component ( I 0 ), the hue component (H) and the saturation component (S) by performing the IHS transform on the three-channel image. The forward transform of IHS is as shown in (2).
I 0 v 1 v 2 = 1 / 3 1 / 3 1 / 3 2 / 6 2 / 6 2 / 6 1 / 2 1 / 2 0 R 0 G 0 B 0
where I 0 and I new individually represent the spatial component before and after the substitution. The inverse transformation is as shown in (3).
R new G new B new = 1 1 / 2 1 / 2 1 1 / 2 1 / 2 1 2 0 I new v 1 v 2
Then, the gain coefficients and injection expression could be determined by the following.
R n e w G n e w B n e w = 1 1 / 2 1 / 2 1 1 / 2 1 / 2 1 2 0 I 0 + ( I n e w I 0 ) v 1 v 2 = R 0 + δ G 0 + δ B 0 + δ
where [ R new , G new , B new ] T are the three channels in the fused image with the IHS method and δ = I new I 0 . The detail injection expression for the IHS transformation is as shown in Equation (5). Thus, it is easy to determine the weights and gain coefficients in the IHS fusion method, i.e., ω k = 1 / 3 , g k = 1 .
M S k ^ = M S k ˜ + g k δ = M S k ˜ + I new 1 3 k = 1 N M S k ˜
For the GIHS method’s N-channel (N > 3) form, the GIHS transformation is as Equation (6). The spatial component’s mean intensity is synthesized as in (7).
v 1 = I 0 v 2 v N = Φ M S 1 ˜ M S 2 ˜ M S N ˜ = φ 11 φ 12 φ 1 N φ 21 φ 22 φ 2 N φ N 1 φ N 2 φ N N M S 1 ˜ M S 2 ˜ M S N ˜
I 0 = k = 1 N φ 1 k M S k ˜
After replacing I 0 with I new , the fused multispectral image could be obtained by Equation (8).
M S 1 ^ M S 2 ^ M S N ^ = Φ 1 I n e w v 2 v N = Φ 1 v 1 v 2 v N + 1 0 0 ( I n e w I 0 ) = M S 1 ˜ M S 2 ˜ M S N ˜ + Φ 1 1 0 0 ( I n e w I 0 )
where Φ 1 is the inverse of the matrix Φ . From (8), it can be seen that the detail injection is only the first column of the matrix Φ 1 in action, which could be written as a new vector g. g = [ g 1 , g 2 , , g N ] T is called the detail gain vector. Then, Equation (8) can be written as a more concise expression for detail injection.
M S 1 ^ M S 2 ^ M S N ^ = M S 1 ˜ M S 2 ˜ M S N ˜ + g I n e w I 0 = M S 1 ˜ M S 2 ˜ M S N ˜ + g 1 g 2 g N δ
Uniform detail injection form can be written as in (10)
M S k ^ = M S k ˜ + g k δ , k ,

3.2. L1-TV Optimization

Due to the fact that the new spatial component could be obtained by the original spatial component and the PAN image, I new could be regarded as the function of I 0 and PAN. As for the problem of spectral distortion, intensity distribution could be the most critical constraint. Combining the need to enhance the spatial detail of multispectral images, we establish the functional of the objective function for pansharpening. Specifically, the new framework (GIHS-TV) proposed in this paper considers the construction task of the new spatial component as an optimization problem.
First, the intensity distribution of the new spatial component I new should be as consistent as possible with the spatial component I 0 , i.e., the difference between the two should be as small as possible. To solve the problem of spectral distortion, we formulated constraints (11) as the fidelity term, which should be small enough.
ε 1 I new = 1 p I new I 0 p p
On the other hand, the PAN image contains more edge information. Due to the significant difference in intensity distribution between the spatial component and PAN image, it is more reasonable to use the gradient rather than the intensity to express the edge information. Therefore, the gradient of I new should be consistent with the PAN image, and it could be regarded as the regularization item, as in Equation (12).
ε 2 I new = 1 q I new P A N q q
The fusion task is then depicted as minimizing the objective function (13) to obtain spatial and spectral fidelity simultaneously.
ε I new = 1 p I new I 0 p p + λ q I new P A N q q
Let p = 1 , q = 1 . The rationale for the choice of the norm is as follows. Firstly, preserving the intensity distribution as much as possible is desirable, i.e., most of the fidelity term, I new I , should be zero. From this objective, a small part of I new I should be large to transfer the gradient information from the PAN image to the new spatial component. The fidelity term should be subject to the Laplacian or impulsive distribution, so the L1 norm is chosen as the constraint. Secondly, the sparsity of gradients is encouraged since natural images are usually piece-wise smooth and their gradients tend to be sparse. The regularization adopted the L1 norm because relevant mathematical theory guarantees that the L1 norm can obtain the sparse solution. This objective function expects most of the difference between the I new and I 0 to be zero, where the non-zero item indicates the gradient information added from the PAN image, which could ensure the sparsity of the fidelity and the consistency of intensity distribution. On the other hand, the L1 norm of the gradient, i.e., the total variation, can also encourage the sparsity of the gradient. As for the solution of the objective function, it can be turned into the L1-TV minimization problem of the variable D i f f with a simple variable substitution as (14) [55], and then D i f f can be solved by IRN [56], ADMM [57] or FISTA [58], etc. The Iteratively Reweighted Norm (IRN, [56]) method was chosen in this paper, whose stream is shown in Algorithm 1.
D i f f = I n e w P A N
Algorithm 1 The solution flow of D i f f in the GIHS-TV algorithm.
  • Input linear operator A , image b
  •     A = E ( E is the identity matrix), p = q = 1 , b = I 0 PAN
  •     Diff ( 0 ) = ( A T A + λ D T D ) 1 A T b ( D is the gradient matrix)
  •     for   k = 0 , 1 , do
  •      W F ( k ) = diag ( τ F , ε F ( Diff ( k ) b ) )
  •      W R ( k ) = diag ( τ R , ε R ( D X Diff ( k ) ) 2 + ( D Y Diff ( k ) ) 2 )
  •      Diff ( k + 1 ) = ( A T W F ( k ) A + λ D T W R ( k ) D ) 1 A T W F ( k ) b
  •    end for
  • Output D i f f
The new spatial component could be obtained from Equation (15). After the inverse transformation is conducted with I n e w , the multispectral image with high spatial resolution is obtained.
I new = D i f f + P A N
As for the new spatial component I n e w , the simplest synthesized approach was adopted in this paper, i.e., the weight ω k = 1 , k . In addition, the detail gain coefficient g k satisfies (16).
g k = k = 1 N ω k 1 , k
Thus, g k = 1 , k . The detail residuals δ can be obtained.
δ = I new I 0 = D i f f + P A N I 0
The fast implementation expression in the method with L1-TV constraints for constructing the new spatial component can be determined as in (18).
M S k ^ = M S k ˜ + D i f f + P A N I 0 , k
It is worth mentioning that the three-channel and the multi-channel form of the GIHS-TV share the same form of injection expression. The difference between them is only in synthesizing the mean intensity (spatial component), which brings great convenience to the two types of fusion.

4. Experiments

4.1. Datasets and Evaluation Metrics

In this paper, “PAirMax” [14], a publicly available fusion dataset produced by Vivone et al., was selected to explore the performance of GIHS-TV. The dataset consists of 14 pairs of images, including MS image (4-band, 8-band) and PAN image from GeoEye-1 (GE), WorldView-2 (W2), WorldView-3 (W3), WorldView-4 (W4), SPOT-7 (S7) and Pleiades (Pl), where the resolution of PAN is four times that of MS images, and the dataset has been registered. The dataset provides three types of scenarios from different countries: urban (Urb), natural (Nat) and mixed urban–natural (Mix). It shows many kinds of urban environments: typical, dense, with long shadows, water or vegetation. The data names consist of the satellite, location and scene type.
It is necessary to compare visual effects and evaluation metrics to evaluate the performance of our method. Typical CS approaches (IHS, PCA, BT, GS, GSA, BDSD, PRACS) and MRA approaches (HPF, SFIM, Indusion, ATWT, AWLP, ATWT-M2, ATWT-M3, MTF-GLP, MTF-GLP-HPM-PP, MTF-GLP-CBD) were chosen to conduct the comparison with GIHS-TV in the full resolution fusion. The retention of spectral and spatial details is quantitatively evaluated by the selected evaluation metrics, such as spectral angle (SAM) [59], spatial correlation coefficient (SCC) [60], D λ [61], D S [61] and QNR [61]. The details of these metrics are as follows.
(1) SAM: SAM reflects the spectral fidelity between the fused image and the reference MS images by calculating the spectral angle between the corresponding pixels of the two images. Denote F { i } and R { i } as the grayscale values at position i in the fused and MS images, respectively. The calculation is as in Equation (19).
S A M F { i } , R { i } = a r c c o s F { i } , R { i } F { i } R { i }
The average of the spectral angle of all pixels is regarded as the SAM of the two images. The smaller the SAM value, the higher the spectral fidelity.
(2) SCC: The high-frequency information is extracted from the PAN and fused images by high-pass filters; then, the correlation between the two is calculated through correlation coefficients. Its definition is as Equation (20).
S C C = i = 1 M j = 1 N ( H F k ( i , j ) H F k ¯ ) ( H P ( i , j ) H P ¯ ) i = 1 M j = 1 N ( H F k ( i , j ) H F k ¯ ) 2 i = 1 M j = 1 N ( H P ( i , j ) H P ¯ ) 2
where H F k stands for the high-frequency information of the kth band in the fused images, and H P is the high-frequency information of the PAN image. The larger the SCC, the better the spatial correlation between the fused and the PAN images.
(3) D λ : D λ characterizes the spectral loss between the fused and the MS images and is defined by the following equation.
D λ = 1 B ( B 1 ) i = 1 B j = 1 j i B Q ( F i , F j ) Q ( M S i , M S j ) p p
where p amplifies the spectral differences and is usually set to 1. The smaller it is, the smaller the spectral distortion is. Q is the universal image quality index (UIQI). UIQI calculates the correlation, brightness and contrast similarity between the fused and the reference images to characterize the comprehensive performance of fusion. It is usually abbreviated as the Q index, and the definition formula is as follows.
Q ( F , R ) = σ F R σ F σ R 2 μ F μ R μ F 2 + μ R 2 2 σ F σ R σ F 2 + σ R 2
where σ F and σ R are the standard deviations of the fused and the reference images, respectively, σ F R is the covariance of the two, and μ F and μ R denote the mean values of the two, respectively. The three fractions represent the correlation, mean brightness and contrast similarity. The higher Q value means that the fused images are more similar to the reference images.
(4) D S : D S is used to calculate the loss of spatial detail between the fused and the PAN images. It is defined by the following.
D S = 1 B i = 1 B Q ( F i , P A N ) Q ( M S i , P A N L P ) q q
where P A N L P is the downsampled PAN image to the same resolution as the original MS images. From a practical point of view, it should be ensured that the P A N L P should be perfectly aligned with the MS image; otherwise, this metric loses its meaning. q serves to amplify the difference in spatial detail distortion. The smaller the D S , the smaller the spatial detail distortion is.
(5) Q N R : Q N R is the most mainstream non-reference index for evaluating the performance of pansharpening methods, which integrates spectral and spatial distortion characterization, as defined below.
Q N R = ( 1 D λ ) α ( 1 D S ) β
where α and β are parameters used to balance the spectral distortion and spatial distortion, and the larger the QNR, the better the fusion performance, with a maximum value of 1.

4.2. Initial Results

In the GIHS-TV framework, the residuals between the new and original spatial components ( I 0 and I new ) stand for the detail injections added to the fused images. The consistency and difference were analyzed in the global and local performance. In addition, the residual by the IHS method was provided for comparison with the one of the GIHS-TV methods.
The experiments in the scene “GE_Lond_Urb” show the spatial component images before and after the fusion. Figure 3 shows that the two images maintain a good consistency in global hue, which indicates the spectral fidelity of the GIHS-TV method. In the boxes, the texture and contour of the buildings in the I new is richer than that in the I 0 . Further, the residuals ( δ ) between I 0 and I new were evaluated in terms of the grayscale and edge. A histogram stretching was performed on the residuals. Another implication of the residuals is the detail injection in our framework, which accounts for the selection of the IHS method for the auxiliary analysis where the two methods share the same I 0 and our method produces the new spatial component by L1-TV.
In Figure 4, clear edges were shown in the residuals from the GIHS-TV method, achieving the goal of adding edge information from the PAN image to I new without altering the grayscale distribution as much as possible. Many artifacts near the edges were generated in the residuals from the IHS method. Especially, the interior of the building was almost filled with white in the red box; a dramatic brightness changes occurred in the new spatial component, and the final fused images showed spectral distortion. Several other local details are present in Figure 4—our method also has good, cleaner edge details.
The proportion of edge information added from the PAN image was changed through hyperparameter λ in the model. The fused effects of λ values were analyzed in terms of both visual effects and evaluation metrics in the “Pl_Sacr_Mix” scene. Figure 5 shows that the results are consistent with the intuition that as the hyperparameter increases, the geometric information becomes richer while the spectral distortion increases. At the same time, there is a tipping point ( λ = 0.7) and a bottleneck point ( λ = 2.8). D S , and QNR achieves the best value at λ = 0.7, but the visual results of the fusion in Figure 6 show a good balance between spectral fidelity and spatial fidelity when λ is equal to 1.

4.3. Fusion Results

Images by EXP [3] stand for the MS image interpolation using a polynomial kernel with 23 coefficients. The hyperparameters λ in the GIHS-TV method were set to 1 and 2, and the parameters of the other methods were kept consistent with the corresponding articles. Experiments on the “PAirMax” dataset show that most fusion methods enhance spatial details compared with the original MS images. Objects that cannot be identified in any single original image can be easily identified and inferred in fused images. However, CS approaches exhibit various degrees of color distortion. Fused images of typical scenes were shown and analyzed.
In the “GE_Lond_Urb” (Figure 7), the fused images all perform well in the enhancement of spatial details. The details within the shadows are enhanced well, such as the cars and the color, contours and texture of the trees and the low buildings. The textures of the tall buildings outside the shadows also become apparent, and even their structures or materials can be inferred. However, PCA, IHS and GS all exhibit a global blue hue that is different from the original MS image. This shows the color distortion that occurred in the CS approaches. In contrast, GIHS-TV performs better in terms of global hue and the spectral construction of local objects.
In the “S7_NewY_Mix” (Figure 8), the most conspicuous feature is the river and the global hue. The spectral construction of the river by the PCA, IHS, BT and GS shows apparent errors. Regarding the spectral construction of vegetation and buildings, only four improved CS methods, GSA, BDSD, PRACS and GIHS-TV, have superior performances.
In the “W2_Miam_Mix” (Figure 9), we mainly reference the color of the grass and trees and the global hue. The PCA, IHS, BT and BDSD methods show poor global hue, and BDSD performs poorly in spatial expression. It is worth pointing out that the GIHS-TV method contains richer spatial details when the hyperparameter is greater than or equal to 2 but also exhibits color distortion, which also occurs in Figure 8. If the fused images were to be used for the visual tasks, a large λ value would work better, λ = 2 , which is recommended as an experience value based on our experimental results.
In the local of “Pl_Hous_Urb” (Figure 10), the fused images of all methods are rich in spatial details, especially the target contours and textures of cargo bins, trucks plants. In addition, road lines are effectively enhanced. GIHS-TV achieved a visual effect comparable to other methods.
The local fused results of “W4_Mexi_Urb” are as shown in Figure 11, where the spatial details in the fused images are rich. It can be seen that, compared with other methods, the PRACS method is slightly weaker in rendering spatial details, such as the poor clarity of buildings’ contours. The BDSD method tends to introduce spatial artifacts. These two improved CS-based methods are centered on spectral fidelity and do not use the PAN directly when constructing the spatial component but use its downsampled PAN or its partial image, and their spatial detail injection is not as good as other methods, which explains the lack of effective improvement in their spatial fidelity. GIHS-TV fused images perform well in both spectral fidelity and spatial fidelity, and it can effectively identify many small targets, such as buildings, cars, containers and objects on building roofs.
The experimental performance with high spatial and spectral fidelity fully illustrates the advantages of the GIHS-TV framework for constructing spatial components with optimized ideas. GIHS-TV achieved a superior performance.
Representative methods of CS and MRA were selected for the comparison and evaluation of fusion methods in the full-resolution scheme provided by Vivone [5,11]. As shown in Figure 12, it can be seen that the evaluation metrics of the GIHS-TV method fusion perform well, with lower D λ and higher QNR in most scenes. The lower D λ indicates good performance in spectral fidelity, and a higher QNR stands for an excellent comprehensive fidelity.
Among the many methods, only the BDSD method can outperform the GIHS-TV metrics in most scenes, but based on the visual analysis above, it is clear that BDSD does not perform as well as GIHS-TV and other methods.
The above visual effects show that the GIHS-TV method effectively improves the spectral distortion in the component substitution class. The analysis of the evaluation metrics shows better performances than these methods. A comparison with MRA-based methods was conducted and shown. Table 1 offers the mean value of six metrics undertaken in the 13 scenes except for the “W3_Muni_Nat”. The GIHS-TV performs best in all full resolution metrics i.e., D λ , D § , QNR. As for the performance at reduced resolution, our method behaves well in both SAM and SCC. In Figure 13, GIHS-TV performs with high spectral fidelity and more spatial details compared with the IHS method and LRMS images, whose performance is consistent with those at full resolution.
Compared with ATWT-M2 and ATWT-M3 methods, GIHS-TV needs a slightly longer computation time due to optimization iterations. It is more meaningful to make comparisons between the same class of methods. Thus, the P+XS methods are selected to compare the running time in the “GE_Tren_Urb”, “W3_Muni_Urb”, “W3_Muni_Mix” and “W4_Mexi_Nat” scenes. As shown in Table 2, the optimization times of the GIHS-TV method are much lower than the P+XS method—approximately one order of magnitude. This is attributed to the sparsity and conciseness of the proposed model.
The analysis involving visual effects and metrics shows that the GIHS-TV method is excellent in terms of fusion with high spectral and spatial fidelity. Compared with traditional TV-based methods, our method performs with better timeliness.

5. Conclusions

This paper adopted an optimization perspective to improve the CS-based pansharpening methods and build the GIHS-TV framework. Faced with the problem of the loss of spectral information in the fused images, we proposed a method with L1-TV that was used to constrain the spectral–spatial information in the new spatial components, effectively improving the spectral distortion. Its fast algorithm was implemented in the framework. Compared with other variational optimization-based methods, the GIHS-TV framework absorbs the advantage of high spatial fidelity from the CS-based methods. Experiments on the “PairMax” dataset show that GIHS-TV can maintain both the spectral and spatial information from the MS and PAN image well. The spectral fidelity in GIHS-TV is greatly improved compared with other CS approaches.

Author Contributions

Conceptualization, methodology: X.Z. (Xuefeng Zhang) and X.D.; software: X.Z. (Xuefeng Zhang) and Y.H.; validation, formal analysis, investigation and data curation: X.D.; writing: X.Z. (Xuefeng Zhang); writing—review and editing: X.D., X.Z. (Xuemin Zhang) and Y.H.; visualization: X.Z. (Xuefeng Zhang) and Y.K.; supervision and funding acquisition: X.D. and G.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Key R&D Program of China (Grant No. 2022YFB3903401) and China Postdoctoral Science Foundation (Grant No. 2020T130479 & 2021M692461).

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

The authors would like to thank Gemine Vivone for providing the “PAirMax” datasets for data support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Edwards, K.; Davis, P.A. The use of intensity-hue-saturation transformation for producing color shaded-relief images. Photogramm. Eng. Remote Sens. 1994, 60, 1369–1374. [Google Scholar]
  2. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  3. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-driven fusion of high spatial and spectral resolution images based on oversampled multiresolution analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  4. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2007, 46, 228–236. [Google Scholar] [CrossRef]
  5. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Restaino, R.; Scarpa, G.; Ulfarsson, M.O.; Alparone, L.; Chanussot, J. A new benchmark based on recent advances in multispectral pansharpening: Revisiting pansharpening with classical and emerging pansharpening methods. IEEE Geosci. Remote Sens. Mag. 2020, 9, 53–81. [Google Scholar] [CrossRef]
  6. Tu, T.M.; Su, S.C.; Shyu, H.C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion 2001, 2, 177–186. [Google Scholar] [CrossRef]
  7. Javan, F.D.; Samadzadegan, F.; Mehravar, S.; Toosi, A.; Khatami, R.; Stein, A. A review of image fusion techniques for pan-sharpening of high-resolution satellite imagery. ISPRS J. Photogramm. Remote Sens. 2021, 171, 101–117. [Google Scholar] [CrossRef]
  8. Vivone, G.; Restaino, R.; Dalla Mura, M.; Licciardi, G.; Chanussot, J. Contrast and error-based fusion schemes for multispectral image pansharpening. IEEE Geosci. Remote Sens. Lett. 2013, 11, 930–934. [Google Scholar] [CrossRef] [Green Version]
  9. Ranchin, T.; Wald, L. Fusion of high spatial and spectral resolution images: The ARSIS concept and its implementation. Photogramm. Eng. Remote Sens. 2000, 66, 49–61. [Google Scholar]
  10. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored multiscale fusion of high-resolution MS and Pan imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  11. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2565–2586. [Google Scholar] [CrossRef]
  12. Otazu, X.; González-Audícana, M.; Fors, O.; Núñez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  13. Liu, P.; Xiao, L.; Tang, S.; Sun, L. Fractional order variational pan-sharpening. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 2602–2605. [Google Scholar]
  14. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Pacifici, F. A benchmarking protocol for pansharpening: Dataset, preprocessing, and quality assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6102–6118. [Google Scholar] [CrossRef]
  15. Li, S.; Yang, B. A new pan-sharpening method using a compressed sensing technique. IEEE Trans. Geosci. Remote Sens. 2010, 49, 738–746. [Google Scholar] [CrossRef]
  16. Li, S.; Yin, H.; Fang, L. Remote sensing image fusion via sparse representations over learned dictionaries. IEEE Trans. Geosci. Remote Sens. 2013, 51, 4779–4789. [Google Scholar] [CrossRef]
  17. Zhong, J.; Yang, B.; Huang, G.; Zhong, F.; Chen, Z. Remote sensing image fusion with convolutional neural network. Sens. Imaging 2016, 17, 10. [Google Scholar] [CrossRef]
  18. Li, Z.; Cheng, C. A CNN-based pan-sharpening method for integrating panchromatic and multispectral images using Landsat 8. Remote Sens. 2019, 11, 2606. [Google Scholar] [CrossRef] [Green Version]
  19. Zeng, Z.; Wang, D.; Tan, W.; Yu, G.; You, J.; Lv, B.; Wu, Z. RCSANet: A Full Convolutional Network for Extracting Inland Aquaculture Ponds from High-Spatial-Resolution Images. Remote Sens. 2020, 13, 92. [Google Scholar] [CrossRef]
  20. Wang, W.; Zhou, Z.; Liu, H.; Xie, G. MSDRN: Pansharpening of multispectral images via multi-scale deep residual network. Remote Sens. 2021, 13, 1200. [Google Scholar] [CrossRef]
  21. Liu, Q.; Han, L.; Tan, R.; Fan, H.; Li, W.; Zhu, H.; Du, B.; Liu, S. Hybrid attention based residual network for pansharpening. Remote Sens. 2021, 13, 1962. [Google Scholar] [CrossRef]
  22. Wu, Y.; Feng, S.; Lin, C.; Zhou, H.; Huang, M. A three stages detail injection network for remote sensing images pansharpening. Remote Sens. 2022, 14, 1077. [Google Scholar] [CrossRef]
  23. Yin, J.; Qu, J.; Sun, L.; Huang, W.; Chen, Q. A Local and Nonlocal Feature Interaction Network for Pansharpening. Remote Sens. 2022, 14, 3743. [Google Scholar] [CrossRef]
  24. Nie, Z.; Chen, L.; Jeon, S.; Yang, X. Spectral-Spatial Interaction Network for Multispectral Image and Panchromatic Image Fusion. Remote Sens. 2022, 14, 4100. [Google Scholar] [CrossRef]
  25. Jian, L.; Wu, S.; Chen, L.; Vivone, G.; Rayhana, R.; Zhang, D. Multi-Scale and Multi-Stream Fusion Network for Pansharpening. Remote Sens. 2023, 15, 1666. [Google Scholar] [CrossRef]
  26. Ciotola, M.; Scarpa, G. Fast Full-Resolution Target-Adaptive CNN-Based Pansharpening Framework. Remote Sens. 2023, 15, 319. [Google Scholar] [CrossRef]
  27. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  28. Zhang, L.; Li, W.; Huang, H.; Lei, D. A pansharpening generative adversarial network with multilevel structure enhancement and a multistream fusion architecture. Remote Sens. 2021, 13, 2423. [Google Scholar] [CrossRef]
  29. Gastineau, A.; Aujol, J.F.; Berthoumieu, Y.; Germain, C. Generative adversarial network for pansharpening with spectral and spatial discriminators. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4401611. [Google Scholar] [CrossRef]
  30. Liu, Q.; Zhou, H.; Xu, Q.; Liu, X.; Wang, Y. PSGAN: A generative adversarial network for remote sensing image pan-sharpening. IEEE Trans. Geosci. Remote Sens. 2020, 59, 10227–10242. [Google Scholar] [CrossRef]
  31. Jozdani, S.; Chen, D.; Chen, W.; Leblanc, S.G.; Lovitt, J.; He, L.; Fraser, R.H.; Johnson, B.A. Evaluating Image Normalization via GANs for Environmental Mapping: A Case Study of Lichen Mapping Using High-Resolution Satellite Imagery. Remote Sens. 2021, 13, 5035. [Google Scholar] [CrossRef]
  32. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
  33. Choi, J.; Yu, K.; Kim, Y. A new adaptive component-substitution-based satellite image fusion by using partial replacement. IEEE Trans. Geosci. Remote Sens. 2010, 49, 295–309. [Google Scholar] [CrossRef]
  34. Carper, W.; Lillesand, T.; Kiefer, R. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  35. Kwarteng, P.; Chavez, A. Extracting spectral contrast in Landsat Thematic Mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  36. Gillespie, A.R.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images. II. Channel ratio and “chromaticity” transformation techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar] [CrossRef]
  37. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS + Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  38. Vivone, G. Robust band-dependent spatial-detail approaches for panchromatic sharpening. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  39. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. An MTF-based spectral distortion minimizing model for pan-sharpening of very high resolution multispectral images of urban areas. In Proceedings of the 2003 2nd GRSS/ISPRS Joint Workshop on Remote Sensing and Data Fusion over Urban Areas, Berlin, Germany, 22–23 May 2003; pp. 90–94. [Google Scholar]
  40. Chavez, P.; Sides, S.C.; Anderson, J.A. Comparison of three different methods to merge multiresolution and multispectral data- Landsat TM and SPOT panchromatic. Photogramm. Eng. Remote Sens. 1991, 57, 295–303. [Google Scholar]
  41. Khan, M.M.; Chanussot, J.; Condat, L.; Montanvert, A. Indusion: Fusion of multispectral and panchromatic images using the induction scaling technique. IEEE Geosci. Remote Sens. Lett. 2008, 5, 98–102. [Google Scholar] [CrossRef] [Green Version]
  42. Liu, J. Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  43. Wald, L.; Ranchin, T. Liu’Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details’. Int. J. Remote Sens. 2002, 23, 593–597. [Google Scholar] [CrossRef]
  44. Dong, W.; Xiao, S.; Li, Y.; Qu, J. Hyperspectral pansharpening based on intrinsic image decomposition and weighted least squares filter. Remote Sens. 2018, 10, 445. [Google Scholar] [CrossRef] [Green Version]
  45. Ballester, C.; Caselles, V.; Igual, L.; Verdera, J.; Rougé, B. A variational model for P+ XS image fusion. Int. J. Comput. Vis. 2006, 69, 43. [Google Scholar] [CrossRef]
  46. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. A new pansharpening algorithm based on total variation. IEEE Geosci. Remote Sens. Lett. 2013, 11, 318–322. [Google Scholar] [CrossRef]
  47. He, X.; Condat, L.; Bioucas-Dias, J.M.; Chanussot, J.; Xia, J. A new pansharpening method based on spatial and spectral sparsity priors. IEEE Trans. Image Process. 2014, 23, 4160–4174. [Google Scholar] [CrossRef] [PubMed]
  48. Tian, X.; Chen, Y.; Yang, C.; Gao, X.; Ma, J. A variational pansharpening method based on gradient sparse representation. IEEE Signal Process. Lett. 2020, 27, 1180–1184. [Google Scholar] [CrossRef]
  49. Fasbender, D.; Radoux, J.; Bogaert, P. Bayesian data fusion for adaptable image pansharpening. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1847–1857. [Google Scholar] [CrossRef]
  50. Möller, M.; Wittman, T.; Bertozzi, A.L.; Burger, M. A variational approach for sharpening high dimensional images. SIAM J. Imaging Sci. 2012, 5, 150–178. [Google Scholar] [CrossRef] [Green Version]
  51. Zhang, G.; Fang, F.; Zhou, A.; Li, F. Pan-sharpening of multi-spectral images using a new variational model. Int. J. Remote Sens. 2015, 36, 1484–1508. [Google Scholar] [CrossRef]
  52. Duran, J.; Buades, A.; Coll, B.; Sbert, C. A nonlocal variational model for pansharpening image fusion. SIAM J. Imaging Sci. 2014, 7, 761–796. [Google Scholar] [CrossRef]
  53. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A deep network architecture for pan-sharpening. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5449–5457. [Google Scholar]
  54. Xu, S.; Zhang, J.; Zhao, Z.; Sun, K.; Liu, J.; Zhang, C. Deep gradient projection networks for pan-sharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1366–1375. [Google Scholar]
  55. Zhang, X.; Dai, X.; Zhang, X.; Jin, G. Joint principal component analysis and total variation for infrared and visible image fusion. Infrared Phys. Technol. 2023, 128, 104523. [Google Scholar] [CrossRef]
  56. Rodríguez, P.; Wohlberg, B. Efficient minimization method for a generalized total variation functional. IEEE Trans. Image Process. 2008, 18, 322–332. [Google Scholar] [CrossRef] [PubMed]
  57. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  58. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef] [Green Version]
  59. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the spectral angle mapper (SAM) algorithm. In JPL, Summaries of the Third Annual JPL Airborne Geoscience Workshop, Volume 1: AVIRIS Workshop; NASA: Washington, DC, USA, 1992. [Google Scholar]
  60. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  61. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The spatial and spectral distortion.
Figure 1. The spatial and spectral distortion.
Remotesensing 15 02945 g001
Figure 2. The GIHS-TV framework.
Figure 2. The GIHS-TV framework.
Remotesensing 15 02945 g002
Figure 3. The original and new spatial components.
Figure 3. The original and new spatial components.
Remotesensing 15 02945 g003
Figure 4. The residuals with details in the IHS and GIHS-TV methods.
Figure 4. The residuals with details in the IHS and GIHS-TV methods.
Remotesensing 15 02945 g004
Figure 5. The evaluation metrics of fusion with different values of λ .
Figure 5. The evaluation metrics of fusion with different values of λ .
Remotesensing 15 02945 g005
Figure 6. The visual effect of fusion with different values of λ .
Figure 6. The visual effect of fusion with different values of λ .
Remotesensing 15 02945 g006
Figure 7. Fusion results in the “GE_Lond_Urb” scene.
Figure 7. Fusion results in the “GE_Lond_Urb” scene.
Remotesensing 15 02945 g007
Figure 8. Fusion results in the “S7_NewY_Mix” scene.
Figure 8. Fusion results in the “S7_NewY_Mix” scene.
Remotesensing 15 02945 g008
Figure 9. Fusion results in the “W2_Miam_Mix” scene.
Figure 9. Fusion results in the “W2_Miam_Mix” scene.
Remotesensing 15 02945 g009
Figure 10. The local fused results in the “Pl_Hous_Urb” scene.
Figure 10. The local fused results in the “Pl_Hous_Urb” scene.
Remotesensing 15 02945 g010
Figure 11. The local fused results in the “W4_Mexi_Urb” scene.
Figure 11. The local fused results in the “W4_Mexi_Urb” scene.
Remotesensing 15 02945 g011
Figure 12. Evaluation metrics of fusion on the “PAirMax” dataset.
Figure 12. Evaluation metrics of fusion on the “PAirMax” dataset.
Remotesensing 15 02945 g012
Figure 13. The visual effect of fusion at reduced resolution. The first row shows the original MS image. Other rows’ images stands for the fused images by IHS and GIHS-TV methods, respectively.
Figure 13. The visual effect of fusion at reduced resolution. The first row shows the original MS image. Other rows’ images stands for the fused images by IHS and GIHS-TV methods, respectively.
Remotesensing 15 02945 g013
Table 1. The mean evaluation metrics of fusion on the “PAirMax” dataset. Numbers in the bold stands for the best values.
Table 1. The mean evaluation metrics of fusion on the “PAirMax” dataset. Numbers in the bold stands for the best values.
Method D λ D S QNRSAMSCC
HPF0.06510.10310.83891.57750.8600
SFIM0.06440.09980.84271.22770.8559
Indusion0.06130.08820.85622.78590.8040
ATWT0.07690.11150.82061.89980.8738
AWLP0.06370.09640.84632.06270.8302
ATWT-M20.05810.13750.81322.02030.8408
ATWT-M30.06660.12210.82031.92970.8606
MTF-GLP0.07940.11490.81522.01130.8789
MTF-GLP-HPM-PP0.10650.13540.77302.49690.8948
MTF-GLP-HPM0.07880.10970.82061.45390.8707
MTF-GLP-CBD0.07730.11350.81842.14970.8742
GIHS-TV ( λ = 1 )0.05500.08760.86240.60900.8760
Table 2. The calculation time of P+XS and GIHS-TV (seconds).
Table 2. The calculation time of P+XS and GIHS-TV (seconds).
MethodsGE_Tren_UrbW3_Muni_UrbW3_Muni_MixW4_Mexi_Nat
P+XS197.7686397.8795420.6308207.6450
GIHS-TV27.650823.016226.507424.2559
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Dai, X.; Zhang, X.; Hu, Y.; Kang, Y.; Jin, G. Improved Generalized IHS Based on Total Variation for Pansharpening. Remote Sens. 2023, 15, 2945. https://doi.org/10.3390/rs15112945

AMA Style

Zhang X, Dai X, Zhang X, Hu Y, Kang Y, Jin G. Improved Generalized IHS Based on Total Variation for Pansharpening. Remote Sensing. 2023; 15(11):2945. https://doi.org/10.3390/rs15112945

Chicago/Turabian Style

Zhang, Xuefeng, Xiaobing Dai, Xuemin Zhang, Yuchen Hu, Yingdong Kang, and Guang Jin. 2023. "Improved Generalized IHS Based on Total Variation for Pansharpening" Remote Sensing 15, no. 11: 2945. https://doi.org/10.3390/rs15112945

APA Style

Zhang, X., Dai, X., Zhang, X., Hu, Y., Kang, Y., & Jin, G. (2023). Improved Generalized IHS Based on Total Variation for Pansharpening. Remote Sensing, 15(11), 2945. https://doi.org/10.3390/rs15112945

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop