Next Article in Journal
Enhanced Crop Leaf Area Index Estimation via Random Forest Regression: Bayesian Optimization and Feature Selection Approach
Previous Article in Journal
Enhancing Algal Bloom Level Monitoring with CYGNSS and Sentinel-3 Data
Previous Article in Special Issue
Unsupervised Domain Adaptation with Contrastive Learning-Based Discriminative Feature Augmentation for RS Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interplay Between Atmospheric Correction and Fusion Techniques Enhances the Quality of Remote Sensing Image Fusion

1
Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei 230031, China
2
Graduate School of Science Island, University of Science and Technology of China, Hefei 230026, China
3
Key Laboratory of Optical Calibration and Characterization, Chinese Academy of Sciences, Hefei 230031, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(21), 3916; https://doi.org/10.3390/rs16213916
Submission received: 28 August 2024 / Revised: 9 October 2024 / Accepted: 11 October 2024 / Published: 22 October 2024
(This article belongs to the Special Issue Advances in Deep Fusion of Multi-Source Remote Sensing Images)

Abstract

:
Remote sensing image fusion technology integrates observational data from multiple satellite platforms to leverage the complementary advantages of the different types of remote sensing images. High-quality fused remote sensing images provide detailed information on surface radiation, climate, and environmental conditions, thereby supporting governmental policies on environmental changes. Improving the quality and quantitative accuracy of fused images is a crucial trend in remote sensing image fusion research. This study investigates the impact of atmospheric correction and five widely applied fusion techniques on remote sensing image fusion. By constructing four fusion frameworks, it evaluates how the choice of fusion method, the implementation of atmospheric correction, the synchronization of atmospheric parameters, and the timing of atmospheric correction influence the outcomes of remote sensing image fusion. Aerial flights using remote sensors were conducted to acquire atmospheric parameter distribution images that are strictly synchronous with the remote sensing images. Comprehensive and systematic evaluations of the fused remote sensing images were performed. Experiments show that for the remote sensing images used, selecting the appropriate fusion method can improve the spatial detail evaluation metrics of the fused images by up to 2.739 times, with the smallest deviation from true reflectance reaching 35.02%. Incorporating synchronous atmospheric parameter distribution images can enhance the spatial detail evaluation metrics by up to 2.03 times, with the smallest deviation from true reflectance reaching 5.4%. This indicates that choosing an appropriate fusion method and performing imaging-based synchronous atmospheric correction before fusion can maximize the enhancement of spatial details and spectral quantification in fused images.

1. Introduction

Advancements in space technology have significantly increased both the diversity and quantity of remote sensors, resulting in an abundant array of remote sensing image resources [1]. Achieving both a multispectral and a wide instantaneous field of view in remote sensing images is challenging while maintaining a high signal-to-noise ratio. This limitation implies that a single remote sensor often cannot simultaneously capture images with both high spatial resolution and multispectral resolution [2]. Remote sensing image fusion technology integrates images with different spatial and spectral resolutions to produce a composite image that combines both high spatial and multispectral resolution. This technique enables the fused image to present detailed information with enhanced spatial clarity and spectral richness [3]. High-quality, quantitative remote sensing image fusion must not only achieve high spatial and multispectral resolution but also ensure accurate spectral quantification. This means that the pixel values must be computable to derive physical information about the Earth’s surface, such as radiance or reflectance, within the response wavelength range of the sensor. Quantitative fusion of panchromatic and multispectral color remote sensing images have widespread applications, offering reliable foundations for research in areas such as land cover change [4], terrestrial ecosystem monitoring [5], and the monitoring and classification of forests and crops [6,7]. Additionally, government agencies can leverage these high-quality fused images to develop policies that address environmental changes and ensure the sustainable management of land resources [8].
To achieve the quantitative fusion of remote sensing images, researchers have undertaken significant efforts. For example, in 2011 [9], an improved additive wavelet transform was introduced for fusion, capable of preserving both radiometric and geometric information. In 2012 [10], a multisensor image fusion method based on a hidden Markov tree and a pulse-coupled neural network (PCNN) was proposed. This method uses a PCNN to select the maximum value for low-pass coefficients and a saliency-based rule for directional coefficients, addressing minor distortions in structural components. In 2013 [11], the authors proposed a fusion method based on variational wavelets, which perform well with highly heterogeneous data. In 2014 [12], the authors explored contourlet representation and introduced an adjustable contourlet transform for effective fusion, averaging the low-pass coefficients and selecting the absolute maximum value for directional coefficients. This method corrects minor structural distortions and radiometric blurring. In 2015 [13], a region division strategy was used in the shearlet domain for pansharpening, applying region-correlation-based fusion rules to all decomposed coefficients. Significant spatial enhancement was observed as a result of these region-based rules. In 2016 [14], a pansharpening method in the shearlet domain was proposed, considering regional correlation metrics and applying local-region-based fusion rules to approximation coefficients and gradient-based fusion rules to directional coefficients. The results demonstrate that this method effectively preserves structural and radiometric information. In 2018, [15] conducted the spatiotemporal fusion study using deep convolutional neural networks (CNNs) in the context of massive remote sensing data. In 2019, [16] presented a region-based fusion scheme for combining panchromatic, multispectral, and synthetic aperture radar images. Temporal data fusion and high spatial methods were used to generate synthetic Landsat imagery by combining Landsat and MODIS data. In 2021 [17], a new region-based fusion method combining non-subsampled contourlet transform (NSCT) and particle swarm optimization (PSO) was proposed. This method applies a maximum-based rule to the approximation layer, separates band-pass coefficients into smooth regions, and uses PSO for edge regions and a maximum-based rule for separated components. It performs well in spatial enhancement and mitigates minor blurring effects.
In recent years, deep-learning-based fusion methods have been categorized into three types based on the supervisory paradigms employed during the training process: unsupervised, self-supervised, and supervised approaches [18]. Supervised methods utilize ground truth values to guide the training processes, while unsupervised approaches construct loss functions by constraining the similarity between the fusion results and the source images. Self-supervised algorithms are commonly associated with the AutoEncoder (AE)-based framework. In 2020, [19] introduced the pansharpening generative adversarial network (Pan-GAN), the first method to explore the unsupervised fusion of multispectral and panchromatic images. This approach incorporates two discriminators that create adversarial relationships between the fusion result and the two source images, each assessing the fidelity of spectral and spatial information, respectively. Ref. [20] introduced a semantic-aware real-time image fusion network (SeAFusion) in the same year. This study employs a cascading approach that integrates a fusion module with a semantic segmentation module, allowing semantic loss to guide high-level semantic information to flow back to the fusion module. Additionally, it proposes methods such as gradient residual dense block (GRDB) to bridge the gap between image fusion and high-level vision tasks. The concept of self-supervised edge-attention guidance for image fusion (EAGIF) was proposed in [21], utilizing a coarse-to-fine deep architecture to learn multiscale features from multimodal images. It also designs an edge-guided attention mechanism based on these multiscale features to focus the fusion process on common structures, thereby enhancing detail recovery while attenuating noise. Supervised fusion methods, as proposed in [22,23,24,25], illustrate the advancements that deep learning has contributed to image fusion techniques. In 2017, [22] proposed the pansharpening deep network architecture (PanNet), which employs residual learning to shift network training to the high-frequency domain. This approach allows the network to focus on learning high-frequency structural information, thereby enhancing the spatial quality of the fusion results. Ref. [23] proposed the super-resolution-guided progressive pansharpening neural network (SRPPNN), which incorporates two specific structural designs: a super-resolution module and progressive learning. These features enable the network to continuously capture spatial details at various scales and progressively integrate them into the upsampled multispectral images. Ref. [24] proposed the gradient projection-based pansharpening neural network (GPPNN), which investigates generative models for panchromatic and multispectral images. This approach explores the spatial and spectral degradation processes and uses them as priors to guide the optimization of neural networks, thereby enhancing fusion performance. In 2022, [25] introduced GTP-PNet, which employs a specialized transformation network (TNet) to model the spectral degradation process. This method establishes a more accurate nonlinear regression relationship between multispectral and panchromatic images in the gradient domain. The nonlinear regression relationship is used as a prior to constrain the preservation of spatial structures, thereby ensuring a balance between spectral and spatial information.
The research discussed above primarily focuses on improving fusion results through various approaches, including advancements in image decomposition methods [9,11,12,14], the establishment of stringent criteria for selecting fusion coefficients [10,13,17], enhancing the accuracy of nonlinear spectral fitting [22,25], and optimizing detail and spectral quantitative guidance [19,20,21,23,24]. However, there are still some problems in quantitative remote sensing image fusion. The information in remote sensing images represents a combination of atmospheric and surface data. Atmospheric effects not only reduce contrast and blur detail textures in panchromatic images [1], but also distort spectral information and decrease the quantitative accuracy of multispectral color images [26]. As a result, the fused images obtained under these conditions exhibit blurred spatial details and errors in color spectral information, failing to meet the requirements for quantitative remote sensing image fusion. Most research considers atmospheric information as part of remote sensing image data and employs mathematical methods to develop updated and more efficient fusion algorithms. However, the impact of atmospheric effects is often overlooked. This neglect prevents significant improvements in the spatial detail and spectral quantification of fused images. Additionally, aerosol observation methods are relatively limited, and aerosols exhibit strong spatiotemporal heterogeneity [27]. This results in aerosol parameters that often do not align temporally and spatially with the remote sensing images, leading to residuals in atmospheric correction results [28].
The primary objective of this study is to explore methods for obtaining highly synchronous aerosol optical depth (AOD) and column water vapor (CWV) measurements and integrating these atmospheric parameters as source images for fusion with panchromatic and multispectral remote sensing images. This approach aims to mitigate the impact of atmospheric effects on the quantitative fusion of remote sensing images. The study examines the interaction between atmospheric correction and five widely used fusion methods: principal component analysis (PCA) [3], intensity-hue-saturation (IHS) [29], Laplacian pyramid (LP) [30], discrete wavelet transform (DWT) [4], and non-subsampled contourlet transform (NSCT) [31]. It investigates how factors such as the choice of fusion methods, atmospheric correction, the synchronization of atmospheric parameters, and the timing of atmospheric correction affect the quality of fused remote sensing images. The findings provide a valuable reference for establishing robust quantitative fusion processes. The data for this study were obtained from airborne flight experiments involving multiple remote sensors, enabling a systematic comparison of spatial detail and spectral quantitative accuracy in the fused images.

2. Materials and Methods

2.1. Image Fusion

This study employed five classic and widely used pixel-level remote sensing image fusion techniques, including the IHS and PCA fusion methods based on spatial domain techniques, as well as the LP, DWT, and NSCT fusion methods based on transform domain techniques. To ensure that the fused images achieved spectral quantitative accuracy and spatial resolution, and to enhance the comparability of different fusion methods, the fusion rules should ideally extract spatial details from the panchromatic images while deriving spectral information from the multispectral images [3].

2.1.1. Spatial-Domain-Based Image Fusion

The spatial domain techniques were based on the projection of a multispectral image into another space using a transformation that separates the spatial structure from the spectral information in different components. Then, the component containing the spatial structure was replaced with the panchromatic image. A greater correlation between the replaced component and the panchromatic image produces less distortion [32].
IHS and PCA are two commonly used spatial domain methods for image fusion. PCA-based image fusion calculates and combines the principal components of the original images to reduce data dimensionality and extract image information. While PCA effectively preserves the overall information of the image, it may lead to color distortion. In contrast, the IHS transformation decomposes a color image into three components: intensity, hue, and saturation, and performs fusion on the intensity component. The IHS method is effective in preserving color information but is prone to a loss of image details.
The fusion rules for the IHS and PCA methods follow the principle of component substitution, as illustrated in Figure 1; Image_P represents the panchromatic image, while Image_M denotes the multispectral image. The specific steps for fusion are as follows: (1) Perform PCA/IHS transformation on the R, G, and B bands of the multispectral image. (2) Conduct histogram matching between the first principal component or intensity component obtained from the PCA/IHS transformation and the panchromatic image, replacing the first principal component or intensity component with the histogram-matched panchromatic image. (3) Finally, apply the inverse PCA/IHS transformation to generate the fused image, designated as Image_F.

2.1.2. Transform-Domain-Based Image Fusion

Transform domain fusion methods involve the multiscale decomposition of images, generating low-frequency subband coefficients that contain most of the spectral information and high-frequency subband coefficients that capture most of the spatial details. LP, DWT, and NSCT are three classical transform domain fusion methods, with their primary difference lying in the level of image decomposition. The LP fusion method is relatively simple, employing Gaussian pyramid decomposition to express images at multiple scales and resolutions. However, LP exhibits insufficient directional decomposition and is prone to matrix artifacts. The DWT fusion method decomposes images into different scales and resolutions using two-dimensional separable discrete wavelets, offering good time-frequency analysis capabilities; however, DWT representation is not sparse. In contrast, NSCT iteratively decomposes images using the non-subsampled Laplacian pyramid (NSP) to generate images at various scales, followed by further multidirectional decomposition through the non-subsampled directional filter bank (NSDFB). As a result, NSCT provides image decomposition with multiscale, multidirectional, and translation-invariant properties.
Although the three fusion methods differ in their approaches and levels of image decomposition, the resulting subband coefficients can all be categorized into high-frequency and low-frequency. To ensure the comparability of the fused images, the multiscale decomposition level was uniformly set to three, meaning that the image underwent three iterations of scale decomposition. For low-frequency, the fusion rule involved retaining the low-frequency subband coefficients of the decomposed multispectral image. For high-frequency, to avoid artifacts and ensure a smooth transition between spatial details and spectral information, a fusion rule based on a matching degree was applied [33]. This rule was determined by the salience and matching degree of the high-frequency subband coefficients.
Figure 2 illustrates the fusion processes for the LP, DWT, and NSCT methods. The specific steps for fusion are as follows: (1) Perform LP/DWT/NSCT decomposition on the panchromatic and multispectral color images to obtain the subband coefficients for each image, with the R, G, and B channels of the multispectral color image decomposed separately. (2) Apply a matching degree-based fusion rule to combine the high-frequency subband coefficients that share the same decomposition scale and direction, resulting in the high-frequency subband coefficients of the fused image. The low-frequency subband coefficients from the color image are retained as the low-frequency subband coefficients of the fused image. (3) Reconstruct the fused high-frequency and low-frequency subband coefficients and perform the inverse LP/DWT/NSCT transformation to obtain the fused image, designated as Image_F.

2.1.3. Fusion of Images and Atmospheric Parameters

The purpose of fusing atmospheric parameters (AOD and CWV) with panchromatic and multispectral remote sensing images was to perform atmospheric correction, thereby mitigating or even eliminating atmospheric effects. The fusion rules were determined by the 6S atmospheric radiative transfer model [34,35], which establishes a representation of visible and near-infrared band reflectance in relation to the ground bidirectional reflectance distribution function (BRDF) at the top of the atmosphere, as shown in Equations (1) and (2):
ρ T O A = ρ 0 + T g [ A B ρ S + A C ρ ¯ + A D ρ ¯ + C D ρ ¯ ¯ + ( B + C ) ( A + D ) S ( ρ ¯ ¯ ) 2 1 S ρ ¯ ¯ ]
A = e τ / μ V ,         B = e τ / μ S ,         C = t d ( μ S ) ,         D = t d ( μ V )
where the variable ρ T O A represents the reflectance observed at the top of the atmosphere. ρ 0 denotes the atmospheric path radiance reflectance, while ρ S represents the surface target reflectance. S is the hemisphere reflectance in the downward direction of the lower atmosphere. μ S corresponds to the cosine of the solar zenith angle, and μ V is the cosine of the satellite zenith angle. The terms B and C signify the atmospheric transmittance for direct solar radiation and atmospheric diffuse radiation reaching the ground, respectively. Similarly, A and D represent the atmospheric transmittance for the direct reflectance from the ground to the sensor and the atmospheric transmittance for the diffuse radiation reaching the sensor from the ground. τ represents the atmospheric optical thickness. Additionally, ρ ¯ , ρ ¯ , and ρ ¯ ¯ stand for the hemisphere reflectance scattered by the atmosphere to the ground, the ground hemisphere reflectance scattered back into the atmosphere, and the ground hemisphere reflectance after two scattering events through both the atmosphere and the ground, respectively. These variables are dependent on atmospheric optical parameters and surface reflection characteristics. T g represents the absorption by atmospheric components such as ozone and water vapor in the visible and near-infrared spectra. It is evident that to determine the ground reflectance ρ S , a series of intermediate parameters must be obtained. Among these, A , B , C , D , S , T g , and ρ 0 are solely related to atmospheric conditions and are calculated by the 6S model based on AOD and CWV. ρ ¯ , ρ ¯ , and ρ ¯ ¯ are related to image pixels and are determined by the 6S model using the BRDF model.
To facilitate the rapid fusion of atmospheric parameter distribution images with remote sensing images, an atmospheric correction lookup table (LUT) was established [36]. The principle of the LUT is to pre-compute and store the results of radiative transfer models under various conditions, enabling quick lookup and application during actual processing. The parameters setting of 6S model are shown in the Table 1. A range of values was defined for observation geometry, AOD, and CWV, with step values set to iterate through the specified parameter values. The 6S model was then run iteratively, traversing through predefined parameter settings to generate a lookup table for parameters a, b, and c. The relationship between image pixel values before and after atmospheric correction can be expressed using a computational formula based on these parameters, as shown in Equation (3):
o u t p u t = a × i n p u t b 1 + ( a × i n p u t b ) × c
The fusion process of atmospheric parameters (AOD and CWV) with the remote sensing image can be illustrated as shown in Figure 3. For each pixel, the corresponding AOD and CWV indices are used to extract the three parameters from the lookup table for fusion calculations, where ⊗ represents Equation (5). The values in the fused image represent the reflectance with atmospheric effects removed or minimized (which can also be expressed in terms of radiance). The accuracy and precision of the atmospheric parameters AOD and CWV determine the spatial detail clarity and spectral quantification of the fused remote sensing image.

2.2. Evaluation Metrics

This study categorized the widely used quality evaluation metrics for remote sensing image fusion into three types based on factors such as calculation methods and their physical significance:
  • Spectral domain: The evaluation metrics in the spectral domain are used to reflect the degree of spectral variation or distortion in the images. This can be represented by comparing the reflectance calculated from the image with the actual reflectance measured through experimental means, namely the deviation from true reflectance (DTR).
  • Spatial domain: The evaluation metrics in the spatial domain assess spatial details such as edges and textures within the image. These can be represented through structural and feature-based measures, including the average gradient (AG) [17], standard deviation (SD) [37], normalized correlation coefficient (NCC) [38], Structural SIMilarity (SSIM) [38], and universal image quality index (UIQI) [39].
  • Informational domain: The evaluation metrics in the information domain are used to infer the amount of information contained in the images. This can be expressed through statistical indicators of the images and relevant measures from information theory, such as the normalized mutual information (NMI) [40] and information entropy (EN) [37].
Next, the calculation methods and significance of each evaluation metric will be introduced. Table 2 presents the attributes of these evaluation metrics.
  • Average Gradient
The definition formula of AG is shown in Equation (4), where F represents the fused image, and the dimensions of the image are M × N. Generally, a higher average gradient in the fused image indicates greater contrast, which is typically associated with better fusion quality.
A G = 1 ( M 1 ) ( N 1 ) i = 1 M 1 j = 1 N 1 [ F ( i + 1 , j ) F ( i , j ) ] 2 + [ F ( i , j + 1 ) F ( i , j ) ] 2 2
2.
Standard Deviation
The calculation formula of SD is shown in Equation (5), where F ¯ represents the mean gray value of the fused image. A larger SD in the fused image suggests a more dispersed distribution of gray levels, which is generally considered indicative of better image quality.
S D = 1 M × N i = 1 M 1 j = 1 N 1 [ F ( i , j ) F ¯ ] 2
3.
Entropy
The calculation formula of EN is shown in Equation (6), where p represents the gray level distribution of the fused image, n is the dynamic range of the pixels, and p i is the probability of a pixel in the fused image having a gray level of i. A higher EN in the fused image indicates a greater amount of information contained within it, which is generally regarded as a sign of better image quality.
E N = i = 1 n p i log 2 p i
4.
Normalized Correlation Coefficient
The definition of NCC is given by Equation (7), where R represents the image before fusion and R ¯ represents the mean value. The NCC compares the two images before and after fusion; a higher NCC value indicates a greater correlation between the two images.
N C C = i = 1 M j = 1 N [ ( R ( i , j ) R ¯ ) ( F ( i , j ) F ¯ ) ] i = 1 M j = 1 N ( R ( i , j ) R ¯ ) 2 i = 1 M j = 1 N ( F ( i , j ) F ¯ ) 2
5.
Structural Similarity
The SSIM evaluates image performance by comparing differences in brightness, contrast, and texture distortion between images. The comparison functions for brightness, contrast, and structure are presented in Equation (8). Constants C 1 , C 2 , and C 3 in Equation (9) are introduced to ensure stability when the denominator approaches zero, and are related to the dynamic range of the remote sensing image pixels, n . The standard deviations of the image before fusion and the fused image are denoted as σ R and σ F , respectively, while σ R F represents the covariance between the image before fusion and the fused image.
l ( R , F ) = 2 R ¯ F ¯ + C 1 R ¯ 2 + F ¯ 2 + C 1 ,         c ( R , F ) = 2 σ R σ F + C 2 σ R 2 + σ F 2 + C 2 ,   s ( R , F ) = 2 σ R F + C 3 σ R σ F + C 3
C 1 = ( 0.01 × n ) 2 ,         C 2 = ( 0.03 × n ) 2 ,         C 3 = C 2 2
Combining Equations (8) and (9), the SSIM function can be derived as shown in Equation (10):
S S I M ( R , F ) = [ l ( R , F ) ] α [ c ( R , F ) ] β [ s ( R , F ) ] γ
α , β , and γ represent the relative importance of each metric and sum to 1. The SSIM compares the two images before and after fusion; a higher SSIM value indicates greater structural similarity between the two images.
6.
Normalized Mutual Information
The definition of MI is given by Equation (11), where q represents the distribution of the image before fusion R , and q i denotes the probability that a pixel in R has a grayscale value of i . Additionally, γ i , j represents the joint probability density of R and F .
M I = i = 1 n j = 1 n γ i , j log 2 γ i , j p i q i
The normalization method of MI is detailed in Equation (12). It compares the two images before and after fusion; a higher NMI value indicates a greater degree of dependence between the two images.
N M I ( R , F ) = 2 M I ( R , F ) i = 1 n p i log 2 p i j = 1 n q j log 2 q j
7.
Universal Image Quality Index
Similar to the SSIM, the UIQI also consists of three components that reflect the loss of correlation between signals, luminance distortion, and contrast distortion. The UIQI between the two images is represented by Equation (13):
U I Q I ( R , F ) = 2 σ R F R ¯ F ¯ ( σ R 2 + σ F 2 ) ( R ¯ 2 + F ¯ 2 )
By conducting a region-based evaluation of the image, the formula can be expressed as Equation (14):
U I Q I ( R , F ) = 1 | W | U I Q I ( R , F | w )
where W represents the sum of all windows, | W | denotes the cardinality of W , and U I Q I ( R , F | w ) indicates the similarity between images R and F within window w , as calculated by Equation (13). Next, by introducing the two images prior to fusion, we calculate the overall similarity between the fused image F and the images P and R before fusion, as shown in Equation (15):
U I Q I ( P , R , F ) = 1 | W | w W [ λ ( w ) U I Q I ( P , F | w ) + ( 1 λ ( w ) ) U I Q I ( R , F | w ) ]
where the formula for calculating the weight λ ( w ) is given in Equation (16):
λ ( w ) = s ( P | w ) s ( P | w ) + s ( R | w )
and s represents the saliency measure of the image within the window, which in this paper, is expressed by the image’s contrast. UIQI compares the three images before and after fusion; a higher value indicates a greater overall similarity between the fused image and the two images prior to fusion.
8.
Deviation from True Reflectance
Define DTR as the percentage difference between the true surface reflectance and the reflectance calculated from the image, as shown in Equation (17):
η ( R s u r f a c e , F ) = | 1 M N i = 1 , j = 1 M , N F ( i , j ) R s u r f a c e | R s u r f a c e × 100 %
Here, R s u r f a c e represents the measured reflectance at the surface, and (i, j) indicates the coordinates of the target in the remote sensing image. This study uses the measurements obtained from the ground-based Analytical Spectral Devices (ASDs) as the true reflectance.

2.3. Data

The aerial flight experiments were conducted to acquire remote sensing images. The instruments utilized in these experiments included the visible near infrared airborne synchronous monitoring atmospheric corrector (VNIR-ASMAC), the short wave infrared ASMAC (SWIR-ASMAC), a high-resolution panchromatic camera, a multispectral color camera, a gyroscope, GPS, the CE318 sun photometer, and ASDs. To achieve synchronous imaging across the various remote sensors, the pulse signals generated during imaging by VNIR-ASMAC and SWIR-ASMAC served as the reference. These signals were subsequently transmitted through hardware to trigger imaging in both the high-resolution panchromatic camera and the multispectral color camera. Figure 4 illustrates some of the images obtained during the aerial flight experiments.

2.3.1. Panchromatic and Multispectral Color Remote Sensing Images

The main cameras used in this experiment were the panchromatic camera, model LBAS-U350-74M, and the multispectral color camera, model LBAS-U350-74C, both manufactured by LUSTER. The main distinction between the two cameras lies in the color camera’s incorporation of the Bayer filter, which enables the acquisition images of the R, G, and B band. Figure 5a,b provides physical diagrams of the cameras. Both cameras underwent geometric calibration, radiometric calibration, and spectral response testing, enabling the conversion of the digital number (DN) values of remote sensing images into radiance values corresponding to their respective wavelengths.
The primary cameras captured panchromatic and multispectral color images of three scenes, as shown in Figure 6. The red stars indicate the locations where ground personnel measured reflectance using ASDs, as shown in Figure 5c and detailed in Table 3.

2.3.2. Atmospheric Parameter Images

Figure 7a,b provides physical diagrams of the VNIR-ASMAC and SWIR-ASMAC. The VNIR-ASMAC included three non-polarized bands (765, 910, and 950 nm) and three polarized bands (490, 670, and 870 nm). Similarly, the SWIR-ASMAC featured two non-polarized bands (950 and 1380 nm) and one polarized band (1610 nm). The multiangle, multiband polarimetric imaging capability of the ASMAC enabled the real-time monitoring of aerosol distribution in the atmosphere. The ASMAC underwent rigorous laboratory calibration, with each pixel subjected to geometric, radiometric, and polarization calibration, effectively treating each pixel as a radiance meter. Therefore, the remote sensing images captured by ASMAC can be inverted into AOD and CWV images [36].
Figure 8 illustrates the distribution images of AOD and CWV (already registered) corresponding to the remote sensing scenes, with their average and extremum values across the entire image plane listed in the Table 4. This indicates that each pixel in the panchromatic and multispectral images was associated with atmospheric parameters that are strictly synchronous in both time and space, thereby minimizing the impact of spatiotemporal heterogeneity in atmospheric parameters.
Single-value atmospheric parameters representing asynchronous data were obtained from the CE318, as illustrated in Figure 7c. The single-value AOD and CWV were selected based on the acquisition time of the image, as shown in Table 5.

2.4. Fusion Frameworks

The atmospheric parameter distribution images were used as source images and fused with panchromatic and multispectral remote sensing images. The fusion framework, as shown in Figure 9, was designed to explore various factors influencing the performance of the fused images. Four fusion experiments were conducted to investigate the effects of different fusion methods, atmospheric correction, the synchronization of atmospheric parameters, and the timing of atmospheric correction:
  • Fusion 1: Disconnect all switches and perform decomposition or transformation only on the panchromatic and color remote sensing images. Then, select the fusion coefficients based on the fusion rules and reconstruct the fused image.
  • Fusion 2: Connect the switch for Fusion 2, while keeping others disconnected. Single-value atmospheric parameter LUTs are used to perform atmospheric correction on the panchromatic and color remote sensing images, followed by decomposition or transformation. Then, select the fusion coefficients based on the fusion rules and reconstruct the fused image.
  • Fusion 3: Connect the switch for Fusion 3, while keeping others disconnected. Synchronous atmospheric parameter distribution LUTs are used to apply imaging-based synchronous atmospheric correction to the panchromatic and color remote sensing images, followed by decomposition or transformation. Then, select the fusion coefficients based on the fusion rules and reconstruct the fused image.
  • Fusion 4: Connect the switch for Fusion 4, while keeping others disconnected. Perform decomposition or transformation only on the panchromatic and color remote sensing images. Then, select the fusion coefficients based on the fusion rules and reconstruct the fused image. Finally, an imaging-based synchronous atmospheric correction is applied to the fused image.
Fusion 1 was derived solely from panchromatic and color remote sensing images. The evaluation results of this fused image can be used to assess the impact of different decomposition or transformation methods on the fusion outcome. The key difference between Fusion 1 and Fusion 2 was the inclusion of atmospheric parameters in the image sources. The evaluation results of these fused images can be used to compare the effect of atmospheric correction on the fusion outcome. The distinction between Fusion 2 and Fusion 3 lies in whether the atmospheric parameter distribution image was temporally and spatially synchronous with the panchromatic and color remote sensing images. These results can be used to analyze the effect of atmospheric parameter synchronization on the fusion outcome. Finally, the difference between Fusion 3 and Fusion 4 was the timing of the inclusion of atmospheric parameter distribution images in the fusion process. The evaluation results of these fused images can be used to compare the impact of atmospheric correction timing on fusion performance.

3. Results

3.1. Fused Image

Using the fusion methods and rules provided in Section 2.1, and following the fusion framework described in Section 2.4, panchromatic images, multispectral color images, and atmospheric parameter images from the three scenes described in Section 2.3 were fused. The resulting fused images are presented in Figure 10, Figure 11, Figure 12 and Figure 13. The figures, from left to right, are the fused images obtained using the PCA, IHS, DWT, NSCT, and LP fusion methods, respectively. Panels a–e depict the artificial target scene, f–j represent the road and grass scene, and k–o illustrate the wheat field scene.

3.2. Spatial Detail Evaluation

This section provides an objective analysis of the fused images and evaluates them using the metrics described in Section 2.2.

3.2.1. Effects of Fusion Methods on Spatial Detail

Table 6 presents the evaluation values of UIQI, NCC, NMI, SSIM, and AG for the fused images obtained through five different fusion methods under Fusion 1. This was performed to compare the impact of the choice of fusion method on the spatial details of the fused images. The evaluation values represent the mean values across the RGB channels. Additionally, the extreme value ratio for each metric was calculated, indicating the ratio of the maximum to minimum evaluation values obtained through the different fusion methods.

3.2.2. Effects of Atmospheric Correction and Parameter Synchronization on Spatial Detail

Table 7 presents the evaluation values of AG, SD, and EN for the fused images obtained through five different fusion methods under Fusion 1, Fusion 2, and Fusion 3. Since the source images in Fusion 1, Fusion 2, and Fusion 3 differ, using source-based evaluation metrics like UIQI, NCC, NMI, and SSIM are not meaningful. Instead, AG, SD, and EN, which are based on the statistical features of the fused images themselves, are employed, making them more comparable. Fusion 2 and Fusion 3 are utilized to assess the impact of atmospheric parameter synchronization on the spatial details of the fused images, while Fusion 1 and Fusion 3 are used to compare the effects of atmospheric correction on these details.

3.2.3. Effects of Atmospheric Correction Timing on Spatial Detail

Similar to Table 7, Table 8 presents the evaluation values of AG, SD, and EN for the fused images obtained through five different fusion methods under Fusion 3 and Fusion 4. These metrics are utilized to compare the impact of the timing of atmospheric correction on the spatial details of the fused images.

3.3. Spectral Quantification Evaluation

3.3.1. Effects of Fusion Methods on Spectral Quantification

Table 9 presents the DTR for the fusion images of the artificial targets and the roads and grassland scenes using five different fusion methods under Fusion 1. These values are used to assess the impact of the fusion methods on the spectral quantification of the fused image. DTR is expressed as a percentage, where smaller values indicate a closer approximation to the true reflectance. Similarly, the extreme values of DTR were also calculated.

3.3.2. Effects of Atmospheric Correction and Parameter Synchronization on Spectral Quantification

Due to the inconsistent performance of IHS and PCA in improving spectral quantification (as discussed in Section 4.1), only transform-domain-based fusion methods are considered in the discussion. Table 10 presents the DTR values for the fused images of artificial targets, roads, and grassland scenes, obtained using three transform-domain-based fusion methods within Fusion 1, Fusion 2, and Fusion 3. These values are used to evaluate the impact of atmospheric parameters and their synchronization on the spectral quantification of the fused images.

3.3.3. Effects of Atmospheric Correction Timing on Spectral Quantification

Similar to Table 10, Table 11 presents the DTR values for the fused images of artificial targets, roads, and grassland scenes, obtained using three transform-domain-based fusion methods within Fusion 3 and Fusion 4. These values are used to evaluate the impact of the timing of atmospheric correction on the spectral quantification of the fused images.

4. Discussion

4.1. Effects of Fusion Methods

Table 6 presents the UIQI, NCC, NMI, SSIM, and AG values for fused images under Fusion 1. These metrics are used to compare the impact of different fusion methods on the spatial details of the fused images. The findings summarized from Table 6 are as follows:
  • In the three scenarios examined, the NSCT-based fusion method consistently demonstrates the best evaluation values. Generally, transform-domain-based fusion methods perform better than spatial-domain-based methods.
  • Among the three transform-domain-based fusion methods, NSCT outperforms DWT, and DWT performs better than LP.
  • The ratio of the extreme values indicates that, for the artificial target scene, selecting an appropriate fusion method can enhance the spatial details of the fused images by a factor of 1.149 to 2.739. For road and grassland scenes, the enhancement ranges from 1.056 to 2.597 times, while for the wheat field scene, the enhancement ranges from 1.057 to 2.111 times.
Spatial-domain fusion methods involve mathematical or statistical computations on pixels, which can result in a decreased contrast in the fused image and may not effectively preserve the detailed features of the original images. This is due to the less refined nature of image decomposition in the spatial domain, making it theoretically challenging to achieve satisfactory fusion results. In contrast, transform-domain fusion methods utilize multiscale decomposition techniques to represent images at various scales, resolutions, and orientations. This approach allows the fusion process to occur in the transform domain, offering a greater variety of optional fusion rules and finer control. These characteristics enable transform-domain fusion to preserve spatial information from the original images as much as possible. The enhancement of spatial details in fused images depends on the “fineness” of the image decomposition. Consequently, the NSCT method, which achieves a higher degree of image decomposition, demonstrates relatively superior fusion performance.
Table 9 presents the DTR values for fused images under the Fusion 1 framework, used to compare the impact of different fusion methods on the spectral quantification of the fused images. The results in Table 9 indicate that PCA-based fused images exhibit the smallest DTR for black targets, while IHS-based fused images show the smallest DTR for white targets and cement surfaces. Conversely, PCA-based fused images have the largest DTR for white targets, and IHS-based fused images demonstrate the largest DTR for black targets. This suggests that the DTR stability of PCA and IHS fusion methods is relatively poor. In contrast, the DTR stability for images fused using the DWT, NSCT, and LP methods is superior, making the evaluation of transform-domain fusion methods for spectral quantification more meaningful and practical. The summarized DTR results for transform-domain fusion methods from Table 9 are as follows:
  • For black targets, selecting an appropriate fusion method can reduce the DTR of the fused images from a maximum of 193.26% to 175.20% (G channel);
  • For white targets, selecting an appropriate fusion method can reduce the DTR from a maximum of 40.77% to 36.70% (B channel);
  • For cement surfaces, selecting an appropriate fusion method can reduce the DTR from a maximum of 41.92% to 35.02% (G channel);
  • For the grassland, selecting an appropriate fusion method can reduce the DTR from a maximum of 68.41% to 51.97% (B channel).
Due to the complexity of image edges and texture information, which involve various land cover components, spectral quantification is often compared using flat, homogeneous land targets. In such cases, the spectral quantification of fused images primarily depends on the low-frequency signals. Transform domain fusion methods effectively separate low-frequency and high-frequency signals in a stable and detailed manner, resulting in better DTR stability for images fused using the DWT, NSCT, and LP methods.

4.2. Effects of Atmospheric Correction and Synchronous of Parameters

Table 7 presents the AG, SD, and EN values for fused images under Fusion 1, Fusion 2, and Fusion 3, which are used to compare the impact of atmospheric parameters and their synchronization on the spatial details of the fused images. The findings summarized from Table 7 are as follows:
  • The evaluation values for Fusion 1 are the lowest among all fusion methods, indicating that introducing atmospheric parameters into the fusion process can effectively enhance the spatial details of the fused images.
  • The highest evaluation values for most fusion methods are observed in Fusion 3, while PCA and IHS fusion images show a few highest values in Fusion 2. This indicates that synchronous atmospheric parameters have the greatest impact on improving the spatial details of the fused images.
  • The ratio of the extreme values indicates that, for the artificial target scene, the synchronous atmospheric parameter can enhance the spatial details of the fused images by 1.04 to 2.03 times. For road and grassland scenes, the enhancement ranges from 1.02 to 1.31 times. For the wheat field scene, the enhancement ranges from 1.070 to 1.380 times.
Atmospheric effects can cause blurring in remote sensing images, leading to unclear spatial details and textures. By incorporating atmospheric parameters into the fusion process, these effects can be mitigated, and the contrast of the fused images can be improved. Synchronous atmospheric parameter maps are particularly effective in addressing the spatiotemporal heterogeneity of aerosols, resulting in a more significant enhancement of spatial details.
Table 10 presents the DTR values for fused images under Fusion 1, Fusion 2, and Fusion 3, which are used to compare the impact of atmospheric parameters and their synchronization on the spectral quantification of the fused images. The results in Table 10 indicate that Fusion 1 exhibits the highest DTR values, while Fusion 3 shows the lowest DTR values for most cases. Fusion 2 has the lowest DTR values for a few grassland scenarios. The summarized results from Table 10 are as follows:
  • For black targets, synchronous atmospheric parameters can reduce the DTR of the fused images from a maximum of 163.7% to 9.4% (G channel);
  • For white targets, synchronous atmospheric parameters can reduce the DTR from a maximum of 36.7% to 6.3% (B channel);
  • For cement surfaces, synchronous atmospheric parameters can reduce the DTR from a maximum of 35.0% to 10.0% (G channel);
  • For the grassland, atmospheric parameters can reduce the DTR of the fused images from a maximum of 51.9% to 5.4% (B channel).
Atmospheric effects are one of the primary factors causing spectral distortion in remote sensing images. Fusion processes can introduce these distorted signals into the fused images, and strict fusion rules alone cannot eliminate atmospheric information. By incorporating atmospheric parameters as sources in the fusion process, these effects can be effectively mitigated or even eliminated. Synchronous atmospheric parameter maps further address the spatiotemporal heterogeneity of aerosols, leading to enhanced spectral quantification in the fused images.

4.3. Effects of the Timing of Atmospheric Correction

Table 8 presents the AG, SD, and EN values for fused images under Fusion 3 and Fusion 4, which are used to compare the impact of the timing of atmospheric correction on the spatial details of the fused images. The results in Table 8 indicate that Fusion 4 consistently shows lower evaluation values than Fusion 3 across all fusion methods, suggesting that performing atmospheric correction before image fusion leads to a greater enhancement of spatial details in the fused images.
Table 11 presents the DTR values for fused images under Fusion 3 and Fusion 4, which are used to compare the impact of the timing of atmospheric correction on the spectral quantification of the fused images. The results in Table 11 indicate that the DTR values for all land cover targets computed in Fusion 3 are lower than those in Fusion 4, suggesting that performing atmospheric correction prior to image fusion leads to a greater improvement in the spectral quantification of the fused images.

5. Conclusions

This study investigates the impact of 6S atmospheric correction combined with five image fusion methods (PCA, IHS, DWT, LP, and NSCT) on the performance of fused images, focusing on four aspects: the choice of fusion method, the implementation of atmospheric correction, the synchronization of atmospheric parameters, and the timing of atmospheric correction.
The experimental results for each fusion method indicate that the choice of fusion technique significantly affects the spatial details of remote sensing fused images. In three experimental scenarios, selecting an appropriate fusion method can enhance the spatial detail evaluation values of the fused images by factors of 1.149 to 2.739, 1.056 to 2.597, and 1.057 to 2.111, respectively. However, the choice of fusion method has a relatively minor impact on the spectral quantification of remote sensing images. The maximum reduction in DTR for black targets, white targets, cement surfaces, and the grassland was observed to be as follows: from 193.26% to 175.20%, from 40.77% to 36.70%, from 41.92% to 35.02%, and from 68.41% to 51.97%, respectively.
Experiments employing different atmospheric parameters for atmospheric correction demonstrate that atmospheric correction improves the performance of fused images across all methods, with synchronized atmospheric parameter maps yielding better results than single-value atmospheric parameters. In the three experimental scenarios, atmospheric correction enhances the spatial details of PCA, IHS, DWT, NSCT, and LP fused images by factors ranging from 1.12 to 1.38, 1.06 to 1.98, 1.03 to 1.89, 1.02 to 1.73, and 1.06 to 2.03, respectively. Additionally, atmospheric correction reduces the DTR for four types of features (black targets, white targets, cement surfaces, and grassland), with maximum reductions of the following: from 163.7% to 9.4%, from 36.7% to 6.3%, from 35.0% to 10.0%, and from 51.9% to 5.4%. Furthermore, experiments conducted at different timings for atmospheric correction indicate that placing the atmospheric correction step before the fusion step yields better results.
In summary, integrating synchronized atmospheric parameter distribution images into the fusion framework prior to remote sensing image fusion, along with selecting fusion methods based on multiscale decomposition, such as NSCT, can significantly enhance both the spatial details and spectral quantification of fused images. The findings of this study provide a technical pathway for achieving the high-quality quantitative fusion of remote sensing images.
While this research demonstrates the impact of 6S atmospheric correction and five fusion techniques on the performance of fused images, future work will explore deep-learning-based methods. Deep learning approaches not only hold promise for image fusion but may also address challenges related to atmospheric correction and aerosol parameter inversion, necessitating further investigation.

Author Contributions

Conceptualization, Z.Q.; methodology, Y.L.; validation, F.C., T.S. and Y.L.; formal analysis, Y.L.; investigation, Y.L., F.C., T.S., W.C. and R.T.; resources, Y.L.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L.; visualization, Y.L.; supervision, Y.L.; project administration, Z.Q. and J.H.; funding acquisition, Z.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Chinese Civil Aerospace Program, grant number D010206.

Data Availability Statement

The dataset is available on request from the authors.

Acknowledgments

Yang Li is very grateful to Rufang Ti for her help on the aerosol retrieval algorithm and Tangyu Sui for his help on the atmospheric correction.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, Y.; Qiu, Z.; Chen, F.; Sui, T.; Ti, R.; Cheng, W.; Hong, J. Remote sensing image registration method based on synchronous atmospheric correction. Opt. Express 2024, 32, 24573–24591. [Google Scholar] [CrossRef]
  2. Zhang, H.; Ma, J. SDNet: A Versatile Squeeze-and-Decomposition Network for Real-Time Image Fusion. Int. J. Comput. Vis. 2021, 129, 2761–2785. [Google Scholar] [CrossRef]
  3. Singh, S.; Singh, H.; Bueno, G.; Deniz, O.; Singh, S.; Monga, H.; Hrisheekesha, P.N.; Pedraza, A. A review of image fusion: Methods, applications and performance metrics. Digit. Signal Process. 2023, 137, 104020. [Google Scholar] [CrossRef]
  4. Chughtai, A.H.; Abbasi, H.; Ismail, R.K. A review on change detection method and accuracy assessment for land use land cover. Remote Sens. Appl. Soc. Environ. 2021, 22, 100482. [Google Scholar] [CrossRef]
  5. Yang, Y.; Anderson, M.C.; Gao, F.; Wood, J.D.; Gu, L.; Hain, C. Studying drought-induced forest mortality using high spatiotemporal resolution evapotranspiration data from thermal satellite imaging. Remote Sens. Environ. 2021, 265, 112640. [Google Scholar] [CrossRef]
  6. Nguyen, T.-A.; Kellenberger, B.; Tuia, D. Mapping forest in the Swiss Alps treeline ecotone with explainable deep learning. Remote Sens. Environ. 2022, 281, 113217. [Google Scholar] [CrossRef]
  7. Kavitha, A.V.; Srikrishna, A.; Satyanarayana, C. Crop image classification using spherical contact distributions from remote sensing images. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 534–545. [Google Scholar] [CrossRef]
  8. Al-Doski, J.; Hassan, F.M.; Norman, M.; Najim, A.A. Interaction of image fusion techniques and atmospheric correction for improve SVM accuracy. Earth Sci. Inform. 2022, 15, 2673–2687. [Google Scholar] [CrossRef]
  9. Kim, Y.; Lee, C.; Han, D.; Kim, Y.; Kim, Y. Improved Additive-Wavelet Image Fusion. IEEE Geosci. Remote Sens. Lett. 2011, 8, 263–267. [Google Scholar] [CrossRef]
  10. Yang, S.; Wang, M.; Jiao, L. Contourlet hidden Markov Tree and clarity-saliency driven PCNN based remote sensing images fusion. Appl. Soft Comput. 2012, 12, 228–237. [Google Scholar] [CrossRef]
  11. Fang, F.; Li, F.; Shen, C.; Zhang, G. A Variational Approach for Pan-Sharpening. IEEE Trans. Image Process. A Publ. IEEE Signal Process. Soc. 2013, 22, 2822–2834. [Google Scholar] [CrossRef] [PubMed]
  12. El-Mezouar, M.C.; Kpalma, K.; Taleb, N.; Ronsin, J. A Pan-Sharpening Based on the Non-Subsampled Contourlet Transform: Application to Worldview-2 Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1806–1815. [Google Scholar] [CrossRef]
  13. Shi, C.; Liu, F.; Miao, Q. Pan-sharpening via regional division and NSST. Multimed. Tools Appl. 2015, 74, 7843–7857. [Google Scholar] [CrossRef]
  14. Poobalasubramanian, M.; Agrawal, A. Fusion of PAN and multispectral remote sensing images in shearlet domain by considering regional metrics. J. Appl. Remote Sens. 2016, 10, 045003. [Google Scholar] [CrossRef]
  15. Song, H.; Liu, Q.; Wang, G.; Hang, R.; Huang, B. Spatiotemporal Satellite Image Fusion Using Deep Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 821–829. [Google Scholar] [CrossRef]
  16. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  17. Singh, H.; Cristobal, G.; Blanco, S.; Bueno, G.; Sanchez, C. Nonsubsampled contourlet transform based tone mapping operator to optimize the dynamic range of diatom shells. Microsc. Res. Tech. 2021, 84, 2034–2045. [Google Scholar] [CrossRef]
  18. Tang, L.; Zhang, H.; Xu, H.; Ma, J. Deep learning-based image fusion: A survey. J. Image Graph. 2023, 28, 3–36. [Google Scholar] [CrossRef]
  19. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  20. Tang, L.; Yuan, J.; Ma, J. Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network. Inf. Fusion 2022, 82, 28–42. [Google Scholar] [CrossRef]
  21. Liu, J.; Fan, X.; Jiang, J.; Liu, R.; Luo, Z. Learning a Deep Multi-Scale Feature Ensemble and an Edge-Attention Guidance for Image Fusion. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 105–119. [Google Scholar] [CrossRef]
  22. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A Deep Network Architecture for Pan-Sharpening. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  23. Cai, J.; Huang, B. Super-Resolution-Guided Progressive Pansharpening Based on a Deep Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1–15. [Google Scholar] [CrossRef]
  24. Xu, S.; Zhang, J.; Zhao, Z.; Sun, K.; Liu, J.; Zhang, C. Deep Gradient Projection Networks for Pan-Sharpening. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  25. Zhang, H.; Ma, J. GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening. ISPRS J. Photogramm. Remote Sens. 2021, 172, 223–239. [Google Scholar] [CrossRef]
  26. Yin, F.; Gomez-Dans, J.; Lewis, P. A Sensor Invariant Atmospheric Correction Method for Satellite Images. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018. [Google Scholar]
  27. Doxani, G.; Vermote, E.; Roger, J.C. Atmospheric Correction Inter-Comparison Exercise. Remote Sens. 2018, 10, 352. [Google Scholar] [CrossRef]
  28. Wang, T.; Zhou, N.; Yi, W. Adjacency effect correction of optical satellite image with sub-meter spatial resolution. Acta Phys. Sin. 2021, 70, 139101-1–139101-8. [Google Scholar] [CrossRef]
  29. Ankarao, V.; Sowmya, V.; Soman, K.P. Multi-sensor data fusion using NIHS transform and decomposition algorithms. Multimed. Tools Appl. 2018, 77, 30381–30402. [Google Scholar] [CrossRef]
  30. Shivsubramani, K.; Soman, K.P. Implementation and Comparative Study of Image Fusion Algorithms. Int. J. Comput. Appl. 2010, 9, 10–20. [Google Scholar]
  31. Alseelawi, N.; Tuama Hazim, H.; Alrikabi, H.T.S. A Novel Method of Multimodal Medical Image Fusion Based on Hybrid Approach of NSCT and DTCWT. Int. J. Online Biomed. Eng. (IJOE) 2022, 18, 114–133. [Google Scholar] [CrossRef]
  32. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  33. Zhang, P. Image Fusion Algorithm Based on Region Energy and Block Matching. Master’s Thesis, Beijing Jiaotong University, Beijing, China, 2018. [Google Scholar]
  34. Tanre, D.; Deuze, J.L.; Herman, M.; Morcette, J.J. Second Simulation of the Satellite Signal in the Solar Spectrum, 6S: An overview. IEEE Trans. Geosci. Remote Sens. 2002, 35, 675–686. [Google Scholar]
  35. Wilson, T.R. Py6S: A Python interface to the 6S radiative transfer model. Comput. Geoences 2013, 51, 166–171. [Google Scholar] [CrossRef]
  36. Fan, Y.; Sun, X.; Ti, R.; Huang, H.; Liu, X.; Yu, H. Aerosol Retrieval Study from a Particulate Observing Scanning Polarimeter Onboard Gao-Fen 5B without Prior Surface Knowledge, Based on the Optimal Estimation Method. Remote Sens. 2023, 15, 385. [Google Scholar] [CrossRef]
  37. Wang, W.; Chang, F. A Multi-focus Image Fusion Method Based on Laplacian Pyramid. J. Comput. 2011, 6, 2559–2566. [Google Scholar] [CrossRef]
  38. Kulkarni, S.C.; Rege, P.P. Pixel level fusion techniques for SAR and optical images: A review. Inf. Fusion 2020, 59, 13–29. [Google Scholar] [CrossRef]
  39. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  40. Tsai, D.Y.; Lee, Y.; Matsuyama, E. Information entropy measure for evaluation of image quality. J. Digit. Imaging 2008, 21, 338–347. [Google Scholar] [CrossRef]
Figure 1. Fusion diagram based on IHS and PCA methods.
Figure 1. Fusion diagram based on IHS and PCA methods.
Remotesensing 16 03916 g001
Figure 2. Fusion diagram based on LP, DWT, and NSCT methods.
Figure 2. Fusion diagram based on LP, DWT, and NSCT methods.
Remotesensing 16 03916 g002
Figure 3. Atmospheric parameters and remote sensing image fusion.
Figure 3. Atmospheric parameters and remote sensing image fusion.
Remotesensing 16 03916 g003
Figure 4. Aerial flight experiments: (a) ground operation; (b) in-helicopter testing; (c) low-altitude aerial experiment; (d) high-altitude aerial experiment; (e) data collection by CE318; (f) data collection by ASDs.
Figure 4. Aerial flight experiments: (a) ground operation; (b) in-helicopter testing; (c) low-altitude aerial experiment; (d) high-altitude aerial experiment; (e) data collection by CE318; (f) data collection by ASDs.
Remotesensing 16 03916 g004
Figure 5. Main cameras devices on the ground: (a) panchromatic camera; (b) multispectral color camera; (c) ASDs.
Figure 5. Main cameras devices on the ground: (a) panchromatic camera; (b) multispectral color camera; (c) ASDs.
Remotesensing 16 03916 g005
Figure 6. Panchromatic (right) and multispectral color (left) image: (a) artificial target scene; (b) road and grass scene; (c) cornfield scene.
Figure 6. Panchromatic (right) and multispectral color (left) image: (a) artificial target scene; (b) road and grass scene; (c) cornfield scene.
Remotesensing 16 03916 g006
Figure 7. Atmospheric measurement instrument: (a) VNIR-ASMAC; (b) SWIR-ASMAC; (c) CE318.
Figure 7. Atmospheric measurement instrument: (a) VNIR-ASMAC; (b) SWIR-ASMAC; (c) CE318.
Remotesensing 16 03916 g007
Figure 8. Synchronous AOD (left) and CWV (right) distribution images: (a) artificial target scene; (b) road and grass scene; (c) cornfield scene.
Figure 8. Synchronous AOD (left) and CWV (right) distribution images: (a) artificial target scene; (b) road and grass scene; (c) cornfield scene.
Remotesensing 16 03916 g008
Figure 9. Experimental diagram of the interaction between atmospheric correction and fusion techniques.
Figure 9. Experimental diagram of the interaction between atmospheric correction and fusion techniques.
Remotesensing 16 03916 g009
Figure 10. The fused images under Fusion 1. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Figure 10. The fused images under Fusion 1. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Remotesensing 16 03916 g010
Figure 11. The fused images under Fusion 2. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Figure 11. The fused images under Fusion 2. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Remotesensing 16 03916 g011aRemotesensing 16 03916 g011b
Figure 12. The fused images under Fusion 3. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Figure 12. The fused images under Fusion 3. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Remotesensing 16 03916 g012
Figure 13. The fused images under Fusion 4. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Figure 13. The fused images under Fusion 4. Each row, from left to right, is a fusion image of PCA, IHS, DWT, NSCT, and LP. (ae) depict the artificial target scene, (fj) represent the road and grass scene, and (ko) illustrate the wheat field scene.
Remotesensing 16 03916 g013aRemotesensing 16 03916 g013b
Table 1. Set the 6S model parameters to establish the LUT.
Table 1. Set the 6S model parameters to establish the LUT.
NameSetting Parameter Values
Molecular atmosphere modelMid latitude winter
Aerosol modelContinental
Surface reflectanceBRDF
Response bandSpectral response of main cameras
Observation geometryIn dataset
AOD at 550 nmRange of 0.20 to 0.80, with a step increment of 0.001.
CWV (g/cm3)Range of 0.80 to 1.80, with a step increment of 0.01.
Table 2. Attributes of the evaluation metrics.
Table 2. Attributes of the evaluation metrics.
NameClassificationMeaningNumber of Images Required
AGSpatialContrast and texture1
SDSpatialDispersion1
ENInformationalInformation content1
NCCSpatialCorrelation2
SSIMSpatialStructural similarity2
NMIInformationalInformation similarity2
UIQISpatialStructural similarity3
DTRSpectralSpectral quantification1
Table 3. Surface reflectance measured by ASDs.
Table 3. Surface reflectance measured by ASDs.
Type of Ground ObjectReflectivity Measured by the ASDs
R (570~690 nm)G (450~660 nm)B (380~570 nm)
Artificial white target0.6740.6600.595
Artificial black target0.0410.0420.043
Grass0.1660.1350.097
Concrete floor0.3640.3650.324
Table 4. The extremum and average values of the synchronous AOD and CWV images inverted from ASMAC.
Table 4. The extremum and average values of the synchronous AOD and CWV images inverted from ASMAC.
SceneAODCWV (g/cm3)
MinimumAverageMaximumMinimumAverageMaximum
Artificial target0.4410.5490.6260.9761.2731.539
Road and grass0.3840.4940.5870.9811.2261.493
Cornfield0.4650.5010.5931.1211.3171.554
Table 5. The asynchronous single values of AOD and CWV measured from CE318.
Table 5. The asynchronous single values of AOD and CWV measured from CE318.
Scene AODCWV (g/cm3)
Artificial target0.5481.436
Road and grass0.5521.442
Cornfield0.5451.445
Table 6. Evaluation metrics for fused images within the Fusion 1 framework.
Table 6. Evaluation metrics for fused images within the Fusion 1 framework.
SceneMethodUIQINCCNMISSIMAGSDEN
Artificial targetPCA0.486 *0.9410.7490.3232.36 × 10−30.0284.408
IHS0.8320.9250.7400.6283.11 × 10−30.0655.229
DWT0.8470.9460.7550.7934.30 × 10−30.0705.538
NSCT0.8680.9490.7640.7975.18 × 10−30.0735.731
LP0.8340.8260.7160.7633.76 × 10−30.0775.403
Extreme value ratio1.7861.1491.0682.4712.1922.7391.300
Road and grassPCA0.4850.7910.7090.3102.61 × 10−30.0244.051
IHS0.8390.8530.7010.6392.96 × 10−30.0435.173
DWT0.8500.9220.7280.8033.55 × 10−30.0425.553
NSCT0.8530.9270.7380.8063.85 × 10−30.0445.747
LP0.8360.8040.6990.5802.81 × 10−30.0415.035
Extreme value ratio1.7601.1711.0562.5971.4751.8271.419
CornfieldPCA0.4670.7200.6620.3391.85 × 10−30.0173.151
IHS0.8080.7580.6520.6402.02 × 10−30.0204.594
DWT0.8120.8780.6870.7031.89 × 10−30.0194.020
NSCT0.8220.8810.6890.7152.10 × 10−30.0193.999
LP0.8070.8370.6730.7092.06 × 10−30.0174.592
Extreme value ratio1.7591.2231.0572.1111.1351.1911.458
* The maximum (optimal) and minimum (worst) evaluation values are highlighted using bold and italicized font in this table.
Table 7. Evaluation metrics for fused images within the Fusion 1, Fusion 2, and Fusion 3 framework.
Table 7. Evaluation metrics for fused images within the Fusion 1, Fusion 2, and Fusion 3 framework.
MethodFrameworkArtificial Target SceneRoad and Grass SceneCornfield Scene
AGSDENAGSDENAGSDEN
PCAFusion 12.36 × 10−30.0284.4082.61 × 10−30.0244.0511.85 × 10−30.0173.151
Fusion 22.93 × 10−30.039 *4.9502.90 × 10−30.0294.5332.45 × 10−30.0203.569
Fusion 32.94 × 10−30.0384.9472.94 × 10−30.0304.4772.53 × 10−30.0203.584
Extreme value ratio1.251.381.121.131.251.121.3691.1681.137
IHSFusion 13.11 × 10−30.0655.2292.96 × 10−30.0435.1732.02 × 10−30.0204.594
Fusion 26.07 × 10−30.0775.6943.66 × 10−30.0515.4342.58 × 10−30.0254.871
Fusion 36.17 × 10−30.0785.6073.88 × 10−30.0525.4732.79 × 10−30.0254.917
Extreme value ratio1.981.201.091.311.211.061.3801.2491.070
DWTFusion 14.30 × 10−30.0705.5383.55 × 10−30.0425.5531.89 × 10−30.0194.020
Fusion 28.07 × 10−30.0825.7804.10 × 10−30.0505.6812.30 × 10−30.0234.313
Fusion 38.12 × 10−30.0845.7854.12 × 10−30.0515.7272.43 × 10−30.0244.389
Extreme value ratio1.891.191.041.161.211.031.2811.2951.092
NSCTFusion 15.18 × 10−30.0735.7313.85 × 10−30.0445.7472.10 × 10−30.0193.999
Fusion 28.93 × 10−30.0846.0743.97 × 10−30.0545.8762.39 × 10−30.0234.292
Fusion 38.97 × 10−30.0846.0784.10 × 10−30.0545.8822.51 × 10−30.0254.364
Extreme value ratio1.731.151.061.071.231.021.1951.3481.091
LPFusion 13.76 × 10−30.0775.4032.81 × 10−30.0415.0352.06 × 10−30.0174.592
Fusion 27.58 × 10−30.0936.1573.38 × 10−30.0495.3342.48 × 10−30.0244.891
Fusion 37.64 × 10−30.0946.1643.53 × 10−30.0495.3512.63 × 10−30.0244.919
Extreme value ratio2.031.231.141.261.221.061.2761.3941.071
* The maximum evaluation values (optimal value) are highlighted using bold and italicized font in this table.
Table 8. Evaluation metrics for fused images within the Fusion 3 and Fusion 4 framework.
Table 8. Evaluation metrics for fused images within the Fusion 3 and Fusion 4 framework.
MethodFrameworkArtificial Target SceneRoad and Grass SceneCornfield Scene
AGSDENAGSDENAGSDEN
PCAFusion 32.94 × 10−30.0384.9472.94 × 10−30.0304.4772.53 × 10−30.0203.584
Fusion 42.02 × 10−30.0244.4182.28 × 10−30.0223.9822.38 × 10−30.0153.344
Extreme value ratio1.4541.6181.1201.2871.3841.1241.0661.3151.072
IHSFusion 36.17 × 10−30.0785.6073.88 × 10−30.0525.4732.79 × 10−30.0254.917
Fusion 45.93 × 10−30.0635.0123.55 × 10−30.0485.3322.52 × 10−30.0234.858
Extreme value ratio1.0411.2351.1191.0931.0861.0261.1061.0981.012
DWTFusion 38.12 × 10−30.0845.7854.12 × 10−30.0515.7272.43 × 10−30.0244.389
Fusion 43.91 × 10−30.0595.6013.03 × 10−30.0395.0392.28 × 10−30.0234.296
Extreme value ratio2.0761.4101.0331.3601.3161.1371.0651.0621.022
NSCTFusion 38.97 × 10−30.0846.0784.10 × 10−30.0545.8822.51 × 10−30.0254.364
Fusion 43.77 × 10−30.0595.5942.91 × 10−30.0395.0342.17 × 10−30.0234.275
Extreme value ratio2.3761.4101.0861.4071.3991.1681.1561.1061.021
LPFusion 37.64 × 10−30.0946.1643.53 × 10−30.0495.3512.63 × 10−30.0244.919
Fusion 44.23 × 10−30.0856.0383.21 × 10−30.0455.2252.40 × 10−30.0214.793
Extreme value ratio1.8051.0991.0211.0991.1001.0241.0931.1451.026
Table 9. DTR for Fusion 1 framework.
Table 9. DTR for Fusion 1 framework.
Ground TypeMethodCalculated ReflectanceDTR
BGRBGR
Black targetPCA0.0170.0230.02459.35%45.29%40.36%
IHS0.1130.1350.156161.67%221.19%280.53%
DWT0.1040.1160.120140.83%175.20% *193.60%
NSCT0.1020.1110.114138.08%163.66%178.17%
LP0.1030.1230.125139.45%193.26%204.85%
White targetPCA0.0890.1640.25485.06%75.14%62.25%
IHS0.4680.5380.62221.34%18.52%7.76%
DWT0.3770.3970.41136.70%39.86%39.02%
NSCT0.3750.4120.42136.92%37.59%37.54%
LP0.3520.3940.42340.77%40.29%37.28%
Concrete floorPCA0.0740.1440.21377.19%60.47%41.60%
IHS0.2660.3390.41017.88%7.16%12.64%
DWT0.2050.2370.23336.87%35.02%36.10%
NSCT0.2100.2370.22935.28%35.18%36.97%
LP0.2030.2120.22837.35%41.92%37.36%
GrassPCA0.0420.0780.12356.95%42.35%25.86%
IHS0.1540.1910.23659.08%41.25%42.34%
DWT0.1470.1750.22951.94%29.60%37.69%
NSCT0.1490.1700.22454.10%26.28%34.81%
LP0.1630.1800.24168.41%33.64%45.00%
* When considering the evaluation metrics for DWT, NSCT, and LP fused images, the maximum and minimum values are highlighted in bold and italicized font in this table.
Table 10. DTR for the Fusion 1, Fusion 2, and Fusion 3 framework.
Table 10. DTR for the Fusion 1, Fusion 2, and Fusion 3 framework.
Ground TypeMethod and DTRBGR
Fusion 1Fusion 2Fusion 3Fusion 1Fusion 2Fusion 3Fusion 1Fusion 2Fusion 3
Black targetDWT0.1040.0560.0510.1160.0520.0490.1200.0610.051
NSCT0.1020.0530.0480.1110.0490.0460.1140.0570.047
LP0.1030.0580.0530.1230.0600.0570.1250.0710.061
DTRDWT140.8%30.5%18.9%175.2%23.1%16.3%193.6%48.4%23.7%
DTRNSCT138.1%23.2%11.6%163.7%16.2%9.4%178.2%39.9%15.2%
DTRLP139.5%34.6%22.9%193.3%42.0%35.3%204.8%72.2%47.6%
White targetDWT0.3770.5450.5570.3970.5420.4940.4110.5160.611
NSCT0.3750.5320.5440.4120.5470.4990.4210.5060.602
LP0.3520.5160.5290.3940.5290.4810.4230.5150.610
DTRDWT36.7%8.5%6.3%39.9%17.8%25.2%39.0%23.5%9.4%
DTRNSCT36.9%10.6%8.5%37.6%17.1%24.4%37.5%24.9%10.7%
DTRLP40.8%13.2%11.1%40.3%19.8%27.2%37.3%23.6%9.5%
Concrete floorDWT0.2050.2610.2820.2370.2930.3290.2330.3130.314
NSCT0.2100.2520.2730.2370.2850.3200.2290.3200.321
LP0.2030.2550.2760.2120.2610.2970.2280.3150.316
DTRDWT36.9%19.6%13.1%35.0%19.7%10.0%36.1%14.0%13.8%
DTRNSCT35.3%22.2%15.8%35.2%21.9%12.3%37.0%12.1%11.9%
DTRLP37.4%21.3%14.8%41.9%28.4%18.7%37.4%13.4%13.2%
GrassDWT0.1470.1020.1060.1750.1490.1540.2290.1850.181
NSCT0.1490.1040.1080.1700.1420.1460.2240.1790.176
LP0.1630.1180.1220.1800.1510.1560.2410.1950.192
DTRDWT51.9%5.4%9.5%29.6%10.7%14.1%37.7%11.2%9.1%
DTRNSCT54.1%6.9%11.1%26.3%5.0%8.4%34.8%7.9%5.8%
DTRLP68.4%21.9%26.1%33.6%12.0%15.4%45.0%17.5%15.4%
Table 11. DTR for the Fusion 3 and Fusion 4 framework.
Table 11. DTR for the Fusion 3 and Fusion 4 framework.
Ground TypeMethod and DTRBGR
Fusion 3Fusion 4Fusion 3Fusion 4Fusion 3Fusion 4
Black targetDWT0.0510.0540.0490.0490.0510.053
NSCT0.0480.0500.0460.0480.0470.052
LP0.0530.0540.0570.0610.0610.066
DTRDWT18.9%25.4%16.3%17.6%23.7%28.4%
DTRNSCT11.6%16.1%9.4%13.7%15.2%25.6%
DTRLP22.9%26.7%35.3%46.1%47.6%60.8%
White targetDWT0.5570.5100.4940.4510.6110.599
NSCT0.5440.5300.4990.4710.6020.589
LP0.5290.5200.4810.4660.6100.563
DTRDWT6.3%14.2%25.2%31.7%9.4%11.1%
DTRNSCT8.5%10.9%24.4%28.7%10.7%12.6%
DTRLP11.1%12.5%27.2%29.5%9.5%16.4%
Concrete floorDWT0.2820.2810.3290.3190.3140.307
NSCT0.2730.2680.3200.3090.3210.317
LP0.2760.2710.2970.2920.3160.312
DTRDWT13.1%13.3%10.0%12.5%13.8%15.6%
DTRNSCT15.8%17.3%12.3%15.2%11.9%12.9%
DTRLP14.8%16.2%18.7%20.1%13.2%14.3%
GrassDWT0.1060.1090.1540.1650.1810.120
NSCT0.1080.1110.1460.1550.1760.191
LP0.1220.1280.1560.1780.1920.123
DTRDWT9.5%12.1%14.1%21.9%9.1%27.7%
DTRNSCT11.1%14.8%8.4%14.5%5.8%15.1%
DTRLP26.1%32.0%15.4%32.1%15.4%25.9%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Chen, F.; Sui, T.; Ti, R.; Cheng, W.; Hong, J.; Qiu, Z. Interplay Between Atmospheric Correction and Fusion Techniques Enhances the Quality of Remote Sensing Image Fusion. Remote Sens. 2024, 16, 3916. https://doi.org/10.3390/rs16213916

AMA Style

Li Y, Chen F, Sui T, Ti R, Cheng W, Hong J, Qiu Z. Interplay Between Atmospheric Correction and Fusion Techniques Enhances the Quality of Remote Sensing Image Fusion. Remote Sensing. 2024; 16(21):3916. https://doi.org/10.3390/rs16213916

Chicago/Turabian Style

Li, Yang, Feinan Chen, Tangyu Sui, Rufang Ti, Weihua Cheng, Jin Hong, and Zhenwei Qiu. 2024. "Interplay Between Atmospheric Correction and Fusion Techniques Enhances the Quality of Remote Sensing Image Fusion" Remote Sensing 16, no. 21: 3916. https://doi.org/10.3390/rs16213916

APA Style

Li, Y., Chen, F., Sui, T., Ti, R., Cheng, W., Hong, J., & Qiu, Z. (2024). Interplay Between Atmospheric Correction and Fusion Techniques Enhances the Quality of Remote Sensing Image Fusion. Remote Sensing, 16(21), 3916. https://doi.org/10.3390/rs16213916

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop