Next Article in Journal
Detection of Crop Seeding and Harvest through Analysis of Time-Series Sentinel-1 Interferometric SAR Data
Next Article in Special Issue
Geometric Recognition of Moving Objects in Monocular Rotating Imagery Using Faster R-CNN
Previous Article in Journal
Identification and Analysis of Microscale Hydrologic Flood Impacts Using Unmanned Aerial Systems
Previous Article in Special Issue
A Deep Learning-Based Robust Change Detection Approach for Very High Resolution Remotely Sensed Images with Multiple Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery

1
College of Resource Environment and Tourism, Capital Normal University, Beijing 100048, China
2
College of Geospatial Information Science and Technology, Capital Normal University, Beijing 100048, China
3
Department of Survey, Ministry of Land Management, Cooperatives and Poverty Alleviation, Government of Nepal, Kathmandu 44600, Nepal
4
Key Laboratory of 3D Information Acquisition and Application, Capital Normal University, Beijing 100048, China
5
National Disaster Reduction Centre, Ministry of Emergency Management, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(10), 1550; https://doi.org/10.3390/rs12101550
Submission received: 20 April 2020 / Revised: 10 May 2020 / Accepted: 12 May 2020 / Published: 13 May 2020

Abstract

:
In recent years, the use of image fusion method has received increasing attention in remote sensing, vegetation cover changes, vegetation indices (VIs) mapping, etc. For making high-resolution and good quality (with low-cost) VI mapping from a fused image, its quality and underlying factors need to be identified properly. For example, same-sensor image fusion generally has a higher spatial resolution ratio (SRR) (1:3 to 1:5) but multi-sensor fusion has a lower SRR (1:8 to 1:10). In addition to SRR, there might be other factors affecting the fused vegetation index (FVI) result which have not been investigated in detail before. In this research, we used a strategy on image fusion and quality assessment to find the effect of image fusion for VI quality using Gaofen-1 (GF1), Gaofen-2 (GF2), Gaofen-4 (GF4), Landsat-8 OLI, and MODIS imagery with their panchromatic (PAN) and multispectral (MS) bands in low SRR (1:6 to 1:15). For this research, we acquired a total of nine images (4 PAN+5 MS) on the same (almost) date (GF1, GF2, GF4 and MODIS images were acquired on 2017/07/13 and the Landsat-8 OLI image was acquired on 2017/07/17). The results show that image fusion has the least impact on Green Normalized Vegetation Index (GNDVI) and Atmospherically Resistant Vegetation Index (ARVI) compared to other VIs. The quality of VI is mostly insensitive with image fusion except for the high-pass filter (HPF) algorithm. The subjective and objective quality evaluation shows that Gram-Schmidt (GS) fusion has the least impact on FVI quality, and with decreasing SRR, the FVI quality is decreasing at a slow rate. FVI quality varies with types image fusion algorithms and SRR along with spectral response function (SRF) and signal-to-noise ratio (SNR). However, the FVI quality seems good even for small SRR (1:6 to 1:15 or lower) as long as they have good SNR and minimum SRF effect. The findings of this study could be cost-effective and highly applicable for high-quality VI mapping even in small SRR (1:15 or even lower).

Graphical Abstract

1. Introduction

Image fusion has received increasing attention as simultaneous purchase of a PAN and MS image from a high-resolution satellite is usually very costly [1]. It is also necessary as many sensors acquire information about the Earth in either only PAN (EROS-A, EROS-B, Worldview-1) or only MS (Rapid Eye) mode [2]. Image fusion has many applications in different fields like remote sensing [3], urban sprawl monitoring, coastal mapping [4], land cover change detection [5], land use classification [6], crop monitoring [7], land use classification on flood areas [8], changes in vegetation cover [9], and vegetation indices [10,11]. A high-resolution and good-quality VI mapping is necessary [1] for monitoring the earth’s vegetative cover as a precise radiometric measure of green vegetation [12]. Such high-resolution VIs derived from satellite images area useful data source for many agricultural, environmental, and climate studies [13]. An image with high quality is preferred for analysis and other applications [3]. However, it is important to assess the quality of the fused image before using it for various applications of remote sensing like VI mapping [14].
Recently, many image fusion algorithms have been developed; for example, the sparse images fusion algorithm (named as Sparse FI, pronounced as “sparsify”) has super-resolution capability and robustness of sparse signal reconstruction algorithms, which gives a higher spatial resolution and, in most cases, less spectral distortion compared with the conventional methods [15]. Similarly, Liu et al. [16] proposed a convolutional sparse representation (CSR)-based image fusion framework, in which each source image is decomposed into a base layer and a detail layer, for multi-focus image fusion and multi-modal image fusion which outperform the SR-based methods in terms of both objective assessment and visual quality. To recover the original information covered by the clouds and the accompanying shadows, a nonnegative matrix factorization (NMF) and error correction method (S-NMF-EC) is proposed [17]. Li et al. [18] also proposed a new algorithm, applied to a reflectance product and a land surface temperature product, in which a qualitative visual effect and the quantitative objectives are effective. However, there are several popular image fusion algorithms which also have good spectral characteristics preservation, and are frequently used in many image fusion applications like GS [6,19,20], Ehlers [21], Principal Component (PC) [6,22,23,24], Modified Intensity Hue Saturation (MIHS) [25,26] and HPF [22,23].
Commonly used image fusion techniques generally produce good results with a standard SRR (ratio of PAN resolution to MS resolution) of 1:4 for images obtained from the same sensor [27]. Ehlers, 2008 [28] showed that SRR can get worse if data from different satellites are used; for example, SRR between IKONOS and SPOT 5 (multispectral mode) is 1:10. Ling et al. [29]’s experimental results indicated that SRR of 1: 10 or higher is desired for optimal multisensory image fusion and also indicated that with decreasing SRR, the fused image quality decreases. However, they were focused on fused image quality, not FVI quality. VIs involve further image transformations of the fused image bands, which can have additional impacts on their spectral and spatial quality, so it is necessary to evaluate the effects that these further transformations have on the fused imagery [13]. Image quality must be inferred from measurements of spatial resolution, calibration accuracy, SNR, contrast, bit error rate, sensor stability, and other factors [30]. This indicates that in addition to SRR, there are other factors which need to be identified to study the effect of image fusion on VI quality properly and to evaluate the relationship between SRR and FVI quality. Only a few studies have evaluated the effects of image fusion methods on VIs quality. For example, Rahaman et al. [31] calculated vegetation greenness and canopy moisture indices at 15 m spatial resolution and validated using their equivalent-values at a spatial resolution of 30 m. Similarly, Brain J. 2014 [13] performed four pansharpening methods and two VIs, Normalized Difference Vegetation Index (NDVI) and Simple Ratio (SR) to analyze the effect. Both researchers were limited to the same sensor fusion (with large SRR); they used Landsat PAN and MS bands to derive 15 m FVI and evaluated their result briefly without considering multi-sensor fusion with different SRF, SNR.
Some past studies have used image fusion methods for VI mappings, like Landsat imagery to spatially downscale 250 m resolution MODIS VI data [32,33]. It has become more significant for researchers to have higher resolution imageries to calculate several vegetation indices since the spectral resolution of the PAN band spans 0.503–0.676 µm with clearly no overlap with the blue, and NIR spectral bands in Landsat-8 OLI satellite platform (operating since February 2013) [31]. The daily surface reflectance products (MOD09GA) at a spatial resolution of 500 m, captured from the MODIS sensor, were used along with the Landsat-8 OLI data to generate a 30-m resolution synthetic image [34]. A new index defined by El-mezouar et al. [1] proposed high-resolution NDVI = (Fused NIR-Fused R)/(Fused NIR + Fused R) which was a good solution in introducing the spatial resolution in computation of the NDVI and showed less confusion in the vegetation detection tested in IKONAS imagery. Similarly, the vegetation changes in the Zayandeh-rud river basin in the period 2001 to 2016 have been investigated based on the combination of the 15 different vegetation indices [9].
It can be seen that many image fusion algorithms and satellite images are used for VI mapping in different application areas. Most of them were focused on making a high-resolution FVI map using same-sensor fusion or multi-sensor fusion using Landsat and MODIS imagery, but not focused on many satellite images (like GF1, GF2, GF4, Landsat-8 OLI, MODIS etc.) and could not explain what happens to the quality when SRR is lower than 1:10 (SRR between GF2 to GF4 is 1:12, IKONAS to SPOT 4 is 1:20), also missing the sensitivity analysis of VI with image fusion and different factors affecting FVI quality. To address this research gap, our main objective was to study the effect of image fusion for the quality of VI mapping. To meet the main objective, we were focused on 1) analyzing the sensitivity of VI with image fusion 2) selecting fusion method with least impact on FVI quality 3) finding out factors of FVI quality and 4) evaluating how SRR and FVI are inter-related. We made a strategy on image fusion and quality assessment for different satellite images; GF1, GF2, GF4, Landsat-8 OLI and MODIS using their PAN and MS bands in different SRR (1:6 to 1:15) using both same-sensor and multi-sensor fusion. Five popular image fusion methods and six VIs were used with subjective and objective quality indices to analyze the result properly.

2. Materials and Methods

2.1. Study Area

We chose the research area located in Hebei province, China. It is located in the western part of Beijing and covers some area of the Guanting reservoir. The reservoir was created by building a dam on Yongding river (drainage area 43402 sq. km) near the town of Guanting and covers areas in the Inner Mongolia Autonomous Region, Beijing, and the provinces of Shanxi and Hebei [35]. However, we used the extent of the study area ranging from 40°15’54"N to 40°30’55"N latitude and 115°26’13’’E to 115°47’13"E longitude with a total area of approximately 155 sq. km which covers mostly forest areas along with river, roads and built-up areas (Figure 1).

2.2. Methodology

This research methodology can be described in three sections: first, image registration, second, FVI calculation and third, quality assessment. The overall methodology involves image acquisition, pre-processing, bringing into a same coordinate system, image registration, image fusion algorithms selection, VI selection, quality assessment and comparative fusion result and analysis. The details are described after the methodology chart (Figure 2).

2.2.1. Image Acquisition

We used a total of 9 (PAN images 4 and MS images 5) scenes from different satellite images. We downloaded the Landsat-8 OLI MS, PAN and Moderate Resolution Imaging Spectroradiometer (MODIS) MS images from the United States Geological Survey’s (USGS) web portal (https://earthexplorer.usgs.gov/). The GF1, GF2 and GF4 MS and PAN images were downloaded from Earth Observation Portal (EOP) (https://directory.eoportal.org/web/eoportal/). As long as the time-lag is small with respect to the time scale of the variations in small-size features, its influence upon the quality of the transformation of the spectral content when enhancing the spatial resolution is low or negligible [36]. Thus, GF1, GF2, GF4 and MODIS images were acquired on the same date, 13 July 2017, but Landsat-8 OLI images were not available on the same date so they were acquired on 17 July 2017 (after four days). All the images were selected carefully (cloud and snow-free) to ensure better image quality. The details of different resolutions, imager types, selected MS and PAN bands with their corresponding spectral ranges are listed in Table 1.

2.2.2. Pre-processing

The MOD09GA Version 6 product provides an estimate of the surface spectral reflectance of Terra MODIS Bands 1 through 7, corrected for the atmospheric condition such as gases, aerosols, and rayleigh scattering [37]. However, we only used four bands (Table 1) with the 500 m surface reflectance observation [38]. Hence, no more correction is necessary for MODIS multispectral product. The GF1, GF2, GF4 and Landsat-8 OLI images were atmospherically corrected using Fast Line-of-sight Atmospheric Analysis of Hypercube (FLAASH) module in ENVI 5.3 software to obtain surface reflectance [39].

2.2.3. Same Coordinate System

As the MODIS image was in a sinusoidal projection system, Landsat-8 OLI was in the projected coordinate system and GF1, GF2, GF4 were in geographic coordinate system WGS 1984 50N (latitude/longitude format). For fusion purposes, all imagery were brought into the same coordinate system, i.e., WGS 1984 50N in projected format.

2.2.4. Image Registration

Image registration is a crucial and fundamental problem in image processing and computer vision, which aims to align two or more images of the same scene acquired from different views or at different times [40]. Precise image registration is essential for image fusion; especially for pixel-based image fusion, the impact of image registration is very high. Feng et al. [41] developed a two-step algorithm combining and localizing the feature and area-based methods to align remote sensing images. Automatic image registration is the process of aligning two or more images of approximately the same scene with minimal human assistance [42]. The steps were first, selecting files for image registration, second, generating tie points automatically and third, reviewing tie points and exporting image registration results [43].Considering all these ideas, automatic tie points were generated and manually added in between warp image and base image carefully to get precise registration before fusion activities started.

2.2.5. Selection of Image Fusion Algorithms

For the image fusion experiments, five popular image fusion algorithms were selected based on their relative advantages found in different types of literature, as listed in Table 2. Interestingly, they are also available in commercial software packages like ENVI and ERDAS Imagine. Thus, they can be used easily for many applications.

2.2.6. Selection of VI

Some popular VIs selected in this research are presented in Table 3. Four bands were selected for image fusion to check the quality of image fusion, including Blue (459-479 nm), Green (530–590 nm), Red (640–670 nm) and Near Infrared (850–880 nm), for all MS satellite images. These spectral bands were used in this study as their combinations are good for calculating different VIs. These bands are also commonly used to calculate VIs such as ARVI, Enhanced Vegetation Index (EVI), GNDVI, NDVI, Optimized Soil Adjusted Vegetation Index (OSAVI ) and Soil Adjusted Vegetation Index (SAVI). They are chosen in such a way that they will contain different bands in their VI equation (Table 3). For example, some vegetation indices like NDVI, OSAVI and SAVI have mainly R and NIR bands in their equations. However, EVI and ARVI have NIR, R and B bands in VI equations, and GNDVI has NIR and G bands in VI equation. In our strategy, RVI and FVI are calculated through band math operation from reference MS and fused MS images using the following VI equations (Table 3).

2.2.7. Strategy on Image Fusion and Quality Assessment

Only the selection of image fusion algorithms, VIs and quality indices is not sufficient to analyze the effect of image fusion on VIs properly, so one strategy was made, as shown in Table 4. In this strategy, we included both same-sensor and multi-sensor image fusion with different SRF, SRR and SNR to analyze the result in detail. Different resolutions PAN and MS images were strategically fused, making the same SRR and different SRR, same SRF and different SRF, good SNR and poor SNR to analyze their effects properly. RVI was calculated using band math operation in different resolutions 1 m, 4 m, 8 m and 50 m from respective original or resampled reference MS images (Table 4).
In same-sensor fusion (GF2 to GF2), as GF2 has 1 m PAN and 4 m MS images, we resampled 4 m MS image into 1 m, 6 m, 8 m, 10 m and 12 m resolutions. Then, we fused a 1 m PAN image with 6 m, 8 m, 10 m and 12 m MS images to get a fused MS image in 1 m resolution, making 1:6, 1:8, 1:10 and 1:12 SRR, respectively. Then, we calculated different FVI (1 m) from fused MS images (1 m) and RVI (1 m) from resampled MS image (1 m) to evaluate the quality. In multi-sensor fusion, for example, in GF1 to GF4, we have GF1 (PAN 2 m and MS 8 m) and GF4 (MS 50 m). Thus, we resampled 2 m PAN into 8 m PAN to make the same resolution for MS and PAN images, fused the image, calculated FVI (8 m) and evaluated the quality with RVI (8 m calculated from original 8 m MS image). Similarly, we made a detailed strategy for all RVI and FVI, as mentioned in Table 4. To evaluate the result, we need quality indices like root mean square error (RMSE), which is a square root of MSE [25]; it is a standard measure of the difference between the reference image and the fused image. The greater the RMSE, the higher the difference between reference and the fused image will be [24,50]. RMSE calculates the amount of change per pixel due to processing. Similarly, the value of correlation coefficient (CC) should be close to +1, which indicates that the reference and fused images are the same. Variation increases when the value of CC is lower [14]. Assessment of objective quality (CC and RMSE) [48] and a variety of visual inspections [26] is necessary for analyzing the effect of image fusion properly.

3. Results

3.1. Single Sensor FVI Quality (Using Resampling)

Single-sensor (GF2 PAN+GF2 MS) FVI quality was tested according to the strategy in Table 4. The original GF2 has a 1 m PAN and a 4 m MS image, so we resampled the 4 m MS image into 6 m, 8 m, 10 m and 12 m MS images so that our SRR became 1:6, 1:8, 1:10 and 1:12 respectively. We fused the 1 m PAN image with 6 m, 8 m, 10 m and 12 m MS images, and calculated FVIs. To evaluate the quality of FVIs, we need RVI in same resolution (i.e., 1 m), but we have an original 4 m MS image. To get 1 m RVI, we again resampled the original 4 m MS image into a 1 m MS image and calculated the RVI, then we compared and evaluated the quality of FVI with respect to RVI using quality indices. All the FVI quality results are expressed in % so that its precise numeric value can be listed. The result for GS fusion has CC (98.85 to 99.05%), RMSE (10.06 to 13.50%) for SRR 1:6, CC (98.93 to 99.08%), RMSE (12.00 to 13.54%) for SRR 1:8, CC (97.39 to 98.73%), RMSE (10.81 to 14.66%) for SRR 1:10 and CC (98.17 to 98.59%), RMSE (12.77 to 15.78%) for SRR 1:12. Similarly, the detailed result of GS, Ehlers, PC, MIHS and the HPF fusion result for different VIs are listed in Table 5. In GF2-GF2 fusion, for most of the VIs like EVI, NDVI, OSAVI and SAVI; GS fusion has a lower effect on FVI quality (high CC and low RMSE values). However, in SRR, 1:6 and 1:8, for ARVI and GNDVI, MIHS fusion is good. However, for GNDVI and ARVI, GS fusion has a better position after MIHS (Table 5). It also shows the range of CC and RMSE values in different SRR and different fusion algorithms (Table 5).

3.2. Multi-Sensor FVI Quality (Using Real Image)

Different types of image and feature fusion exist like multi-sensor, multi-resolution, multi-temporal, multi-angular, and model-based feature fusion [51]. Image fusion can be applied both to data collected from different sensors (inter-sensor), or from the same sensor (intra-sensor), as well as to data collected at the same time under the same conditions (single-dated) or at different dates (multi-date) [26]. We have already found a single-sensor fusion result in Table 5; our result of multi-sensor fusion are calculated and listed in Table 6.
For GF1 (PAN) to GF4 (MS) fusion with SRR 1:6, in ARVI objective quality (CC, RMSE) is 84.27%, 33.97% for GS fusion which is better (higher CC and lower RMSE) than other fusion methods. For EVI, GNDVI, NDVI, OSAVI and SAVI, Ehlers fusion has higher CC but GS has lower RMSE but the value of CC for GS is very close to the Ehlers method. The result for GS fusion has CC (89.57 to 92.56%), RMSE (21.35 to 29.56%) for SRR 1:8, CC (79.46 to 89.93%), RMSE (22.86 to 33.74%) for SRR 1:10 and CC (88.24 to 91.75%), RMSE (24.86 to 34.67%) for SRR 1:12. Similarly, the detailed result of GF1 (PAN)+GF4 (MS), GF2 (PAN)+Landsat-8 OLI (MS), GF4 (PAN)+MODIS (MS) for GS, Ehlers, PC, MIHS and HPF algorithms for FVI is listed in Table 6. In multi-sensor fusion, most of the VIs in SRR 1:6 (GF1+GF4) and 1:10 (GF4+MODIS), the quality of FVI result is inconsistent with fusion algorithms.

3.3. Multi-Sensor FVI Quality (Resampled)

One further experiment was made by fusing GF2 PAN (4 m) with Landsat-8 OLI MS (30 m, 40 m, 48 m and 60 m) using resampling technique to get 1:8, 1:10, 1:12 and 1:15 SRR respectively. The corresponding FVI was compared with RVI (4 m) from the GF2 MS image (calculated from its 4 m original MS bands) and the result is listed in Table 7.
We found the result of GF2 (PAN 4 m) fusion with Landsat-8 OLI MS (30 m) in SRR 1:8 in the previous section (Table 6). Thus, from this new fusion, we analyzed continuous FVI quality with different SRR. The result for GS fusion has CC (88.92 to 92.11%), RMSE (21.84 to 30.58%) for SRR 1:10, CC (88.59 to 92.54%), RMSE (22.10 to 30.79%) for SRR 1:12 and CC (88.19 to 92.40%), RMSE (22.30 to 31.32%) for SRR 1:15. Similarly, the detail fusion result for GS, Ehlers, PC, MIHS and HPF algorithms for FVI is listed in Table 7. It shows that both CC and RMSE values are consistent and uniform for all image fusion algorithms in SRR 1:8 to 1:15 but HPF has a lower quality (Table 7).

3.4. Visual Quality Evaluation

Subjective evaluation is evaluating the quality of the images with the naked eyes and making judgments according to the subjective feelings of image quality [22]. In this analysis, a team of observers compared the fused image with the input images and valued the quality of the fused image by employing various optical parameters such as spatial details, geometric pattern, size of objects, and color [14]. The subjective evaluation method is simple and intuitive, and it can make a quick and convenient evaluation of the images [22]. For visual quality evaluation, we need a good RVI with the same spatial resolution of FVI, which would only give proper judgment, as we already calculated FVI and RVI in different resolutions during objective quality evaluation in previous section. Then, we compared them with their visual quality for different features in different parts of the images.
A total of 384 VI maps were prepared, including 360 FVI (6 VIs @ 5 fusion algorithms @ 4 SRR @ 3 types of image fusion) maps and 24 RVI maps (6 GF2 (1 m) + 6 GF1 (8 m) + 6 GF2 (4 m) +6 GF4 (50 m)). Due to such large numbers, individually, it is difficult to show their visual quality, so subjective quality was analyzed in different parts of the images selectively. To analyze the FVI quality more clearly, water area, forest areas, built-up areas, lakes, roads, etc., were compared with their respective RVI. Different image fusion algorithms showed different influences on different VIs. Some image fusion algorithms had a good performance at the edges and some had good results in vegetation areas, rivers, etc. The following table (Table 8) shows the visual quality of FVI results from same-sensor, multi-sensor and multi-sensor resampled abbreviated as Same-Sen., Multi-Sen. and Multi-Sen(R), respectively.
The visual quality analysis was based on a selective way to include different FVI maps for all types of image fusion. For example, the same-sensor FVI result is shown for ARVI, NDVI and OSAVI for SRR 1:6 and 1:8 and multi-sensor FVI results are shown for EVI, GNDVI and SAVI in GF1 PAN (8 m) with GF4 MS (50 m) in SRR 1:6, GF2 PAN (4 m) with Landsat-8 OLI (30 m) in SRR 1:8. In SRR 1:6, the same-sensor FVI results are very similar in color, texture and shape. The roads, rivers, built-up, forest areas in FVI are very similar to RVI. For the images acquired at the same time by the same sensor, most algorithms for pan-sharpening provide very good results, i.e., they retain the high spatial resolution of the panchromatic image and the spectral information from the multispectral image (single-sensor, single-date fusion) [2]. For multi-sensor fusion in SRR 1:6, EVI shows the river area well but not the built-up areas. Moreover, GNDVI shows that some fusion algorithms have good results for edge detection like PC but not good for built-up areas. Ehlers does not have a good result for road detection. However, the GS method has a better visual quality than the other methods for these features (Table 8).
Some VIs (ARVI, NDVI and OSAVI) have a very similar visual quality with their corresponding RVI in single-sensor fusion with SRR 1:8. However, other VIs (EVI, GNDVI and SAVI) have a better quality for all features in multi-sensor fusion with SRR 1:8. Same-sensor fusion still has a good visual quality in SRR 1:10 like in SRR 1:6 and SRR 1:8. However, in multi-sensor fusion, for GF4 PAN(50 m) with MODIS MS (500 m) fusion, the visual quality is degraded. In EVI and OSAVI, the size and boundary of the river are not properly mapped and other details are also very poor. Thus, we made other multi-sensor fusion by resampling the Landsat-8 OLI image. NDVI and SAVI have a better visual quality for built-up, road, river and forest areas. In SRR 1:12, the same-sensor FVI has a good visual result. In SRR1:12, for multi-sensor fusion, GF2 PAN (4 m) with GF4 MS (50 m), the quality slightly improved compared to SRR 1:10. EVI and OSAVI showed a better visual quality for the GS and PC methods (Table 8).

4. Discussion

4.1. Quality Assessment

Ren et al. [52] indicated that sometimes, the visual effect and the objective result can be different. To assess this, we made qualitative and quantitative evaluations in our research. Ehlers, 2008, showed that SRR can get worse [28] if data from different satellites are used; for example, SRR between IKONOS and SPOT 5 (multispectral mode) is 1:10, for SPOT 4 even 1:20, as mentioned in the Section 1. The SRR varies with different multi-sensor fusions and their quality assessment is always important to find out. As image fusion has proven to be a potential tool in remote sensing applications which integrates the information from combinations of panchromatic, multispectral and hyperspectral images; this is intended to result in a composite image having both higher spatial and higher spectral resolutions [53]. We show the FVI quality results (subjective and objective) from different satellite images GF1, GF2, GF4, Landsat-8 OLI and MODIS in the previous Section 3. The fusion quality often depends on the user’s experience, the fusion method, and on the dataset being fused [54]. Thus, our fusion strategy (Table 8) was very carefully prepared, including different resolutions, SRF, SNR and SRR and also carefully performed. Almost the same date images (Landsat-8 OLI on 2017/07/17 and others on 2017/07/13) were used on a clear-sky day with favorable weather conditions during the dry season to get a better fusion result [34].

4.2. Influence of SRR (with Same SRF and Good SNR)

In Section 3, we found that single-sensor fusion has a good visual quality for VI. For objective quality, the pattern was almost uniform for CC for all fusions, and RMSE was quite different for different image fusions and VIs. However, the general trend is that by increasing SRR, the CC graph is falling slowly and the RMSE graph is increasing slowly (Figure 3). This good quality might be due to the high-resolution PAN (1 m) and MS (6,8,10,12 m) images, the good SNR and the same SRF. The CC and RMSE values are almost the same for all image fusion methods, which indicates that image fusion is insensitive with VI. However, the HPF has a scattered graph and seems to be sensitive. ARVI and GNDVI have quite low RMSE values and are well separated in all image fusion algorithms and all SRR. However, CC seems to be almost the same for all VI and are almost overlapped. Vibhute et al. [55] found that the GS spectral sharpening algorithm gives better results than PC spectral sharpening, according to the classification results. Similarly, in our result, the GS fusion has a good quality (low RMSE and high CC) for all VIs in all SRR (1:6 to 1:12) (Figure 3). In reality, such same-sensor fusion does not exist in such large SRR (1:6 to 1:12). We did this arbitrary image fusion to analyze how much SRF and SRR could affect the fusion result as long as there is good SNR between PAN and MS images. In general, the FVI quality is decreasing with decreasing SRR very slowly as long as there is a good SNR and a minimum SRF effect (Figure 3).

4.3. Influence of SRR (with different SRF and SNR (Good to Poor))

Multi-sensor fusion showed different visual quality for different algorithms for different VI (Table 4). Some image fusion algorithms had a good visual result in preserving all types of details and some algorithms did not. In the objective quality results, the pattern was fluctuating (both CC and RMSE) (Figure 4) depending on SRR, SRF and SNR. For SRR 1:6 in GS fusion, for ARVI CC is 84.27% and RMSE is 33.97% but when the SRR is 1:8, for ARVI CC is 92.56% and RMSE is 22.59% (Table 6), remaining VI has a similar trend (Figure 4). With decreased SRR, the image fusion accuracy was assumed to be decreased but the result showed an increased accuracy(better quality). This might be due to two reasons: one is the difference in SRF and the other is the resolution of the MS image. Belgiu and Stein [56] indicated that for fusing spatiotemporal images, spectral responses of input images have to be unified, so the difference in SRF between two sensors has an impact on the quality of FVI. In SRR 1:8; the MS image has a 30 m resolution and PAN image has a 4 m resolution which is a better resolution than SRR 1:6 (PAN (8 m) with MS (50 m)). SRR 1:10 fusion is the result of a low-resolution PAN (50 m) image and a very low-resolution MS (500 m) images. The SNR is a measure of the purity of a signal, the ratio between the information and noise of the fused image [57]. MODIS (500 m) and GF4 (50 m) has very low resolution; both of them have poor SNR, so FVI quality is poor for them. The influence of SRF is also prominent along with SNR for multi-sensor FVI. However, the SRR effect is not very high. The visual analysis showed that FVI has a different quality depending on its SRF, SNR and SRR (Table 8). Not only does GS have a good result but also Ehlers has a good result for ARVI and their quality indices are very close. This shows that very low-resolution image fusion has inconsistent results due to poor SNR. For SRR 1:12 in GS fusion; for ARVI CC is 91.59% and RMSE is 25.25% (Table 6) which is a better result than SRR 1:6, although it has a higher SRR than 1:8 and 1:10. The reason might be the same, due to the high-resolution (4 m) PAN image used in SRR 1:12, although it has lower SRR. Not only does GS fusion have this trend for VI mappings but all the remaining fusion methods have the same pattern for all FVI (Figure 4). We changed the fusion algorithms (keeping same VI) but the CC values are very close to each other. Similarly, the change in image fusion algorithms did not make a large difference in RMSE values. This indicates that image fusion is insensitive to VI (i.e., change in image fusion algorithms did not make a large difference in VI quality). However, the HPF has a scattered graph and seems to be sensitive. ARVI and GNDVI have quite lower RMSE, high CC values so they are well separated in all image fusion and in all SRR (Figure 4). However, for other VI (except ARVI and GNDVI), CC and RMSE, the graph is almost same and overlapped. The GS pan-sharpen method has become one of the most widely used high-quality methods since its publication [20]. In multi-sensor FVI, also the GS fusion has a good quality (low RMSE and high CC) for all VI and in all SRR. In general, the quality of FVI is decreasing with decreasing SRR (1:6 to 1:12). However, the higher effect on FVI quality is due to SNR and SRF rather than SRR (Figure 4).

4.4. Influence of SRR (with constant SRF and Good SNR)

One further experiment was made to check the pattern of FVI properly. GF2 PAN (4 m) is fused with Landsat-8 OLI MS (30 m, 40 m, 48 m and 60 m). In GF2 to Landsat-8 OLI (30 m, 40 m, 48 m and 60 m) fusion, there is a uniform relationship between SRR and FVI quality (CC and RMSE) for ARVI (92.56%, 22.59%), (92.11%, 23.05%), (91.88% and 23.43%) and (91.17% and 23.45%) in SRR 1:8, 1:10, 1:12 and 1:15 respectively (Table 7). By increasing SRR from 1:8 to 1:15; all image fusion techniques showed a uniform result (Figure 5). This uniformity could be due to the constant SRF (Landsat-8 OLI and GF2) and better SNR than the previous multi-sensor fusion. However, the HPF still has a lower quality and seems to be sensitive. ARVI and GNDVI have quite lower RMSE, high CC so they are well separated in all image fusion and all SRR (Figure 5). However, for other VI (except ARVI and GNDVI), the CC and RMSE graph is almost same and overlapped (except for HPF). The quantitative evaluation results show that the brightness and information retention of GS fusion image was the best [52]. In the fusion of the near-infrared band, the GS method can be considered as the best [58]. All VI equations contain at least one NIR band and GS fusion has a good result in NIR band which might be the result behind the good quality (low RMSE and high CC) in all FVI and for all SRR. In general, the quality of FVI is decreasing with decreasing SRR (1:8 to 1:15) (Figure 5).
However, if we compare all three figures (Figure 3, Figure 4 and Figure 5) together, we find that the higher effect on FVI quality is due to SNR and SRF rather than SRR. With decreasing SRR (keeping the effect of SRF and SNR minimum), the quality decreases at a very slow rate. The VI having R and NIR bands in their VI equations has a similar pattern. Their CC and RMSE are almost identical (NDVI, OSAVI, SAVI, EVI) with a similar pattern (Figure 5). The remaining VI, including other bands in their VI equation (B and G bands), has also a similar pattern. The CC and RMSE values have a similar pattern for ARVI and GNDVI but different from NDVI, OSAVI, SAVI and EVI. In short, ARVI and GNDVI have the least impact in all cases with all fusion methods. In general, image fusion seems to be insensitive with VI and SRR. The HPF seems to be sensitive with VI and SRR. However, this indicates that the influence of SRF and SNR is higher than SRR.

4.5. Limitations and Possible Future Study

We used GF1, GF2, GF4, Landsat-8 OLI and MODIS satellite images for our research, but still, it could be better if more satellite images like Sentinel-2, World View-4, IKONAS, SPOT were used for strategic image fusion. Our fusion result was focused on one season, that is, 2017 July 13 (for GF1, GF2, GF4 and MODIS imagery) and 2017 July 17 (for Landsat-8 OLI image) to compare and analyze the result, but it would be much better if we had used different satellite images in different seasons. Similarly, further investigation of FVI quality using more VI like AFRI, GLI, NGRDI, RVI, TVI etc. [9] could give more detail results. We were limited to five (GS, Ehlers, PC, MIHS and HPF) image fusion algorithms but NNDiffuse [59] and other algorithms could be used to investigate the result in a better way. Besides CC and RMSE, there are other quality indices like PSNR [50], entropy [14], and SSIM [60] which can give more quality results and a more robust conclusion. Similarly, in the future, we can use combined bands from different fusion methods (NIR from GS and R, G, B bands from PC or Ehlers and MIHS fusion) depending on their degree of goodness for each band to get a better FVI quality.

5. Conclusions

Many image fusion techniques have developed and applied in various field but the quality assessment is always a key factor before their application. As a preliminary work towards image fusion, different satellite images (GF2, GF1, Landsat-8 OLI, GF4 and MODIS) with PAN and MS bands were acquired, preprocessed and precisely registered. The main objective of this research was to study the effect of image fusion for the quality of VI mapping. Using a strategy of image fusion and quality assessment (in different SRR, SRF and SNR) we came to a robust conclusion. First, image fusion has the least impact on GNDVI and ARVI compared to other VI. The quality of FVI is mostly insensitive to image fusion. However, for HPF, the quality of FVI is sensitive. Second, the value of CC and RMSE shows that the GS method has the least impact on FVI quality. For example, ARVI in GF1-GF4 with CC and RMSE (84.27%, 33.97%), GF2-Landsat-8 OLI (92.56%, 22.59%), GF4-MODIS (82.20%, 28.94%) and GF2-GF4 (91.59%, 25.25%) proves that GS has a better quality in FVI. Third, some factors affect FVI results; they are SRF, SRR and SNR. For example, in GF1-GF4 (SRR 1:6), the influence of image fusion is higher than GF2-Lansat-8 OLI (SRR 1:8) and GF2-GF4 (SRR 1:12) due to the SNR factor. Moreover, GF4-MODIS (SRR 1:10) fusion has a quite different result due to the poor SNR. Fourth, with decreasing SRR, the FVI quality is decreasing at a slow rate. For example, in GF2 to Landsat-8 OLI (30 m, 40 m, 48 m and 60 m) fusion, there is a uniform relationship between SRR and FVI quality (CC and RMSE) for ARVI (92.56%, 22.59%), (92.11%, 23.05%), (91.88% and 23.43%) and (91.17% and 23.45%) in SRR 1:8, 1:10, 1:12 and 1:15, respectively.
Hence, this research shows how image fusion influences VI quality with a critical analysis. The research confirms that the quality of FVI is good even for small SRR (1:6 to 1:15) as long as they have a minimum SRF effect and a good SNR. It would be very helpful for making high-resolution VI map using a lower SRR (not limited to SRR 1:8 or higher) in different applications like agricultural, environmental, and climate studies. Even in small SRR, VI mapping has a good quality, which means we can use high-resolution PAN with low-resolution MS band for image fusion to get high-resolution MS bands to calculate FVI. Although the PAN band should have a high resolution, we can use low-resolution MS bands to get high-resolution MS band and FVI, which will minimize the cost for data acquisition. However, the use of a higher number of fusion algorithms, more VI and more quality indices could give a better result for a comparative FVI quality analysis. Similarly, testing the result in different images like Sentinel-2, World View-4, IKO NAS, SPOT and making a decision could have a better result. In the future, minimizing these limitations, and the developing image fusion algorithms with the least effect on SRR, SRF and SNR for band–band fusion and FVI could be new research topics to gain a better quality FVI.

Author Contributions

P.G. and D.L. designed and developed the research idea. N.J. downloaded the data (GF1, GF2 and GF4) and pre-processed. P.G. downloaded and processed all remaining data and wrote the manuscript. P.G. and D.L. contributed to result and data interpretation, discussion and revision of the manuscript. All the authors revised and approved the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research & Development Program of China (no.2018YFC0706004).

Acknowledgments

The authors are very thankful for Ms Zou Han Yue and Ms Qiao Dan Yu for their valuable support.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. El-mezouar, M.C.; Taleb, N.; Kpalma, K.; Ronsin, J. A high-resolution index for vegetation extraction in IKONOS images. In Proceedings of the SPIE on Remote Sensing for Agriculture, Ecosystems, and Hydrology, Toulouse, France, 22 October 2010; Volume 78242, pp. 1–9. [Google Scholar]
  2. Ehlers, M.; Klonus, S.; Astrand, P.J.; Rosso, P. Multi-sensor image fusion for pansharpening in remote sensing. Int. J. Image Data Fusion 2010, 1, 25–45. [Google Scholar] [CrossRef]
  3. Swathika, R.; Sharmila, T.S. Image fusion for MODIS and Landsat images using top hat based moving technique with FIS. Clust. Comput. 2019, 22, 12939–12947. [Google Scholar] [CrossRef]
  4. Jiang, D.; Zhuang, D.; Huang, Y. Investigation of Image Fusion for Remote Sensing Application. Available online: www.intechopen.com/books/new-advances-in-image-fusion/investigation-of-image-fusion-for-remote-sensing-application (accessed on 2 May 2020).
  5. Zeng, Y.; Zhang, J.; Genderen, J.L.; Zhang, Y. Image fusion for land cover change detection. Int. J. Image Data Fusion 2010, 1, 193–215. [Google Scholar] [CrossRef]
  6. Sarp, G. Spectral and spatial quality analysis of pan-sharpening algorithms: A case study in Istanbul. Eur. J. Remote Sens. 2014, 47, 19–28. [Google Scholar] [CrossRef] [Green Version]
  7. Sagan, V.; Maimaitijiang, M.; Sidike, P.; Maimaitiyiming, M.; Erkbol, H.; Hartling, S.; Peterson, K.T.; Peterson, J.; Burken, J.; Fritschi, F. Uav/satellite multiscale data fusion for crop monitoring and early stress detection. In Proceedings of the The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Enschede, The Netherlands, 10–14 June 2019; Volume XLII, pp. 715–722. [Google Scholar]
  8. Ghimire, P.; Deng, L. Image Fusion Technique: Algorithm to Application. Nepal. J. Geoinform. 2019, 18, 27–34. [Google Scholar]
  9. Gholinejad, S.; Fatemi, S.B. Optimum indices for vegetation cover change detection in the Zayandeh-rud river basin: A fusion approach. Int. J. Image Data Fusion 2019, 10, 199–216. [Google Scholar] [CrossRef]
  10. Roy, D.P.; Kovalskyy, V.; Zhang, H.K.; Vermote, E.F.; Yan, L.; Kumar, S.S.; Egorov, A. Characterization of Landsat-7 to Landsat-8 reflective wavelength and normalized difference vegetation index continuity. Remote Sens. Environ. 2016, 185, 57–70. [Google Scholar] [CrossRef] [Green Version]
  11. Xue, J.; Su, B. Significant Remote Sensing Vegetation Indices: A Review of Developments and Applications. J. Sens. 2017, 2017, 1–17. [Google Scholar] [CrossRef] [Green Version]
  12. Ahmad, F. Spectral vegetation indices performance evaluated for Cholistan Desert. J. Geogr. Reg. Plan. 2012, 5, 165–172. [Google Scholar]
  13. Johnson, B. Effects of pansharpening on vegetation indices. ISPRS Int. J. Geo-Inf. 2014, 3, 507–522. [Google Scholar] [CrossRef]
  14. Jagalingam, P.; Hegde, A.V. A Review of Quality Metrics for Fused Image. In Proceedings of the International Conference on Water Resources, Coastal and Ocean Engineering (ICWRCOE 2015), Mangalore, India, 12–14 March 2015; Elsevier: Amsterdam, The Netherlands, 2015; Volume 4, pp. 133–142. [Google Scholar]
  15. Zhu, X.X.; Bamler, R. A Sparse Image Fusion Algorithm With Application to Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
  16. Liu, Y.; Chen, X.; Ward, R.K.; Wang, Z.J. Image Fusion With Convolutional Sparse Representation. IEEE Signal. Process. Lett. 2016, 23, 1882–1886. [Google Scholar] [CrossRef]
  17. Li, X.; Wang, L.; Cheng, Q.; Wu, P.; Gan, W.; Fang, L. Cloud removal in remote sensing images using nonnegative matrix factorization and error correction. ISPRS J. Photogramm. Remote Sens. 2019, 148, 103–113. [Google Scholar] [CrossRef]
  18. Li, X.; Shen, H.; Zhang, L.; Zhang, H.; Yuan, Q.; Yang, G. Recovering Quantitative Remote Sensing Products Contaminated by Thick Clouds and Shadows Using Multitemporal Dictionary Learning. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7086–7098. [Google Scholar]
  19. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery using Pan-Sharpening. U.S. Patent 6011875, 4 January 2000. [Google Scholar]
  20. Klonus, S.; Ehlers, M. Performance of evaluation methods in image fusion. In Proceedings of the 12th International Conference on Information Fusion, Seattle, WA, USA, 6–9 July 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1409–1416. [Google Scholar]
  21. Klonus, S.; Ehlers, M. Image fusion using the Ehlers spectral characteristics preservation algorithm. Giscience Remote Sens. 2007, 44, 93–116. [Google Scholar] [CrossRef]
  22. Kang, T.; Zhang, X.; Wang, H. Assessment of the fused image of multispectral and panchromatic images of SPOT5 in the investigation of geological hazards. Sci. China Ser. E Technol. Sci. 2008, 51, 144–153. [Google Scholar] [CrossRef]
  23. Wu, J.; Jiang, P. A complete no-reference image quality assessment method based on local feature. Int. J. Image Data Fusion 2019, 10, 165–176. [Google Scholar] [CrossRef]
  24. Sahu, D.K.; Parsai, M.P. Different Image Fusion Techniques-A Critical Review. Int. J. Mod. Eng. Res. 2012, 2, 4298–4301. [Google Scholar]
  25. Jawak, S.D.; Luis, A.J. A Comprehensive Evaluation of PAN-Sharpening Algorithms Coupled with Resampling Methods for Image Synthesis of Very High Resolution Remotely Sensed Satellite Data. Adv. Remote Sens. 2013, 2, 332–344. [Google Scholar] [CrossRef] [Green Version]
  26. Nikolakopoulos, K.; Oikonomidis, D. Quality assessment of ten fusion techniques applied on Worldview-2. Eur. J. Remote Sens. 2015, 48, 141–167. [Google Scholar] [CrossRef]
  27. Sekrecka, A.; Kedzierski, M. Integration of Satellite Data with High Resolution Ratio: Improvement of Spectral Quality with Preserving Spatial Details. Sensors 2018, 18, 4418. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Ehlers, M. Multi-image Fusionin Remote Sensing: Spatial Enhancementvs. Spectral Characteristics Preservation. In Advances in Visual Computing, Proceedings of the 4th International Symposium, ISVC 2008,Part II, Las Vegas, NV, USA, 1–3 December 2008; Bebis, G., Parvin, B., Koracin, D., Boyle, R., Porikli, F., Peters, J., Klosowski, J., Rhyne, T.-M., Ams, L., Chun, Y.K., et al., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 75–84. [Google Scholar]
  29. Ling, Y.; Ehlers, M.; Usery, E.L.; Madden, M. Effects of spatial resolution ratio in image fusion. Int. J. Remote Sens. 2008, 29, 2157–2167. [Google Scholar] [CrossRef]
  30. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  31. Rahaman, K.R.; Hassan, Q.K.; Ahmed, M.R. Pan-Sharpening of Landsat-8 Images and Its Application in Calculating Vegetation Greenness and Canopy Water Contents. ISPRS Int. J. Geo-Inf. 2017, 6, 168. [Google Scholar] [CrossRef] [Green Version]
  32. Hwang, T.; Song, C.; Bolstad, P.V.; Band, L.E. Downscaling real-time vegetation dynamics by fusing multi-temporal MODIS and Landsat NDVI in topographically complex terrain. Remote Sens. Environ. 2011, 115, 2499–2512. [Google Scholar] [CrossRef]
  33. Hassan, Q.K.; Bourque, C.P.-A.; Meng, F.-R. Application of Landsat-7 ETM+ and MODIS products in mapping seasonal accumulation of growing degree days at an enhanced resolution. J. Appl. Remote Sens. 2007, 1, 013539. [Google Scholar] [CrossRef]
  34. Dao, P.D.; Mong, N.T.; Chan, H.-P. Landsat-MODIS image fusion and object-based image analysis for observing flood inundation in a heterogeneous vegetated scene. Giscience Remote Sens. 2019, 56, 1148–1169. [Google Scholar] [CrossRef]
  35. He, G.; Fang, H.; Bai, S.; Liu, X.; Chen, M.; Bai, J. Application of a three-dimensional eutrophication model for the Beijing Guanting Reservoir, China. Ecol. Model. 2011, 222, 1491–1501. [Google Scholar] [CrossRef]
  36. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  37. Vermote, E.F.; Roger, J.C.; Ray, J.P. MODIS Surface Reflectance User’s Guide. Available online: https://lpdaac.usgs.gov/documents/306/MOD09_User_Guide_V6.pdf (accessed on 5 April 2020).
  38. Che, X.; Feng, M.; Sexton, J.O.; Channan, S.; Yang, Y.; Sun, Q. Assessment of MODIS BRDF/Albedo Model Parameters (MCD43A1 Collection 6) for Directional Reflectance Retrieval. Remote Sens. 2017, 9, 1123. [Google Scholar] [CrossRef] [Green Version]
  39. ENVI Atmospheric Correction Module: QUAC and FLAASH User’s Guide. Available online: http://www.harrisgeospatial.com/portals/0/pdfs/envi/flaash_module.pdf (accessed on 2 May 2020).
  40. Zhang, Z.; Han, D.; Dezert, J.; Yang, Y. A New Image Registration Algorithm Based on Evidential Reasoning. Sensors 2019, 19, 1091. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Feng, R.; Du, Q.; Li, X.; Shen, H. Robust registration for remote sensing images by combining and localizing feature- and area-based methods. ISPRS J. Photogramm. Remote Sens. 2019, 151, 15–26. [Google Scholar] [CrossRef]
  42. Murphy, J.M.; Moigne, J.; Le Harding, D.J.; Space, G. Automatic Image Registration of Multi-Modal Remotely Sensed Data with Global Shearlet Features. IEEE Trans. Geosci. Remote Sens. 2017, 54, 1685–1704. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Jin, X. ENVI Automated Image Registration Solutions. Available online: https://www.harrisgeospatial.com/portals/0/pdfs/ENVI_Image_Registration_Whitepaper.pdf (accessed on 5 April 2020).
  44. Kaufman, Y.J.; Tanre, D. Atmospherically Resistant Vegetation Index (ARVI) for EOS-MODIS. IEEE Trans. Geosci. Remote Sens. 1992, 30, 260–271. [Google Scholar] [CrossRef]
  45. Huete, A.; Didan, K.; Miura, T.; Rodriguez, E.P.; Gao, X.; Ferreira, L.G. Overview of the radiometric and biophysical performance of the MODIS vegetation indices. Remote Sens. Environ. 2002, 83, 195–213. [Google Scholar] [CrossRef]
  46. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a Green Channel in Remote Sensing of Global Vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  47. Cheng, W.-C.; Chang, J.-C.; Chang, C.-P.; Su, Y.; Tu, T.-M. A Fixed-Threshold Approach to Generate High-Resolution Vegetation Maps for IKONOS Imagery. Sensors 2008, 8, 4308–4317. [Google Scholar] [CrossRef]
  48. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-Driven Fusion of High Spatial and Spectral Resolution Images Based on Oversampled Multiresolution Analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  49. Huete, A.R. A Soil-Adjusted Vegetation Index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  50. Budhewar, S.T. Wavelet and Curvelet Transform based Image Fusion Algorithm. Int. J. Comput. Sci. Inf. Technol. 2014, 5, 3703–3707. [Google Scholar]
  51. Gamba, P. Image and data fusion in remote sensing of urban areas: Status issues and research trends. Int. J. Image Data Fusion 2014, 5, 2–12. [Google Scholar] [CrossRef]
  52. Ren, J.; Yang, W.; Yang, X.; Deng, X.; Zhao, H.; Wang, F.; Wang, L. Optimization of Fusion Method for GF-2 Satellite Remote Sensing Images based on the Classification Effect. Earth Sci. Res. J. 2019, 23, 163–169. [Google Scholar] [CrossRef]
  53. Pandit, V.R.; Bhiwani, R.J. Image Fusion in Remote Sensing Applications: A Review. Int. J. Comput. Appl. 2015, 120, 22–32. [Google Scholar]
  54. Fonseca, L.; Namikawa, L.; Castejon, E.; Carvalho, L.; Pinho, C.; Pagamisse, A. Image Fusion for Remote Sensing Applications. Available online: http://www.intechopen.com/books/image-fusion-and-its-applications/image-fusion-for-remote-sensing-applications (accessed on 2 May 2020).
  55. Vibhute, A.D.; Dhumal, R.; Nagne, A.; Gaikwad, S.; Kale, K.V.; Mehrotra, S.C. Multi-Sensor, Multi-Resolution and Multi-Temporal Satellite Data Fusion for Soil Type Classification. In Proceedings of the International Conference on Cognitive Knowledge Engineering, Aurangabad, India, 21–23 December 2016; pp. 27–32. [Google Scholar]
  56. Belgiu, M.; Stein, A. Spatiotemporal Image Fusion in Remote Sensing. Remote Sens. 2019, 11, 818. [Google Scholar] [CrossRef] [Green Version]
  57. Al-Wassai, F.A.; Kalyankar, N.V.; Al-Zaky, A.A. Spatial and Spectral Quality Evaluation Based on Edges Regions of Satellite Image fusion. In Proceedings of the 2012 Second International Conference on Advanced Computing and Communication Technologies (ACCT), Haryana, India, 7–8 January 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 265–275. [Google Scholar]
  58. Zhang, D.; Xie, F.; Zhang, L. Preprocessing and fusion analysis of GF-2 satellite Remote-sensed spatial data. In Proceedings of the 2018 International Conference on Information Systems and Computer Aided Education (ICISCAE), Changchun, China, 6–8 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 24–29. [Google Scholar]
  59. Sun, W.; Chen, B.; Messinger, D.W. Nearest-neighbor diffusion-based pan-sharpening algorithm for spectral images. Opt. Eng. 2014, 53, 013107. [Google Scholar] [CrossRef] [Green Version]
  60. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Location of study area.
Figure 1. Location of study area.
Remotesensing 12 01550 g001
Figure 2. Methodological flow chart of research.
Figure 2. Methodological flow chart of research.
Remotesensing 12 01550 g002
Figure 3. Trend of FVI quality with Image Fusion and SRR (same SRF and good SNR).
Figure 3. Trend of FVI quality with Image Fusion and SRR (same SRF and good SNR).
Remotesensing 12 01550 g003
Figure 4. Trend of FVI quality with Image Fusion and SRR(Different SRF and SNR (Good to Poor)).
Figure 4. Trend of FVI quality with Image Fusion and SRR(Different SRF and SNR (Good to Poor)).
Remotesensing 12 01550 g004
Figure 5. Trend of FVI quality with Image Fusion and SRR (Constant SRF and Good SNR).
Figure 5. Trend of FVI quality with Image Fusion and SRR (Constant SRF and Good SNR).
Remotesensing 12 01550 g005
Table 1. Different images used for the research.
Table 1. Different images used for the research.
Collection DateSatellite ImagesResolutionImager TypeAvailable BandsSelected Bands for This Research
2017/07/13Gaofen-1PAN-2 m, MS-8 mPushbroom with TDI capability4B1/blue   0.45-0.52 µm
B2/green  0.52-0.59 µm
B3/red  0.63-0.69 µm
B4/NIR  0.77-0.89 µm
PAN  0.45-0.90 µm
Gaofen-2PAN-1 m, MS-4 m
Gaofen-4PAN-50 m, MS-50 m
2017/07/17Landsat-8 OLIMS-30 mOperational Land Imager (OLI)11B1/blue  0.45-0.51 µm
B2/green  0.53-0.59 µm
B3/red  0.64-0.67 µm
B4/NIR  0.85-0.88 µm
2017/07/13MODISMS-500 mMODIS Payload Imaging Sensor7B3/blue  0.45-0.47 µm
B4/green  0.54-0.56µm
B1/red  0.62-0.67 µm
B2/NIR  0.84-0.87 µm
Table 2. Different image fusion algorithms for the research.
Table 2. Different image fusion algorithms for the research.
AlgorithmsDetailsAdvantagesReferences
GSA generalization of PCA, in which PC1 may be arbitrarily chosen and the remaining components are calculated to be orthogonal or uncorrelated to one another.Spectral characteristics of lower-spatial-resolution MS data are preserved.[6,19,20]
EhlersEnhances high-frequency changes such as edges and grey level discontinuities in an image.Spectral characteristic preserving.[21]
PCReplacement of PC1 by high-resolution image, which contains the most information of the image resembles.Gives high-resolution MS images.[6,22,23,24]
MIHSMaking forward PCA transform for both intensity component(I) and high-resolution panchromatic images getting the first component as I‘. Matching the histogram of I’ component with I to make I’’ and finally transferring I’’, H, S to RGB true color space.Involves both HIS and PCA methods, Preserves both spatial and spectral details.[25,26]
HPF High-frequency spatial content of the pan image is extracted using a high-pass filter and transferred to the resampled MS image.Preserves a high percentage of the spectral characteristics of the MS image.[22,23]
Table 3. Different VI for the research.
Table 3. Different VI for the research.
VIsCharacteristics DefinitionReferences
ARVIThe improvement is much better for vegetated surfaces than for soils. (NIR-2*R-B)/(NIR+2*R-B)[44]
EVIEstimates vegetation LAI, biomass and water content, and improve sensitivity in high biomass region.2.5*[(NIR-R)/(NIR+6*R-7.5*B+1)][45]
GNDVIHas a wider dynamic range than the NDVI and is, on average, at least five times more sensitive to chlorophyll concentration.(NIR-G)/(NIR+G)[44,46]
NDVIDetects change in the amount of green biomass efficiently in vegetation with low to moderate density.(NIR-R)/(NIR+R)[9,11,12,13,47]
OSAVIOSAVI does not depend on the soil line and can eliminate the influence of the soil background effectively. Where L is around 0.16.(L+1)*[(NIR-R)/(NIR+R+L+1)[11,48]
SAVIMinimizes soil brightness influences from spectral vegetation indices involving red and near-infrared (NIR) wavelengths. Where L is around 0.5.(L+1)*[(NIR-R)/(NIR+R+L+1)[22,49]
Table 4. Strategy for image fusion and quality assessment.
Table 4. Strategy for image fusion and quality assessment.
SRRFusion TypeOriginal ResolutionResampled ResolutionActivity/ ComputationQuality Indices
1:6Same-Sensor
(GF2+GF2)
GF2 (PAN-1 m)
GF2 (MS-4 m)
GF2 (MS-1 m)
GF2 (MS-6 m)
1. Compute 1 m RVI and 1 m FVI
2. Assess RVI and FVI
Multi-Sensor
(GF1+GF4)
GF1(PAN-2 m)
GF1(MS-8 m)
GF4( MS-50 m)
GF1(PAN-8 m)
GF4( MS-50 m)
1. Compute 8 m RVI and 8 m FVI
2. Assess RVI and FVI
1:8Same-Sensor
(GF2+GF2)
GF2 (PAN-1 m)
GF2 (MS-4 m)
GF2 (MS-1 m)
GF2 (MS-8 m)
1. Compute 1 m RVI and 1 m FVI
2. Assess RVI and FVI
Multi-Sensor
(GF2+Landsat-8 OLI)
GF2 (PAN-1 m)
GF2 (MS-4 m)
LS-8 (MS-30 m)
GF2 (PAN-4 m) 1. Compute 4 m RVI and 4 m FVI
2. Assess RVI and FVI
1:10Same-Sensor
(GF2+GF2)
GF2 (PAN-1 m)
GF2 (MS-4 m)
GF2 (MS-1 m)
GF2 (MS-10 m)
1. Compute 1 m RVI and 1 m FVI
2. Assess RVI and FVI
1.Objective (CC, RMSE)
Multi-Sensor
(GF4+MODIS)
GF4 (PAN-50 m)
GF4 (MS-50 m)
MODIS (MS-500 m)
--1. Compute 50 m RVI and 50 m FVI
2. Assess RVI and FVI
Multi-Sensor
(GF2+Landsat-8 OLI)
GF2 (PAN-1 m)
GF1(MS-4 m)
LS-8( MS-30 m)
GF1(PAN-4 m)
LS-8 (MS-40 m)
1. Compute 4 m RVI and 4 m FVI
2. Assess RVI and FVI
1:12Same-Sensor
(GF2+GF2)
GF2 (PAN-1 m)
GF2 (MS-4 m)
GF2 (MS-1 m)
GF2 (MS-12 m)
1. Compute 1 m RVI and 1 m FVI
2. Assess RVI and FVI
2.Subjective
(Visual Inspection)
Multi-Sensor
(GF2+GF4)
GF2 (PAN-1 m)
GF2 (MS-4 m)
GF4( MS-50 m)
GF2 (PAN-4 m) 1. Compute 4 m RVI and 4 m FVI
2. Assess RVI and FVI
Multi-Sensor
(GF2+Landsat-8 OLI)
GF2 (PAN-1 m)
GF1(MS-4 m)
LS-8( MS-30 m)
GF2(PAN-4 m)
LS-8 (MS-48 m)
1. Compute 4 m RVI and 4 m FVI
2. Assess RVI and FVI
Table 5. Quality of FVI result in single-sensor fusion (values in bold font indicate good quality).
Table 5. Quality of FVI result in single-sensor fusion (values in bold font indicate good quality).
SRRIndices (%)Image FusionARVIEVIGNDVINDVIOSAVISAVI
1:6
GF2+GF2
CCGS99.0598.8599.0398.8698.8798.86
Ehlers97.7697.3298.1397.3697.3797.36
PC97.6294.4796.3497.3997.4097.40
MIHS98.2897.6298.7397.8897.8997.91
HPF92.3596.4694.3696.2896.2996.33
RMSEGS12.5812.2610.0613.4813.5013.11
Ehlers16.3016.6511.9616.1916.1916.00
PC13.3616.1018.0826.5128.0531.09
MIHS10.9014.759.8514.1614.1414.00
HPF57.4339.3946.7141.1349.3055.96
1:8
GF2+GF2
CCGS99.0898.9499.0498.9398.9498.93
Ehlers97.1595.4897.1895.9395.5395.83
PC94.1696.6194.0497.4197.4197.42
MIHS97.5696.4498.0596.7796.7796.78
HPF63.9762.9182.3087.5384.2075.19
RMSEGS13.40 12.0012.19 13.5413.3512.66
Ehlers16.14 24.30 16.02 24.30 24.30 23.97
PC40.94 29.58 32.28 25.16 26.93 30.33
MIHS12.9717.76 11.8617.17 17.17 17.06
HPF53.22 54.16 46.54 61.62 60.91 37.50
1:10
GF2+GF2
CCGS98.7397.5798.6897.5097.3997.40
Ehlers96.8195.5997.0495.7295.7295.72
PC97.1396.6895.896.6096.6196.61
MIHS97.3196.0789.2697.0497.0492.11
HPF89.2284.7786.9982.1182.1082.11
RMSEGS12.8811.3310.8114.6014.5614.66
Ehlers18.6821.9816.7322.0022.2121.75
PC29.3417.9719.1331.2217.1218.11
MIHS13.6018.8413.1727.1527.1536.74
HPF63.0061.1246.6566.1765.6565.38
1:12
GF2+GF2
CCGS98.5998.1598.5398.1798.1798.17
Ehlers95.7293.4195.8893.5793.5793.57
PC96.9596.0695.5896.2596.2496.20
MIHS96.2994.2996.9194.7894.7894.78
HPF93.7692.5593.5792.3693.3792.43
RMSEGS13.8614.7012.7715.7815.7815.43
Ehlers20.1927.6018.0426.9826.9826.70
PC14.8524.7019.2447.4720.4523.85
MIHS15.9722.7215.2021.8221.8221.73
HPF74.7471.2726.6526.1725.6525.38
Table 6. Quality of FVI result in multi-sensor fusion (values in bold font indicate good quality).
Table 6. Quality of FVI result in multi-sensor fusion (values in bold font indicate good quality).
SRRIndices (%)Image FusionARVIEVIGNDVINDVIOSAVISAVI
1:6
(GF1+GF4)
CCGS84.2778.2184.2478.93 78.93 78.92
Ehlers83.3478.5484.6579.7279.7279.70
PC77.0070.4773.2171.0471.0471.04
MIHS81.7678.16 84.1979.5579.5579.55
HPF82.8077.5581.8478.4778.4778.47
RMSEGS33.9742.0733.6041.6941.6841.68
Ehlers34.42 44.23 39.47 42.79 42.79 42.79
PC40.16 49.22 41.92 48.21 48.21 48.21
MIHS61.33 43.06 36.61 41.77 41.79 42.52
HPF36.59 42.97 36.62 41.73 41.74 41.74
1:8
(GF2+Landsat-8)
CCGS92.5689.4593.1389.5789.5789.57
Ehlers92.2689.0292.9889.4489.4489.44
PC91.0885.9990.0086.1586.1586.15
MIHS91.8287.2392.3887.8187.8287.82
HPF90.6987.9788.7883.2376.0676.08
RMSEGS22.5929.6921.3529.5729.5729.56
Ehlers22.6830.2821.5929.6129.6029.60
PC24.4733.9725.5133.8133.8133.81
MIHS23.3632.5922.6431.7131.7131.71
HPF48.3232.1329.3952.8846.3351.47
1:10
(GF4+MODIS)
CCGS82.2079.2989.9379.4779.5079.46
Ehlers82.8680.0088.3580.1381.4080.14
PC81.5675.8489.6876.4576.6476.77
MIHS83.7180.1787.1379.8379.8479.81
HPF75.5777.1287.2676.6476.6576.65
RMSEGS28.9433.7622.8633.3533.3633.00
Ehlers28.3533.0724.3432.6931.6831.65
PC29.2635.9823.1435.4035.1435.30
MIHS27.7233.2725.4333.2033.2133.21
HPF35.1135.8325.3336.1236.1236.11
1:12
(GF2+GF4)
CCGS91.5990.9191.7588.2388.2588.24
Ehlers90.4490.5190.7888.2788.2688.27
PC90.0090.2189.2186.2486.2286.24
MIHS76.6590.7590.8888.6185.5788.62
HPF79.9584.1888.6887.4083.8187.40
RMSEGS25.2525.2324.8634.6733.6634.67
Ehlers27.0925.2326.3234.6734.3434.66
PC27.5842.7728.4437.3341.7637.32
MIHS42.7731.9426.1034.1834.3934.18
HPF38.4140.9431.3835.8340.1235.83
Table 7. Quality of FVI result in multi-sensor fusion (values in bold font indicate good quality).
Table 7. Quality of FVI result in multi-sensor fusion (values in bold font indicate good quality).
SRRIndices (%)Image FusionARVIEVIGNDVINDVIOSAVISAVI
1:10
(GF2+Landsat-8 (40 m))
CCGS92.1188.7792.7388.9388.9288.93
Ehlers91.6587.4692.1287.6587.6687.65
PC90.7485.5889.8185.7685.7685.77
MIHS90.7285.7091.4786.4486.4286.42
HPF89.3386.9388.1282.3375.8375.85
RMSEGS23.0530.5821.8430.3430.3330.32
Ehlers23.4232.2422.7231.6531.6331.65
PC24.7934.4325.6534.2634.2134.26
MIHS26.1836.1225.0134.9434.9334.93
HPF51.3234.1331.3953.8953.3653.47
1:12
(GF2+Landsat-8 (48 m))
CCGS91.8888.4392.5488.6088.5988.60
Ehlers91.3486.5291.5287.1587.1587.16
PC90.5485.2189.6585.4485.4385.47
MIHS89.5484.9290.5285.0085.0085.11
HPF88.6386.1387.5281.8375.1375.15
RMSEGS23.4331.0222.1030.7530.7930.75
Ehlers24.0333.4023.5532.4532.4532.43
PC25.0034.8225.8434.5534.5534.44
MIHS25.2034.9324.1733.8433.0433.05
HPF52.5235.6332.1954.0954.2654.42
1:15
(GF2+Landsat-8 (60 m))
CCGS91.1788.1992.4088.3688.3588.36
Ehlers90.5585.7991.5287.1587.1587.16
PC90.4284.9889.5585.2585.2485.24
MIHS89.5484.9290.5285.0085.0085.11
HPF88.1485.4386.0281.2474.5374.65
RMSEGS23.4531.3222.3031.0431.0831.05
Ehlers24.8734.2123.9633.2433.2433.25
PC25.0034.8225.8434.5534.5534.56
MIHS26.1836.1225.0134.9434.9334.92
HPF53.2136.1333.2156.2156.3656.43
Table 8. Visual quality of FVI result in single/multi-sensor fusion.
Table 8. Visual quality of FVI result in single/multi-sensor fusion.
RVI and Fusion AlgorithmsSame-Sen. ARVI(1 m)Multi-Sen. EVI(8 m)Same-Sen. NDVI(1 m)Multi-Sen. GNDVI(8 m)Same-Sen. OSAVI(1 m)Multi-Sen. SAVI(8 m)
RVI (1 m, 8 m) Remotesensing 12 01550 i001 Remotesensing 12 01550 i002 Remotesensing 12 01550 i003 Remotesensing 12 01550 i004 Remotesensing 12 01550 i005 Remotesensing 12 01550 i006
GS (1:6) Remotesensing 12 01550 i007 Remotesensing 12 01550 i008 Remotesensing 12 01550 i009 Remotesensing 12 01550 i010 Remotesensing 12 01550 i011 Remotesensing 12 01550 i012
Ehlers (1:6) Remotesensing 12 01550 i013 Remotesensing 12 01550 i014 Remotesensing 12 01550 i015 Remotesensing 12 01550 i016 Remotesensing 12 01550 i017 Remotesensing 12 01550 i018
PC ( 1:6) Remotesensing 12 01550 i019 Remotesensing 12 01550 i020 Remotesensing 12 01550 i021 Remotesensing 12 01550 i022 Remotesensing 12 01550 i023 Remotesensing 12 01550 i024
MIHS (1:6) Remotesensing 12 01550 i025 Remotesensing 12 01550 i026 Remotesensing 12 01550 i027 Remotesensing 12 01550 i028 Remotesensing 12 01550 i029 Remotesensing 12 01550 i030
HPF (1:6) Remotesensing 12 01550 i031 Remotesensing 12 01550 i032 Remotesensing 12 01550 i033 Remotesensing 12 01550 i034 Remotesensing 12 01550 i035 Remotesensing 12 01550 i036
RVI and Fusion AlgorithmsSame-Sen. ARVI(1 m)Multi-Sen. EVI(4 m)Same-Sen. NDVI(1 m)Multi-Sen. GNDVI(4 m)Same-Sen. OSAVI(1 m)Multi-Sen. SAVI(4 m)
RVI (1 m, 4 m) Remotesensing 12 01550 i037 Remotesensing 12 01550 i038 Remotesensing 12 01550 i039 Remotesensing 12 01550 i040 Remotesensing 12 01550 i041 Remotesensing 12 01550 i042
GS (1:8) Remotesensing 12 01550 i043 Remotesensing 12 01550 i044 Remotesensing 12 01550 i045 Remotesensing 12 01550 i046 Remotesensing 12 01550 i047 Remotesensing 12 01550 i048
Ehlers (1:8) Remotesensing 12 01550 i049 Remotesensing 12 01550 i050 Remotesensing 12 01550 i051 Remotesensing 12 01550 i052 Remotesensing 12 01550 i053 Remotesensing 12 01550 i054
PC ( 1:8) Remotesensing 12 01550 i055 Remotesensing 12 01550 i056 Remotesensing 12 01550 i057 Remotesensing 12 01550 i058 Remotesensing 12 01550 i059 Remotesensing 12 01550 i060
MIHS (1:8) Remotesensing 12 01550 i061 Remotesensing 12 01550 i062 Remotesensing 12 01550 i063 Remotesensing 12 01550 i064 Remotesensing 12 01550 i065 Remotesensing 12 01550 i066
HPF (1:8) Remotesensing 12 01550 i067 Remotesensing 12 01550 i068 Remotesensing 12 01550 i069 Remotesensing 12 01550 i070 Remotesensing 12 01550 i071 Remotesensing 12 01550 i072
RVI and Fusion AlgorithmsSame-Sen. ARVI(1 m)Multi-Sen. EVI(50 m)Multi-Sen(R) NDVI(4 m)Same-Sen. GNDVI(1 m)Multi-Sen. OSAVI(50 m)Multi-Sen(R) SAVI(4 m)
RVI (1, 50, 4 m) Remotesensing 12 01550 i073 Remotesensing 12 01550 i074 Remotesensing 12 01550 i075 Remotesensing 12 01550 i076 Remotesensing 12 01550 i077 Remotesensing 12 01550 i078
GS (1:10) Remotesensing 12 01550 i079 Remotesensing 12 01550 i080 Remotesensing 12 01550 i081 Remotesensing 12 01550 i082 Remotesensing 12 01550 i083 Remotesensing 12 01550 i084
Ehlers (1:10) Remotesensing 12 01550 i085 Remotesensing 12 01550 i086 Remotesensing 12 01550 i087 Remotesensing 12 01550 i088 Remotesensing 12 01550 i089 Remotesensing 12 01550 i090
PC ( 1:10) Remotesensing 12 01550 i091 Remotesensing 12 01550 i092 Remotesensing 12 01550 i093 Remotesensing 12 01550 i094 Remotesensing 12 01550 i095 Remotesensing 12 01550 i096
MIHS (1:10) Remotesensing 12 01550 i097 Remotesensing 12 01550 i098 Remotesensing 12 01550 i099 Remotesensing 12 01550 i100 Remotesensing 12 01550 i101 Remotesensing 12 01550 i102
HPF (1:10) Remotesensing 12 01550 i103 Remotesensing 12 01550 i104 Remotesensing 12 01550 i105 Remotesensing 12 01550 i106 Remotesensing 12 01550 i107 Remotesensing 12 01550 i108
RVI and Fusion AlgorithmsSame-Sen. ARVI(1 m)Multi-Sen. EVI(4 m)Multi-Sen(R) NDVI(4 m)Same-Sen. GNDVI(1 m)Multi-Sen. OSAVI(4 m)Multi-Sen(R) SAVI(4 m)
RVI (1 m, 4 m) Remotesensing 12 01550 i109 Remotesensing 12 01550 i110 Remotesensing 12 01550 i111 Remotesensing 12 01550 i112 Remotesensing 12 01550 i113 Remotesensing 12 01550 i114
GS (1:12) Remotesensing 12 01550 i115 Remotesensing 12 01550 i116 Remotesensing 12 01550 i117 Remotesensing 12 01550 i118 Remotesensing 12 01550 i119 Remotesensing 12 01550 i120
Ehlers (1:12) Remotesensing 12 01550 i121 Remotesensing 12 01550 i122 Remotesensing 12 01550 i123 Remotesensing 12 01550 i124 Remotesensing 12 01550 i125 Remotesensing 12 01550 i126
PC ( 1:12) Remotesensing 12 01550 i127 Remotesensing 12 01550 i128 Remotesensing 12 01550 i129 Remotesensing 12 01550 i130 Remotesensing 12 01550 i131 Remotesensing 12 01550 i132
MIHS (1:12) Remotesensing 12 01550 i133 Remotesensing 12 01550 i134 Remotesensing 12 01550 i135 Remotesensing 12 01550 i136 Remotesensing 12 01550 i137 Remotesensing 12 01550 i138
HPF (1:12) Remotesensing 12 01550 i139 Remotesensing 12 01550 i140 Remotesensing 12 01550 i141 Remotesensing 12 01550 i142 Remotesensing 12 01550 i143 Remotesensing 12 01550 i144

Share and Cite

MDPI and ACS Style

Ghimire, P.; Lei, D.; Juan, N. Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery. Remote Sens. 2020, 12, 1550. https://doi.org/10.3390/rs12101550

AMA Style

Ghimire P, Lei D, Juan N. Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery. Remote Sensing. 2020; 12(10):1550. https://doi.org/10.3390/rs12101550

Chicago/Turabian Style

Ghimire, Prakash, Deng Lei, and Nie Juan. 2020. "Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery" Remote Sensing 12, no. 10: 1550. https://doi.org/10.3390/rs12101550

APA Style

Ghimire, P., Lei, D., & Juan, N. (2020). Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery. Remote Sensing, 12(10), 1550. https://doi.org/10.3390/rs12101550

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop