1. Introduction
The Normalized Difference Vegetation Index (NDVI) is a fundamental remote-sensing measure of land surface greenness that is widely used to assess ecosystem properties such as vegetation phenology, biomass production, soil moisture, and carbon sequestration [
1,
2,
3]. NDVI also supports the further analyses like classification and regionalization [
4], evaluating temporal dynamics and trends [
5], and change detection [
6]. However, the quality of NDVI products is often compromised by factors such as atmospheric conditions (e.g., clouds, dust, etc.), snow cover, varying sun-sensor-surface viewing geometries, and sensor faults. Noise from these factors must be carefully processed to avoid unreliable analyses [
7]. Thus, employing data reconstruction (or noise reduction/restoration) techniques is crucial to enhance NDVI quality and ensure the reliability of its applications.
Over the past several decades, numerous techniques have been developed to reconstruct NDVI data. For instance, the resulting NDVI products must account for the bidirectional reflectance distribution function (BRDF) effect, and BRDF normalization is commonly used to mitigate noise introduced by observation and illumination angular effects [
8]. Based on the use of temporal and spatial information, NDVI reconstruction methods can be generally categorized into (1) time-series-based methods, which reconstruct NDVI time series at each pixel individually without considering spatial associations between NDVI values at different locations, and (2) spatio-temporal methods, which simultaneously leverage temporal and spatial interrelations for reconstruction. Time-series-based methods have been well developed and widely applied in NDVI reconstruction, remaining prevalent in the field. These methods typically apply a mathematical filter to NDVI time series, either in the temporal domain or the frequency domain, to reduce noise. Specifically, methods in the temporal domain include the 4253H twice filter [
9]; the best index slope extraction algorithm (BISE) [
10,
11]; asymmetric Gaussian (AG) function fitting [
12]; the Savitzky–Golay (SG) filter [
13]; the Mean-Value Iteration (MVI) filter [
7]; double logistic function (DL) fitting [
14]; the Whittaker smoother (WS) [
15]; the changing-weight filter method [
16]; and the RMMEH filter [
17]. Methods in the frequency domain include Fourier analysis [
18]; harmonic analysis (HA) [
19]; and wavelet transformation [
20]. Despite their widespread use, time-series-based methods have significant limitations due to their dependence on incomplete information. If an NDVI time series is dominated by data noise (imprecise information) or data gaps (missing information), particularly continuous noisy entries, reconstruction becomes challenging due to the lack of essential information. Unfortunately, low-quality NDVI time-series data are common in NDVI products. To enhance the effectiveness of time-series-based methods, researchers have recently explored additional information from the spatial domain of NDVI data for reconstruction. These are spatio-temporal NDVI-reconstruction methods. For example, Poggio et al. [
21] proposed a hybrid Generalized Additive Model (GAM)–geostatistical space-time model (denoted as GAM-Geo) that combines a multidimensional GAM fitting the spatio-temporal trends of NDVI with a geostatistical approach interpolating the residual components of NDVI over space. Oliveira et al. [
22] developed a window regression (WR) method, which replaces noisy NDVI values with new values regressed from the relationship between the NDVI time series at the target pixel and one selected from the nearest eight neighbors. Other methods, such as the correction based on spatial and temporal continuity (CSaTC) method [
23] and the temporal-spatial iteration (TSI) method [
24], iteratively restore some noisy NDVI values using data from high-quality pixels selected either temporally or spatially close to the noisy pixel.
In theory, spatio-temporal methods are expected to outperform time-series-based methods due to the additional information they utilize; however, these emerging methods are still immature, and each has its own drawbacks. For instance, the GAM-Geo method utilizes a multidimensional GAM model to incorporate spatio-temporal information, but fitting this model could be subjective due to the tradeoff between function smoothness and fitting accuracy. Moreover, neither the smooth function used in the GAM model nor the geostatistical approach (specifically, simple kriging interpolation) can precisely model non-continuous patterns of NDVI variability across space. Both the TSI method and the CSaTC method require supplementary data (such as land cover types or ecological zone maps) in addition to NDVI data. They also use spatial and temporal information separately, meaning that for each noisy pixel, the reconstructed NDVI values are determined using either spatial or temporal information alone. This approach can lead to an imbalance in the ratio between the pixel counts of the two aspects. Additionally, the WR method has been shown to perform even worse than time-series-based methods and is practically incapable of reconstructing all noisy pixels [
24].
Despite previous efforts, two major obstacles hinder the effective use of spatial information in NDVI data reconstruction. First, distinct from the temporal dimension, NDVI’s spatial information lies in a two-dimensional space, adding complexity to its exploitation. To address this, some methods (e.g., WR, CsaTC, and TSI) extract spatial information from a very limited number of pixels, typically just one, selected within a small sliding window. Second, NDVI variability often exhibits unsmooth patterns over space, especially in high-resolution NDVI products. Abrupt transitions in land cover types or topographical features lead to uneven textures and irregular structures in NDVI spatial profiles. It is also worth mentioning that the introduction of spatial information into time-series reconstruction depends on the ground sampling distance and the related spatial resolution of the input data. Consequently, many common spatial data-processing techniques, such as smoothing, filtering, and interpolation, may fail in NDVI reconstruction. Due to the lack of effective solutions for such issues, methods like WR, CsaTC, and TSI must carefully select the most suitable pixel that is homogeneous to the noisy pixel based on land cover type, phenological pattern, or other criteria.
Thus, the objective of this study is to develop a novel NDVI reconstruction method that can take full advantage of the spatio-temporal interrelations in NDVI data. Inspired by recent advancements in graph signal processing, we applied a temporal-difference graph (TDG) method [
25] to NDVI reconstruction. Originally designed to reconstruct partially-sampled time-varying graph signals, the TDG method is particularly well suited for NDVI reconstruction because it addresses the two major obstacles identified in previous efforts. First, rather than relying on local spatial windows, this method models spatial interrelations among data pixels using a graph structure, where every two pixels are directly or indirectly connected by edges. Second, the TDG method does not rely on the spatial smoothness of the data. Instead, it operates on the assumption that the temporal differences of the data (the first-order derivative of data versus time) should be smooth across space. This assumption is more realistic for NDVI data while NDVI values may not be smooth over space; the changes in NDVI over short periods tend to correlate strongly at adjacent locations. Based on these two principles, the TDG method effectively utilizes both temporal and spatial information in an asymmetric and rational manner.
2. Methodology
This section provides a clear description of the method used for reconstructing NDVI data, and
Figure 1 illustrates a flowchart of the research methodology.
2.1. Graph Theory and Graph Signal Processing
Graph theory is a fundamental branch of mathematics that investigates structures consisting of vertices and edges, collectively known as graphs. A graph conceptualizes pairwise relations between objects; each vertex represents an object, and each edge represents the relationship between two linked objects. This generalized structure is versatile and has been used to model various types of relations and processes across a wide range of domains, including computer science, physics, biology, and engineering [
26,
27].
For clarity, we introduce some formal notation. An undirected and weighted graph is denoted as , where denotes the vertex set , denotes the edge set , and denotes the adjacency matrix of the graph. Each individual vertex and edge are denoted as and , respectively, with and representing the total number of vertices and edges, respectively. Each edge links two vertices, and we may alternatively denote an edge as , indicating the edge that links vertices and . These vertices are called the end vertices of edge . The adjacency matrix is an matrix defined as follows. If there exists an edge that connects two vertices and , the corresponding entry (the entry in the ith and the jth row) equals the weight associated with edge ; otherwise, equals zero. The setting of edge weights is flexible and depends on the problem at hand. A common strategy involves assigning edge weights based on the similarity between the two vertices connected by the edge, and this approach is prevalent in studies focusing on large-scale graph modeling of real-world systems, such as social networks, computer networks, and biological networks. In addition to the adjacency matrix, another common matrix representation for a graph is the Laplacian matrix , defined as , where D is a diagonal matrix with the th diagonal element equal to the degree of vertex , i.e., the sum of the weights of all the edges directly linked with vertex . Note that is positive-semidefinite, which follows from its symmetry and diagonal dominance.
Graph theory has been developed for centuries, and even nowadays, with the expansion of its applications, new methods and techniques that merge graph theory with other domains are continuously emerging. Among these subdomains, an emerging field known as graph signal processing (GSP), or signal processing on graphs, has received great attention in recent years [
28]. Graph signal processing combines graph theory and signal processing to deal with signal-processing problems such as filtering, compression, and reconstruction, specifically for data whose inherent associations or structures are represented by a graph. These data are collectively referred to as a graph signal, with each data sample attached to a vertex of the graph. Intuitively, the vertices in the graph act as signal sources, while the edges depict the relationships of these signal sources. We used the vector
to denote a graph signal built on a graph
. The
th component of
, denoted by
, is the data sample at the
th vertex
in the graph. If the data sample at each vertex is not a single value but a time series, this signal is referred to as a time-varying graph signal. In these cases, the signal is represented as the matrix
, where
is the number of time instances. Each column of this matrix denotes the samples at all vertices in the graph at a single time instance, and each row represents the time-series samples at a single vertex throughout time. It is worth mentioning that we only consider graph signals with synchronous time instances for the time series at each vertex.
2.2. Time-Varying Graph Signal Reconstruction: The Temporal-Difference Graph Method
The realm that inspires our solution to the NDVI reconstruction problem is graph signal reconstruction, one of the core research topics in graph signal processing. A graph signal
is considered partially sampled when some of its entries are not observed, prompting the need for methods to recover the missing entries based on the known entries. A fundamental approach to reconstructing graph signals involves assuming that the signal is smooth on the graph [
29], meaning the missing entries are assigned to the values that maximize the overall smoothness of the signal. In most of the literature [
25,
28,
30], the smoothness of a graph signal is typically measured as follows:
where
is the Laplacian matrix of the related graph. The smaller
is, the better the smoothness that the graph signal exhibits. If we unfold the equation, we obtain a more intuitive form, as follows:
which means the smoothness of a graph signal is measured as the weighted squared sum of the signal difference between every vertex pair linked by an edge. For a time-varying graph signal
, its smoothness can be calculated as the sum of the smoothness values at every time instance, as follows:
where
calculates the trace of a given matrix, and
denotes the
th column of
, i.e., the graph signal at time instance
.
One way to reconstruct time-varying graph signals is to directly minimize
. However, this method essentially reconstructs the graph signal at each time instance separately, disregarding the relationship between temporally-adjacent signals. Moreover, some signals in nature, including NDVI, exhibit non-smooth behavior over space and, therefore, reconstructing signals based solely on spatial smoothness may lead to significant deviations from reality. Due to the influence of geographic conditions on vegetation growth, NDVI signals would inherit spatial patterns from various factors such as topography, land cover, and soil and water variability, resulting in uneven textures and irregular structures in NDVI spatial profiles. To reconstruct such non-smooth time-varying graph signals, Qiu et al. [
25] proposed a new method based on the fact that for many time-varying graph signals in nature, the temporal differences of the signal often exhibit better smoothness than the original signals. We refer to this method as the TDG method, and the follow-up Equations (4) and (5) can also be found in Qiu et al. [
25].
The temporal differences of a time-varying graph signal
are expressed as:
where
is a
matrix used for compact notation. The TDG method recovers missing entries in
by maximizing the smoothness of the temporal difference signal
rather than the smoothness of the original signal
. Mathematically, the reconstruction problem can be interpreted as a constrained optimization problem; determining the optimal
that minimizes
(maximize smoothness) while keeping the sampled entries in
unchanged. This can be formulated as follows:
where
is the objective function,
is the decision variable, and
are given constant matrices. The constraint
represents the sampling of
, where
is a sampling operator of the same size as
, and
denotes the entry-wise product. Entries of
are equal to 1 if the corresponding entries in
are observed, and equal to 0 if the corresponding entries in
are missing. In other words,
represents the observed or reliable part of
, which is determined to be fixed. Since
is positive-semidefinite and
is in a variant quadratic form, it can be inferred that
is a convex (downward) function. Qiu et al. [
25] solved this optimization problem using a gradient projection method, employing a backtracking line search approach to determine the step size. According to convex optimization theory, this algorithm theoretically converges to the global optimal solution after sufficient iterations.
The TDG method has great potential for reconstructing NDVI data because it aligns with a key characteristic of NDVI data—the short-term changes in NDVI between adjacent regions tend to be highly consistent. Even in heterogeneous areas, the temporal differences in NDVI often show greater smoothness than the original data. This phenomenon likely arises because weather conditions—the primary factor influencing vegetation growth—tend to exhibit spatial smoothness and act as a “soft” filter applied to the original NDVI data.
2.3. Applying the TDG Model to NDVI Reconstruction
2.3.1. Grid-to-Graph Conversion
To apply the TDG method to NDVI reconstruction, the NDVI data must first be interpreted as a time-varying graph signal. As the most widely used NDVI products, MODIS data are commonly applied in global evaluations of reconstruction methods [
31]. MODIS NDVI data are typically gridded, with each pixel positioned at the cells of a regular grid. Hence, we propose a method to establish graph structures on the grid. In this method, each pixel of the original grid serves as a vertex of the target graph directly. To determine the edges, we used a
k-nearest-neighbor rule, linking each vertex only to its
k nearest neighboring vertices.
Figure 2 shows two examples (with four and eight nearest neighbors). The weights of edges are determined by the distances between connected vertices. When two connected vertices are close, they tend to exhibit high similarity in NDVI variability, so we assign a larger weight to their connecting edge. By default, we set the weight of each edge to be inversely proportional to the distance between the two connected vertices. Particularly, when using the four-nearest-neighbor rule, all edges have the same weight.
Accordingly, a weighted graph is established on the original grid, and the NDVI data can be viewed as a time-varying graph signal represented by a matrix . Each column of , denoted by , represents the spatial profile of NDVI over the region of interest at a specific time instance .
2.3.2. Sampling of the NDVI Signal
The NDVI time-varying graph signal is partially sampled due to the presence of noise and data gaps, which we treat as missing entries in a graph signal. Thus, the NDVI reconstruction problem is framed as a time-varying graph signal reconstruction problem. This is further formulated as a constrained optimization problem in Equation (5) using the TDG method. The optimal solution to this problem yields the final reconstructed NDVI data. In this approach, spatial correlations within NDVI data are modeled by the graph structure, while temporal correlations are captured by the temporal difference signals. The utilization of spatial and temporal information for NDVI reconstruction is global, asymmetric, and reasonable.
The identification of noise in NDVI data is not within the scope of this study. Nowadays, many NDVI products provide pixel-level quality assessment information along with the NDVI data. It is straightforward to distinguish between reliable observations and noise based on such information.
2.3.3. Algorithm to Solve the Problem
We employed the gradient-projection algorithm with backtracking line search, as described by Qiu et al. [
25], to solve the problem. Given
, and an initial signal
, the algorithm updates the signal along the negative gradient direction step by step, gradually approaching the optimal solution after sufficient iterations. For a detailed description of the procedures of this algorithm, readers can refer to Qiu et al. [
25]. There is one slight difference in our approach; instead of initializing all noisy entries in the NDVI signal to zero, as this algorithm normally does, we used the preprocessed NDVI values (described in
Section 3) as the initial signal
. This benefits solving the optimization problem by providing prior estimations of the true NDVI values.
The graph involved in this study is generally large in scale because graphs built on a NDVI grid are very dense. Specifically, we are dealing with graphs consisting of
vertices and even more edges, which is much larger in scale than the application examples demonstrated by Qiu et al. [
25], by about two orders of magnitude. Therefore, solving the problem could be time consuming. To accelerate the computation, we adopted two strategies in practice: First, we utilized vector/matrix operations instead of entry-wise operations as much as possible in our implementation of the solving algorithm. Such operations can be significantly accelerated using Basic Linear Algebra Subprograms (BLAS, see Blackford et al. [
32]). Second, we simplified the formula for calculating the smoothness of temporal difference signals. The counterpart of Equation (2) for temporal difference signals is as follows:
where
denotes the signal sample on vertex
at time instance
. Note that the smoothness related to reliable entries is always constant during optimizing iterations, whereas what really matters is only the smoothness related to noise. Thus, we only need to calculate the noise-related portion of smoothness, that is
where
is a subset of the entire edge set
. Each of the edges in
has at least one end vertex that is diagnosed as noise at time instance
or
. By doing so, the quantity of computation is reduced substantially.
2.4. Savitzky–Golay Filter Method
For comparison purposes, we implemented a time-series-based NDVI reconstruction method known as the SG filter method [
13]. While the basic idea of the SG filter method is to smooth NDVI time series, Chen et al. [
13] designed an iterative procedure to replace incompatible depressed values in the NDVI time series with corresponding larger values generated by SG filtering. Since its proposal, the SG filter method has been widely used to preprocess NDVI data in numerous studies [
33,
34,
35].
Recent studies comparing various NDVI reconstruction methods suggest that the SG filter method represents the current state-of-the-art. Michishita et al. [
36] compared seven NDVI reconstruction methods and identified the SG filter method and the RMMEH method [
17] as the top choices for the Poyang Lake area in China. Zhou et al. [
37] revealed that the SG filter outperformed four other methods in tropical and subtropical regions, while the AG method [
12] excelled in high latitude boreal regions. Overall, although no single method consistently emerges as superior, the SG filter method is widely recognized as a state-of-art approach. In this study, the SG filter method serves as the baseline model for NDVI reconstruction. We aimed to interpret the performance comparison between the TDG method and the SG filter method as a comparison between an advanced spatio-temporal method and the prevailing time-series-based methods.
2.5. Model Evaluation
To assess the performance of the NDVI reconstruction methods, particularly to compare the TDG method and the SG method, we designed a straightforward evaluation framework. First, some artificial noise was introduced into the original NDVI data. Subsequently, each NDVI reconstruction method was separately used to recover the noise-introduced NDVI data. The results were then evaluated by comparing the recovered NDVI values with their original values, considered as ground-truthing. The greater the proximity between the recovered values and the original ones, the better the method’s performance.
In the noise introduction phase, random entries from the reliable data were selected to be contaminated by artificial noise. These chosen entries were then assigned new random values following a uniform distribution within the valid range. Moreover, these entries were labeled as noise (by changing the Pixel Reliability Index (PRI) value as shown in
Table 1), ensuring they are targeted for recovery by the NDVI reconstruction methods. To evaluate the accuracy of the reconstructed results, we employed the root-mean-square error (RMSE) metric. The RMSE between the reconstructed NDVI data matrix
and the original NDVI
is defined as
where the noise-introduction operator
is a matrix whose entries equal 1 if the corresponding NDVI entries are contaminated with artificial noise, or, otherwise, equal to 0. The scalar
is the number of non-zero entries in
, namely, the number of noise-introduced entries. The notation
denotes the Frobenius norm of a given matrix.
5. Conclusions
This study proposed a TDG method for NDVI reconstruction and demonstrated its usefulness in the Three-River Headwaters Region of Northwest China. By converting NDVI data into a time-varying graph signal, the TDG method leverages an optimization algorithm to maximize the spatial smoothness of temporal differences in the original signal. A comparative performance analysis between the TDG method and the SG filter method was conducted, which involved introducing artificial noise into the data and evaluating the RMSE between the recovered NDVI values and the actual values. The results revealed a substantial superiority in the TDG method over the SG filter method for NDVI reconstruction. The average RMSE of reconstructed NDVI using the TDG method was 0.026, regardless of whether the noise was positively biased, negatively biased, or linearly interpolated. In contrast, the SG filter method’s average RMSEs were 0.387, 0.074, and 0.038 for the three types of noise, respectively. Moreover, the TDG method exhibited remarkable robustness to noise intensity, requiring only 10% to 20% of the data to achieve high-accuracy reconstruction. Thus, the TDG method shows excellent performance with no preference for noise types and yields good results from both a time-series perspective and spatial-profile perspective.
The TDG method is a very useful tool for reconstructing noisy NDVI products, primarily due to its comprehensive utilization of both the temporal and spatial information in NDVI data. This is achieved through two key aspects—modeling spatial interrelations between data pixels using a graph structure, and exploiting temporal associations of NDVI values through temporal differences, based on the assumption that the temporal-difference signal tends to be spatially smooth. Moreover, the TDG method exhibits universal applicability in its performance for NDVI reconstruction, regardless of noise characteristics, land cover types, vegetation phenology patterns, and other factors. Thus, the proposed method’s potential for extended application to datasets beyond NDVI is highly prospective, and we hope our efforts could provide new insights into remote-sensing data processing. Note that the principle of the TDG method is also applicable to a wide range of remote-sensing-based products. Nevertheless, many issues still need to be addressed. Considering the spatial smoothness in the TDG method, edge effects due to both the temporal and spatial borders would introduce additional uncertainty. Although the computation efficiency in this study is relatively high due to the use of a high-performance server, more tests and simulations are necessary to demonstrate the operational applicability of the TDG method. Furthermore, various combinations of different types of noise should be investigated, including measurement noise, and, also, the principle, configuration, and algorithm of the proposed approach need to be further improved and validated based on various datasets and ground observations in the future.