Enhancing Soil Moisture Forecasting Accuracy with REDF-LSTM: Integrating Residual En-Decoding and Feature Attention Mechanisms
Abstract
:1. Introduction
2. Materials and Methods
2.1. Data Description
2.2. REDF-LSTM Model
2.2.1. Long Short-Term Memory Network (LSTM)
2.2.2. Feedforward Attention Mechanism
- Feature Weight Allocation: In the context of soil moisture prediction, the feedforward attention mechanism first assesses the importance of each feature within the input feature set. This is accomplished by calculating each feature’s contribution to the prediction target. The mechanism generates a weight value for each feature, indicating the relative importance of various features in predicting soil moisture.
- Dynamic Adjustment: A key characteristic of the feedforward attention mechanism is its dynamism. It can automatically adjust feature weights based on different input data and environmental conditions. For instance, if a particular area experiences sudden rainfall, the weights of moisture-related features (such as recent rainfall amounts and the current moisture state of the soil) might increase, as these factors become more critical in subsequent soil moisture predictions.
- Weighted Input Synthesis: After weight assignment, the feedforward attention mechanism applies these weights to the corresponding features to generate an “attention-weighted feature representation”. This representation focuses on those features most crucial to the prediction outcome, allowing the model to concentrate more on these key pieces of information.
- Enhancing Prediction Accuracy and Interpretability: By emphasizing important features and suppressing less significant information, the feedforward attention mechanism not only enhances prediction accuracy but also improves the model’s interpretability. The model can explicitly indicate which factors are key to influencing the prediction outcomes, which is invaluable for formulating management strategies and decision support in practical applications.
2.2.3. REDF-LSTM
- Residual Learning Encoder–Decoder Structure: In Equation (1), x(t) includes current atmospheric conditions and land surface state data. x(t) is a vector containing current precipitation, temperature, specific humidity, wind speed, and surface radiation. These inputs are multiplied by the weight matrix Wix and combined with the previous moment’s hidden state h(t − 1) and cell state c(t − 1) to compute the activation values of the input gate. In Equation (2), similar to the input gate, x(t) provides information about the current environmental state, assisting the model in deciding what information should be forgotten from the cell state. In Equation (3), the update of the cell state is directly controlled by the input and forget gates, combining new input data with the previous moment’s cell state to form a new cell state. In Equations (4) and (5), the output gate determines which information will be passed to the hidden state h(t) at the next time step, affecting the model’s predictive output and the long-term maintenance of internal memory. Equation (6) generates the model’s prediction of soil moisture at the current moment, based on the hidden state, reflecting the cumulative impact of all past inputs. This structure, by introducing residual connections, allows direct transmission of information across different layers of the model, effectively reducing the loss of information during transmission. The encoding stage deeply mines the features of the input data and forms a high-level feature representation, while the decoding stage uses these advanced features in conjunction with original input features to participate in the final prediction, enhancing the model’s understanding of the intrinsic structure and dynamic changes in the data, thus reducing the uncertainty in prediction.
- Feedforward Attention Mechanism: The soil moisture time series, as referenced in Equation (6), is fed into the feedforward attention mechanism. By dynamically allocating feature weights across different time steps, the feedforward attention mechanism enables the model to focus on the most influential features for the prediction outcomes. Unlike traditional LSTM models that treat all input features equally, the feedforward attention mechanism adjusts weights based on the actual impact of the data, thus improving the accuracy and efficiency of predictions. This mechanism is particularly important in dealing with nonlinear and nonstationary characteristics in time series, effectively enhancing the model’s responsiveness to sudden or significant events.
- Input layer: As the entry point of the model, the input layer is responsible for handling complex multi-dimensional time series data. This layer transforms the input data into a three-dimensional array format suitable for the LSTM structure, with dimensions (number of samples S, time steps T = {T1, T2, … Ti}, feature dimensions X = {X1, X2, … Xj}). Here, S represents the total number of samples in the dataset, while i and j denote specific indices for time steps and feature dimensions, respectively.
- Encoder–decoder LSTM layer: The encoder deeply encodes multi-dimensional input features to create a new comprehensive feature representation. Then, the decoder phase combines this comprehensive representation with the original input features. Through this strategy, the model reveals the inherent connections in the data, enhancing understanding of the input data’s structure and dynamics. This enhanced feature set significantly reduces the model’s uncertainty in predicting soil moisture, ensuring closer correspondence of predictions to actual observational data.
- Fully connected LSTM layer: The objective of this layer is to integrate the output from the encoder–decoder LSTM segment with the initial input soil moisture characteristics, mitigating overfitting and adjusting for any predictive discrepancies originating from the encoder–decoder LSTM structure.
- Attention layer: In this layer, a feature attention mechanism is used to calculate weights for each time step in the input sequence, denoted as , where m is the total number of time steps and represents the feature weight for the mth time step. By using these dynamically adjusted weights, the model dynamically adjusts these weights through integrating time series data, thus allocating more attention to important data and reducing attention to less important information.
- Output layer: The task of this layer is to output the model’s prediction results, with an output format of (output sample number O, time steps H). The design of this layer ensures that the model can make precise predictions for future sequences based on historical and current input information.
- Model Optimization and Simplification: Investigating more efficient network architectures or simplifying the REDF-LSTM model through techniques like model pruning could reduce its computational resource demands. This would make the model more suitable for deployment in environments with limited computing capabilities, broadening its application scope.
- Enhancing Data Efficiency: Developing new training strategies or data augmentation techniques to reduce the model’s reliance on large amounts of training data. For example, utilizing data from different but related domains through transfer learning or semi-supervised learning methods could enhance the model’s training process.
- Improving Model Adaptability and Robustness: Introducing more adaptive and self-regulating mechanisms to better handle extreme weather events and other unconventional inputs. For instance, integrating graph neural networks could improve the understanding and prediction of interactions within complex climate networks.
- Exploring Cross-Domain Applications: By testing and validating the REDF-LSTM model in different environmental and application contexts, its versatility and effectiveness can be further assessed and optimized. This not only promotes the model’s application beyond soil moisture prediction to other environmental monitoring tasks but also deepens the understanding of its performance in various practical scenarios.
2.2.4. Model Setting, Training
2.2.5. Model Evaluation
3. Results
3.1. Box Plot Comparison of Model Performances
3.2. Visual Comparison of Model Performance
3.3. Time Series Plots Comparison of Model Performance
3.4. Comparative Error Bar Graph of Model Performance
3.5. Forecasting the Seventh Day Ahead and Predictive Results Surface_Sensible_Heat_Flux
3.6. Ablation Experiments
4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Denissen, J.M.C.; Orth, R.; Wouters, H.; Miralles, D.G.; van Heerwaarden, C.C.; de Arellano, J.V.-G.; Teuling, A.J. Soil moisture signature in global weather balloon soundings. npj Clim. Atmos. Sci. 2021, 4, 13. [Google Scholar] [CrossRef]
- Koster, R.D.; Dirmeyer, P.A.; Guo, Z.; Bonan, G.; Chan, E.; Cox, P.; Gordon, C.T.; Kanae, S.; Kowalczyk, E.; Lawrence, D.; et al. Regions of strong coupling between soil moisture and precipitation. Science 2004, 305, 1138–1140. [Google Scholar] [CrossRef]
- Norbiato, D.; Borga, M.; Esposti, S.D.; Gaume, E.; Anquetin, S. Flash flood warning based on rainfall thresholds and soil moisture conditions: An assessment for gauged and ungauged basins. J. Hydrol. 2008, 362, 274–290. [Google Scholar] [CrossRef]
- Dirmeyer, P.A.; Gao, X.; Zhao, M.; Guo, Z.; Oki, T.; Hanasaki, N. GSWP-2: Multimodel Analysis and Implications for Our Perception of the Land Surface. Bull. Am. Meteorol. Soc. 2006, 87, 1381–1397. [Google Scholar] [CrossRef]
- Li, Q.; Li, Z.; Shangguan, W.; Wang, X.; Li, L.; Yu, F. Improving soil moisture prediction using a novel encoder-decoder model with residual learning. Comput. Electron. Agric. 2022, 195, 106816. [Google Scholar] [CrossRef]
- Rosenbaum, U.; Bogena, H.; Herbst, M.; Huisman, J.A.; Peterson, T.; Weuthen, A.; Western, A.W.; Vereecken, H. Seasonal and event dynamics of spatial soil moisture patterns at the small catchment scale. Water Resour. Res. 2012, 48, W10544. [Google Scholar] [CrossRef]
- Li, Q.; Zhao, Y.; Yu, F. A Novel Multichannel Long Short-Term Memory Method with Time Series for Soil Temperature Modeling. IEEE Access 2020, 8, 182026–182043. [Google Scholar] [CrossRef]
- Martinez, C.; Hancock, G.R.; Kalma, J.D.; Wells, T. Spatio-temporal distribution of near-surface and root zone soil moisture at the catchment scale. Hydrol. Process. 2008, 22, 2699–2714. [Google Scholar] [CrossRef]
- Zhu, S.; Chen, H.; Dong, X.; Wei, J. Influence of persistence and oceanic forcing on global soil moisture predictability. Clim. Dyn. 2020, 54, 3375–3385. [Google Scholar] [CrossRef]
- Shoaib, M.; Shamseldin, A.Y.; Melville, B.W.; Khan, M.M. A comparison between wavelet based static and dynamic neural network approaches for runoff prediction. J. Hydrol. 2016, 535, 211–225. [Google Scholar] [CrossRef]
- Henderson-Sellers, A.; Yang, Z.-L.; Dickinson, R.E. The Project for Intercomparison of Land-surface Parameterization Schemes. Bull. Am. Meteorol. Soc. 1993, 74, 1335–1349. [Google Scholar] [CrossRef]
- Rasp, S.; Dueben, P.D.; Scher, S.; Weyn, J.A.; Mouatadid, S.; Thuerey, N. WeatherBench: A benchmark data set for data-driven weather forecasting. J. Adv. Model. Earth Syst. 2020, 12, e2020MS002203. [Google Scholar] [CrossRef]
- Sabzipour, B.; Arsenault, R.; Troin, M.; Martel, J.-L.; Brissette, F.; Brunet, F.; Mai, J. Comparing a long short-term memory (LSTM) neural network with a physically-based hydrological model for streamflow forecasting over a Canadian catchment. J. Hydrol. 2023, 627, 130380. [Google Scholar] [CrossRef]
- Li, Q.; Wang, Z.; Shangguan, W.; Li, L.; Yao, Y.; Yu, F. Improved daily SMAP satellite soil moisture prediction over China using deep learning model with transfer learning. J. Hydrol. 2021, 600, 126698. [Google Scholar] [CrossRef]
- Lei, G.; Zeng, W.; Yu, J.; Huang, J. A comparison of physical-based and machine learning modeling for soil salt dynamics in crop fields. Agric. Water Manag. 2023, 277, 108115. [Google Scholar] [CrossRef]
- Jung, C.; Lee, Y.; Cho, Y.; Kim, S. A Study of Spatial Soil Moisture Estimation Using a Multiple Linear Regression Model and MODIS Land Surface Temperature Data Corrected by Conditional Merging. Remote Sens. 2017, 9, 870. [Google Scholar] [CrossRef]
- Shrestha, N.K.; Shukla, S. Support vector machine based modeling of evapotranspiration using hydro-climatic variables in a sub-tropical environment. Bioresour. Technol. 2013, 128, 351–358. [Google Scholar] [CrossRef]
- Pan, J.; Shangguan, W.; Li, L.; Yuan, H.; Zhang, S.; Lu, X.; Wei, N.; Dai, Y. Using data-driven methods to explore the predictability of surface soil moisture with FLUXNET site data. Hydrol. Process. 2019, 33, 2978–2996. [Google Scholar] [CrossRef]
- Li, Q.; Hao, H.; Zhao, Y.; Geng, Q.; Liu, G.; Zhang, Y.; Yu, F. GANs-LSTM Model for Soil Temperature Estimation from Meteorological: A New Approach. IEEE Access 2020, 8, 59427–59443. [Google Scholar] [CrossRef]
- Li, P.; Zha, Y.; Shi, L.; Tso, C.-H.; Zhang, Y.; Zeng, W. Comparison of the use of a physical-based model with data assimilation and machine learning methods for simulating soil water dynamics. J. Hydrol. 2020, 584, 124692. [Google Scholar] [CrossRef]
- Gumiere, S.J.; Camporese, M.; Botto, A.; Lafond, J.A.; Paniconi, C.; Gallichand, J.; Rousseau, A.N. Machine Learning vs. Physics-Based Modeling for Real-Time Irrigation Management. Front. Water 2020, 2, 8. [Google Scholar] [CrossRef]
- Zhang, G. Synergistic advantages of deep learning and reinforcement learning in economic forecasting. Int. J. Glob. Econ. Manag. 2023, 1, 89–95. [Google Scholar] [CrossRef]
- Hou, X.; Feng, Y.; Wu, G.; He, Y.; Chang, D.; Yang, H. Application research on artificial neural network dynamic prediction model of soil moisture. Water Sav. Irrigation. 2016, 7, 70–72. [Google Scholar]
- Li, N.; Zhang, Q.; Yang, F.X.; Deng, Z.L. Research of adaptive genetic neural network algorithm in soil moisture prediction. Comput. Eng. Appl. 2018, 54, 54–59+69. [Google Scholar]
- Tesch, T.; Kollet, S.; Garcke, J. Causal deep learning models for studying the Earth system. Geosci. Model Dev. 2023, 16, 2149–2166. [Google Scholar] [CrossRef]
- Chen, Z.; Zhang, R.; Song, Y.; Wan, X.; Li, G. Advancing Visual Grounding with Scene Knowledge: Benchmark and Method. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
- Ding, J.; Wen, L.; Zhong, C.; Loffeld, O. Video SAR Moving Target Indication Using Deep Neural Network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7194–7204. [Google Scholar] [CrossRef]
- Li, Q.; Zhang, C.; Shangguan, W.; Li, L.; Dai, Y. A novel local-global dependency deep learning model for soil mapping. Geoderma 2023, 438, 116649. [Google Scholar] [CrossRef]
- Fang, K.; Shen, C. Full-flow-regime storage-streamflow correlation patterns provide insights into hydrologic functioning over the continental US. Water Resour. Res. 2017, 53, 8064–8083. [Google Scholar] [CrossRef]
- Fang, K.; Shen, C. Near-Real-Time Forecast of Satellite-Based Soil Moisture Using Long Short-Term Memory with an Adaptive Data Integration Kernel. J. Hydrometeorol. 2020, 21, 399–413. [Google Scholar] [CrossRef]
- Lal, P.; Shekhar, A.; Gharun, M.; Das, N.N. Spatiotemporal evolution of global long-term patterns of soil moisture. Sci. Total. Environ. 2023, 867, 161470. [Google Scholar] [CrossRef]
- Datta, P.; Faroughi, S.A. A multihead LSTM technique for prognostic prediction of soil moisture. Geoderma 2023, 433, 116452. [Google Scholar] [CrossRef]
- Li, Q.; Zhu, Y.; Shangguan, W.; Wang, X.; Li, L.; Yu, F. An attention-aware LSTM model for soil moisture and soil temperature prediction. Geoderma 2022, 409, 115651. [Google Scholar] [CrossRef]
- Li, L.; Dai, Y.; Shangguan, W.; Wei, N.; Wei, Z.; Gupta, S. Multistep Forecasting of Soil Moisture Using Spatiotemporal Deep Encoder–Decoder Networks. J. Hydrometeorol. 2022, 1, 337–350. [Google Scholar] [CrossRef]
- Wang, Y.; Shi, L.; Hu, Y.; Hu, X.; Song, W.; Wang, L. A comprehensive study of deep learning for soil moisture prediction. Hydrol. Earth Syst. Sci. 2024, 28, 917–943. [Google Scholar] [CrossRef]
- Reichstein, M.; Camps-Valls, G.; Stevens, B.; Jung, M.; Denzler, J.; Carvalhais, N.; Prabhat, F. Deep learning and process understanding for data-driven Earth system science. Nature 2019, 566, 195–204. [Google Scholar] [CrossRef]
- Khanal, S.; Lutz, A.F.; Immerzeel, W.W.; de Vries, H.; Wanders, N.; Hurk, B. The Impact of Meteorological and Hydrological Memory on Compound Peak Flows in the Rhine River Basin. Atmosphere 2019, 10, 171. [Google Scholar] [CrossRef]
- Raffel, C.; Ellis, D.P. Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems. arXiv 2015, arXiv:1512.08756. [Google Scholar]
- Anshuman, A.; Eldho, T. A parallel workflow framework using encoder-decoder LSTMs for uncertainty quantification in contaminant source identification in groundwater. J. Hydrol. 2023, 619, 129296. [Google Scholar] [CrossRef]
- Yang, Y.; Gao, P.; Sun, Z.; Wang, H.; Lu, M.; Liu, Y.; Hu, J. Multistep ahead prediction of temperature and humidity in solar greenhouse based on FAM-LSTM model. Comput. Electron. Agric. 2023, 213, 108261. [Google Scholar] [CrossRef]
- Zhai, W.; Mo, G.; Xiao, Y.; Xiong, X.; Wu, C.; Zhang, X.; Xu, Z.; Pan, J. GAN-BiLSTM network for field-road classification on imbalanced GNSS recordings. Comput. Electron. Agric. 2024, 216, 108457. [Google Scholar] [CrossRef]
- Cao, B.; Gruber, S.; Zheng, D.; Li, X. The ERA5-Land soil temperature bias in permafrost regions. Cryosphere 2020, 14, 2581–2595. [Google Scholar] [CrossRef]
- Zhang, J.; Zhu, Y.; Zhang, X.; Ye, M.; Yang, J. Developing a Long Short-Term Memory (LSTM) based model for predicting water table depth in agricultural areas. J. Hydrol. 2018, 561, 918–929. [Google Scholar] [CrossRef]
- Chen, H.; Huang, J.J.; Dash, S.S.; Wei, Y.; Li, H. A hybrid deep learning framework with physical process description for simulation of evapotranspiration. J. Hydrol. 2022, 606, 127422. [Google Scholar] [CrossRef]
- Guo, M.-H.; Xu, T.-X.; Liu, J.-J.; Liu, Z.-N.; Jiang, P.-T.; Mu, T.-J.; Zhang, S.-H.; Martin, R.R.; Cheng, M.-M.; Hu, S.-M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media. 2022, 8, 331–368. [Google Scholar] [CrossRef]
- Wei, H.; Zhu, M.; Wang, B.; Wang, J.; Sun, D. Two-Level Progressive Attention Convolutional Network for Fine-Grained Image Recognition. IEEE Access 2020, 8, 104985–104995. [Google Scholar] [CrossRef]
- Li, L.; Shangguan, W.; Deng, Y.; Mao, J.; Pan, J.; Wei, N.; Yuan, H.; Zhang, S.; Zhang, Y.; Dai, Y. A Causal Inference Model Based on Random Forests to Identify the Effect of Soil Moisture on Precipitation. J. Hydrometeorol. 2020, 21, 1115–1131. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, TO, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Li, Q.; Zhang, C.; Shangguan, W.; Wei, Z.; Yuan, H.; Zhu, J.; Li, X.; Li, L.; Li, G.; Liu, P.; et al. LandBench 1.0: A benchmark dataset and evaluation metrics for data-driven land surface variables prediction. Expert Syst. Appl. 2023, 243, 122917. [Google Scholar] [CrossRef]
- Hao, H.; Yu, F.; Li, Q. Soil Temperature prediction using convolutional neural network based on ensemble empirical mode decomposition. IEEE Access 2021, 9, 4084–4096. [Google Scholar] [CrossRef]
- Zhao, Z.; Yao, X.; Xu, K.; Song, J.; Chen, X. Water yield of mine analysis and prediction method based on EEMD-PSO- ELM-LSTM model. arxiv 2023. [Google Scholar] [CrossRef]
Learning Rate | Hidden Size | Batch Size | Epoch | Niter | R |
---|---|---|---|---|---|
1 × 10−2 | 256 | 64 | 1000 | 400 | 0.8587 |
1 × 10−3 | 256 | 64 | 1000 | 400 | 0.9506 |
1 × 10−4 | 256 | 64 | 1000 | 400 | 0.9383 |
1 × 10−3 | 256 | 64 | 800 | 400 | 0.9465 |
1 × 10−3 | 256 | 128 | 1000 | 400 | 0.9412 |
1 × 10−3 | 256 | 32 | 1000 | 400 | 0.9492 |
1 × 10−3 | 256 | 64 | 1200 | 400 | 0.9407 |
1 × 10−3 | 128 | 64 | 1000 | 400 | 0.9453 |
1 × 10−3 | 512 | 64 | 1000 | 400 | 0.9467 |
1 × 10−3 | 256 | 64 | 1000 | 300 | 0.9451 |
1 × 10−3 | 256 | 64 | 1000 | 500 | 0.9483 |
Method | R | KGE | RMSE | BIAS |
---|---|---|---|---|
LSTM | 0.917 | 0.749 | 0.024 | 0.019 |
ED-LSTM | 0.934 | 0.856 | 0.020 | 0.015 |
FAM-LSTM | 0.943 | 0.860 | 0.019 | 0.014 |
REDF-LSTM | 0.951 | 0.869 | 0.013 | 0.013 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, X.; Zhang, Z.; Li, Q.; Zhu, J. Enhancing Soil Moisture Forecasting Accuracy with REDF-LSTM: Integrating Residual En-Decoding and Feature Attention Mechanisms. Water 2024, 16, 1376. https://doi.org/10.3390/w16101376
Li X, Zhang Z, Li Q, Zhu J. Enhancing Soil Moisture Forecasting Accuracy with REDF-LSTM: Integrating Residual En-Decoding and Feature Attention Mechanisms. Water. 2024; 16(10):1376. https://doi.org/10.3390/w16101376
Chicago/Turabian StyleLi, Xiaoning, Ziyin Zhang, Qingliang Li, and Jinlong Zhu. 2024. "Enhancing Soil Moisture Forecasting Accuracy with REDF-LSTM: Integrating Residual En-Decoding and Feature Attention Mechanisms" Water 16, no. 10: 1376. https://doi.org/10.3390/w16101376
APA StyleLi, X., Zhang, Z., Li, Q., & Zhu, J. (2024). Enhancing Soil Moisture Forecasting Accuracy with REDF-LSTM: Integrating Residual En-Decoding and Feature Attention Mechanisms. Water, 16(10), 1376. https://doi.org/10.3390/w16101376