Crude Oil Price Forecasting Model Based on Neural Networks and Error Correction
Abstract
:1. Introduction
2. Methodology
2.1. Structure of the Hybrid Model
- A.
- Data preprocessing and prediction: The raw data are divided into three parts: training sets, validation sets, and test sets. Then, EWT is applied to decompose the three parts adaptively. The base predictors, the TCN, GRU, and ESN, predict the decomposed sub-series. The details of EWT and predictors are given in Section 2.2 and Section 2.3, respectively.
- B.
- Prediction results ensemble: In this paper, the ensemble method is completed by setting weights of the forecasting results of different predictors. The calculation method is shown in Equation (1). The SARSA algorithm, a reinforcement learning method, is applied to optimize the weights. The SARSA-based ensemble learning method is detailed in Section 2.4.
- C.
- Calculating error and error correction: The forecasting residuals still have predictable components after ensemble learning. Hence, there is still room for further improvement in forecasting accuracy. An extreme learning machine (ELM) is used for error correction. The final crude oil price prediction results are obtained by combining error correction results and model ensemble results. The ECM is detailed in Section 2.5.
2.2. Empirical Wavelet Transform
2.3. Forecasting Methods
2.3.1. Temporal Convolutional Network
2.3.2. Gated Recurrent Unit
2.3.3. Echo State Network
2.4. Multi-Predictor Ensemble
Algorithm 1 Ensemble module based on the SARSA |
Input: Forecasting results for the TCN, GRU, and ESN: . The maximum number of episodes: . The maximum step of each episode: . Discount coefficient: . Learning rate: . Output: Weights of the three predictors: . Algorithm: 1: Initialize all parameters 2: for z = 1: Z do 3: for k = 1: K do 4: Construct loss function and reward : 5: Select through the policy 6: Compute loss function and reward and update table: 7: end for 8: end for |
2.5. Error Correction Module
3. Experiments
3.1. Crude Oil Price Datasets
3.2. Evaluation Indicators
3.3. Experimental Evaluation of the Proposed Model
3.3.1. Comparison with Diverse Benchmark Models
- a.
- The forecasting precision of the deep networks was superior to that of other traditional algorithms in all cases. This shows that the deep networks can better identify the volatility shape of raw data. The probable reason is that deep networks have rich hidden layers, which improves their performance to deal with non-stationary data and could mine the deep information more effectively.
- b.
- The TCN, GRU, and ESN showed better prediction performance than the LSTM and CNN on three datasets. Possible reasons are as follows. First, the TCN incorporates dilated convolution and residual structures. These features enable a large receptive field in a relatively shallow network and maintain stable gradient propagation while capturing long-range dependencies. Second, the GRU employs a simplified gating mechanism that lowers model complexity. Hence, it achieves better training efficiency and faster convergence when handling random fluctuations and long-term dependencies. Third, the ESN utilizes a reservoir with randomly sparse connections, which helps capture dynamic features in crude oil price data more effectively. It also maintains robust performance under highly random conditions. In contrast, the LSTM and CNN have more complex structures or limited receptive fields. When they face high noise and strong non-stationarity, they often encounter unstable training or inadequate long-term dependence capture. Therefore, the TCN, GRU, and ESN offer stronger adaptability and prediction ability for non-stationary and random crude oil price data.
3.3.2. Comparison with Models Utilizing Diverse Ensemble Strategies
- a.
- The prediction errors of all ensemble methods were lower than those of the TCN, GRU, and ESN. Hybrid model 4 improved the accuracy of the single prediction algorithm by 4–12%. This shows that ensemble learning predicted the trend of crude oil price data more accurately than the single predictor. The possible reason is the ensemble method made excellent optimization decisions on ensemble weights according to the volatility of crude oil price series. Hence, ensemble approaches could combine the strengths for different single models to reduce prediction errors.
- b.
- Compared with other models, model 4 showed optimal forecasting precision. This shows that RL can optimize the ensemble weights more effectively than the traditional heuristic method, and improve the precision for the ensemble method. It could be seen that during the process for optimal decision-making, agents adjusted themselves through continuous interaction with their surroundings, which makes the RL more intelligent and the prediction results more accurate.
3.3.3. Comparison with Diverse Decomposition Methods
- a.
- The prediction performance of the proposed model utilizing the decomposition algorithm is more excellent than that of the model without the algorithm. The forecasting performance of models using the algorithm is improved by more than 40%. This shows decomposition methods greatly reduce the high fluctuation of raw data and the prediction errors.
- b.
- In all experiments, the prediction errors of EWT are the lowest among the three decomposition algorithms. Compared with two other decomposition algorithms, EWT could effectually decrease the nonlinearity for raw data and enhance forecasting performance. The main reason is that EWT could adaptively decompose the raw series into multiple subsequences, which enhances the ability of the ensemble method to analyze volatility characteristics.
3.3.4. Comparison with Models Using Error Correction Methods
- a.
- ECM decomposition can effectively correct prediction residuals. In three different datasets, model 9 reduces the , and of model 8 by more than 15%. This indicates that the ECM could well excavate the predictable components hidden in the residuals and improve forecasting performance.
- b.
- Each module of our proposed model can effectively reduce prediction errors. For example, the MAEs of TCN, GRU, ESN, model 4, mode 8, and model 9 were USD 0.7462, USD 0.7680, USD 0.8262, USD 0.7119, USD 0.1728, and USD 0.1239, respectively. This indicates that the RL-based ensemble method in model 4 can make the optimal weight decision when combining different predictors, which can realize the complementary advantages of base predictors. The decomposition method in model 8 reduces the non-stationarity and randomness of raw data and improves the prediction accuracy. Model 9 could correct the prediction residuals based on the first two modules, which could minimize the prediction errors of our hybrid model.
3.4. Supplementary Experiments
- a.
- In all cases, the forecasting precision of the hybrid model was better than that of the base predictor. This indicates that because the crude oil price data showed obvious chaos and fluctuation, it was difficult for a single model to accurately capture the changes in crude oil price, especially at mutation points. Hence, it was necessary to adopt an effective hybrid model to achieve precise forecasting for the crude oil price.
- b.
- Compared with other state-of-the-art models, the EWT-SARSA-TGE-ELM captured the changes at the mutation points more accurately. The main reason is that our model fully combines the advantages of single algorithms, which makes the model more robust and generalizable. EWT adaptively decomposed the original data into multiple sub-sequences to decrease the nonlinearity and randomness of raw data. In addition, the RL-based ensemble method made the optimal weight decision and combined different predictors to achieve the complementary advantages of base predictors. In addition, the ECM well extracted the predictable components hidden in residuals. In the context of the geopolitical conflict, the ECM focused on the leftover signals in residuals that reflected abrupt changes in price trends. A more accurate prediction could be obtained by overlaying the correction output from the ECM with the prediction from the ensemble module. Hence, our proposed model has potential application in crude oil price forecasting.
4. Application Analysis
4.1. Real-Time Adaptability
- a.
- Hybrid models require longer computation time than single models. This is primarily due to hybrid models’ intricate structure, which prolongs computation time compared with individual models. Moreover, using RL for ensemble optimization is slower than employing heuristic algorithms. This occurs because RL requires extensive training iterations during ensemble optimization to explore optimal strategies. During each decision-making step, RL not only evaluates the action values of the current state but also predicts the state and action values of the next step, thereby updating the current strategy. These procedures increase computational complexity. In contrast, heuristic algorithms optimize directly using rules or prior knowledge, requiring fewer iterations and thus achieving higher computational efficiency. However, the maximum time for RL was 105.36 s, which was considerably shorter than the model’s time interval (1 day). Therefore, despite the longer computation time for RL, such durations remain acceptable given the more excellent forecasting performance achieved.
- b.
- A highly adaptive real-time model can rapidly respond to market fluctuations and promptly update forecast outcomes, thus providing investors and decision-makers with timely and accurate references. As shown in the accompanying table, the model’s computational time was in the range of [663.46 s, 1297.93 s] across three datasets. The longest recorded computation time was 300 s. Notably, the shortest dataset interval in this study was one day, indicating that the model’s computation time is considerably shorter than the minimum data update interval. Consequently, the proposed model demonstrated robust real-time adaptability. Furthermore, the proposed model shows substantial potential for real-world applications, providing strong technical support for accurate crude oil price forecasting.
4.2. Application
5. Conclusions and Future Researches
5.1. Conclusions
5.2. Future Research
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
Abbreviations | |
AI | Artificial intelligence |
ARIMA | Autoregressive integrated moving average |
BHA | Black-hole optimization |
BPNN | Back propagation neural network |
CEEMD | Complementary ensemble empirical mode decomposition |
DNN | Deep neural network |
ECM | Error correction method |
EELM | Extended extreme learning machine |
ELM | Extreme learning machine |
EEMD | Ensemble empirical mode decomposition |
ESN | Echo state network |
EWT | Empirical wavelet transform |
GA | Genetic algorithm |
GARCH | Generalized autoregressive conditional heteroskedasticity |
GM | Gray model |
GRNN | Generalized regression neural network |
GWO | Gray wolf optimization |
GRU | Gated recurrent unit |
HA | Historical average |
IMF | Intrinsic mode function |
LSSVM | Least squares support vector machine |
LSTM | Long short-term memory network |
MAE | Mean absolute error |
MAPE | Mean absolute percentage error |
NAR | Nonlinear auto regressive |
OLS | Ordinary least squares |
PSO | Particle swarm optimization |
RL | Reinforcement learning |
RMSE | Root mean square error |
RNN | Recurrent neural network |
SARSA | State action reward state action |
TCN | Temporal convolutional network |
VMD | Variational modal decomposition |
WPD | Wavelet packet decomposition |
Appendix A
Stage | Model/Algorithm | Parameter | Value |
---|---|---|---|
Data pre- processing | - | Split ratio | 3:1:1 |
Sliding window | 5 | ||
Decomposition | EWT | Detection method | scalespace |
Degree for the polynomial interpolation | 6 | ||
Maximum number of bands | 25 | ||
Sampling rate | 1 | ||
Filter width | 10 | ||
Forecasting | ESN | Reservoir size | 400 |
Spectral radius | 0.95 | ||
Input scaling | 0.5 | ||
Reservoir connectivity | 0.1 | ||
TCN | Size of epochs | 100 | |
Size of layers | 4 | ||
Kernel size | 5 | ||
Dropout | 0.15 | ||
Batch size | 16 | ||
Learning rate | 0.01 | ||
GRU | Size of epochs | 100 | |
Size of layers | 3 | ||
Batch size | 16 | ||
Size of hidden units | 100 | ||
Learning rate | 0.01 | ||
Ensemble | SARSA | Maximum number of episodes | 100 |
Maximum step of each episode | 100 | ||
Learning rate | 0.3 | ||
Discount coefficient | 0.95 | ||
Error correction | ELM | Size of hidden neurons | 500 |
Activation function | ReLU |
References
- Kaymak, Ö.Ö.; Kaymak, Y. Prediction of crude oil prices in COVID-19 outbreak using real data. Chaos Solitons Fractals 2022, 158, 111990. [Google Scholar] [CrossRef] [PubMed]
- Miao, H.; Ramchander, S.; Wang, T.; Yang, D. Influential factors in crude oil price forecasting. Energy Econ. 2017, 68, 77–88. [Google Scholar] [CrossRef]
- Bisoi, R.; Dash, P.; Mishra, S. Modes decomposition method in fusion with robust random vector functional link network for crude oil price forecasting. Appl. Soft Comput. 2019, 80, 475–493. [Google Scholar] [CrossRef]
- Wang, D.; Luo, H.; Grunder, O.; Lin, Y.; Guo, H. Multi-step ahead electricity price forecasting using a hybrid model based on two-layer decomposition technique and BP neural network optimized by firefly algorithm. Appl. Energy 2017, 190, 390–407. [Google Scholar] [CrossRef]
- Tang, L.; Wu, Y.; Yu, L. A non-iterative decomposition-ensemble learning paradigm using RVFL network for crude oil price forecasting. Appl. Soft Comput. 2018, 70, 1097–1108. [Google Scholar] [CrossRef]
- Zhang, G.P. Time series forecasting using a hybrid ARIMA and neural network model. Neurocomputing 2003, 50, 159–175. [Google Scholar] [CrossRef]
- Wang, Q.; Song, X. Forecasting China’s oil consumption: A comparison of novel nonlinear-dynamic grey model (GM), linear GM, nonlinear GM and metabolism GM. Energy 2019, 183, 160–171. [Google Scholar] [CrossRef]
- Zhou, X.; Zhou, H.; Long, H. Forecasting the equity premium: Do deep neural network models work? Mod. Financ. 2023, 1, 1–11. [Google Scholar] [CrossRef]
- Chen, J.; Wei, X.; Liu, Y.; Zhao, C.; Liu, Z.; Bao, Z. Deep Learning for Water Quality Prediction—A Case Study of the Huangyang Reservoir. Appl. Sci. 2024, 14, 8755. [Google Scholar] [CrossRef]
- Chen, L.; Pelger, M.; Zhu, J. Deep learning in asset pricing. Manag. Sci. 2024, 70, 714–750. [Google Scholar] [CrossRef]
- Fan, X.; Wang, R.; Yang, Y.; Wang, J. Transformer–BiLSTM Fusion Neural Network for Short-Term PV Output Prediction Based on NRBO Algorithm and VMD. Appl. Sci. 2024, 14, 11991. [Google Scholar] [CrossRef]
- Cen, Z.; Wang, J. Crude oil price prediction model with long short term memory deep learning based on prior knowledge data transfer. Energy 2019, 169, 160–171. [Google Scholar] [CrossRef]
- Busari, G.A.; Lim, D.H. Crude oil price prediction: A comparison between AdaBoost-LSTM and AdaBoost-GRU for improving forecasting performance. Comput. Chem. Eng. 2021, 155, 107513. [Google Scholar] [CrossRef]
- Lytvynenko, V.; Wojcik, W.; Fefelov, A.; Lurie, I.; Savina, N.; Voronenko, M.; Boskin, O.; Smailova, S. Hybrid methods of GMDH-neural networks synthesis and training for solving problems of time series forecasting. In Proceedings of the Lecture Notes in Computational Intelligence and Decision Making: Proceedings of the XV International Scientific Conference “Intellectual Systems of Decision Making and Problems of Computational Intelligence”(ISDMCI’2019), Salisnyj Port, Ukraine, 21–25 May 2019; pp. 513–531. [Google Scholar]
- Sobolewski, Ł.; Miczulski, W. Methods of constructing time series for predicting local time scales by means of a GMDH-type neural network. Appl. Sci. 2021, 11, 5615. [Google Scholar] [CrossRef]
- Foroutan, P.; Lahmiri, S. Deep learning systems for forecasting the prices of crude oil and precious metals. Financ. Innov. 2024, 10, 111. [Google Scholar] [CrossRef]
- Li, J.; Zhu, S.; Wu, Q. Monthly crude oil spot price forecasting using variational mode decomposition. Energy Econ. 2019, 83, 240–253. [Google Scholar] [CrossRef]
- Wang, J.; Zhang, T.; Lu, T.; Xue, Z. A hybrid forecast model of EEMD-CNN-ILSTM for crude oil futures price. Electronics 2023, 12, 2521. [Google Scholar] [CrossRef]
- Liu, W.; Wang, C.; Li, Y.; Liu, Y.; Huang, K. Ensemble forecasting for product futures prices using variational mode decomposition and artificial neural networks. Chaos Solitons Fractals 2021, 146, 110822. [Google Scholar] [CrossRef]
- Lin, H.; Sun, Q. Crude oil prices forecasting: An approach of using CEEMDAN-based multi-layer gated recurrent unit networks. Energies 2020, 13, 1543. [Google Scholar] [CrossRef]
- Zhou, Y.; Wang, J.; Lu, H.; Zhao, W. Short-term wind power prediction optimized by multi-objective dragonfly algorithm based on variational mode decomposition. Chaos Solitons Fractals 2022, 157, 111982. [Google Scholar] [CrossRef]
- Zeng, L.; Ling, L.; Zhang, D.; Jiang, W. Optimal forecast combination based on PSO-CS approach for daily agricultural future prices forecasting. Appl. Soft Comput. 2023, 132, 109833. [Google Scholar] [CrossRef]
- Qu, Z.; Li, Y.; Jiang, X.; Niu, C. An innovative ensemble model based on multiple neural networks and a novel heuristic optimization algorithm for COVID-19 forecasting. Expert Syst. Appl. 2023, 212, 118746. [Google Scholar] [CrossRef]
- Alruqimi, M.; Di Persio, L. Multistep Brent oil price forecasting with a multi-aspect aeta-heuristic optimization and ensemble deep learning model. Energy Inform. 2024, 7, 130. [Google Scholar] [CrossRef]
- Ding, M.; Zhou, H.; Xie, H.; Wu, M.; Nakanishi, Y.; Yokoyama, R. A gated recurrent unit neural networks based wind speed error correction model for short-term wind power forecasting. Neurocomputing 2019, 365, 54–61. [Google Scholar] [CrossRef]
- Gilles, J. Empirical wavelet transform. IEEE Trans. Signal Process. 2013, 61, 3999–4010. [Google Scholar] [CrossRef]
- Hong, S.; Zhou, Z.; Zio, E.; Wang, W. An adaptive method for health trend prediction of rotating bearings. Digit. Signal Process. 2014, 35, 117–123. [Google Scholar] [CrossRef]
- Kong, Y.; Wang, T.; Chu, F. Meshing frequency modulation assisted empirical wavelet transform for fault diagnosis of wind turbine planetary ring gear. Renew. Energy 2019, 132, 1373–1388. [Google Scholar] [CrossRef]
- Bhattacharyya, A.; Singh, L.; Pachori, R.B. Fourier–Bessel series expansion based empirical wavelet transform for analysis of non-stationary signals. Digit. Signal Process. 2018, 78, 185–196. [Google Scholar] [CrossRef]
- Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv preprint 2018, arXiv:1803.01271. [Google Scholar]
- Wang, Y.; Chen, J.; Chen, X.; Zeng, X.; Kong, Y.; Sun, S.; Guo, Y.; Liu, Y. Short-term load forecasting for industrial customers based on TCN-LightGBM. IEEE Trans. Power Syst. 2020, 36, 1984–1997. [Google Scholar] [CrossRef]
- Wang, J.; Chen, Y.; Zhu, S.; Xu, W. Depth feature extraction-based deep ensemble learning framework for high frequency futures price forecasting. Digit. Signal Process. 2022, 127, 103567. [Google Scholar] [CrossRef]
- Ke, K.; Hongbin, S.; Chengkang, Z.; Brown, C. Short-term electrical load forecasting method based on stacked auto-encoding and GRU neural network. Evol. Intell. 2019, 12, 385–394. [Google Scholar] [CrossRef]
- Wang, B.; Wang, J. Energy futures and spots prices forecasting by hybrid SW-GRU with EMD and error evaluation. Energy Econ. 2020, 90, 104827. [Google Scholar] [CrossRef]
- Qin, L.; Li, W.; Li, S. Effective passenger flow forecasting using STL and ESN based on two improvement strategies. Neurocomputing 2019, 356, 244–256. [Google Scholar] [CrossRef]
- Wang, L.; Hu, H.; Ai, X.-Y.; Liu, H. Effective electricity energy consumption forecasting using echo state network improved by differential evolution algorithm. Energy 2018, 153, 801–815. [Google Scholar] [CrossRef]
- Li, B.; Wu, G.; He, Y.; Fan, M.; Pedrycz, W. An Overview and Experimental Study of Learning-Based Optimization Algorithms for the Vehicle Routing Problem. IEEE/CAA J. Autom. Sin. 2022, 9, 1115–1138. [Google Scholar] [CrossRef]
- Kiran, B.R.; Sobh, I.; Talpaert, V.; Mannion, P.; Al Sallab, A.A.; Yogamani, S.; Pérez, P. Deep reinforcement learning for autonomous driving: A survey. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4909–4926. [Google Scholar] [CrossRef]
- Mi, X.; Liu, H.; Li, Y. Wind speed prediction model using singular spectrum analysis, empirical mode decomposition and convolutional support vector machine. Energy Convers. Manag. 2019, 180, 196–205. [Google Scholar] [CrossRef]
- Rummery, G.A.; Niranjan, M. On-Line Q-Learning Using Connectionist Systems; University of Cambridge: Cambridge, UK, 1994; Volume 37. [Google Scholar]
- Wang, L.; He, Y.; Liu, X.; Li, L.; Shao, K. M2TNet: Multi-modal multi-task Transformer network for ultra-short-term wind power multi-step forecasting. Energy Rep. 2022, 8, 7628–7642. [Google Scholar] [CrossRef]
- Song, W.; Fujimura, S. Capturing combination patterns of long-and short-term dependencies in multivariate time series forecasting. Neurocomputing 2021, 464, 72–82. [Google Scholar] [CrossRef]
- Hae, H.; Kang, S.-J.; Kim, T.O.; Lee, P.H.; Lee, S.-W.; Kim, Y.-H.; Lee, C.W.; Park, S.-W. Machine Learning-Based prediction of Post-Treatment ambulatory blood pressure in patients with hypertension. Blood Press. 2023, 32, 2209674. [Google Scholar] [CrossRef] [PubMed]
- Liu, H.; Yu, C.; Yu, C.; Chen, C.; Wu, H. A novel axle temperature forecasting method based on decomposition, reinforcement learning optimization and neural network. Adv. Eng. Inform. 2020, 44, 101089. [Google Scholar] [CrossRef]
- Huang, Y.; Deng, Y. A new crude oil price forecasting model based on variational mode decomposition. Knowl.-Based Syst. 2021, 213, 106669. [Google Scholar] [CrossRef]
Methods | Refs. | Models | Year | Advantages | Drawbacks |
Traditional methods | [6] | ARIMA | 2003 |
|
|
[7] | GM | 2019 | |||
Deep learning methods | [12] | LSTM | 2019 |
|
|
[13] | GRU | 2021 | |||
[8] | DNN | 2024 | |||
[15] | GMDH | 2021 | |||
[16] | TCN | 2024 | |||
Hybrid methods | [18] | EEMD-CNN-ILSTM | 2023 |
|
|
[19] | VMD-ANN | 2021 | |||
[20] | CEEMD-ML-GRU | 2020 | |||
[22] | PSO-based ensemble | 2023 | |||
[23] | SCWOA-based ensemble | 2023 | |||
[24] | GWO-based ensemble | 2024 |
Crude Oil Price Data | Dataset #1 (USD) | Dataset #2 (USD) | Dataset #3 (USD) |
---|---|---|---|
Minimum | 26.21 | 10.79 | 11.22 |
Mean | 66.05 | 50.21 | 46.86 |
Maximum | 110.53 | 145.29 | 140 |
Standard derivation | 22.83 | 28.92 | 28.87 |
Name | Description |
---|---|
Model 1 | GA-TCN-GRU-ESN |
Model 2 | BHA-TCN-GRU-ESN |
Model 3 | GWO-TCN-GRU-ESN |
Model 4 | SARSA-TCN-GRU-ESN |
Model 5 | PSO-TCN-GRU-ESN |
Model 6 | WPD-SARSA-TCN-GRU-ESN |
Model 7 | VMD-SARSA-TCN-GRU-ESN |
Model 8 | EWT-SARSA-TCN-GRU-ESN |
Model 9 | EWT-SARSA-TCN-GRU-ESN-ELM |
Series | Predictors | MAE (USD) | MAPE (%) | RMSE (USD) |
---|---|---|---|---|
#1 | TCN | 0.7462 | 1.1634 | 1.0092 |
GRU | 0.7680 | 1.1910 | 1.0048 | |
ESN | 0.8262 | 1.2775 | 1.0821 | |
BPNN | 2.7229 | 4.3405 | 2.9994 | |
GRNN | 3.6410 | 5.6739 | 4.0624 | |
RBFNN | 1.3925 | 2.1287 | 1.8051 | |
LSTM | 0.9536 | 1.4394 | 1.1935 | |
CNN | 0.8698 | 1.3729 | 1.1409 | |
GMDH | 1.1592 | 1.8026 | 1.4410 | |
#2 | TCN | 2.2485 | 4.7553 | 3.2635 |
GRU | 2.1264 | 4.6699 | 2.8677 | |
ESN | 2.1818 | 4.7152 | 2.9101 | |
BPNN | 2.8466 | 6.3487 | 3.6113 | |
GRNN | 3.5657 | 7.9179 | 4.6912 | |
RBFNN | 3.5109 | 7.8512 | 4.8379 | |
LSTM | 2.3411 | 5.0723 | 3.1779 | |
CNN | 2.3602 | 5.1736 | 3.2895 | |
GMDH | 2.9150 | 6.5246 | 3.8918 | |
#3 | TCN | 5.6431 | 12.0686 | 7.7571 |
GRU | 6.0148 | 13.9647 | 8.5027 | |
ESN | 5.5273 | 12.0189 | 7.4268 | |
BPNN | 7.1119 | 14.7113 | 9.2021 | |
GRNN | 7.7206 | 18.4653 | 10.3241 | |
RBFNN | 7.3445 | 18.0054 | 10.1406 | |
LSTM | 6.0466 | 14.1694 | 8.4556 | |
CNN | 6.2123 | 14.1316 | 8.6501 | |
GMDH | 7.6269 | 18.3832 | 11.0227 |
Series | Models | Ensemble Methods | MAE (USD) | MAPE (%) | RMSE (USD) |
---|---|---|---|---|---|
#1 | Model 1 | GA | 0.7327 | 1.1419 | 0.9950 |
Model 2 | BHA | 0.7162 | 1.1142 | 0.9739 | |
Model 3 | GWO | 0.7218 | 1.1239 | 0.9826 | |
Model 4 | SARSA | 0.7119 | 1.1082 | 0.9691 | |
#2 | Model 1 | GA | 2.0895 | 4.4896 | 2.7733 |
Model 2 | BHA | 2.1079 | 4.5164 | 2.8199 | |
Model 3 | GWO | 2.0513 | 4.4849 | 2.7391 | |
Model 4 | SARSA | 2.0144 | 4.3907 | 2.6651 | |
#3 | Model 1 | GA | 5.3891 | 11.8152 | 7.1542 |
Model 2 | BHA | 5.4202 | 12.0044 | 7.1597 | |
Model 3 | GWO | 5.4539 | 11.8065 | 7.2562 | |
Model 4 | SARSA | 5.2533 | 11.8783 | 6.9218 |
Models | Indices | Series #1 | Series #2 | Series #3 |
---|---|---|---|---|
Model 4 vs. TCN | PMAE (%) | 4.5966 | 10.4114 | 6.9076 |
PMAPE (%) | 4.7445 | 7.6672 | 1.5768 | |
PRMSE (%) | 3.9734 | 18.3111 | 10.7682 | |
Model 4 vs. GRU | PMAE (%) | 7.3047 | 5.2671 | 12.6604 |
PMAPE (%) | 6.9521 | 5.9787 | 14.9405 | |
PRMSE (%) | 3.5529 | 7.0649 | 18.5929 | |
Model 4 vs. ESN | PMAE (%) | 11.8343 | 7.6726 | 4.9572 |
PMAPE (%) | 13.2524 | 6.8820 | 1.1698 | |
PRMSE (%) | 10.4427 | 7.0649 | 6.7997 |
Series | Models | Decomposition Methods | MAE (USD) | MAPE (%) | RMSE (USD) |
---|---|---|---|---|---|
#1 | Model 4 | - | 0.7119 | 1.1082 | 0.9691 |
Model 6 | WPD | 0.2239 | 0.3538 | 0.2671 | |
Model 7 | VMD | 0.4131 | 0.6228 | 0.5415 | |
Model 8 | EWT | 0.1728 | 0.2640 | 0.2207 | |
#2 | Model 4 | - | 2.0144 | 4.3907 | 2.6651 |
Model 6 | WPD | 0.6911 | 1.5565 | 0.9437 | |
Model 7 | VMD | 1.1031 | 2.4394 | 1.3172 | |
Model 8 | EWT | 0.3387 | 0.7625 | 0.4701 | |
#3 | Model 4 | - | 5.2533 | 11.8783 | 6.9218 |
Model 6 | WPD | 1.7792 | 4.2214 | 2.4301 | |
Model 7 | VMD | 3.2493 | 6.6851 | 3.6351 | |
Model 8 | EWT | 1.6131 | 3.5695 | 1.9352 |
Models | Indices | Series #1 | Series #2 | Series #3 |
---|---|---|---|---|
Model 6 vs. Model 4 | PMAE (%) | 68.5490 | 65.6920 | 66.1318 |
PMAPE (%) | 68.0744 | 64.5501 | 64.4612 | |
PRMSE (%) | 72.4383 | 64.5904 | 64.8920 | |
Model 7 vs. Model 4 | PMAE (%) | 41.9722 | 45.2392 | 38.1474 |
PMAPE (%) | 43.8008 | 44.4417 | 43.7201 | |
PRMSE (%) | 44.1234 | 50.5760 | 47.4833 | |
Model 8 vs. Model 4 | PMAE (%) | 75.7269 | 83.1861 | 69.2936 |
PMAPE (%) | 76.1776 | 82.6337 | 69.9494 | |
PRMSE (%) | 77.2263 | 82.3609 | 72.0420 |
Series | Decomposition Methods | MSE | ||
---|---|---|---|---|
TCN | GRU | ESN | ||
#1 | WPD VMD EWT | 0.1353 | 0.1619 | 0.1527 |
0.3731 | 0.4492 | 0.4913 | ||
0.0975 | 0.1013 | 0.1434 | ||
#2 | WPD VMD EWT | 1.1731 | 1.0752 | 1.5963 |
2.3832 | 1.8923 | 2.4435 | ||
0.9816 | 0.4352 | 0.7369 | ||
#3 | WPD VMD EWT | 7.1536 | 6.7949 | 6.8753 |
15.4031 | 16.7098 | 15.2039 | ||
5.0524 | 4.1983 | 4.7721 |
Series | Models | MAE (USD) | MAPE (%) | RMSE (USD) |
---|---|---|---|---|
#1 | TCN | 0.7462 | 1.1634 | 1.0092 |
GRU | 0.7680 | 1.1910 | 1.0048 | |
ESN | 0.8262 | 1.2775 | 1.0821 | |
Model 4 | 0.7119 | 1.1082 | 0.9691 | |
Model 8 | 0.1728 | 0.264 | 0.2207 | |
Model 9 | 0.1239 | 0.1914 | 0.1582 | |
#2 | TCN | 2.2485 | 4.7553 | 3.2635 |
GRU | 2.1264 | 4.6699 | 2.8677 | |
ESN | 2.1818 | 4.7152 | 2.9101 | |
Model 4 | 2.0144 | 4.3907 | 2.6651 | |
Model 8 | 0.3387 | 0.7625 | 0.4701 | |
Model 9 | 0.2529 | 0.5493 | 0.3375 | |
#3 | TCN | 5.6431 | 12.0686 | 7.7571 |
GRU | 6.0148 | 13.9647 | 8.5027 | |
ESN | 5.5273 | 12.0189 | 7.4268 | |
Model 4 | 5.2533 | 11.8783 | 6.9218 | |
Model 8 | 1.6131 | 3.5695 | 1.9352 | |
Model 9 | 1.3224 | 2.7004 | 1.6415 |
Indices | Model 9 vs. Model 8 | ||
---|---|---|---|
Series #1 | Series #2 | Series #3 | |
PMAE (%) | 28.2986 | 25.3322 | 18.0212 |
PMAPE (%) | 27.5000 | 27.9607 | 24.3479 |
PRMSE (%) | 28.3190 | 28.2068 | 15.1767 |
Models | MAE (USD) | MAPE (%) | RMSE (USD) |
---|---|---|---|
TCN | 2.3003 | 2.8941 | 3.3531 |
GRU | 2.9344 | 3.4514 | 5.1989 |
ESN | 2.4249 | 2.9076 | 4.4565 |
Liu’s model | 0.4926 | 0.5575 | 1.1428 |
Mi’s model | 0.5827 | 0.6596 | 1.3803 |
Huang’s model | 0.4270 | 0.5446 | 0.7404 |
Model 9 | 0.3510 | 0.4420 | 0.5818 |
Forecasting Models | Training Time (s) | ||
---|---|---|---|
#1 | #2 | #3 | |
TCN | 45.32 | 47.81 | 26.93 |
GRU | 19.21 | 18.71 | 7.34 |
ESN | 8.13 | 7.25 | 4.15 |
PSO-TCN-GRU-ESN | 87.75 | 91.54 | 53.51 |
GA-TCN-GRU-ESN | 79.82 | 81.61 | 45.32 |
SARSA-TCN-GRU-ESN | 100.23 | 105.36 | 58.86 |
EWT- SARSA-TCN-GRU-ESN | 1185.91 | 1290.49 | 659.19 |
EWT-SARSA-TGE-ELM | 1194.16 | 1297.93 | 663.46 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zheng, G.; Li, Y.; Xia, Y. Crude Oil Price Forecasting Model Based on Neural Networks and Error Correction. Appl. Sci. 2025, 15, 1055. https://doi.org/10.3390/app15031055
Zheng G, Li Y, Xia Y. Crude Oil Price Forecasting Model Based on Neural Networks and Error Correction. Applied Sciences. 2025; 15(3):1055. https://doi.org/10.3390/app15031055
Chicago/Turabian StyleZheng, Guangji, Ye Li, and Yu Xia. 2025. "Crude Oil Price Forecasting Model Based on Neural Networks and Error Correction" Applied Sciences 15, no. 3: 1055. https://doi.org/10.3390/app15031055
APA StyleZheng, G., Li, Y., & Xia, Y. (2025). Crude Oil Price Forecasting Model Based on Neural Networks and Error Correction. Applied Sciences, 15(3), 1055. https://doi.org/10.3390/app15031055