Multidimensional Feature-Based Graph Attention Networks and Dynamic Learning for Electricity Load Forecasting
Abstract
:1. Introduction
- We performed data preprocessing on the original data to extract the trend terms and seasonal features of the load data. For the load data series with natural cyclical characteristics, subsequent load forecasting based on trend terms and seasonal features can improve the accuracy of the forecasts.
- We construct the differential features of the load data based on the GRU module, combine the original features of the load data and the differential features of the load data for post modeling and solve the problem of the influence of the temporal difference information of the load data on load forecasting.
- We consider the load series incorporating multidimensional features such as seasonal trends and holidays as a complete graph and model the correlation of different variable features of the load series by constructing a multidimensional feature graph attention layer where each node in the graph represents a feature of the load sequence, and the edge between adjacent nodes represents the relationship between two features.
2. Related Works
2.1. Time-Series Prediction Model Based on RNN
2.2. Time-Series Prediction Model Based on Transformer
2.3. Time-Series Prediction Model Based on Graph
3. TCN_DLGAT Model
3.1. Differential Learning Network
3.2. Time-Series Convolutional Network
3.3. Multidimensional Feature Graph Attention Network
3.4. TCN_DLGAT Model Framework
4. Methods
4.1. Problem Definition
4.2. STL Decomposition to Extract Trend and Seasonal Features
4.3. Extract the Year, Month, Day and Holiday Characteristics
4.4. Data Normalization
4.5. Training Loss Function
5. Experiments
5.1. Experimental Setup
- LSTM: A temporal recurrent neural network suitable to be used to process and predict events with relatively long intervals and delays in the time series.
- CNN_LSTM: A combination of the convolutional neural network and long and short-term memory network.
- LSTNet [30]: The LSTM model combined with a temporal attention mechanism.
- Transformer:an encoder–decoder network combining attention mechanisms.
- TCN: An advanced TCN model called WaveNet where an additional gating mechanism is applied after expansion causal convolution.
5.2. Metrics
5.3. Experimental Results
5.4. Ablation Experiments
- TCN_GAT: Omitting the differential learning network, the input containing only the raw load data information is used as one input channel of the temporal convolutional network to extract the long series time dependence of the load data. The spatial correlation between the multidimensional feature variables is also extracted using a dimensional feature map attention network, and then the two parts of the feature outputs are connected through a convolutional neural network to obtain the predicted expression form.
- TCN_DL: Omitting the multidimensional feature map attention network, the original and differential sequences of the load data are fed into a time-series convolutional network, as well as a differential learning network, to model the dynamic variation characteristics of the load sequences and the time-dependence of the long time sequences. The two parts of the feature outputs are connected through a convolutional neural network to obtain the predicted expressions.
- DLGAT: Omitting the time-series convolutional network, the original and differential sequences of the load data are first fed into the differential learning network, as well as the multidimensional feature map attention network. Then, the two parts of the feature outputs are connected through a convolutional neural network to obtain the predicted expressions.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
RNN | Recurrent neural networks |
GRU | Gate recurrent unit |
LSTM | Long–short-term memory |
TCN | Time series convolution network |
GAT | Graph attention networks |
GCN | Graph convolutional network |
References
- Zhu, X.C. Research on short-term power load forecasting method based on IFOA-GRNN. Power Syst. Prot. Control 2020, 48, 121–127. [Google Scholar]
- Zheng, L.; Shuxin, T.; Dawei, Y. A medium- and long-term load forecasting method based on ARIMA-TARCH-BP neural network model. Electron. Devices 2020, 43, 175–179. [Google Scholar]
- Wang, Y.; Du, X.; Lu, Z.; Duan, Q.; Wu, J. Improved LSTM-Based Time-Series Anomaly Detection in Rail Transit Operation Environments. IEEE Trans. Ind. Inform. 2022, 18, 9027–9036. [Google Scholar] [CrossRef]
- Aamir, M.; Shabri, A. Modelling and forecasting monthly crude oil price of Pakistan: A comparative study of ARIMA, GARCH and ARIMA Kalman model. AIP Conf. Proc. 2016, 1750, 060015. [Google Scholar]
- Wan, R.; Mei, S.; Wang, J.; Liu, M.; Yang, F. Multivariate Temporal Convolutional Network: A Deep Neural Networks Approach for Multivariate Time Series Forecasting. Electronics 2019, 8, 876. [Google Scholar] [CrossRef]
- Wang, X.; Liu, H.; Du, J.; Dong, X.; Yang, Z. A long-term multivariate time series forecasting network combining series decomposition and convolutional neural networks. Appl. Soft Comput. 2023, 139, 110214. [Google Scholar] [CrossRef]
- Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Lewandowski, M.; Makris, D.; Velastin, S.A.; Nebel, J.C. Structural Laplacian Eigenmaps for Modeling Sets of Multivariate Sequences. IEEE Trans. Cybern. 2014, 44, 936–949. [Google Scholar] [CrossRef] [PubMed]
- Salinas, D.; Flunkert, V.; Gasthaus, J. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
- Rangapuram, S.S.; Seeger, M.; Gasthaus, J.; Stella, L.; Wang, Y.; Januschowski, T. Deep State Space Models for Time Series Forecasting. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 7796–7805. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Wu, N.; Green, B.; Xue, B.; O’Banion, S. Deep Transformer Models for Time Series Forecasting: The Influenza Prevalence Case. arXiv 2020, arXiv:2001.08317. [Google Scholar]
- Zheng, C.; Fan, X.; Wang, C.; Qi, J. GMAN: A Graph Multi-Attention Network for Traffic Prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1234–1241. [Google Scholar]
- Lim, B.; ArÄśk, S.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast. 2021, 37, 1748–1764. [Google Scholar] [CrossRef]
- Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021. [Google Scholar]
- Du, S.; Li, T.; Yang, Y.; Horng, S.J. Multivariate time series forecasting via attention-based encoder—decoder framework. Neurocomputing 2020, 388, 269–279. [Google Scholar] [CrossRef]
- Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph Wavenet for Deep Spatial-Temporal Graph Modeling. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019; pp. 1907–1913. [Google Scholar]
- Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Systems. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
- Chen, F.; Pan, S.; Jiang, J.; Huo, H.; Long, G. DAGCN: Dual Attention Graph Convolutional Networks. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
- Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language Modeling with Gated Convolutional Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 933–941. [Google Scholar]
- Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
- Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Fan, J.; Zhang, K.; Huang, Y.; Zhu, Y.; Chen, B. Parallel spatio-temporal attention-based TCN for multivariate time series prediction. Neural Comput. Appl. 2023, 35, 13109–13118. [Google Scholar] [CrossRef]
- Cirstea, R.G.; Guo, C.; Yang, B. Graph Attention Recurrent Neural Networks for Correlated Time Series Forecasting. arXiv 2021, arXiv:2103.10760. [Google Scholar]
- Cleveland, R.B.; Cleveland, W.S. STL: A seasonal-trend decomposition procedure based on Loess. J. Off. Stat. 1990, 6, 3–73. [Google Scholar]
- Li, J.; Wei, S.; Dai, W. Combination of Manifold Learning and Deep Learning Algorithms for Mid-Term Electrical Load Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 2584–2593. [Google Scholar] [CrossRef] [PubMed]
- Lai, G.; Chang, W.C.; Yang, Y.; Liu, H. Modeling Long and Short-Term Temporal Patterns with Deep Neural Networks. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, Ann Arbor, MI, USA, 8–12 July 2018; pp. 95–104. [Google Scholar]
Model Parameter | Parameter Meaning | Parameter Meaning |
---|---|---|
input_seq | Sliding window size | 672 |
input_size | Dimension of input data | 3 |
hidden_size | Dimension of hidden layer state | 96 |
output_size | Size of output data | 96 |
batch_size | Number of batch size | 32 |
num_layers | Number of layers in the LSTM stack | 1 |
learning_rate | learning_rate | 0.001 |
dropout | dropout | 0.2 |
encoder_layers | Number of layers of encoder in Transformer | 3 |
decoder_layers | Number of layers of decoder in Transformer | 3 |
n_heads | Number of heads of attention in Transformer | 3 |
LSTM | CNN_LSTM | LSTNet | Transformer | TCN | TCN_DLGAT | |
---|---|---|---|---|---|---|
MAE | 0.039 (+15%) | 0.037 (+9%) | 0.036 (+5%) | 0.043 (+26%) | 0.036 (+5%) | 0.034 |
MSE | 0.0026 (+30%) | 0.0023 (+15%) | 0.0022 (+10%) | 0.0024 (+20%) | 0.0027 (+35%) | 0.0020 |
MAPE | 13.098 (+355%) | 14.923 (+418%) | 3.562 (+24%) | 8.095 (+181%) | 14.838 (+415%) | 2.881 |
TCN_GAT | TCN_DL | DLGAT | TCN_DLGAT | |
---|---|---|---|---|
MAE | 0.044 (+29%) | 0.037 (+8%) | 0.043 (+26%) | 0.034 |
MSE | 0.0031 (+55%) | 0.0043 (+115%) | 0.0029 (+45%) | 0.0020 |
MAPE | 3.818 (+33%) | 22.477 (+330%) | 3.038 (+5%) | 2.881 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Huang, C.; Du, N.; He, J.; Li, N.; Feng, Y.; Cai, W. Multidimensional Feature-Based Graph Attention Networks and Dynamic Learning for Electricity Load Forecasting. Energies 2023, 16, 6443. https://doi.org/10.3390/en16186443
Huang C, Du N, He J, Li N, Feng Y, Cai W. Multidimensional Feature-Based Graph Attention Networks and Dynamic Learning for Electricity Load Forecasting. Energies. 2023; 16(18):6443. https://doi.org/10.3390/en16186443
Chicago/Turabian StyleHuang, Chaokai, Ning Du, Jiahan He, Na Li, Yifan Feng, and Weihong Cai. 2023. "Multidimensional Feature-Based Graph Attention Networks and Dynamic Learning for Electricity Load Forecasting" Energies 16, no. 18: 6443. https://doi.org/10.3390/en16186443
APA StyleHuang, C., Du, N., He, J., Li, N., Feng, Y., & Cai, W. (2023). Multidimensional Feature-Based Graph Attention Networks and Dynamic Learning for Electricity Load Forecasting. Energies, 16(18), 6443. https://doi.org/10.3390/en16186443