Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting
Abstract
:1. Introduction
- (1)
- To effectively capture the long-term temporal features, a TFE-HCM module is proposed to capture the dynamic correlation of different nodes over time with a GTF block and LTF block. Specifically, the GTF block utilizes an attention mechanism to capture the global dynamic correlation of various nodes across different time steps, while the LTF block relies on the attention scores to encode the impact of temporally adjacent nodes for future predictions.
- (2)
- To overcome the limitation of the GCN-based models using the manually predefined graph structure to capture the spatial features, an SFE-DCM is devised to extract dynamic spatial correlations by an SDM block and an ASM block. The SDM block uses an attention mechanism to explore and quantify the influence nodes in the spatial dimension, and the ASM block is used to capture the dynamic correlations for all nodes at each time step.
- (3)
- An STM-GM module is designed to effectively fuse the spatial features and temporal features, where a gated fusion unit is explored to effectively learn and leverage the inherent spatial–temporal relationships within the traffic data.
2. Related Works
2.1. Traffic Flow Forecasting
2.2. GCN-Based Traffic Forecasting
3. Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network
3.1. Preliminaries
3.2. The Proposed Model
3.2.1. Spatial–Temporal Embedding Encoder Module (STEM)
3.2.2. Temporal Feature Encoder Module with Heterogeneous Time Series Correlation Modeling
3.2.3. Spatial Feature Encoder Module with Dynamic Multi-Graph Modeling (SFE-DGM)
3.2.4. Spatial–Temporal Feature Fusion Module Based on Gating Fusion Mechanism (STM-GM)
4. Experiments
4.1. Datasets and Evaluation Metrics
4.2. Experimental Settings
4.3. Some State-of-the-Art Methods
- (1)
- ARIMA [5] (auto-regressive integrated moving average): an auto-regressive model that utilizes auto-regression and a moving average model for time series data prediction.
- (2)
- VAR [29] (vector auto-regression): an auto-regressive model that uses a statistical method to model the time correlation of traffic data.
- (3)
- LSTM [30] (long short-term memory): a recurrent neural network (RNN)-based traffic flow forecasting model that utilizes multi-layer LSTM and an auto-regressive model to capture time series dependency of traffic signals.
- (4)
- GRU-ED [31] (gated recurrent unit encoder–decoder): a variant of the LSTM model that construct an encoder–decoder model with multiple GRU units for time series forecasting.
- (5)
- ST-GCN [16] (spatial–temporal graph convolution network): a graph convolution network (GCN)-based model that combines the gated 1-D convolution layer and spectral GCN to capture the spatial and temporal dependencies of traffic signals.
- (6)
- DCRNN [18] (diffusion convolutional recurrent neural network): a GCN-based model that adopts a seq-to-seq framework for forecasting, it combines diffusion GCN with GRU to extract the spatial and temporal dependencies of traffic signals.
- (7)
- GWN [19] (Graph WaveNet): an adaptive GCN-based model that combines spatial GCN and gated TCN with an adaptive adjacency matrix to capture the spatial and temporal dependencies.
- (8)
- AGCRN [27] (adaptive graph convolutional recurrent network): an adaptive GCN-based model that combines the adaptive graph and GRU units with a node-adaptive parameter learning module for capturing the spatial–temporal correlation of traffic flow data.
- (9)
- DSANet [32] (dual self-attention network): an attention-based model that utilizes a dual self- attention mechanism to capture spatial dependencies and extracts temporal correlations using a convolution with a kernel size of 3.
- (10)
- ASTGCN [20] (attention-based spatial–temporal graph convolutional network): an attention-based model that combines a self-attention mechanism with GCN to capture spatial–temporal dependencies for traffic data.
- (11)
- STG-NCDE [21] (spatial–temporal graph neural controlled differential equation): an ST-GCN-based model that combines the neural controlled differential equations with GCN for processing the time series data.
- (12)
- STGODE [28] (spatial–temporal graph ordinary differential equation networks): an ST-GCN-based model that utilizes ordinary differential equations to overcome the over-smooth issue in GCN, where a spatial adjacency matrix, semantic adjacency matrix, and two temporal dilation convolutions are designed to capture the long-term temporal dependencies of traffic flow data.
4.4. Ablation Study
4.4.1. Ablation Study for STC-DGCN
- (1)
- w/o STEM: to verify the effects of STEM, we remove the STEM module, only using the original traffic data for prediction.
- (2)
- w/o TFE-HCM: the TFE-HCM module is replaced by the traditional TCN for capturing temporal correlation.
- (3)
- w/o SFE-DGM: the SFE-DGM module is replaced by a pre-defined graph structure to capture the spatial features, letting the model rely only on a fixed adjacency matrix to extract neighborhood node features.
- (4)
- w/o STM-GM: removing the STM-GM module and directly using the last layer output of the STC-DGCN model for traffic flow forecasting.
4.4.2. The Effect of TFE-HCM
- (1)
- w/o TCN: this variant uses the traditional TCN for capturing temporal correlation instead of the TFE-HCM module.
- (2)
- w/o GTF: the TFE-HCM module is replaced by only using the designed GTF block to capture the temporal correlation of nodes in road network.
- (3)
- w/o LTF: the TFE-HCM module is replaced by only using the designed LTF block to capture temporal correlation.
4.4.3. The Effect of SFE-DGM
- (1)
- w/o SCN: this variant uses the traditional pre-defined graph structure with spectral convolution to capture the spatial feature instead of the SFE-DGM module, where in Equation (11) measures the spatial correction between nodes.
- (2)
- w/o SDM: the SFE-DGM module is replaced by only using the designed SDM block to capture spatial correlation of nodes in the road network.
- (3)
- w/o ASM: the SFE-DGM module is replaced by only using the designed ASM block.
4.5. Comparison with SOTA Methods
4.5.1. Results on PEMS04
4.5.2. Results on PEMS07
4.5.3. Results on PEMS08
4.5.4. Visualization of the Prediction Error with Different Time Steps
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Wang, F.; Liang, Y.; Lin, Z.; Zhou, J.; Zhou, T. SSA-ELM: A Hybrid Learning Model for Short-Term Traffic Flow Forecasting. Mathematics 2024, 12, 1895. [Google Scholar] [CrossRef]
- Yang, H.H.; Wen, J.M.; Wu, X.J.; He, L. An Efficient Edge Artificial Intelligence Multi-pedestrian Tracking Method with Rank Constraint. IEEE Trans. Ind. Inform. 2019, 15, 4178–4188. [Google Scholar] [CrossRef]
- Zhao, Y.; Lin, Y.; Wen, H.; Wei, T.; Jin, X.; Wan, H. Spatial-Temporal Position-Aware Graph Convolution Networks for Traffic Flow Forecasting. IEEE Trans. Intell. Transp. Syst. 2023, 24, 8650–8666. [Google Scholar] [CrossRef]
- Liu, X.; Wang, W. Deep Time Series Forecasting Models: A Comprehensive Survey. Mathematics 2024, 12, 1504. [Google Scholar] [CrossRef]
- Smith, B.L.; Williams, B.M.; Oswald, R.K. Comparison of parametric and nonparametric models for traffic flow forecasting. Transp. Res. C Emerg. Technol. 2002, 10, 303–321. [Google Scholar] [CrossRef]
- Chandra, S.R.; Al-Deek, H. Predictions of freeway traffic speeds and volumes using vector autoregressive models. J. Intell. Transp. Syst. 2009, 13, 53–72. [Google Scholar] [CrossRef]
- Lippi, M.; Bertini, M.; Frasconi, P. Short-term traffic flow forecasting: An experimental comparison of time-series analysis and supervised learning. IEEE Trans. Intell. Transp. Syst. 2013, 14, 871–882. [Google Scholar] [CrossRef]
- Johansson, U.; Boström, H.; Löfström, T.; Linusson, H. Regression conformal prediction with random forests. Mach. Learn. 2014, 97, 155–176. [Google Scholar] [CrossRef]
- Zheng, Z.; Su, D. Short-term traffic volume forecasting: A k-nearest neighbor approach enhanced by constrained linearly sewing principal component algorithm. Transp. Res. Part C Emerg. Technol. 2014, 43, 143–157. [Google Scholar] [CrossRef]
- Liu, A.; Zhang, Y. Spatial–Temporal Dynamic Graph Convolutional Network with Interactive Learning for Traffic Forecasting. IEEE Trans. Intell. Transp. Syst. 2024, 25, 7645–7660. [Google Scholar] [CrossRef]
- Ma, X.; Tao, Z.; Wang, Y.; Yu, H.; Wang, Y. Long short-term memory neural network for traffic speed prediction using remote microwave sensor data. Transp. Res. Part C Emerg. Technol. 2015, 54, 187–197. [Google Scholar] [CrossRef]
- Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. In Proceedings of the 2014 NIPS Workshop on Deep Learning, San Diego, CA, USA, 8–11 December 2014; pp. 1–9. [Google Scholar]
- Cui, Z.; Ke, R.; Pu, Z.; Wang, Y. Stacked bidirectional and uni directional lstm recurrent neural network for forecasting network wide traffic state with missing values. Transp. Res. Part C Emerg. Technol. 2020, 118, 102674. [Google Scholar] [CrossRef]
- Yao, H.; Tang, X.; Wei, H.; Zheng, G.; Li, Z. Revisiting spatial temporal similarity: A deep learning framework for traffic prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 5668–5675. [Google Scholar]
- Jiang, W.W.; Luo, J.Y.; He, M.; Gu, W. Graph Neural Network for Traffic Forecasting: The research progress. ISPRS Int. J. Geo-Inf. 2023, 12, 100. [Google Scholar] [CrossRef]
- Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; ACM: New York, NY, USA, 2018; pp. 3634–3640. [Google Scholar]
- Weng, W.C.; Fan, J.; Wu, H.F.; Hu, Y.J.; Tian, H.; Zhu, F.; Wu, J. A Decomposition Dynamic graph convolutional recurrent network for traffic forecasting. Pattern Recognit. 2023, 142, 109670. [Google Scholar] [CrossRef]
- Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. In Proceedings of the 6th International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–16. [Google Scholar]
- Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph wavenet for deep spatial-temporal graph modeling. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence IJCAI, Macao, China, 10–16 August 2019. [Google Scholar]
- Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention based spatial-temporal graph convolutional networks for traffic flow forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 922–929. [Google Scholar]
- Choi, J.; Choi, H.; Hwang, J.; Park, N. Graph neural controlled differential equations for traffic forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 6367–6374. [Google Scholar]
- Li, Z.; Ren, Q.; Chen, L.; Sui, X.; Li, J. Multi-Hierarchical Spatial-Temporal Graph Convolutional Networks for Traffic Flow Forecasting. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4913–4919. [Google Scholar]
- Lea, C.; Flynn, M.D.; Vidal, R.; Reiter, A.; Hager, G.D. Temporal convolutional networks for action segmentation and detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 156–165. [Google Scholar]
- Liu, Z.; Lin, Y.T.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
- Song, C.; Lin, Y.; Guo, S.; Wan, H. Spatial-temporal syn chronous graph convolutional networks: A new framework for spatial-temporal network data forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 914–921. [Google Scholar]
- Chen, C.; Petty, K.; Skabardonis, A.; Varaiya, P.; Jia, Z. Freeway performance measurement system: Mining loop detector data. Transp. Res. Rec. 2001, 1748, 96–102. [Google Scholar] [CrossRef]
- Bai, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive Graph Convolutional Recurrent Network for Traffic Forecasting. In Proceedings of the 34th Neural Information Processing Systems (NIPS), San Diego, CA, USA, 6–12 December 2020; pp. 17804–17815. [Google Scholar]
- Fang, Z.; Long, Q.; Song, G.; Xie, K. Spatial-temporal Graph ode Networks for Traffic Flow Forecasting. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 364–373. [Google Scholar]
- Zivot, E.; Wang, J. Vector autoregressive models for multi variate time series. In Modeling Financial Time Series with S-PLUS®; Springer: New York, NY, USA, 2006; pp. 385–429. [Google Scholar]
- Cui, Z.; Ke, R.; Pu, Z.; Wang, Y. Deep bidirectional and unidirectional lstm recurrent neural network for network-wide traffic speed prediction. arXiv 2018, arXiv:1801.02143. [Google Scholar]
- Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. arXiv 2014, arXiv:1409.3215. [Google Scholar] [CrossRef]
- Huang, S.; Wang, D.; Wu, X.; Tang, A. Dsanet: Dual Self-attention Network for Multivariate Time Series Forecasting. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 2129–2132. [Google Scholar]
Dataset | PEMS04 | PEMS07 | PEMS08 |
---|---|---|---|
Nodes | 307 | 883 | 170 |
Time steps | 16,992 | 28,224 | 17,856 |
Time interval | 5 min | 5 min | 5 min |
Start time | 1 January 2018 | 1 May 2017 | 1 July 2016 |
End time | 28 February 2018 | 31 August 2017 | 31 August 2016 |
Type | traffic flow | traffic flow | traffic flow |
Area | San Francisco | San Francisco | San Bernardino |
Models | PEMS04 | PEMS07 | PEMS08 | |||||||
---|---|---|---|---|---|---|---|---|---|---|
MAE | RMSE | MAPE/% | MAE | RMSE | MAPE/% | MAE | RMSE | MAPE/% | ||
Statistical-based models | ARIMA | 33.73 | 48.80 | 24.18 | 38.17 | 59.27 | 19.46 | 31.09 | 44.32 | 22.73 |
VAR | 24.54 | 38.61 | 17.24 | 50.22 | 75.63 | 32.22 | 19.19 | 29.81 | 13.10 | |
RNN-based models | LSTM | 26.81 | 40.74 | 22.33 | 29.71 | 45.32 | 14.14 | 22.19 | 33.59 | 18.74 |
GRU-ED | 23.68 | 39.27 | 16.44 | 27.66 | 43.49 | 12.20 | 22.00 | 36.22 | 13.33 | |
GCN-based models | ST-GCN | 21.16 | 34.89 | 13.83 | 25.33 | 39.34 | 11.21 | 17.50 | 27.09 | 11.29 |
DCRNN | 22.74 | 36.58 | 14.75 | 23.64 | 36.52 | 12.28 | 18.18 | 28.18 | 11.23 | |
Adaptive GCN-based models | GWN | 24.89 | 39.66 | 17.29 | 26.39 | 41.50 | 11.97 | 18.28 | 30.05 | |
AGCRN | 19.83 | 32.26 | 12.97 | 22.37 | 36.55 | 9.12 | 15.95 | 25.22 | 10.09 | |
Attention-mechanism-based models | DSANet | 22.79 | 35.77 | 16.03 | 31.36 | 49.11 | 14.43 | 17.14 | 26.96 | 11.32 |
ASTGCN | 22.42 | 34.25 | 15.87 | 21.22 | 34.10 | 9.05 | 15.07 | 24.80 | 9.51 | |
Differential-equation-controlled GCN-based models | STG-NCDE | 19.21 | 31.09 | 12.76 | 20.53 | 33.84 | 8.80 | 15.45 | 24.81 | 9.92 |
STGODE | 20.84 | 32.82 | 13.77 | 22.59 | 37.54 | 10.14 | 16.81 | 25.97 | 10.62 | |
Our model | STC-DGCN | 18.27 | 30.37 | 12.12 | 19.36 | 32.79 | 8.06 | 14.32 | 23.74 | 9.34 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ge, Y.; Wang, J.; Zhang, B.; Peng, F.; Ma, J.; Yang, C.; Zhao, Y.; Liu, M. Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting. Mathematics 2024, 12, 3159. https://doi.org/10.3390/math12193159
Ge Y, Wang J, Zhang B, Peng F, Ma J, Yang C, Zhao Y, Liu M. Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting. Mathematics. 2024; 12(19):3159. https://doi.org/10.3390/math12193159
Chicago/Turabian StyleGe, Yajun, Jiannan Wang, Bo Zhang, Fan Peng, Jing Ma, Chenyu Yang, Yue Zhao, and Ming Liu. 2024. "Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting" Mathematics 12, no. 19: 3159. https://doi.org/10.3390/math12193159
APA StyleGe, Y., Wang, J., Zhang, B., Peng, F., Ma, J., Yang, C., Zhao, Y., & Liu, M. (2024). Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting. Mathematics, 12(19), 3159. https://doi.org/10.3390/math12193159