Next Article in Journal
Gender Imbalance in the Marriage Market and Housing Demand: Evidence from China
Previous Article in Journal
Examining Crop Yield Losses in Iğdır Plain Irrigation Systems in Türkiye Amidst Water Constraints
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatiotemporal Dynamic Multi-Hop Network for Traffic Flow Forecasting

1
School of Computer Science and Technology, Guangdong University of Technology, Guangzhou 510006, China
2
School of Cyberspace Security, Hainan University, Haikou 570228, China
3
College of Engineering, Shantou University, Shantou 515063, China
4
School of Philosophy, Fudan University, Shanghai 200433, China
5
Yangtze Delta Region Institute, University of Electronic Science and Technology of China, Quzhou 324003, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sustainability 2024, 16(14), 5860; https://doi.org/10.3390/su16145860
Submission received: 24 May 2024 / Revised: 3 July 2024 / Accepted: 8 July 2024 / Published: 9 July 2024

Abstract

:
Accurate traffic flow forecasting is vital for intelligent transportation systems, especially with urbanization worsening traffic congestion, which affects daily life, economic growth, and the environment. Precise forecasts aid in managing and optimizing transportation systems, reducing congestion, and improving air quality by cutting emissions. However, predicting outcomes is difficult due to intricate spatial relationships, nonlinear temporal patterns, and the challenges associated with long-term forecasting. Current research often uses static graph structures, overlooking dynamic and long-range dependencies. To tackle these issues, we introduce the spatiotemporal dynamic multi-hop network (ST-DMN), a Seq2Seq framework. This model incorporates spatiotemporal convolutional blocks (ST-Blocks) with residual connections in the encoder to condense historical traffic data into a fixed-dimensional vector. A dynamic graph represents time-varying inter-segment relationships, and multi-hop operation in the encoder’s spatial convolutional layer and the decoder’s diffusion multi-hop graph convolutional gated recurrent units (DMGCGRUs) capture long-range dependencies. Experiments on two real-world datasets METR-LA and PEMS-BAY show that ST-DMN surpasses existing models in three metrics.

1. Introduction

As urban vehicle populations increase and traffic data proliferate, the advancement of intelligent transportation systems (ITS) is becoming increasingly crucial [1,2]. Predicting traffic flow accurately is vital for these systems, as it involves analyzing historical traffic data to forecast current and future road conditions [3]. Achieving these goals is essential for improving traffic management, reducing congestion, and enhancing the overall efficiency of ITS, which directly contributes to the improvement of human well-being [4,5]. Optimized traffic management reduces commute times and driver stress [6]. Additionally, enhanced public transportation systems promote a healthier, more sustainable lifestyle for urban residents by improving efficiency and reliability [7,8].
Traffic flow is influenced by both time and the interconnections of roadways, resulting in a complex network structure [9,10]. Hence, forecasting real-time and accurate traffic flow is highly challenging because of the complex spatiotemporal dependencies inherent in traffic flow. The challenges are twofold: Firstly, sophisticated spatial correlations are influenced by both nearby and distant nodes. Secondly, historical traffic data exhibit complex nonlinear relationships, complicating long-term forecasts. Long-term forecasts (over 30 min) are crucial for traffic planning and management, helping predict and respond to events like congestion and accidents, while short-term forecasts (5–30 min) address immediate conditions.
Advancing industrial technology has enabled the widespread use of sensors and data collection devices in transportation networks, collecting large amounts of crucial data for research. Methods such as the historical average (HA), exponential smoothing, and autoregressive integrated moving average (ARIMA) [11], which are commonly used in traditional traffic forecasting, rely on the assumption of stationarity. However, this assumption is not typically applicable to the dynamic nature of traffic data. These methods frequently fall short of providing accurate traffic forecasts because they cannot capture the complex dynamic characteristics of traffic data.
Recently, deep learning traffic flow forecasting methods utilize various neural network architectures to model spatiotemporal features [12,13]. Inspired by sequential learning, some neural networks simulate the temporal impact of traffic fluctuations [14,15]. Convolutional neural networks (CNNs) are used to capture spatial relationships between adjacent regions [16], while recurrent neural networks (RNNs) address the temporal aspects [17]. SHARE [18] predicts parking space availability in cities by employing a hierarchical graph convolutional structure to model spatial autocorrelation between parking lots and an RNN to capture dynamic temporal dependencies. Additionally, PredRNN [19] is a new recurrent neural network that forms a unified representation of complex environments with decoupled memory units and uses zigzag memory flow to transfer information between layers, effectively capturing the dynamics of spatiotemporal data.
Graph neural networks (GNNs) are widely employed in research to capture spatiotemporal dependencies among variables [20,21]. Early approaches, such as those by [22,23], utilized predefined graph structures to represent dependencies, but often struggled to capture complex interconnections, resulting in inaccuracies. Park et al. [24] employed a self-attention mechanism for dynamic correlation computation between sensor pairs, but scaling to large-scale graphs is hindered by quadratic computational complexity. Wang et al. [25] proposed D-TGNM, a dynamic temporal graph neural network designed to predict urban traffic flow robustly even with incomplete data. This model incorporates Traffic BERT to capture dynamic spatial relationships in road networks, alongside a temporal graph neural network (TGNM) that analyzes traffic flow patterns while accounting for missing data. Despite the progress in GNNs, current methods encounter two primary challenges.
First, current studies often use static graph models to describe spatial correlations in traffic networks, which limits traditional traffic prediction methods to direct relationships between adjacent road segments [26,27,28]. However, actual traffic systems are influenced by complex spatial interdependencies due to external factors such as weather changes, incidents, regulations, and temporal dynamics. Even geographically distant road segments may exhibit similar traffic fluctuation patterns over time. Focusing only on local spatial interdependencies may overlook information from remote segments, leading to increased forecast uncertainty. Figure 1 depicts the alterations in relationships among road networks, nodes, and vehicles influenced by traffic control measures. To better model dynamic interactions between road segments, it is crucial to incorporate dynamic graph models. Existing graph neural networks (GNNs) typically propagate information from immediate neighbors during message passing, thereby constraining their capability to capture long-range spatial dependencies [29].
Secondly, current neural network architectures face difficulties in modeling multi-step predictions for non-stationary time series data. Traffic data often show autocorrelation at adjacent time points and periodic fluctuations due to human activities, which vary in pattern and intensity across different periods and weekdays [30]. Researchers have explored various strategies to handle these complex temporal dependencies in multi-step forecasting. One approach involves temporal convolutional networks (TCNs) [31,32], which use convolution operations along the temporal dimension to process time series data efficiently, avoiding issues like gradient vanishing or exploding. While TCNs perform well at single-step predictions with parallel processing, they may not be as effective for multi-step forecasting scenarios. Another method uses sequence-to-sequence models based on recurrent neural networks (RNN) (Seq2Seq) [22,33], which are adept at capturing the sequential nature of time series data. However, Seq2Seq models require multiple iterations to calculate long-distance dependencies [34], and their training phase lacks parallel computing, which could affect computational efficiency and graph-related structural information in pre-processing [35]. Hybrid methods that combine time convolution with recurrent structures may capture time dependencies more accurately and efficiently. Further research is required to confirm the efficacy of such hybrid architectures.
This paper introduces a novel model architecture, named the spatiotemporal dynamic multi-hop network (ST-DMN), designed to tackle the aforementioned challenges. At the core of our approach is a dynamic graph learning algorithm tailored for each time step, adept at updating the graph structure in real time to reflect evolving temporal relationships. To address long-distance spatial dependencies, we have integrated multi-hop operation into our framework. Additionally, our model utilizes a transformer layer to introduce position coding to the spatiotemporal sequence generated by ST-Blocks. This approach not only improves the model’s ability to capture the global temporal structure of the data but also emphasizes the interaction of relative positions within the sequence. In the domain of temporal correlation modeling, the ST-DMN employs an encoder–decoder architecture. The encoder in this architecture consists of L ST-Blocks designed to extract spatiotemporal features from historical traffic data. Subsequently, the decoder utilizes multiple diffusion multi-hop graph convolutional gated recurrent units (DMGCGRUs) as its foundational elements. These units enable the model to achieve multi-step prediction by performing autoregressive decoding of the extracted features. This research offers several key contributions:
  • We have created an advanced encoder–decoder architecture tailored for multi-step prediction. The encoder extracts important spatiotemporal features from historical data by using multiple ST-Blocks. The decoder then uses DMGCGRUs to decode these features and produce multi-step prediction results.
  • We present a dynamic graph learning algorithm to better capture the complex and evolving topology of traffic networks. This algorithm utilizes an iterative updating mechanism to dynamically construct and adjust the topological graph of road networks.
  • ST-DMN combines the multi-hop operation of dynamic graphs with the diffusion convolution technique to effectively capture the inherent long-distance spatial dependence in traffic data. Furthermore, the transformer layer enhances the model’s comprehension and perception of the overall temporal structure within the spatiotemporal embeddings generated by the ST-Blocks.
  • Experimental results on publicly available traffic speed datasets, including METR-LA and PEMS-BAY, indicate that ST-DMN achieves competitive or even superior performance compared to various baseline models.
The structure of the paper is as follows: Section 2 provides a review of existing research and graph structure learning methods in the domain of traffic prediction. In Section 3, we provide a detailed explanation of the architecture of ST-DMN and its underlying algorithms. Section 4 and Section 5 present a comprehensive comparison of the prediction accuracy of the ST-DMN model against several baseline models. It evaluates the influence of different parameters on the model’s outcomes and compares training times, thereby demonstrating its superior performance. Section 6 concludes with the findings and discusses avenues for future research.

2. Related Works

2.1. Traffic Flow Forecasting

Pioneering techniques for anticipating traffic dynamics, including the historical average (HA), autoregressive integrated moving average (ARIMA) [11], and vector autoregression (VAR) [36], were founded on principles of mathematical statistics and rudimentary machine learning algorithms. However, these models frequently faced challenges in addressing the complexity and dynamics of real-world traffic conditions. Traditional traffic research has often oversimplified forecasting by treating it primarily as a time series problem, overlooking the substantial impact of other network nodes on traffic conditions. To improve predictive performance, researchers have explored innovative methods to capture the spatial characteristics of traffic networks. For example, ConvLSTM [37] uses convolutional operations to capture spatial features within temporal sequences. Graph convolutional neural networks (GCNs) extend traditional convolution concepts to non-Euclidean data structures. Spatiotemporal graph neural networks (STGNNs), such as STGCN [23], integrate graph convolution with temporal convolution or recurrent neural networks to adeptly capture spatiotemporal features within road networks.
In recent years, advanced models like GraphWaveNet [31] and AGCRN [38] have gained attention for their ability to dynamically learn the graph structure of road networks from traffic data. These models achieve predictive performance comparable to models based on predefined graphs. Deep neural networks, known for their powerful predictive capabilities, have been integrated with graph neural networks (GNNs) to construct spatiotemporal graph neural networks, enhancing prediction performance [22,39]. These models can be categorized by temporal modeling approaches into RNN-based, CNN-based, and attention-based models. For example, DCRNN [22] combines diffusion graph convolution with GRU for traffic forecasting, and STGCN [23] captures spatiotemporal correlations using 1D CNN and graph convolution. Transformer-based models like STAEFormer [40] and HTVGNN [41] have made significant strides in predictive accuracy and performance by introducing adaptive embeddings and time-varying graph structures.

2.2. Graph Structure Learning

Graph structure learning is a crucial field in machine learning that aims to extract structural information from data. In traffic forecasting, the graph structure models the road network, with nodes representing road intersections or segments and edges indicating traffic capacity or distance relationships. One of the primary challenges in learning graph structures is effectively capturing the intricate interdependencies between nodes.
Lately, academic research has been concentrated on improving the performance of graph neural networks (GNN) by optimizing graph structures [42,43,44]. Current graph structure learning methods can generally be categorized into three main groups: (1) Metric learning strategies, which assess inter-node relationships using various metrics [45,46]. (2) Probabilistic modeling approaches involve constructing graphs by sampling from probability distributions and modeling edge generation probabilities using trainable parameters [47,48]. (3) Direct optimization techniques optimize the entire graph alongside GNN parameters, as explored in recent studies [49]. In traffic forecasting, recent studies have aimed to learn adaptable graph adjacency matrices [31]. However, many methods directly engineer trainable parameters for graph structures, often overlooking node attributes. This oversight can complicate model optimization, especially in cases with limited and sparse training data. Node attributes that provide details on traffic conditions are crucial for constructing a more accurate graph structure.
Recent academic studies, such as [33,44,50], have been concentrating on developing methods to accurately uncover dependencies between nodes. Advanced techniques involve extracting features from large traffic datasets to create graphs, which are then improved with prediction models to enhance prediction performance. As demonstrated by [51], this structured learning approach significantly enhances prediction accuracy. Our proposed structured learning framework goes beyond the analysis of current traffic conditions to deeply investigate traffic patterns. This comprehensive approach ensures the universality and suitability of the framework across different cities and regions, thereby improving its accuracy and scalability.

3. Materials and Methods

3.1. Problem Formulation

Traffic flow forecasting aims to accurately predict future traffic flow based on historical data. To accomplish this, we represent the road network as a weighted directed graph G = ( V , E , W ) . Here, V denotes the set of nodes, e.g., sensors used to collect traffic flow data, with | V | = N . E represents the road segments, and W R N × N is a weighted adjacency matrix representing the proximity of a node. The element W i , j signifies the distance between nodes, typically determined by the road network or topological adjacency.
W i , j = exp dist ( v i , v j ) 2 σ 2 if dist ( v i , v j ) κ , otherwise 0 ,
let W i , j indicate the weight of the edge connecting sensor v i to sensor v j , and dist ( v i , v j ) represent the distance between sensor v i and sensor v j within the road network. Here, σ stands for the standard deviation of the distances, and κ is the designated threshold.
The traffic flow on G is represented by a graph signal X R N × D , where D is the number of features of each node, i.e., speed and flow rate. Traffic flow characteristics encompass various attributes, including traffic flow rate, average speed, and flow density. In this paper, the feature used is speed. The task of predicting traffic data for the next Q time intervals, given the historical data from the previous P time intervals, can be formulated as follows:
X t P + 1 , , X t ; G f θ Y ^ t + 1 , , Y ^ t + Q ,
where θ denotes the learnable model parameters.

3.2. Model Architecture

In this section, we introduce our spatiotemporal dynamic multi-hop network (ST-DMN), which is designed to capture spatiotemporal dependencies in traffic data. Our model employs an encoder–decoder architecture for multi-step traffic flow forecasting, as illustrated in Figure 2.
The encoder comprises L spatiotemporal blocks (ST-Blocks). Each ST-Block is composed of two temporal convolution layers and one spatial convolution layer. The temporal convolution layer captures temporal features, while the spatial convolution layer extracts spatial features. We have integrated residual connections [52] to enhance information flow and address the vanishing gradient issue. Subsequently, the spatiotemporal embedding extracted by the L ST-Blocks is fed to a transformer layer, which enables a better understanding and perception of the global temporal structure of spatiotemporal features.
The decoder is made up of Q diffusion multi-hop graph convolutional gated recurrent units (DMGCGRUs). A DMGCGRU is an improved version of the gated recurrent unit (GRU) that replaces the traditional matrix multiplication in GRUs with diffusion convolution. This modification enhances the efficiency of managing spatiotemporal information transfer. By integrating multi-hop dynamic graph convolution, the DMGCGRUs can better understand relationships across different locations and times. The first DMGCGRU in the sequence takes the feature vector encoded by the encoder, ensuring that the spatiotemporal features captured during encoding smoothly transition into the decoding phase. The decoder interprets this vector to reconstruct the output. At each time step i within the interval [ t + 1 , , t + Q ] , the DMGCGRUs receive the context vector generated from the previous step and produce a prediction using diffusion multi-hop convolution. Integrating diffusion convolution substantially improves the model’s capacity to capture long-term spatiotemporal dependencies. During the training phase, we employ scheduled sampling [53] to alleviate the error accumulation that can arise from multi-step output predictions.

3.3. Encoder Architecture

3.3.1. ST-Blocks

The ST-Blocks integrate temporal convolution and spatial convolutions on graph-structured data to extract intricate dynamic spatial and temporal patterns in traffic data, as shown in Figure 3. Each ST-Block includes two temporal convolution layers and one spatial convolution layer. The initial temporal convolution layer uses convolutional kernels that slide along the temporal dimension to extract temporal features of the input traffic data. Subsequently, the spatial convolution layer learns the spatial dependencies by utilizing spatial convolution operations on the weighted adjacency matrix and performing multi-hop operations on the dynamic adjacency matrix to infer the spatial relationships between nodes. After the spatial convolution layer, the R e L U activation function is applied to enhance the neural network’s nonlinear modeling capability. Further, the second temporal convolution layer strengthens the time-varying relationship across the road segments and extracts hidden temporal features. A residual connection is implemented to capture and transmit important information, thereby avoiding the issue of gradient disappearance in deep networks. Finally, layer normalization is used to ensure the stability of the network training process.
ST-Blocks integrate temporal and spatial convolutions, making them well-suited for complex spatiotemporal tasks that require extracting intricate spatiotemporal dependencies, and they exhibit strong training stability and generalization ability.

3.3.2. Dynamic Graph Learning

In a real traffic network, the interdependence among road segments varies and continuously evolves due to changing conditions such as time and weather. Traditional traffic forecasting methods often assume fixed interdependence, ignoring dynamic characteristics, and limiting prediction accuracy. To overcome this limitation, the study introduces a new method that no longer relies solely on a static adjacency matrix to represent the relationship between nodes. We incorporate the dynamically updated adjacency matrix into the model, as shown in Figure 4.
This matrix primarily includes the dynamic graph from the previous time step, the adjacency matrix at the current time step, and the dynamic graph at the current time step. The dynamic graph learning process will be detailed below. This enhancement aims to enhance the accuracy and applicability of traffic flow forecasting, making it more adaptable to changes in traffic flow data.
Our proposed algorithm takes into account the nonlinear changes in adjacency relationships over time, which are observed in real traffic networks. It involves using a time-varying adjacency matrix and node embeddings to capture this dynamic interaction. Distinct from existing studies, our approach utilizes a three-dimensional tensor D t R T × N × N to represent the temporal changes in adjacency relations. Here T denotes the time steps and N denotes the number of nodes in the dataset. By embedding the time series X through a specific layer, we obtain graph signal embeddings, denoted as X t ( 1 ) and X t ( 2 ) . This facilitates the learning of dynamic interactions among traffic network nodes at individual time steps. The specific learning process is carried out using the following dynamic spatial dependency computation formula:
D t = S o f t M a x R e L U X t ( 1 ) · ( X t ( 2 ) ) ,
X t ( 1 ) and X t ( 2 ) represent embeddings of one-hop neighboring nodes, and D t R T × N × N represents the adjacency matrix at time t.
During this computational process, the model accurately captures and updates the strength of interactions between nodes at different time slots. It is important to note that the adjacency matrix of a node obtained through this method represents the embedding of graph signals at the current time slot. As a result, the adjacency relationships can be dynamically adjusted based on optimization objectives.
In addition, the study proposes an iterative updating mechanism to preserve spatial relationships across consecutive time intervals. This mechanism continuously improves the adjacency matrix over successive time steps, allowing the model to not only represent the current traffic data but also smoothly transition to the next moment. The approach ensures that the model fully considers continuity within the temporal dimension. Thus, at each time step t, the core updating process of the dynamic graph updating algorithm can be represented by the following steps:
A 0 = D 0 , U t = σ ( θ ( D t + A t 1 ) ) ,
initialize dynamic the adjacency matrix A and set it to the initial state of D at time step 0. Compute the update gate U t , which is a parameterized transformation activated by the S i g m o i d activation function. It combines the information from the current time step D t and the prior dynamic adjacency matrix state A t 1 . ⊙ denotes the hadamard product, and θ represents the learnable parameters. Adjust the state of the dynamic adjacency matrix A at time step t by employing the update gate U t to maintain a balance between the stability and the dynamism of the dynamic graph:
A t = U t D t + ( 1 U t ) A t 1 ,
Through this approach, the model proposed in this study not only excels in capturing time-varying traffic characteristics but also demonstrates outstanding scalability and robustness in long-term time series forecast tasks.

3.3.3. Temporal Convolution Layer

When analyzing and forecasting spatiotemporal data, it is crucial to consider the temporal dependency within time series data. In time series analysis, RNN-based methods are commonly used but often encounter challenges such as high computational complexity and issues with exploding or vanishing gradients, especially when dealing with long-range dependencies in real-world applications. To tackle these challenges, this study proposes a strategy that involves stacking temporal convolutional layers within the encoder. This approach improves the model’s sensitivity and its capability to model spatiotemporal information. The temporal convolutional layers effectively capture and model the temporal dependencies within the input data sequences through convolution operations and gated linear units (GLU) [54].
In the traffic network G , each node undergoes processing by the temporal convolutional layer, where padding is applied to keep the time step length P unchanged following the convolution operation. This helps preserve the temporal sequence of the data. Given an input tensor X R B × P × N × D , the temporal convolutional layer uses convolutional kernels of size k t , with an input dimension of D and an output dimension of 2 D . These kernels execute temporal convolution operations, efficiently mapping the input data features into a space of higher dimensionality. The convolution operation ∗ is conducted across multiple channels, with each channel undergoing independent convolution. The outcomes of all channels are then aggregated to produce an output with the dimension of 2 D . This output is subsequently divided into two equal parts along the feature dimension and is subsequently input into the G L U . The mathematical expression for the temporal convolutional layer is formulated as:
C = W X + b ,
where C represents the output of dimension 2 D after convolution, W denotes the convolutional kernel weights, and b stands for the bias term.
G L U ( X ) = σ ( C 1 ) C 2 ,
T ( X ) = ( G L U ( C ) + X ) · θ ,
where σ represents the activation function S i g m o i d applied to the first half channels of C, while C 2 denotes the last half channels of C. The symbol ⊙ indicates element-wise multiplication. The output of the temporal convolutional layer is scaled by θ to enhance stability. Additionally, the gated linear unit (GLU) is utilized for non-linear transformation, aiding in accelerating model training and improving generalization capability.

3.3.4. Spatial Convolution Layer

The dynamics of traffic conditions can vary significantly across different roads and change constantly. To effectively capture long-distance spatial relationships, we have decided to move away from using the traditional multilayer graph neural network (GNN) model. Previous research [55] has pointed out that traditional multi-layer GNNs suffer from over-smoothing, where node representations in locally connected subgraphs become overly similar, resulting in decreased prediction accuracy. We have developed an innovative spatial convolutional layer to adaptively capture the interrelationships between sensors in the road network, addressing this challenge effectively. We use a single-layer approach in this study to update node representations and integrate information from multi-hop neighbors. This method not only comprehensively captures the traffic topology of the road network, but also effectively mitigates the over-smoothing phenomenon. As part of this innovative approach, we have constructed a multi-hop graph model to operationalize our methodology:
A ( k + 1 ) = ( 1 α ) α A ˜ + ( 1 α ) A ( k ) A ˜ ,
where α ( 0 , 1 ) is the attenuation factor used to control the degree of weight attenuation transmitted through the multi-hop path. A ˜ is the one-hop matrix, which is the dynamic adjacency matrix, and k is the number of hops.
In addition to learnable dynamic graphs, predefined adjacency matrices are also crucial. The matrix represents the fixed links among various parts of the urban network. Spatial proximity information is usually provided by the dataset itself or can be accessed through city maps. We model spatial dependencies by correlating time series to diffusion processes, explicitly capturing the random nature of traffic dynamics. An explicit graph S represents these correlations, with the diffusion process characterized by random walks on graph G .
One limitation of basic graph convolution is its restriction to undirected graphs, which does not suit the directed nature of traffic networks. To address this, Li et al. [22] introduced forward and backward diffusion processes for graph signals with K finite steps, enabling convolution on directed graphs. Studies show that the bidirectional diffusion convolution technique successfully models both upstream and downstream traffic impacts in predictive models. The diffusion convolution operation is defined as follows:
S ( X , W ) = k = 0 K 1 W k , 1 D O 1 W k + W k , 2 D I 1 W k X ,
where W k , 1 and W k , 2 are the learnable filter parameters. K denotes the number of diffusion steps. D O 1 W , D I 1 W represent the transfer matrices for the diffusion process (output) and the reverse diffusion process (input), respectively.
In this study, we couple the static graph generated by prior relationships with the dynamic graph produced through multi-hop processing, ultimately proposing the spatial convolution layer of this paper, represented as follows:
Z t ( k + 1 ) = S ( X , W ) + D ( t ) ( k + 1 ) ( X , A ˜ ) ,
where Z t ( k + 1 ) represents the output signal of the spatial convolution layer. S ( · ) and D ( · ) denote the outputs of the static and dynamic graphs after the diffusion convolution layer, respectively.
The spatial convolution layer uses both static graphs and dynamic multi-hop graphs to handle spatial relationships. It makes use of static graphs to address underlying spatial dependencies, while dynamic multi-hop graphs are particularly effective at identifying long-distance spatial connections. We have employed an integrated strategy that combines dynamic graph modeling with multi-hop operation. This fusion allows the network to flexibly map dynamic traffic topology changes and simulate interactions between different road network sections over various time points. This integrated approach ensures a comprehensive understanding of spatial relationships. It enhances the network’s adaptability and responsiveness to fluctuations in traffic patterns.

3.3.5. Transformer Layer

Recently, there has been increasing interest in neural network architectures that effectively model the internal dependencies of spatiotemporal data. In this paper, we introduce a transformer layer module specifically designed to handle spatiotemporal information. Our module integrates a self-attention mechanism, a convolutional layer, and positional encoding to capture complex patterns in a flexible and scalable manner.
The positional encoding augments the sequence information of the spatiotemporal embedding, enabling the model to accurately distinguish information across different time steps and effectively manage long-term dependencies. Enhancing sensitivity to spatiotemporal relationships, the self-attention mechanism enables the model to dynamically focus on information from different locations when processing sequence data. Additionally, the convolution layer is used to extract local features, thereby enhancing the model’s capability to perceive the local structure of the input data space. To further stabilize and expedite the model’s convergence during training, particularly with long sequence data, we have introduced a layer normalization mechanism. This module uses as input the spatiotemporal embedding H R B × T × N × D processed by ST-Blocks, expressed as follows:
H ( P E ) = H + P E ( H ) ,
Q = C o n v 1 H ( P E ) , K = C o n v 2 H ( P E ) , V = L i n e a r H ( P E ) ,
A = S o f t M a x Q K C ,
H o u t = L a y e r N o r m ( A V + H ( P E ) ) ,
where P E is the matrix calculated based on the positional encoding, the input H ( P E ) R B × T × N × D of the transformer block. C o n v 1 and C o n v 2 are convolution operations, and L i n e a r is a linear transformation. A denotes the self-attention matrix, H o u t R B × T × N × D represents the output of this module.

3.4. Decoder Architecture

We utilize Q diffusion multi-hop graph convolutional gated recurrent units (DMGCGRUs) as decoders, as illustrated in Figure 5, to independently capture static and multi-hop spatial dependencies. Specifically, we replace the matrix multiplication of the traditional gated recurrent unit (GRU) with diffusion convolutions. This modification allows us to input the context vector at each time step t within the range t = [ t + 1 , , t + Q ] . Multi-hop dynamic graph convolution, G Z in Figure 5, is introduced in the process of dynamic spatial dependency computation, and this design broadens the receptive field of the target node and adaptively captures the fluctuations in the connectivity of the road network. The computational process for the DMGCGRUs at time t is as follows:
u t = S i g m o i d ( G Z ( X t H t 1 ) θ u + b u ) , r t = S i g m o i d ( G Z ( X t H t 1 ) θ r + b r ) , C t = T a n h ( G Z ( X t ( r t H t 1 ) ) θ C + b C ) , H t = ( 1 u t ) C t + u t H t 1 ,
among them, X t and H t denote the input signal and hidden state signal at time t, respectively. r t and u t denote the reset gate and update gate at time t, respectively. The symbol ⊙ denotes the Hadamard product. The vector concatenation is denoted by ⊕. G Z represents the multi-hop dynamic graph convolution layer as given in Equation (11), with θ u , θ r , θ C being the parameters of the corresponding filters.

3.5. Loss Function

In this study, the loss function used is the mean absolute error (MAE), formulated as shown in Equation (17).
L ( θ ) = 1 T 1 N t = 1 t = T i = 1 i = N | Y i , t Y ^ i , t ( θ ) |
where θ denotes all trainable parameters, Y ^ i , t denotes the predicted output, and Y i , t represents the ground truth of the i-th node at time t.

4. Experiments

4.1. Datasets

Experiments are performed on two real-world large-scale traffic datasets: METR-LA and PEMS-BAY. METR-LA consists of 207 nodes, while PEMS-BAY comprises 325 nodes. Specific statistics can be found in Table 1.
To ensure a fair comparison with other benchmark methods, we adhered to the data processing procedure outlined in the paper [22]. The dataset was split chronologically, with 70 % used for training, 20 % for testing, and the remaining 10 % for validation. Furthermore, all input speed data were normalized using the Z-score method [22,31,56].

4.2. Baselines

To assess the effectiveness of ST-DMN, we benchmarked it against various traditional and advanced model methodologies. Our focus was on the prediction errors associated with the 15, 30, and 60 min forecasts. Here is a refined overview of the comparative baselines we considered:
  • HA: employs the mean of past data as a basis for forecasting subsequent traffic volumes.
  • ARIMA [57]: combines autoregression, differencing, and moving average to forecast non-stationary time series.
  • FC-LSTM [58]: combines fully connected layers and LSTM for enhanced time series prediction.
  • DCRNN [22]: it is a deep learning model that integrates a graph convolutional network with the recurrent neural network and enhances the model’s understanding of spatial correlation by a diffusion convolution operation.
  • Graph WaveNet [31]: uses graph convolutional networks and dilated convolutions to capture spatiotemporal traffic patterns.
  • GMAN [59]: adopts an encoder–decoder architecture with spatiotemporal attention blocks for dynamic traffic prediction.
  • CCRNN [60]: integrates spatial and temporal features using coupled graph convolutions and gated recurrent units.
  • GTS [33]: learns probabilistic graph structures for multiple time series prediction.
  • PM-MemNet [61]: uses a memory network with pattern matching to predict traffic in complex road networks.

4.3. Experiment Settings

All deep learning-based models, including our ST-DMN, were implemented using Python 3.8 and PyTorch 1.10.0 and executed on a GPU server equipped with one RTX 3090/24 GB. We conducted 5 experiments for each model, and the final results were averaged to ensure reliability and statistical significance.
During our experiments, we optimized all hyperparameters by minimizing the MAE metric. According to previous work [22], the number of input steps equals the number of output steps. For instance, we used P = 12 steps (equivalent to 60 min) to predict the next Q = 12 steps (60 min). The model included L = 2 ST-Blocks, allowed for a maximum diffusion of a K = 1 step, used an embedding dimension of DGL equal to 16, and employed a batch size of 64. During training, the Adam optimizer [62] was used with a fixed learning rate of 0.01 , decayed by 0.5 every 10 epoch. Training sessions lasted for 100 epochs, with early stopping implemented to prevent overfitting.

4.4. Evaluation Metrics

Suppose Y denotes the ground truth, Y ^ denotes the predicted values, and N denotes the total number of samples. The evaluation metrics used in this study include: mean absolute error (MAE), measured in mph, root mean square error (RMSE), also measured in mph, and mean absolute percentage error (MAPE), measured in %. They are widely used as an evaluation index in traffic state prediction tasks.
M A E = 1 N i = 1 N | Y Y ^ |
R M S E = 1 N i = 1 N ( Y Y ^ )
M A P E = 1 N i = 1 N | Y Y ^ Y |

5. Results

5.1. Experiment Results and Analysis

The results presented in Table 2 compare the ST-DMN model with other baseline models based on predictions made 15 min (3 steps), 30 min (6 steps), and 60 min (12 steps) in advance using the METR-LA and PEMS-BAY datasets. Superior predictive performance is highlighted in bold, while suboptimal performance is underlined. Our model demonstrates superior performance, indicated by bold highlighting. Based on these experimental findings, the following conclusions can be drawn:
(1)
Deep learning methods demonstrate superior performance over traditional time series methods and machine learning models. Traditional approaches often face challenges in achieving high data stationarity, whereas deep neural networks excel in modeling the nonlinear dynamics of traffic data.
(2)
Among deep learning models, graph-based architectures like DCRNN, GW-Net, and GMAN consistently outperform FC-LSTM models. This emphasizes the significance of integrating road network data into traffic flow forecasting models, indicating that spatial connectivity plays a critical role in accurate prediction.
(3)
The CCRNN model initializes its learnable graph using a 0–1 adjacency matrix of the road network, while the GTS model transforms the problem into learning a probabilistic graphical model by optimizing the performance averaged across the graph distribution. These models leverage dynamic graph structures to enhance predictive performance over earlier methods.
(4)
The PM-MemNet model innovatively uses a key-value memory structure to associate input data with representative patterns and identify the best pattern for predicting future traffic conditions based on given spatiotemporal features.
At present, the PM-MemNet model represents the forefront of traffic forecasting and shows superior performance.
We compared the performance of the ST-DMN with the latest baseline model using the relative error rate metric. On the METR-LA dataset, our ST-DMN model showed improvements in the mean absolute percentage error (MAPE) of 5.41 % for 3 steps, 4.21 % for 6 steps, and 3.53 % for 12 steps. When tested on the PEMS-BAY dataset, we observed improvements of 2.93 % , 1.09 % , and 0.67 % for the 3, 6, and 12 steps ahead predictions, respectively. Overall, the experimental results demonstrate that ST-DMN performs strongly and competitively against both traditional and state-of-the-art baseline models. The ability of the model to leverage spatiotemporal information significantly enhances prediction accuracy, making it a promising candidate for practical applications, including traffic management and urban planning.
Figure 6 compares the time series prediction performance of the four latest models on the METR-LA and PEMS-BAY datasets. As illustrated, the prediction errors for all models tend to increase as the forecasted time horizon extends. This trend is due to the increased uncertainty inherent in longer prediction spans. Overall, the ST-DMN model performs better than the others on both datasets, with suboptimal results for the one-hour forecast on METR-LA. This may be because the factors to be considered become more complex and uncertain as the prediction time increases, such as with accidents and weather conditions. The PM-MemNet model also demonstrates robust performance, particularly in short-term forecasting, likely owing to its efficient memory mechanism for handling time series data. Conversely, the CCRN model exhibits significant errors across all time domains, possibly due to its simpler structure struggling to capture the intricacies of spatiotemporal relationships.
In summary, the ST-DMN model shows more prominent performance on both datasets, likely due to its ability to effectively handle spatiotemporal dependencies in time series data. However, errors for all models increase as the forecast time interval lengthens, highlighting the inherent challenge of time series forecasting: longer forecasts come with greater uncertainty.

5.2. Model Configuration Analysis

Throughout our study, we conducted several experiments to assess the impact of different parameter settings on model performance. Table 3 shows the average values of MAE, RMSE, and MAPE for various parameters.
We started by analyzing the parameter k in Equation (9). The experimental results were notably influenced by the chosen value of k. As k increases, the average values of the three metrics decrease, indicating an overall improvement in predictive performance. For instance, when k = 3, the MAE decreased by 0.04 compared to when k = 1, and the RMSE decreased by 0.1 . This aligns with the findings in [63], demonstrating that by considering more distant neighbors through multi-hop paths, the model’s predictive outcomes are positively affected. However, we also noticed that when k reaches 4, the performance declined. This suggests that while a larger k captures a broader graph structure information, it also increases computational complexity. Therefore, striking a balance between model performance and computational efficiency is crucial. Furthermore, it became apparent that the additional improvements in predictive performance become limited with further increases in k.
Next, we explored how the decay factor α in Equation (9) affects experimental outcomes. Our study revealed that a smaller α has a positive impact on the model, yielding the best results when α = 0.15 . The parameter α influences the amplification of low-frequency signals in the graph structure. A smaller α enhances low-frequency signals while suppressing high-frequency ones [63]. Selecting an appropriate α value allows the model to effectively capture graph structure information and improves prediction performance.
Lastly, we discussed the critical parameter of node embedding dimensions in DGL. Table 3 reveals that neither excessively large nor small embedding dimensions achieve optimal performance. Smaller embedding dimensions may limit the expressive power of node features, while larger dimensions increase computational complexity and risk overfitting. The model achieves the best results with an embedding dimension of 16, demonstrating its robustness.
Analyzing these parameters provides a better understanding of their impact on model performance, which facilitates more informed decisions in practical applications.

5.3. Model Efficiency

Table 4 presents a comparison of the computation times between ST-DMN and other state-of-the-art baseline models on the METR-LA dataset. We recorded the average training time per epoch for each model. One advantage of ST-DMN over DCRNN is that its encoder does not use a recurrent neural network, resulting in better computational performance. PM-MemNet utilizes graph convolutional memory networks (GCMem), which can increase computational costs when dealing with large-scale datasets. While GTS simplifies the optimization process with a single-stage optimization approach, the search for an optimal solution among many possible graph structures can increase computational demands. Despite the lack of a dedicated decoder, ST-DMN outperforms Graph WaveNet in predictive performance, although it is slightly slower in terms of computation speed. Therefore, for researchers looking for a balance between computational cost and performance, ST-DMN is an ideal choice, offering enhanced performance while maintaining high computational efficiency.

5.4. Ablation Study

To thoroughly validate our proposed model design, we conducted a series of ablation experiments to systematically analyze the impact of various components on performance. These experiments were evaluated using the METR-LA and PEMS-BAY datasets. The results, detailed in Table 5, offer a comprehensive overview of the model’s performance across different prediction intervals: 15 min (3 steps), 30 min (6 steps), and 60 min (12 steps). The distinctive features of these variants are as follows:
  • “w/o Transformer Layer” excludes the transformer layer.
  • “w/o DGL” excludes the dynamic graph learning.
  • “w/o Multi-Hop” excludes the multi-hop operation.
In our evaluation of the METR-LA dataset, we initially assessed the effects of excluding the transformer layer from the model. The results indicated a slight increase in MAE, RMSE, and MAPE across all prediction horizons for the model lacking the transformer layer. This finding emphasizes the critical role of the transformer layer in capturing spatiotemporal relationships within traffic data. Similarly, we investigated the performance impact of removing the dynamic graph learning (DGL) component. The results indicated a decrease in performance, highlighting the significance of dynamic graphs in learning spatial dependencies between traffic nodes. Furthermore, we examined the contribution of multi-hop operation to the model. The exclusion of multi-hop operation resulted in higher prediction errors, highlighting their essential role in modeling extended spatial dependencies. The model’s ability to thoroughly capture traffic topology information within the road network and mitigate the issue of over-smoothing further highlights the importance of these operations.
The complete model, ST-DMN, incorporating all proposed components, demonstrates the best performance across all evaluation metrics and prediction periods of the METR-LA dataset. Similar trends as those on the METR-LA dataset were observed in the ablation experiments conducted on the PEMS-BAY. The performance decline of the ST-DMN model when excluding the transformer layer, DGL, and multi-hop operations are graphically represented in Figure 7. Consistently, the comprehensive ST-DMN model outperforms its counterparts lacking certain components across all evaluation metrics and different prediction horizons.
In summary, the ablation experiments highlight the effectiveness of the proposed components in the ST-DMN model and emphasize the synergistic effect of Transformer layers, DGL, and multi-hop operations in improving the overall prediction accuracy on the METR-LA and PEMS-BAY datasets. This detailed analysis further validates the importance of our proposed model architecture in capturing the spatiotemporal dependencies necessary for accurate traffic flow forecasting.

6. Conclusions

This study introduces the spatiotemporal dynamic multi-hop network (ST-DMN), an innovative architectural model designed to address the complexity of traffic flow forecasting. The core of ST-DMN is a dynamic graph learning algorithm that iteratively updates the traffic network graph to accurately represent the evolving traffic patterns. By integrating multi-hop operation with diffusion convolutional techniques, our model effectively captures long-distance spatial dependencies, an important aspect of traffic data that existing models fail to adequately address. In addition, the incorporation of transformer layers enables ST-DMN to effectively understand the global temporal structure, consequently enhancing its prediction ability. ST-DMN successfully captures the intricate spatiotemporal dependence of traffic speed and outperforms existing models in terms of prediction accuracy. In future work, we plan to explore the application of extended models, including validation on different cities or traffic network structures to evaluate its generalization performance.

Author Contributions

Conceptualization, W.C. and T.Z.; methodology, W.C. and Q.L.; software, Q.L. and Z.L.; validation, Z.L. and J.Y.; investigation, J.Y.; writing—original draft preparation, J.Z. and W.C.; supervision, J.Z. and T.Z.; project administration, T.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Guangdong Province (No. 2022A1515011590, 2024A1515011766), the China Postdoctoral Science Foundation (No. 2024M750562), the National Postdoctoral Fellowship Program (No. GZC20230549), the State key laboratory major special projects of Jilin Province Science and Technology Development Plan (No. SKL202402024), the Natural Science Foundation of China (No. 82020108016, 61902232), the National Key Research and Development Program (No. 2021YFB2700600, 2022YFB3104600), the Guangdong Provincial University Innovation Team Project (No. 2020KCXTD012), the Municipal Government of Quzhou (No. 2022D018, 2022D029, 2023D007, 2023D015, 2023D033, 2023D034, 2023D035), and the Natural Science Foundation of Zhejiang Provincial (No. LGF22G010009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset and source code generated during and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Acknowledgments

The authors express their gratitude to the reviewers and editors for their valuable feedback and contributions to refining this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, T.; Zhao, Y.; Lin, Z.; Zhou, J.; Li, H.; Wang, F. Moral and Formal Model-based Control Strategy for Autonomous Vehicles at Traffic-light-free intersections. Smart Constr. Sustain. Cities 2024, 2, 11. [Google Scholar] [CrossRef]
  2. Kaffash, S.; Nguyen, A.T.; Zhu, J. Big data algorithms and applications in intelligent transportation system: A review and bibliometric analysis. Int. J. Prod. Econ. 2021, 231, 107868. [Google Scholar] [CrossRef]
  3. Wang, F.; Liang, Y.; Lin, Z.; Zhou, J.; Zhou, T. SSA-ELM: A Hybrid Learning Model for Short-Term Traffic Flow Forecasting. Mathematics 2024, 12, 1895. [Google Scholar] [CrossRef]
  4. Voukelatou, V.; Gabrielli, L.; Miliou, I.; Cresci, S.; Sharma, R.; Tesconi, M.; Pappalardo, L. Measuring objective and subjective well-being: Dimensions and data sources. Int. J. Data Sci. Anal. 2021, 11, 279–309. [Google Scholar] [CrossRef]
  5. Park, C.L.; Kubzansky, L.D.; Chafouleas, S.M.; Davidson, R.J.; Keltner, D.; Parsafar, P.; Conwell, Y.; Martin, M.Y.; Hanmer, J.; Wang, K.H. Emotional well-being: What it is and why it matters. Affect. Sci. 2023, 4, 10–20. [Google Scholar] [CrossRef]
  6. Chui, K.T. Driver stress recognition for smart transportation: Applying multiobjective genetic algorithm for improving fuzzy c-means clustering with reduced time and model complexity. Sustain. Comput. Inform. Syst. 2022, 35, 100668. [Google Scholar] [CrossRef]
  7. Xu, W.; Liu, J.; Yan, J.; Yang, J.; Liu, H.; Zhou, T. Dynamic Spatiotemporal Graph Wavelet Network for Traffic Flow Prediction. IEEE Internet Things J. 2024, 19, 8019–8029. [Google Scholar] [CrossRef]
  8. Ghafouri-Azar, M.; Diamond, S.; Bowes, J.; Gholamalizadeh, E. The sustainable transport planning index: A tool for the sustainable implementation of public transportation. Sustain. Dev. 2023, 31, 2656–2677. [Google Scholar] [CrossRef]
  9. Li, Z.; Zhou, J.; Lin, Z.; Zhou, T. Dynamic spatial aware graph transformer for spatiotemporal traffic flow forecasting. Knowl.-Based Syst. 2024, 297, 111946. [Google Scholar] [CrossRef]
  10. Ishak, S.; Al-Deek, H. Performance evaluation of short-term time-series traffic prediction model. J. Transp. Eng. 2002, 128, 490–498. [Google Scholar] [CrossRef]
  11. Isufi, E.; Loukas, A.; Simonetto, A.; Leus, G. Autoregressive moving average graph filtering. IEEE Trans. Signal Process. 2016, 65, 274–288. [Google Scholar] [CrossRef]
  12. Wang, X.; Ma, Y.; Wang, Y.; Jin, W.; Wang, X.; Tang, J.; Jia, C.; Yu, J. Traffic flow prediction via spatial temporal graph neural network. In Proceedings of the WWW ’20: The Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 1082–1092. [Google Scholar] [CrossRef]
  13. Shi, X.; Qi, H.; Shen, Y.; Wu, G.; Yin, B. A spatial-temporal attention approach for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2020, 22, 4909–4918. [Google Scholar] [CrossRef]
  14. Liu, Q.; Wu, S.; Wang, L.; Tan, T. Predicting the next location: A recurrent model with spatial and temporal contexts. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar] [CrossRef]
  15. Yu, R.; Li, Y.; Shahabi, C.; Demiryurek, U.; Liu, Y. Deep learning: A generic approach for extreme condition traffic forecasting. In Proceedings of the 2017 SIAM International Conference on Data Mining (SIAM 2017), Houston, TX, USA, 27–29 April 2017; pp. 777–785. [Google Scholar] [CrossRef]
  16. Zhang, J.; Zheng, Y.; Qi, D. Deep spatio-temporal residual networks for citywide crowd flows prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar] [CrossRef]
  17. Yao, H.; Wu, F.; Ke, J.; Tang, X.; Jia, Y.; Lu, S.; Gong, P.; Ye, J.; Li, Z. Deep multi-view spatial-temporal network for taxi demand prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar] [CrossRef]
  18. Zhang, W.; Liu, H.; Liu, Y.; Zhou, J.; Xiong, H. Semi-supervised hierarchical recurrent graph neural network for city-wide parking availability prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1186–1193. [Google Scholar] [CrossRef]
  19. Wang, Y.; Wu, H.; Zhang, J.; Gao, Z.; Wang, J.; Philip, S.Y.; Long, M. Predrnn: A recurrent neural network for spatiotemporal predictive learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2208–2225. [Google Scholar] [CrossRef]
  20. He, S.; Luo, Q.; Du, R.; Zhao, L.; He, G.; Fu, H.; Li, H. STGC-GNNs: A GNN-based traffic prediction framework with a spatial–temporal Granger causality graph. Phys. A Stat. Mech. Appl. 2023, 623, 128913. [Google Scholar] [CrossRef]
  21. Wang, Q.; Liu, W.; Wang, X.; Chen, X.; Chen, G.; Wu, Q. GMHANN: A Novel Traffic Flow Prediction Method for Transportation Management Based on Spatial-Temporal Graph Modeling. IEEE Trans. Intell. Transp. Syst. 2023, 25, 386–401. [Google Scholar] [CrossRef]
  22. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. arXiv 2017, arXiv:1707.01926. [Google Scholar] [CrossRef]
  23. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar] [CrossRef]
  24. Park, C.; Lee, C.; Bahng, H.; Tae, Y.; Jin, S.; Kim, K.; Ko, S.; Choo, J. ST-GRAT: A novel spatio-temporal graph attention networks for accurately forecasting dynamically changing road speed. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Virtual, 19–23 October 2020; pp. 1215–1224. [Google Scholar] [CrossRef]
  25. Wang, P.; Zhang, Y.; Hu, T.; Zhang, T. Urban traffic flow prediction: A dynamic temporal graph network considering missing values. Int. J. Geogr. Inf. Sci. 2023, 37, 885–912. [Google Scholar] [CrossRef]
  26. Geng, X.; Li, Y.; Wang, L.; Zhang, L.; Yang, Q.; Ye, J.; Liu, Y. Spatiotemporal multi-graph convolution network for ride-hailing demand forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 3656–3663. [Google Scholar] [CrossRef]
  27. Lin, Z.; Feng, J.; Lu, Z.; Li, Y.; Jin, D. Deepstn+: Context-aware spatial-temporal neural network for crowd flow prediction in metropolis. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 1020–1027. [Google Scholar] [CrossRef]
  28. Pan, Z.; Liang, Y.; Wang, W.; Yu, Y.; Zheng, Y.; Zhang, J. Urban traffic prediction from spatio-temporal data using deep meta learning. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1720–1730. [Google Scholar] [CrossRef]
  29. Zhong, Z.; Li, C.T.; Pang, J. Hierarchical message-passing graph neural networks. Data Min. Knowl. Discov. 2023, 37, 381–408. [Google Scholar] [CrossRef]
  30. Zhang, K.; Zhou, F.; Wu, L.; Xie, N.; He, Z. Semantic understanding and prompt engineering for large-scale traffic data imputation. Inf. Fusion 2024, 102, 102038. [Google Scholar] [CrossRef]
  31. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph wavenet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar] [CrossRef]
  32. Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention based spatial-temporal graph convolutional networks for traffic flow forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 922–929. [Google Scholar] [CrossRef]
  33. Shang, C.; Chen, J.; Bi, J. Discrete graph structure learning for forecasting multiple time series. arXiv 2021, arXiv:2101.06861. [Google Scholar] [CrossRef]
  34. Gehring, J.; Auli, M.; Grangier, D.; Yarats, D.; Dauphin, Y.N. Convolutional sequence to sequence learning. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; PMLR: London, UK, 2017; pp. 1243–1252. [Google Scholar] [CrossRef]
  35. Chen, W.; Chen, L.; Xie, Y.; Cao, W.; Gao, Y.; Feng, X. Multi-range attentive bicomponent graph convolutional network for traffic forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3529–3536. [Google Scholar] [CrossRef]
  36. Nguyen, H.A.T.; Nguyen, H.D.; Do, T.H. An Application of Vector Autoregressive Model for Analyzing the Impact of Weather And Nearby Traffic Flow On The Traffic Volume. In Proceedings of the 2022 RIVF International Conference on Computing and Communication Technologies (RIVF), Ho Chi Minh City, Vietnam, 20–22 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 328–333. [Google Scholar] [CrossRef]
  37. Liu, Y.; Zheng, H.; Feng, X.; Chen, Z. Short-term traffic flow prediction with Conv-LSTM. In Proceedings of the 2017 9th International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 11–13 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar] [CrossRef]
  38. Bai, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive graph convolutional recurrent network for traffic forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17804–17815. [Google Scholar] [CrossRef]
  39. Zhong, W.; Meidani, H.; Macfarlane, J. Attention-based Spatial-Temporal Graph Neural ODE for Traffic Prediction. arXiv 2023, arXiv:2305.00985. [Google Scholar] [CrossRef]
  40. Liu, H.; Dong, Z.; Jiang, R.; Deng, J.; Deng, J.; Chen, Q.; Song, X. Spatio-temporal adaptive embedding makes vanilla transformer sota for traffic forecasting. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; pp. 4125–4129. [Google Scholar] [CrossRef]
  41. Dai, B.A.; Ye, B.L. A novel hybrid time-varying graph neural network for traffic flow forecasting. arXiv 2024, arXiv:2401.10155. [Google Scholar] [CrossRef]
  42. Zhu, Y.; Xu, W.; Zhang, J.; Du, Y.; Zhang, J.; Liu, Q.; Yang, C.; Wu, S. A survey on graph structure learning: Progress and opportunities. arXiv 2021, arXiv:2103.03036. [Google Scholar] [CrossRef]
  43. Xue, G.; Zhong, M.; Li, J.; Chen, J.; Zhai, C.; Kong, R. Dynamic network embedding survey. Neurocomputing 2022, 472, 212–223. [Google Scholar] [CrossRef]
  44. Cao, D.; Wang, Y.; Duan, J.; Zhang, C.; Zhu, X.; Huang, C.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; et al. Spectral temporal graph neural network for multivariate time-series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17766–17778. [Google Scholar] [CrossRef]
  45. Zhou, Z.; Zhou, S.; Mao, B.; Zhou, X.; Chen, J.; Tan, Q.; Zha, D.; Wang, C.; Feng, Y.; Chen, C. Opengsl: A comprehensive benchmark for graph structure learning. arXiv 2023, arXiv:2306.10280. [Google Scholar] [CrossRef]
  46. Li, R.; Wang, S.; Zhu, F.; Huang, J. Adaptive graph convolutional neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar] [CrossRef]
  47. Franceschi, L.; Niepert, M.; Pontil, M.; He, X. Learning discrete structures for graph neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; PMLR: London, UK, 2019; pp. 1972–1982. [Google Scholar] [CrossRef]
  48. Zheng, C.; Zong, B.; Cheng, W.; Song, D.; Ni, J.; Yu, W.; Chen, H.; Wang, W. Robust graph representation learning via neural sparsification. In Proceedings of the International Conference on Machine Learning, Virtual, 12–18 July 2020; PMLR: London, UK, 2020; pp. 11458–11468. [Google Scholar]
  49. Yang, L.; Kang, Z.; Cao, X.; Jin, D.; Yang, B.; Guo, Y. Topology Optimization based Graph Convolutional Network. In Proceedings of the International Joint Conference on Artificial Intelligence, Macau, China, 10–16 August 2019; pp. 4054–4061. [Google Scholar] [CrossRef]
  50. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Chang, X.; Zhang, C. Connecting the dots: Multivariate time series forecasting with graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 753–763. [Google Scholar] [CrossRef]
  51. Zügner, D.; Aubet, F.X.; Satorras, V.G.; Januschowski, T.; Günnemann, S.; Gasthaus, J. A study of joint graph inference and forecasting. arXiv 2021, arXiv:2109.04979. [Google Scholar] [CrossRef]
  52. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  53. Bengio, S.; Vinyals, O.; Jaitly, N.; Shazeer, N. Scheduled sampling for sequence prediction with recurrent neural networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  54. Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language modeling with gated convolutional networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; PMLR: London, UK, 2017; pp. 933–941. [Google Scholar] [CrossRef]
  55. Cai, C.; Wang, Y. A note on over-smoothing for graph neural networks. arXiv 2020, arXiv:2006.13318. [Google Scholar] [CrossRef]
  56. Yu, H.; Li, T.; Yu, W.; Li, J.; Huang, Y.; Wang, L.; Liu, A. Regularized graph structure learning with semantic knowledge for multi-variates time-series forecasting. arXiv 2022, arXiv:2210.06126. [Google Scholar] [CrossRef]
  57. Williams, B.M.; Hoel, L.A. Modeling and forecasting vehicular traffic flow as a seasonal ARIMA process: Theoretical basis and empirical results. J. Transp. Eng. 2003, 129, 664–672. [Google Scholar] [CrossRef]
  58. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar] [CrossRef]
  59. Zheng, C.; Fan, X.; Wang, C.; Qi, J. Gman: A graph multi-attention network for traffic prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1234–1241. [Google Scholar] [CrossRef]
  60. Ye, J.; Sun, L.; Du, B.; Fu, Y.; Xiong, H. Coupled layer-wise graph convolution for transportation demand prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 4617–4625. [Google Scholar] [CrossRef]
  61. Lee, H.; Jin, S.; Chu, H.; Lim, H.; Ko, S. Learning to remember patterns: Pattern matching memory networks for traffic forecasting. arXiv 2021, arXiv:2110.10380. [Google Scholar] [CrossRef]
  62. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  63. Wang, G.; Ying, R.; Huang, J.; Leskovec, J. Multi-hop attention graph neural network. arXiv 2020, arXiv:2009.14332. [Google Scholar] [CrossRef]
Figure 1. The relationship between the road grid, nodes, and vehicles. At time = 0, sensors A and B exhibit a strong correlation due to road connectivity; at time = 1, this correlation weakens as a result of traffic control measures.
Figure 1. The relationship between the road grid, nodes, and vehicles. At time = 0, sensors A and B exhibit a strong correlation due to road connectivity; at time = 1, this correlation weakens as a result of traffic control measures.
Sustainability 16 05860 g001
Figure 2. The framework of the spatiotemporal dynamic multi-hop network (ST-DMN).
Figure 2. The framework of the spatiotemporal dynamic multi-hop network (ST-DMN).
Sustainability 16 05860 g002
Figure 3. The architecture of ST-Block consists of two temporal convolutional layers and one spatial convolutional layer, with the utilization of residual connections for the output.
Figure 3. The architecture of ST-Block consists of two temporal convolutional layers and one spatial convolutional layer, with the utilization of residual connections for the output.
Sustainability 16 05860 g003
Figure 4. The process of dynamic graph learning demonstrates how the dynamic graph at the current time step can be generated from the dynamic graph of the previous time step through the use of an update gate.
Figure 4. The process of dynamic graph learning demonstrates how the dynamic graph at the current time step can be generated from the dynamic graph of the previous time step through the use of an update gate.
Sustainability 16 05860 g004
Figure 5. The architecture of DMGCGRU. It contains dynamic graph multi-hop operations and static graph coupling processes G Z and GRU units.
Figure 5. The architecture of DMGCGRU. It contains dynamic graph multi-hop operations and static graph coupling processes G Z and GRU units.
Sustainability 16 05860 g005
Figure 6. Prediction error at each horizon on METR-LA and PEMS-BAY.
Figure 6. Prediction error at each horizon on METR-LA and PEMS-BAY.
Sustainability 16 05860 g006
Figure 7. Comparison of ST-DMN variants on the METR-LA and PEMS-BAY dataset.
Figure 7. Comparison of ST-DMN variants on the METR-LA and PEMS-BAY dataset.
Sustainability 16 05860 g007
Table 1. Description and statistics of METR-LA and PEMS-BAY datasets.
Table 1. Description and statistics of METR-LA and PEMS-BAY datasets.
DatasetNodesEdgesTime StepsData Points
METR-LA207151534,2726,519,002
PEMS-BAY325236952,11616,937,179
Table 2. Comparison of forecasting performance between ST-DMN and other baseline models on the METR-LA and PEMS-BAY datasets.
Table 2. Comparison of forecasting performance between ST-DMN and other baseline models on the METR-LA and PEMS-BAY datasets.
DatasetModels15 min/Horizon 330 min/Horizon 660 min/Horizon 12
MAERMSEMAPEMAERMSEMAPEMAERMSEMAPE
METR-LAHA4.167.8013.004.167.8013.004.167.8013.00
ARIMA [57]3.998.129.605.1510.4512.76.9013.2317.40
FC-LSTM [58]3.446.309.603.777.2310.904.378.6913.20
DCRNN [22]2.775.387.303.156.458.803.607.6010.50
GW-Net [31]2.695.156.903.076.228.373.537.3710.01
GMAN [59]2.805.557.413.126.498.733.447.3510.07
CCRNN [60]2.855.547.503.246.548.903.737.6510.59
GTS [33]2.655.226.833.096.348.453.597.299.83
PM-MemNet [61]2.655.297.013.036.298.423.467.5610.26
ST-DMN2.635.096.653.016.158.083.457.329.91
PEMS-BAYHA2.885.596.802.885.596.802.885.596.80
ARIMA [57]1.623.303.502.334.765.403.386.508.30
FC-LSTM [58]2.054.194.802.204.555.202.374.965.70
DCRNN [22]1.382.952.901.743.973.902.074.744.90
GW-Net [31]1.302.742.701.633.703.701.954.524.60
GMAN [59]1.352.902.871.653.823.741.924.494.52
CCRNN [60]1.382.902.901.743.873.902.074.654.87
GTS [33]1.392.952.881.784.063.982.245.175.35
PM-MemNet [61]1.342.822.811.653.763.711.954.494.54
ST-DMN1.302.742.731.623.733.671.894.464.51
Table 3. Average performance metrics for different model configurations.
Table 3. Average performance metrics for different model configurations.
Configurationsk(hops) α EmbeddingMAERMSEMAPE (%)
k(hops)10.15163.016.128.02
20.15163.006.118.08
30.15162.976.027.98
40.15162.986.108.03
α 30.1162.986.047.94
30.15162.976.027.98
30.3162.996.057.98
30.4162.996.077.99
30.5163.006.058.01
Embedding30.1583.006.127.99
30.15162.976.027.98
30.15323.016.107.99
30.15643.006.128.19
Table 4. Comparison of computation time of different models on the METR-LA dataset.
Table 4. Comparison of computation time of different models on the METR-LA dataset.
ModelsST-DMNPM-MemNetGTSDCRNNGW-Net
Time (s/epoch)82.44131.38727.60249.3153.68
Table 5. The results of ablation experiments on METR-LA and PEMS-BAY datasets.
Table 5. The results of ablation experiments on METR-LA and PEMS-BAY datasets.
DatasetModels15 min/Horizon 330 min/Horizon 660 min/Horizon 12
MAERMSEMAPEMAERMSEMAPEMAERMSEMAPE
METR-LAw/o Transformer Layer2.645.126.753.026.208.203.487.399.94
w/o DGL2.665.146.713.046.248.143.507.449.87
w/o Multi-Hop2.665.176.723.036.238.173.477.329.89
ST-DMN2.635.096.653.016.158.083.457.329.91
PEMS-BAYw/o Transformer Layer1.312.752.771.643.773.761.924.524.59
w/o DGL1.312.772.771.633.773.731.924.544.60
w/o Multi-Hop1.322.772.771.653.783.731.944.534.57
ST-DMN1.302.742.731.623.733.671.894.464.51
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chai, W.; Luo, Q.; Lin, Z.; Yan, J.; Zhou, J.; Zhou, T. Spatiotemporal Dynamic Multi-Hop Network for Traffic Flow Forecasting. Sustainability 2024, 16, 5860. https://doi.org/10.3390/su16145860

AMA Style

Chai W, Luo Q, Lin Z, Yan J, Zhou J, Zhou T. Spatiotemporal Dynamic Multi-Hop Network for Traffic Flow Forecasting. Sustainability. 2024; 16(14):5860. https://doi.org/10.3390/su16145860

Chicago/Turabian Style

Chai, Wenguang, Qingfeng Luo, Zhizhe Lin, Jingwen Yan, Jinglin Zhou, and Teng Zhou. 2024. "Spatiotemporal Dynamic Multi-Hop Network for Traffic Flow Forecasting" Sustainability 16, no. 14: 5860. https://doi.org/10.3390/su16145860

APA Style

Chai, W., Luo, Q., Lin, Z., Yan, J., Zhou, J., & Zhou, T. (2024). Spatiotemporal Dynamic Multi-Hop Network for Traffic Flow Forecasting. Sustainability, 16(14), 5860. https://doi.org/10.3390/su16145860

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop