Next Article in Journal
Information Entropy in Chimera States of Human Dynamics
Previous Article in Journal
Adaptive CoCoLasso for High-Dimensional Measurement Error Models
Previous Article in Special Issue
A Cross-Entropy Approach to the Domination Problem and Its Variants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions

National Center for Applied Mathematics in Chongqing, Chongqing Normal University, Chongqing 401331, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(2), 99; https://doi.org/10.3390/e27020099
Submission received: 25 November 2024 / Revised: 16 January 2025 / Accepted: 16 January 2025 / Published: 21 January 2025

Abstract

:
Spatio-temporal prediction is crucial in intelligent transportation systems (ITS) to enhance operational efficiency and safety. Although Transformer-based models have significantly advanced spatio-temporal prediction performance, recent research underscores the importance of learning dynamic spatio-temporal dependencies for these tasks. This paper introduces the Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L), a framework specifically designed to address the complex demands of spatio-temporal prediction. The DG3L model includes a memory-based graph learning module capable of generating dynamic graphs to accurately reflect ongoing changes in spatio-temporal dependencies. By integrating the strengths of Transformer and Graph Convolutional Recurrent Unit (GCRU) technologies within its Dual-Gated Graph Convolutional Recurrent Unit architecture, DG3L provides a mechanism for fusing Transformer features with contextual features from recurrent units. In practical applications, DG3L acts as an advanced representation learning module, delivering highly accurate context features for complex downstream tasks in ITS.

1. Introduction

Traffic flow prediction is a crucial component of modern urban management and Intelligent Transportation Systems (ITSs). It involves forecasting the movement of vehicles or pedestrians through a network of roads or routes over time. This process is shaped by both spatial and temporal dimensions; ‘spatial’ refers to various locations within the network such as intersections and road segments, while ‘temporal’ pertains to time intervals, often measured in minutes or hours. The data used for this prediction typically come in a time series format, where each point detailing traffic volume or density at a specific time and place provides essential insights. These data are meticulously collected through sensors, cameras, or GPS devices, offering a comprehensive, multidimensional view of urban traffic patterns. Such detailed data collection is fundamental as it sets the groundwork for effective traffic flow prediction [1], which is a pivotal task for enhancing road management, travel planning, and vehicle navigation. Accurate traffic prediction can improve system efficiency, reduce congestion, save energy, and support sustainable urban development [2]. However, traffic flow data often exhibit complex spatio-temporal heterogeneity, influenced by factors like weather, accidents, and road structure changes. This complexity makes traffic prediction a challenging task. Traditional statistical models often fail to capture the complex spatio-temporal dependencies in traffic flow, resulting in less accurate predictions.
In recent years, with the continuous development of big data technology and deep learning models, data-driven spatio-temporal forecasting methods have made significant progress. Initially, most models evolved primarily in two directions: methods based on Recurrent Neural Networks (RNNs) and those based on Convolutional Neural Networks (CNNs). Although RNN-based methods capture spatio-temporal dependencies by filtering inputs through Graph Convolutional Networks (GCNs) and passing the hidden states to recurrent units, they exhibit low efficiency when handling long sequences and are prone to gradient explosion when combined with GCN [3]. CNN-based methods, on the other hand, improve computational efficiency by combining graph convolution with standard one-dimensional convolution [2]. However, they still require stacking multiple layers or using global pooling to expand the model’s receptive field. As these methods have developed, research in the spatio-temporal forecasting field has gradually shifted toward Transformer-based models [4], which demonstrate significant advantages in capturing long-term dependencies. Unlike RNN models, which are more suitable for short-sequence forecasting, Transformer models utilize the Self-Attention Mechanism to overcome the limitations of RNN in long-term dependency modeling, thereby showing exceptional performance in long-term forecasting tasks. This self-attention mechanism enables Transformer models to capture complex, long-term spatio-temporal dependencies more efficiently, bringing a new breakthrough to spatio-temporal forecasting [5].
Most traditional spatio-temporal forecasting models are designed based on static graphs. However, in many real-world applications, the graph evolves over time, known as a dynamic graph. As nodes, attributes, and edges change over time, the complexity and challenges of spatio-temporal forecasting tasks also increase. To address this, researchers have proposed several methods to handle dynamic graphs and integrate them with spatio-temporal forecasting. For example, a time graph embedding method based on random walks automatically samples subgraphs with connections within a specific time range, effectively capturing spatio-temporal dependencies [6]. Another method, DyANE [7], transforms the time graph into a static graph representation (hyper-adjacency representation), preserving time path information for subsequent embedding learning. Additionally, point process models have been applied to time graph representation learning, capturing node changes in dynamic graphs by modeling discrete time event sequences. DyRep [8] uses point processes to generate dynamic node embeddings and estimates the likelihood of edge occurrence at specific time stamps based on these embeddings. These dynamic graph learning methods demonstrate how to effectively capture and represent spatio-temporal dependencies in time-varying graphs, advancing spatio-temporal forecasting models to better adapt to more complex dynamic environments, thus improving both prediction accuracy and efficiency. Recent advancements include GW-Net [9], AdapGL [10], and STGM [11], which introduced an adaptive graph learning method using two learnable embedding matrices to dynamically generate graph structures based on spatio-temporal data. In the area of temporal dynamic graph generation, as shown in Figure 1, despite the innovations and certain achievements of DGCRN [12], challenges still arise from the inherent contradictions between the dynamic spatial dependencies of spatio-temporal data and the increased difficulty in model convergence due to the deepening complexity of graph learning models.
Although the aforementioned learning models generally perform well in spatio-temporal forecasting, they need to resolve the contradiction between representing complex spatio-temporal dependencies and achieving convergence stability in graph learning. Additionally, through multiple experiments, we believe that leveraging the short-term forecasting strengths of RNN structures combined with the long-term forecasting capabilities of Transformers is key to enhancing spatio-temporal prediction accuracy. To address these challenges, this paper introduces a new framework of Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning. Our framework aims to learn the dynamic graph under low-rank conditions from the complex representations derived from the Transformer structure, while an improved GRU structure is used to integrate these complex representations to enhance contextual features, thereby capturing complex dependencies within spatio-temporal data. The specific contributions are as follows:
  • We provide a framework that effectively combines the Transformer and GRU structures, while also outputting the spatio-temporal dependencies’ feature matrix and spatio-temporal features as results.
  • Novel Dual-Gated Graph Convolutional Recurrent Unit (DG-GCRU): We design a new Graph Convolutional Recurrent Unit that integrates long-term information, short-term information, and adaptive embeddings for gated selection.
  • New Memory Mechanism for Dynamic Graph Generation: We introduce a memory mechanism to generate a learnable dynamic graph adjacency matrix to optimize the representation learning of the DG-GCRU.
  • We conduct multi-step and single-step traffic flow forecasting experiments on six real-world public datasets. The results demonstrate that our model achieves excellent performance on these datasets.

2. Related Work

Traffic Forecasting. As a typical task in multivariate time series forecasting, traffic prediction has long attracted extensive attention from researchers [1]. Early studies primarily relied on statistical models, such as the Autoregressive (AR) model, Vector Autoregression (VAR) [13], and the Autoregressive Integrated Moving Average (ARIMA) model [14]. These methods, which model historical data based on linear assumptions, achieved some success in the early stages. However, as the complexity of traffic data increased, these traditional methods struggled to effectively capture the nonlinear characteristics and spatio-temporal dependencies in traffic flow. To address this limitation, deep learning models have gradually gained attention, particularly long short-term memory (LSTM) [15] and Gated Recurrent Units (GRUs) [16], which effectively capture both short- and long-term dependencies through gating mechanisms and are widely used in traffic flow forecasting. Subsequently, the emergence of temporal convolutional networks, such as Graph WaveNet, further enhanced the ability to handle long-sequence dependencies [9]. Similarly, the ASTGCN [17] introduces a spatial–temporal attention mechanism to dynamically capture both spatial and temporal correlations of traffic data, significantly enhancing the forecasting accuracy. Following this, MAEGCLSTM [18] adds a Memory Attention module that captures global spatial dependencies and integrates a novel simplified GCLSTM with an encoder–decoder architecture, considerably improving multi-step traffic forecasting. In recent years, transformer models, owing to their superior performance in long-sequence modeling via the Self-Attention Mechanism, have driven advancements in traffic prediction technology [4]. Transformer models can flexibly model long-distance dependencies while adapting to complex spatio-temporal features. Consequently, various Transformer-based models have emerged to address the more complex spatio-temporal dependencies in traffic flow forecasting.
Dynamic Graph Learning. In traffic flow forecasting, early methods often relied on the natural topology of road networks or static graph structures based on certain predefined metrics [2,19]. However, these static graph structures struggle to adapt to the dynamic changes in traffic flow over time. To address this issue, GW-Net [9] introduced an adaptive graph learning method, using two learnable embedding matrices to dynamically generate the graph structure. This enables the model to adjust node relationships based on real-time traffic data, thus better capturing spatio-temporal dependencies. The adaptive graph concept in GW-Net has provided significant inspiration for subsequent studies. Building on this, AGCRN [20] introduced node-specific convolution filters, allowing each node’s characteristics to be better represented during the convolution process. CCRNN [21] further enhanced adaptability by learning multiple adaptive graphs through multi-layer graph convolution, capturing spatio-temporal dependencies at different levels. Additionally, StemGNN [22] combined the Self-Attention Mechanism to dynamically learn latent graph structures from input data, offering more flexible spatio-temporal dependency modeling capabilities. Expanding on dynamic graph structures, the DST-Trans [23] model utilizes a spatio-temporal transformer combined with a gated temporal convolutional network and Graph Convolutional Networks to effectively capture the dynamic spatio-temporal complexity of traffic flow, further enhancing the adaptability and predictive performance of traffic forecasting systems. Recent advancements include MTGNN [24], which utilizes mix-hop propagation layers and dilated inception layers to effectively capture both spatial and temporal dependencies in multivariate time series data. Similarly, D2STGNN [25] introduces a decoupled spatial–temporal framework that models diffusion and inherent signals within traffic data separately, enhancing the prediction of traffic flow dynamics. The introduction of DGCRN [12] and its hyper-networks for dynamic adjacency matrix generation marks a significant advancement in this field, further enhancing the dynamic nature and predictive performance of the models. Similarly, the integration of reinforcement learning in the Dynamic Graph Convolutional Network by Peng et al. [26] for long-term traffic flow prediction represents another innovative step towards addressing data deficiencies and improving spatio-temporal predictions. These methods showcase the substantial potential of dynamic graph learning in traffic flow forecasting, enabling models to flexibly adjust graph structures to accommodate continuously changing traffic data. Compared to early methods that rely on static topologies, adaptive graph structures significantly improve models’ adaptability to complex traffic environments, laying a solid foundation for high-precision traffic forecasting. These advancements also provide critical insights for future dynamic graph modeling, such as STGM [11], emphasizing that dynamically updating node relationships is key to enhancing spatio-temporal forecasting model performance.
Gated Recurrent Unit. In traffic forecasting, the Gated Recurrent Unit (GRU) is an efficient Recurrent Neural Network that regulates information flow through update and reset gates, effectively capturing both short-term and long-term temporal dependencies [27]. GRU is computationally efficient with a simple structure, making it widely used for tasks like traffic flow forecasting and travel time estimation. However, GRU primarily focuses on temporal dependencies and has limited ability to model spatial relationships within traffic networks, making it inadequate for handling dynamic spatial variations in complex traffic networks [28]. To improve spatio-temporal modeling, researchers have explored combining GRUs with Graph Convolutional Networks (GCNs). DCRNN [19] is a classic model in this area, which combines diffusion convolution with GRUs to capture diffusion dynamics in non-Euclidean spaces, better modeling the spatio-temporal propagation relationships among neighboring nodes. The GRU component handles temporal dependencies, significantly improving prediction performance in complex traffic networks. Additionally, Saravanan et al. proposed an improved deep hybrid model (CNN-rGRU) [29], combining convolutional layers with GRU for residual learning, which adaptively learns the spatio-temporal features of traffic congestion. Zhao et al. introduced a Temporal Graph Convolutional Network (T-GCN) [30], where GCN is used to learn complex spatial topologies and GRU captures the dynamic changes in traffic flow, making it suitable for city-road-network-based traffic forecasting. Building on this, MegaCRN [31] introduced a meta-learning framework that dynamically generates graph convolution weight matrices and combines them with recurrent networks to handle temporal dependencies, enhancing the model’s ability to personalize its predictions. MegaCRN adaptively adjusts graph convolution weights to reflect the characteristics of different traffic nodes, improving its ability to capture spatial dynamics.
Transformer. In traffic flow forecasting, the Transformer model has gradually become an important tool for handling spatio-temporal data due to its strong parallel processing and sequence modeling capabilities [4]. Traffic data often exhibit significant dynamism and complex spatial dependencies, which traditional statistical models, such as ARIMA and VAR, struggle to capture effectively. In contrast, the Transformer model leverages the self-attention mechanism to globally model dependencies in long time sequences while integrating multiple spatio-temporal dimensions, making it applicable to tasks like traffic flow, travel time prediction, and urban congestion pattern analysis. For example, Traffic Transformer introduces seven different temporal encoding mechanisms to handle continuity and periodicity in time series data [5]. These encodings help the model more accurately capture daily and weekly repeating patterns and, combined with Graph Convolutional Networks (GCN), model spatial dependencies within traffic networks. The attention mechanism can also identify the importance of data features, helping the model to understand the impact of sudden events, such as traffic accidents or holidays, on traffic flow and thereby improve prediction accuracy. GMAN [32] incorporates spatio-temporal embeddings for both input and output sequences and lies in the introduction of the Graph Multi-Attention Network (GMAN), which utilizes an encoder–decoder architecture with multiple spatio-temporal attention blocks to effectively model the impact of spatio-temporal factors on traffic conditions. STAEformer [33] offers a new approach based on spatio-temporal adaptive embedding. It uses an adaptive embedding layer to dynamically adjust embedding weights according to the spatio-temporal characteristics of traffic data, better capturing the dynamic relationships among nodes in the traffic network [31]. Unlike Traffic Transformer, which relies on fixed temporal encoding, STAEformer employs an adaptive mechanism to learn spatio-temporal pattern changes in real time within the traffic network. This dynamic adjustment allows STAEformer to perform more stably when processing complex traffic data, especially in large-scale long-term forecasting tasks, by reducing error accumulation in multi-step forecasting and improving prediction outcomes.

3. Problem Definition

The primary objective of traffic flow prediction is to forecast the flow within a traffic system over future time periods based on historical observation data. To ensure broad applicability, we define the problem as a multi-step prediction task as follows:
X t ( α 1 ) , , X t ; G θ f ( · ) X t + 1 , , X t + β
Let X i R N × C , where N represents the number of spatial units (e.g., nodes, regions, roads), and C denotes the number of information channels. Additionally, G which is provided before the prediction task begins, represents the connectivity between these N spatial nodes. Given the observations from the previous α time steps X t ( α 1 ) , , X t , our goal is to infer the traffic flow for the next β time steps X t + 1 , , X t + β by training a predictive model f with parameters θ .

4. Methodology

Our DG3L framework, as illustrated in Figure 2, primarily consists of a graph learning module and a Dual-Gated Graph Convolutional Recurrent Unit (DG-GCRU). The graph learning module is employed to extract spatio-temporal dependency information, while the Dual-Gated GCRU is utilized for sequence prediction.

4.1. Transformer-Based Memory Graph Bank for Graph Learning

Our Graph Learning module is composed of two parts: the Spatio-Temporal Transformer and the Memory Graph Bank.
Embedding Layer. In common spatio-temporal prediction tasks, spatio-temporal embeddings [32] related to time and space are crucial. Leveraging the adaptive embedding module from the STAEformer [33], known for its effectiveness, we incorporate this module along with time embedding across all our embedding layers. This integration is crucial for preserving the integrity of the raw data. We employ a linear layer to derive the feature embedding E f R T × N × d f , a spatio-temporal adaptive embedding E a R T × N × d a , and the periodicity embedding E p t R T × d p :
E f = Linear ( X t T + 1 : t )
Γ 0 = E f E p E a
Here, d f represents the dimension of the feature embedding and Linear ( · ) denotes a linear transformation layer that operates on the last dimension of the input, mapping R d i n R d f . By concatenating these embeddings, we obtain a hidden spatio-temporal representation Γ 0 R T × N × d h , where the hidden dimension d h equals d f + d p + d a .
Temporal Transformer. In the Temporal Transformer, we transpose the spatio-temporal representation Γ to Γ R N × T × d h . The transformation is governed by:
SelfAttention T ( Γ ) = A V ( Γ ) = Softmax Q ( Γ ) K ( Γ ) d h V ( Γ )
Γ l * = LayerNorm ( SelfAttention T ( Γ l ) + Γ l ) Γ l + 1 = LayerNorm ( FeedForward ( Γ l * ) + Γ l * )
where Q ( Γ ) , K ( Γ ) , V ( Γ ) are linear layers, and the matrix multiplication Q ( Γ ) K ( Γ ) acts on the last dimension, enabling A R N × T × T to capture the temporal relations across different spatial nodes.
Spatio Transformer. This component mirrors the structure of the Temporal Transformer, with the input spatio-temporal representation Γ transposed to Γ R T × N × d h . Consequently, A R T × N × N captures the spatial relations across different time steps.
Spatio-Temporal Transformer. We achieve our framework by sequentially connecting the Temporal and Spatio Transformers:
F = SpatioTransformer ( TemporalTransformer ( Γ 0 ) )
Memory Graph Bank Section. Inspired by the innovative concept of the Meta-Node Bank (MegaCRNN [31]), which effectively generates a context vector, we propose utilizing time-series data within memory network structures to construct dynamic time-series graphs. To this end, we have developed a Dynamic Graph Memory Bank, denoted as B R ϕ × d , where ϕ represents the number of memory nodes and d denotes the dimension of each memory node. The primary functionalities of this memory bank are outlined as follows:
Q 1 = F · W 1 + b 1 Q 2 = F · W 2 + b 2 E 1 = j = 1 ϕ exp ( Q 1 · B [ j ] ) k = 1 ϕ exp ( Q 1 · B [ k ] ) B [ j ] E 2 = j = 1 ϕ exp ( Q 2 · B [ j ] ) k = 1 ϕ exp ( Q 2 · B [ k ] ) B [ j ]
Here, Q 1 and Q 2 are query vectors derived from the hidden spatio-temporal representation F through two distinct linear layers. Both E 1 and E 2 , which reside in R T × N × ϕ , serve as memory vectors extracted from the memory bank B via two sophisticated attention mechanisms. These vectors facilitate the formation of a dynamic graph, D y n A d j R T × N × N defined by:
D y n A d j = softmax ( relu ( E 1 · E 2 ) )
This dynamic graph is generated from the memory vectors through a softmax function, which reflects an entropy-driven approach to understanding graph structures, capturing the spatio-temporal relationships within the dataset.
Contrastive Loss and Consistency Loss. To further refine our model’s ability to discern and differentiate between various temporal patterns in the data, we have adopted the InfoNCE loss [34], inspired by recent advancements in graph generation methodologies [35]. The InfoNCE loss is particularly effective in enhancing the discriminative capabilities of models by maximizing mutual information between closely related samples while minimizing it among less related ones. The loss is defined as follows:
Loss InfoNCE = t T 0 log exp ( Q 1 , t · B [ p ] / τ ) n = 1 N exp ( Q 1 , t · B [ n ] / τ ) log exp ( Q 2 , t · B [ p ] / τ ) n = 1 N exp ( Q 2 , t · B [ n ] / τ )
Here, Q 1 , t and Q 2 , t are the query vectors derived from the hidden spatio-temporal representations at time step t. B [ p ] represents the positive memory item within the memory bank, which is semantically similar to the query vectors at the corresponding time steps and is selected based on the highest similarity scores to enhance learning from relevant contrasts. Conversely, B [ n ] denotes the set of negative samples, which are chosen based on lower similarity scores to provide diverse comparisons, thereby improving the discriminative training of the model. τ is the temperature parameter that scales the dot product within the softmax function in the InfoNCE loss, controlling the sharpness of the distribution. T 0 specifically refers to the set of the first time steps across all output sequences, crucial for establishing the initial context for the model’s learning process. The InfoNCE loss discriminates node information in the graph by maximizing mutual information between similar samples and minimizing it among dissimilar ones. This concept extends Cross-Entropy Loss to an unsupervised context, focusing on distinguishing complex patterns rather than predicting exact categories. To further understand this idea, consider the following reformulation of the Cross-Entropy Loss:
Loss Cross - Entropy = log exp ( z true ) j = 1 K exp ( z j ) = log exp ( sim ( ϕ i , ϕ i + ) ) j = 1 K exp ( sim ( ϕ i , ϕ j ) )
In the unsupervised task, sim ( ϕ i , ϕ i + ) is denoted as z true , representing the logit of the true class. The parameter K refers to the number of classes in the Cross-Entropy context, which corresponds to the number of samples in a batch for InfoNCE, indicating the range of summation. Additionally, sim ( ϕ i , ϕ j ) quantifies the dissimilarity between the target sample ϕ i and a negative sample ϕ j , serving a role similar to the logits of non-target classes in Cross-Entropy. InfoNCE loss includes a temperature parameter τ to control the sharpness of the similarity distribution, thereby influencing the model’s sensitivity to differences in similarity.
Similar to the approach used in [36], we also apply a consistency loss to ensure node stability across a temporal graph:
Loss consistency = t T 0 | | Q 1 , t B [ p ] | | 2 + | | Q 2 , t B [ p ] | | 2
This loss minimizes the variance in embeddings over time, promoting temporal coherence and ensuring that B [ p ] serves as a stable reference point within the memory bank. To effectively train our model, we combine multiple loss components into a unified training objective:
Loss train = Loss prediction + α · Loss InfoNCE + β · Loss consistency
where Loss prediction is the primary function for accuracy, and α and β are coefficients that regulate the impact of the InfoNCE and consistency losses on the training process.
In these regularization terms, our decision to compute the contrastive and consistency losses only at the first time step T 0 of the output sequence is based on our understanding of recurrent networks. This approach allows the network to establish a strong initial context, which is crucial for accurately modeling the subsequent dynamics of the sequence.

4.2. Dual-Gated Graph Convolutional Recurrent Unit

To enhance the model’s performance, we directly use the output H 0 of a single layer of the Spatio Transformer as the input of decoder, rather than concatenating it with the output of the GCRU-Encoder. This decision is based on our belief that when the input time series is lengthy, the Transformer’s self-attention mechanism can effectively process all time steps in parallel. Referring to (4) and (5), the initial transformation is given by
H 0 = SpatioTransformer * ( Γ 0 )
Here, Γ 0 is a reshaped version of Γ 0 , where the temporal dimension T and the feature dimension d h are merged. This transformation is mathematically represented as Γ 0 R N × ( T d h ) . The SpatioTransformer * module is a SpatioTransformer with a fully connected layer (FC) preceding it, refining feature extraction. This transformation can be formally represented as a function f : R N × ( T d h ) R N × d h .
Since the Spatio-Temporal Transformer has already learned spatial–temporal representations for better capturing spatio-temporal relationships, we aim for the GCRU module to leverage these representations fully. This allows the model to harness the strengths of the Transformer in long-sequence prediction tasks and the GCRU in short-sequence prediction tasks. Therefore, we enhance the GCRU module and introduce the Dual-Gated Graph Convolutional Recurrent Unit (DG-GCRU).
GCN (Chebyshev GCN). Before detailing the DG-GCRU, it is crucial to understand the underlying graph convolution operation. The graph convolution function [37], denoted as G C N , is defined as follows:
G C N ( V , A d j ) = σ ( k = 0 K A d j k V W k )
Here, V denotes the input features, and A d j represents a snapshot of the dynamic adjacency matrix, known as D y n A d j , at a specific time point. This adjacency matrix, A d j R N × N , encapsulates the spatial relationships and dependencies among features at that time instant. The graph convolution operation employs Chebyshev polynomials up to order K to approximate the parameters W k of the kernel. An activation function σ is subsequently applied to the output feature matrix, facilitating the effective processing of connected features through the graph structure.
Fusion Gate. In traditional GRU modules, the interaction between the candidate states and intermediate states is essential for effective sequence modeling. To improve feature integration from the Transformer, we introduce a novel gating mechanism, termed Fusion Gate. This gate effectively balances the contextual information from previous time steps in the GRU with the current time step information processed by the Transformer’s attention mechanisms, as described below:
r t = sigmoid ( W r · [ X t , F t , H t 1 ] + b r ) M t = r t H t 1 + ( 1 r t ) F t
Here, r t is the fusion gates, W r is the weight matrix applied to the concatenated inputs and the previous hidden state to compute the reset gate, X t is the input at time step t, F t is the feature input, H t 1 is the hidden state from the previous time step, and b r is the bias term for the fusion gate. M t is the new intermediate state, ⊙ denotes element-wise multiplication and ( 1 r t ) is the complement of the fusion gate.
GCRU (Graph Convolutional Recurrent Unit). Although the Transformer can output the feature input F t in parallel with the input X t , we naturally believe that the representation learned by F t is not entirely accurate, and it should also be controlled by a reset gate. Therefore, we design
[ z 1 t , z 2 t , q t ] = sigmoid ( G C N ( [ X t , F t , H t 1 ] , D y n A d j t ) + b z )
where z 1 t and z 2 t are reset gates, q t is the update gate, X t is the input at time step t, and b z is the bias term for the update gate. D y n A d j t R N × N , which is a dynamic adjacency matrix at time step t. The sigmoid function ensures that the values of z 1 t , z 2 t , and q t are between 0 and 1, controlling the flow of information.
C t = tanh ( G C N ( [ X t , z 1 t H t 1 , z 2 t F t ] , D y n A d j t ) + b C )
C t is the candidate state, and b C is the bias term for the candidate state. The integration of the candidate state C t with the update mechanism through q t allows for a dynamic adjustment of the hidden state, reflecting both new inputs and historical data.
H t = q t M t + ( 1 q t ) C t Y ^ = FC ( H )
Here, H t represents the hidden state at time step t, and ( 1 q t ) acts as the complement of the update gate. This equation integrates the intermediate state M t and the candidate state C t based on the update gate q t , determining the final hidden state. The collection of hidden states across all time steps is denoted by H = [ H t ( α 1 ) , , H t ] , where H is a tensor in R T × N × d h , encapsulating the results from all time steps. Y ^ denotes the output predictions. The function FC , a fully connected layer, performs a linear transformation mapping R d h R , projecting high-dimensional hidden states to scalar outputs for sequence modeling predictions.

5. Experiment

In the experimental section of this study, all model data were obtained from the BasicTS library [38]. BasicTS provides a unified framework and standardized data processing, allowing for more objective and accurate comparisons between different models in multivariate time series forecasting tasks. In our experimental design, we relied on BasicTS’s implementations of each model to ensure that all models were evaluated under the same conditions, guaranteeing fairness and consistency in the results.

5.1. Experimental Setup

Datasets. Our method was comprehensively validated on six benchmark datasets for traffic flow forecasting: PEMS03, PEMS04, PEMS07, PEMS08, METR-LA, and PEMS-BAY. The first two datasets were proposed by DCRNN [19] and contain traffic speed data collected from 207 sensors in Los Angeles and 325 sensors in the Bay Area, respectively. The latter four datasets were introduced by STSGCN [39]. All datasets are sampled at 5-min intervals, resulting in 12 frames per hour. Detailed information on the datasets is provided in Table 1.
Settings. Our model implementation utilizes PyTorch and operates on a Linux server with a GeForce RTX 4080 GPU. The dataset partitioning is as follows: METR-LA and PEMS-BAY datasets are split into training, validation, and test sets at a 7:1:2 ratio, while the PEMS03, PEMS04, PEMS07, and PEMS08 datasets are divided at a 6:2:2 ratio. The dimension d h , comprising components d f , d p , and d a , maintains a ratio of 2:1:1. For the PEMS03, PEMS04, PEMS07, and PEMS08 datasets, d h is configured to 32, whereas for PEMS-BAY and METR-LA, d h is set to 28. Both spatial and temporal Transformers have 1 layer, with an attention head count of 4. Additionally, the number of memory nodes for all datasets is set to 20, and the memory dimension is configured to 32. The input and prediction durations are each set to 1 h. During training, the batch size is generally 64, though it is reduced to 16 for the PEMS07 dataset due to the higher node count. We utilize the Adam optimizer with an initial learning rate of 0.008, which gradually decreases during training. Training is conducted over 100 epochs, with early stopping applied if the validation error stabilizes over 50 consecutive steps to prevent overfitting. To ensure reliability, each experimental result is averaged across three independent training runs.
Metrics. We use two widely adopted metrics for the traffic forecasting task, namely MAE and RMSE. We evaluate the average performance of all 12 forecasting time steps across the PEMS03, PEMS04, PEMS07, PEMS08, METR-LA, PEMS-BAY datasets. For the METR-LA and PEMS-BAY datasets, we compare the performance at the 3rd, 6th, and 12th time steps (15, 30, and 60 min) to provide a more detailed evaluation.

5.2. Performance Evaluation

Table 2 presents a comprehensive comparison of the performance of our model, DG3L, against several state-of-the-art baseline models across six traffic flow forecasting datasets: PEMS03, PEMS04, PEMS07, PEMS08, METR-LA, and PEMS-BAY. The evaluation metrics include MAE (Mean Absolute Error), RMSE (Root Mean Square Error), and MAPE (Mean Absolute Percentage Error). The results illustrate that our method consistently achieve superior performance on most datasets and metrics, particularly excelling in minimizing MAE and RMSE values. In the more challenging PEMS03 dataset, DG3L outperforms other models, with an MAE of 14.51 and an RMSE of 25.69, showcasing its effectiveness even in larger and more complex datasets. Similarly, in the PEMS04 dataset, DG3L continues to lead, with an MAE of 18.22 and an RMSE of 29.82. It is noteworthy that STWave and STAEformer also show strong performances across various horizons, indicating their potential in handling multi-step predictions. The PEMS07 dataset, known for its high number of nodes, sees our DG3L model maintaining a competitive edge with an MAE of 19.59 and an RMSE of 33.07. This dataset highlights the strengths of the D2STGNN and STAEformer models, which achieve the best and second-best performances, respectively, suggesting their suitability for complex network structures. Finally, on the PEMS08 dataset, DG3L achieves the best MAE of 13.72 and an excellent RMSE of 23.09, closely followed by STAEformer with an MAE of 13.43 and an RMSE of 23.31. These results underscore the adaptability and efficiency of our DG3L model across different traffic conditions and forecasting horizons. On the METR-LA dataset, our model, DG3L, demonstrates robust performance with an MAE of 2.90 and an RMSE of 5.98, which are competitive figures close to the best results shown by D2STGNN with an MAE of 2.87 and an RMSE of 5.90. This indicates a strong predictive accuracy in urban traffic contexts. For the PEMS-BAY dataset, DG3L also shows excellent results, achieving the lowest errors with an MAE of 1.54 and an RMSE of 3.57, closely followed by D2STGNN, which recorded an MAE of 1.52 and an RMSE of 3.53. Overall, our DG3L model not only consistently achieves top-tier performance but also demonstrates significant improvements over other models like DCRNN and STGCN, which struggle with larger datasets. The comparative analysis with models such as GWNet, STNorm, and particularly D2STGNN and STAEformer, highlights the nuanced capabilities of these models in specific scenarios, reinforcing the importance of tailored approaches in traffic flow forecasting.

5.3. Ablation Study

To evaluate the effectiveness of each component in our model, we conducted an ablation study using three variants of the model, as shown in Table 3.
  • w/o D u a l - G a t e . This variant removes the Fusion Gate and the reset gate z 2 of the F vector from the Spatio-Temporal Transformer.
  • w/o M e m o r y G r a p h - B a n k . This variant removes the memory-based dynamic graph generation mechanism.
  • w/o S p a t i o T r a n s f o r m e r - E n c o d e r . This variant replaces the Spatio Transformer (one layer) with a GCRU as the encoder.

5.4. Hyperparameter Sensitivity

Our studies indicate that while the DG3L model presents some susceptibility to variations in hyperparameters like d h and memory nodes, these do not radically alter overall model performance, except beyond certain thresholds. These insights stress the need for nuanced hyperparameter tuning to optimize performance without overscaling the model’s complexity. We conducted an investigation into the hyperparameter sensitivity of the DG3L model, as depicted in Figure 3. Our focus was on understanding the effects of the dimension of hidden units ( d h ) and the number of memory nodes.
Through this study, we discovered that setting the hidden dimension d h to approximately 30 achieves an effective equilibrium between processing capacity and model complexity. We directly define d h by the proportional relationship d f : d p : d a = 2:1:1. This edit retains the essential information and smoothly transitions into the dimension ratio while properly referencing the related equation as requested. The impact of the number of memory nodes varied notably across the PEMS03, PEMS08, and METR-LA datasets. A moderate increase typically improves the model performance; however, exceeding a certain threshold leads to performance degradation due to escalating complexity. Each dataset demonstrated unique responses, identifying distinct optimal memory configurations beyond which further increases diminished returns on performance. These findings underscore the importance of meticulously adjusting hyperparameters like the number of memory nodes in accordance with the specific characteristics and computational constraints of the dataset. Such customization is vital to harnessing the full potential of the DG3L model in diverse data environments.

5.5. Efficiency Study

In our efficiency study, we evaluated the DG3L model’s performance across four datasets, PEMS03, PEMS04, PEMS08, and METR-LA, comparing it with other state-of-the-art models. According to the results illustrated in Figure 4, for the PEMS03 dataset, DG3L, with 332,556 parameters, outperformed MegaCRN in terms of overall MAE and exhibited strong parameter efficiency. On the PEMS04 dataset, it maintained competitive accuracy with 311,285 parameters, showing a lower MAE than MegaCRN and demonstrating considerable efficiency. The PEMS08 dataset highlighted DG3L’s superior accuracy with 306,957 parameters, once again surpassing MegaCRN in overall performance. In the METR-LA dataset, while DG3L achieved excellent parameter efficiency with 286,733 parameters, D2STGNN managed a marginally better MAE. Throughout these analyses, DG3L consistently exhibits a favorable balance of parameter efficiency and competitive accuracy across diverse settings without compromising its performance.

6. Conclusions

In this paper, we introduced the DG3L framework, a sophisticated approach designed for predicting traffic flow within intelligent transportation systems. The DG3L framework incorporates a memory-based dynamic graph learning module coupled with a dual-gated mechanism. This configuration enables the effective capture of spatio-temporal characteristics of traffic flows, facilitating the generation of dynamic dependency matrices and spatio-temporal feature vectors. These elements are crucial for accurately modeling and predicting traffic flow patterns. Our experimental evaluations, as detailed in the performance comparison tables, affirm that DG3L consistently outperforms traditional methods, particularly in monitoring traffic at critical nodes and forecasting traffic flows. The framework’s flexibility in constructing dynamic spatio-temporal correlations is evident, with its memory mechanism playing a pivotal role in capturing long-term dependencies. This enhances both the prediction accuracy and the utility of contextual features, rendering DG3L highly suitable for real-time traffic forecasting and alert scenarios in intelligent transportation systems. However, it is important to acknowledge certain limitations. The primary datasets used in testing DG3L were existing spatio-temporal datasets, which may not encompass the full complexity of real-world spatio-temporal networks or unstructured data. To overcome these limitations and ensure broader applicability, further validation in more diverse and complex spatio-temporal scenarios is essential. As the field of dynamic graph learning progresses, there is potential to extend DG3L’s functionality by integrating newer datasets and models, thereby enhancing its adaptability to a wider array of spatio-temporal data challenges. We encourage researchers and practitioners interested in advancing traffic flow prediction models within intelligent transportation systems to utilize the DG3L framework by accessing the code https://github.com/cjwyx/DG3LFrame (accessed on 24 November 2024).

Author Contributions

Conceptualization, Y.W. and Z.Z.; Data curation, Z.Z.; Formal analysis, Z.Z.; Funding acquisition, J.P.; Investigation, Y.W. and Z.Z.; Methodology, Y.W.; Project administration, J.P.; Resources, Z.Z.; Software, Z.Z.; Supervision, Y.W. and J.P.; Validation, Z.Z., S.P. and H.Z.; Visualization, Z.Z.; Writing—original draft, Y.W.; Writing—review and editing, Y.W., Z.Z. and J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Natural Science Foundation of Chongqing (Grant Nos.CSTB2022NSCQ-LZX0040, Grant Nos.CSTB2023NSCQ-LZX0012, Grant Nos.CSTB2023NSCQ-LZX0160).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank the members of the research team for their hard work and professional contributions, which provided a solid foundation for the successful implementation of the study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiang, R.; Yin, D.; Wang, Z.; Wang, Y.; Deng, J.; Liu, H.; Cai, Z.; Deng, J.; Song, X.; Shibasaki, R. Dl-traff: Survey and benchmark of deep learning models for urban traffic prediction. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Queensland, Australia, 1–5 November 2021; pp. 4515–4525. [Google Scholar]
  2. Yu, B.; Yin, H.; Zhu, Z. Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI-18. International Joint Conferences on Artificial Intelligence Organization, Stockholm, Sweden, 13–19 July 2018; pp. 3634–3640. [Google Scholar] [CrossRef]
  3. Seo, Y.; Defferrard, M.; Vandergheynst, P.; Bresson, X. Structured sequence modeling with graph convolutional recurrent networks. In Proceedings of the Neural Information Processing: 25th International Conference, ICONIP 2018, Siem Reap, Cambodia, 13–16 December 2018; Proceedings, Part I 25. Springer: Berlin/Heidelberg, Germany, 2018; pp. 362–373. [Google Scholar]
  4. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  5. Cai, L.; Janowicz, K.; Mai, G.; Yan, B.; Zhu, R. Traffic transformer: Capturing the continuity and periodicity of time series for traffic forecasting. Trans. GIS 2020, 24, 736–755. [Google Scholar] [CrossRef]
  6. Wang, Y.; Chang, Y.Y.; Liu, Y.; Leskovec, J.; Li, P. Inductive representation learning in temporal networks via causal anonymous walks. arXiv 2021, arXiv:2101.05974. [Google Scholar]
  7. Sato, K.; Oka, M.; Barrat, A.; Cattuto, C. Dyane: Dynamics-aware node embedding for temporal networks. arXiv 2019, arXiv:1909.05976. [Google Scholar]
  8. Trivedi, R.; Farajtabar, M.; Biswal, P.; Zha, H. Dyrep: Learning representations over dynamic graphs. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  9. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph WaveNet for Deep Spatial–temporal Graph Modeling. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19. International Joint Conferences on Artificial Intelligence Organization, Macao, China, 10–16 August 2019; pp. 1907–1913. [Google Scholar] [CrossRef]
  10. Zhang, W.; Zhu, F.; Lv, Y.; Tan, C.; Liu, W.; Zhang, X.; Wang, F.Y. AdapGL: An Adaptive Graph Learning Algorithm for Traffic Prediction Based on Spatiotemporal Neural Networks. Transp. Res. Part C Emerg. Technol. 2022, 139, 103659. [Google Scholar] [CrossRef]
  11. Lablack, M.; Shen, Y. Spatio-temporal graph mixformer for traffic forecasting. Expert Syst. Appl. 2023, 228, 120281. [Google Scholar] [CrossRef]
  12. Li, F.; Feng, J.; Yan, H.; Jin, G.; Yang, F.; Sun, F.; Jin, D.; Li, Y. Dynamic Graph Convolutional Recurrent Network for Traffic Prediction: Benchmark and Solution. Acm Trans. Knowl. Discov. Data Tkdd 2023, 17, 9. [Google Scholar] [CrossRef]
  13. Hamilton, J.D.; Susmel, R. Autoregressive conditional heteroskedasticity and changes in regime. J. Econom. 1994, 64, 307–333. [Google Scholar] [CrossRef]
  14. Pan, B.; Demiryurek, U.; Shahabi, C. Utilizing real-world transportation data for accurate traffic prediction. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining, Brussels, Belgium, 10–13 December 2012; pp. 595–604. [Google Scholar]
  15. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. In Neural Computation; MIT-Press: Cambridge, MA, USA, 1997; Volume 9, pp. 1735–1780. [Google Scholar] [CrossRef]
  16. Cho, K.; van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches. In Proceedings of the SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, Doha, Qatar, 25 October 2014; Wu, D., Carpuat, M., Carreras, X., Vecchi, E.M., Eds.; Association for Computational Linguistics: Doha, Qatar, 2014; pp. 103–111. [Google Scholar] [CrossRef]
  17. Guo, S.; Lin, Y.; Feng, N.; Song, C.; Wan, H. Attention based spatial–temporal Graph Convolutional Networks for traffic flow forecasting. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; AAAI Press: Washington, DC, USA, 2019. AAAI’19/IAAI’19/EAAI’19. [Google Scholar] [CrossRef]
  18. Qin, Y.; Zhao, F.; Fang, Y.; Luo, H.; Wang, C. Memory attention enhanced graph convolution long short-term memory network for traffic forecasting. Int. J. Intell. Syst. 2022, 37, 6555–6576. [Google Scholar] [CrossRef]
  19. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting. In Proceedings of the International Conference on Learning Representations (ICLR ’18), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  20. Bai, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive graph convolutional recurrent network for traffic forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17804–17815. [Google Scholar]
  21. Ye, J.; Sun, L.; Du, B.; Fu, Y.; Xiong, H. Coupled layer-wise graph convolution for transportation demand prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 4617–4625. [Google Scholar]
  22. Cao, D.; Wang, Y.; Duan, J.; Zhang, C.; Zhu, X.; Huang, C.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; et al. Spectral temporal graph neural network for multivariate time-series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17766–17778. [Google Scholar]
  23. Zhang, H.; Wang, H.; Zhang, X.; Gong, L. Research on Traffic Flow Forecasting Based on Dynamic Spatial–temporal Transformer. Transp. Res. Rec. 2024, 2678, 301–313. [Google Scholar] [CrossRef]
  24. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Chang, X.; Zhang, C. Connecting the Dots: Multivariate Time Series Forecasting with Graph Neural Networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 6–10 July 2020; KDD ’20. pp. 753–763. [Google Scholar] [CrossRef]
  25. Shao, Z.; Zhang, Z.; Wei, W.; Wang, F.; Xu, Y.; Cao, X.; Jensen, C.S. Decoupled dynamic spatial–temporal graph neural network for traffic forecasting. Proc. VLDB Endow. 2022, 15, 2733–2746. [Google Scholar] [CrossRef]
  26. Peng, H.; Du, B.; Liu, M.; Liu, M.; Ji, S.; Wang, S.; Zhang, X.; He, L. Dynamic Graph Convolutional Network for Long-Term Traffic Flow Prediction with Reinforcement Learning. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; p. 9. [Google Scholar] [CrossRef]
  27. Dey, R.; Salem, F.M. Gate-variants of Gated Recurrent Unit (GRU) neural networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600. [Google Scholar]
  28. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated Recurrent Neural Networks on sequence modeling. In Proceedings of the NIPS 2014 Workshop on Deep Learning, Montreal, QU, Canada, 13 December 2014. [Google Scholar]
  29. Saravanan, S.; Venkatachalapathy, K. A deep hybrid model for traffic flow prediction using CNN-rGRU. Int. J. Anal. Exp. Modal Anal. 2019, 11, 1554–1560. [Google Scholar]
  30. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-GCN: A Temporal Graph Convolutional Network for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3848–3858. [Google Scholar] [CrossRef]
  31. Jiang, R.; Wang, Z.; Yong, J.; Jeph, P.; Chen, Q.; Kobayashi, Y.; Song, X.; Fukushima, S.; Suzumura, T. Spatio-temporal meta-graph learning for traffic forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, Canada, 7–14 February 2023; Volume 37, pp. 8078–8086. [Google Scholar]
  32. Zheng, C.; Fan, X.; Wang, C.; Qi, J. GMAN: A Graph Multi-Attention Network for Traffic Prediction. Proc. AAAI Conf. Artif. Intell. 2020, 34, 1234–1241. [Google Scholar] [CrossRef]
  33. Liu, H.; Dong, Z.; Jiang, R.; Deng, J.; Chen, Q.; Song, X. STAEformer: Spatio-Temporal Adaptive Embedding Makes Vanilla Transformers SOTA for Traffic Forecasting. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM), Birmingham, UK, 21–25 October 2023; pp. 21–25. [Google Scholar]
  34. van den Oord, A.; Li, Y.; Vinyals, O. Representation Learning with Contrastive Predictive Coding. arXiv 2018, arXiv:1807.03748. [Google Scholar]
  35. Zhang, Q.; Huang, C.; Xia, L.; Wang, Z.; Li, Z.; Yiu, S. Automated Spatio-Temporal Graph Contrastive Learning. In Proceedings of the ACM Web Conference 2023, Austin, TX, USA, 30 April–4 May 2023; pp. 295–305. [Google Scholar] [CrossRef]
  36. Park, H.; Noh, J.; Ham, B. Learning Memory-Guided Normality for Anomaly Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 14360–14369. [Google Scholar]
  37. Kipf, T.N.; Welling, M. Semi-supervised classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  38. Shao, Z.; Wang, F.; Xu, Y.; Wei, W.; Yu, C.; Zhang, Z.; Yao, D.; Sun, T.; Jin, G.; Cao, X.; et al. Exploring Progress in Multivariate Time Series Forecasting: Comprehensive Benchmarking and Heterogeneity Analysis. IEEE Trans. Knowl. Data Eng. 2024, 37, 291–305. [Google Scholar] [CrossRef]
  39. Song, C.; Lin, Y.; Guo, S.; Wan, H. Spatial–temporal synchronous Graph Convolutional Networks: A new framework for spatial–temporal network data forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 914–921. [Google Scholar]
Figure 1. A typical framework for spatio-temporal forecasting using temporal dynamic graphs. Each time step introduces unique spatial dependencies, enhancing the representation of dynamic states. The dynamic graph learning module significantly reduces the main network’s pressure to accurately capture spatio-temporal variations, yet demands higher capabilities from the graph learning module.
Figure 1. A typical framework for spatio-temporal forecasting using temporal dynamic graphs. Each time step introduces unique spatial dependencies, enhancing the representation of dynamic states. The dynamic graph learning module significantly reduces the main network’s pressure to accurately capture spatio-temporal variations, yet demands higher capabilities from the graph learning module.
Entropy 27 00099 g001
Figure 2. The framework of DG3L: For a spatio-temporal data input, the process begins with an Embedding module that concatenates time and adaptive embedding vectors. This is followed by a spatio-temporal transformer that outputs hidden vectors matching the output sequence length. On one side, it queries a Memory Graph Bank to generate graph sequences, and on the other, it feeds into the DG-GCRU module for representation blending. Finally, the DG-GCRU performs graph convolution operations within the recurrent network based on the generated graph sequences. The output sequence from DG-GCRU is then processed through a fully connected layer to produce the desired predictive output.
Figure 2. The framework of DG3L: For a spatio-temporal data input, the process begins with an Embedding module that concatenates time and adaptive embedding vectors. This is followed by a spatio-temporal transformer that outputs hidden vectors matching the output sequence length. On one side, it queries a Memory Graph Bank to generate graph sequences, and on the other, it feeds into the DG-GCRU module for representation blending. Finally, the DG-GCRU performs graph convolution operations within the recurrent network based on the generated graph sequences. The output sequence from DG-GCRU is then processed through a fully connected layer to produce the desired predictive output.
Entropy 27 00099 g002
Figure 3. The variation in Mean Absolute Error (MAE) across the PEMS03, PEMS08, and METR-LA datasets as influenced by the number of memory nodes and the dimension of hidden units.
Figure 3. The variation in Mean Absolute Error (MAE) across the PEMS03, PEMS08, and METR-LA datasets as influenced by the number of memory nodes and the dimension of hidden units.
Entropy 27 00099 g003
Figure 4. This scatter plot illustrates the comparison of Mean Absolute Error (MAE) across various models, indicated by different markers, in relation to the total number of parameters within each dataset.
Figure 4. This scatter plot illustrates the comparison of Mean Absolute Error (MAE) across various models, indicated by different markers, in relation to the total number of parameters within each dataset.
Entropy 27 00099 g004
Table 1. Summary of datasets.
Table 1. Summary of datasets.
Dataset#Spatial UnitsTime Interval#TimestepsStart TimeEnd Time
METR-LA2075 min34,2723/20126/2012
PEMS-BAY3255 min52,1161/20175/2017
PEMS033585 min26,2095/20127/2012
PEMS043075 min16,9921/20182/2018
PEMS078835 min28,2245/20178/2017
PEMS081705 min17,8567/20168/2016
Table 2. Comparison of various forecasting models using the metrics MAE, RMSE, and MAPE across multiple datasets and forecasting horizons, highlighting the top and second-best performances with bold/underlined and bold formatting, respectively.
Table 2. Comparison of various forecasting models using the metrics MAE, RMSE, and MAPE across multiple datasets and forecasting horizons, highlighting the top and second-best performances with bold/underlined and bold formatting, respectively.
DatasetsMethodsOverallHorizon 3Horizon 6Horizon 12
MAERMSEMAPEMAERMSEMAPEMAERMSEMAPEMAERMSEMAPE
PEMS03DCRNN15.47627.57915.80%14.23925.09615.10%15.48627.79915.77%17.49230.77817.61%
STGCN15.94727.32415.58%14.85425.52014.44%15.88427.37915.46%17.76329.90517.14%
GWNet14.52225.11115.36%13.37923.10714.64%14.52525.16415.30%16.27327.80517.05%
STNorm15.34125.90914.42%14.30423.93013.61%15.57326.36214.27%16.86028.40316.01%
MTGNN14.81225.07414.97%13.77723.62914.28%14.82025.16114.77%16.48527.75816.53%
STWave14.90426.12215.41%13.68324.08814.36%14.78625.92915.14%16.47428.53916.82%
DGCRN14.79426.52515.13%13.69524.47514.68%14.88826.81915.19%16.53929.26216.15%
MegaCRN14.68326.10115.80%13.57224.34615.16%14.82126.38315.88%16.46728.73317.37%
D2STGNN14.62025.89115.36%13.50123.84314.46%14.63526.14815.24%16.36428.50716.80%
STAEformer14.70626.04615.39%13.69223.54214.38%14.97226.13315.35%16.83129.28817.11%
DG3L14.51025.68914.38%13.20123.53313.48%14.52125.62314.39%16.32428.33116.06%
PEMS04DCRNN19.65231.17413.73%18.45729.40512.91%19.66331.19213.73%21.57133.79215.22%
STGCN19.72931.44513.66%18.78829.88113.27%19.70231.41913.78%21.36933.84914.39%
GWNet18.82130.15713.18%17.87928.77712.49%18.82830.28913.06%20.33732.21214.61%
STNorm19.19632.19512.96%18.42430.52112.56%19.30332.56413.08%20.48534.37813.63%
MTGNN19.19631.46313.41%18.32529.86412.78%19.29431.67913.55%20.60533.75814.25%
STWave18.22829.97512.15%17.46928.70311.67%18.17929.96312.06%19.60831.76312.93%
DGCRN18.96130.89912.94%17.96229.10312.33%19.00230.98412.98%20.59733.45314.00%
MegaCRN18.78030.27513.13%17.74328.67612.54%18.82530.33913.21%20.44632.65114.23%
D2STGNN18.32930.04312.52%17.56528.62612.13%18.36730.30112.64%19.60032.07813.44%
STAEformer18.20730.39112.39%17.47528.99811.95%18.23230.51212.40%19.32032.27913.14%
DG3L18.21629.81712.41%17.47328.83211.83%18.24130.23512.39%19.52032.59313.44%
PEMS07DCRNN21.43334.9109.02%19.61331.6018.29%21.41034.8988.97%24.43139.56210.32%
STGCN22.06635.6699.41%20.54932.8008.84%22.00135.5399.36%24.68839.94210.44%
GWNet20.35633.3468.67%18.85530.8148.08%20.37433.4098.61%22.77736.9639.77%
STNorm20.64434.9968.72%19.24031.7578.15%20.78335.2248.80%22.80538.9539.70%
MTGNN21.34334.3269.44%19.44431.2548.34%21.25834.2249.23%24.50938.76711.43%
STWave19.91933.8768.40%18.58530.7577.84%19.91733.1988.41%21.90536.3929.40%
DGCRN23.32836.49110.51%19.54831.2758.48%22.20234.8669.97%30.62645.67114.28%
MegaCRN22.28834.95410.45%20.27931.7899.65%22.24134.90310.51%25.45239.40211.88%
D2STGNN19.56632.6318.19%18.16430.1117.68%19.69132.6788.20%21.55536.2459.06%
STAEformer19.39432.7248.10%18.09130.2477.57%19.39832.7458.07%21.44636.1978.99%
DG3L19.59333.0688.17%18.07730.1617.59%19.57833.0218.13%21.79837.0129.09%
PEMS08DCRNN15.19924.19910.23%14.14022.1989.51%15.21724.26410.19%16.88626.92511.51%
STGCN16.17125.39210.47%15.10123.4709.88%16.06625.33810.44%18.02028.19911.40%
GWNet14.68423.6109.74%13.69621.7649.02%14.67523.5979.77%16.18126.10910.59%
STNorm15.41324.9129.84%14.45722.8089.11%15.48425.0499.96%16.91027.61311.07%
MTGNN15.23124.0629.88%14.25622.2779.10%15.20024.1119.71%16.83126.57711.25%
STWave13.89624.1759.14%13.02122.3168.63%13.80424.2469.09%15.02126.2379.94%
DGCRN14.88423.7759.92%13.69321.6958.96%14.84323.7919.81%16.81126.64511.45%
MegaCRN16.24425.26511.01%14.64122.7389.93%16.12725.14310.65%18.91029.01112.32%
D2STGNN14.15123.5839.11%13.20621.5398.50%14.16423.5679.10%15.49826.10310.05%
STAEformer13.43123.3138.97%12.54521.4298.41%13.43023.3158.92%14.78725.8289.74%
DG3L13.72023.0888.91%12.74221.2298.38%13.68623.0788.87%15.10725.6329.92%
METR-LADCRNN3.0396.2488.34%2.6765.1886.88%3.0766.2918.43%3.5607.49010.41%
STGCN3.0936.2688.35%2.7425.2687.08%3.1336.3218.48%3.5877.43410.09%
GWNet3.0316.1218.14%2.6895.1436.89%3.0726.1768.28%3.5107.2579.88%
STNorm3.1446.4758.77%2.8175.5237.51%3.2046.5909.00%3.5947.54010.42%
MTGNN3.0216.1608.18%2.6855.1756.88%3.0566.1948.28%3.4927.29410.00%
STWave3.1026.4658.79%2.7945.5097.42%3.1446.5278.86%3.5037.47110.44%
DGCRN3.0676.3338.08%2.6785.1736.75%3.1016.3718.19%3.6067.6369.90%
MegaCRN2.9626.0438.00%2.6114.9966.68%2.9986.0738.12%3.4617.2479.84%
D2STGNN2.8695.8957.83%2.5584.9536.54%2.9045.9387.92%3.3367.0329.71%
STAEformer2.9565.9997.93%2.6985.2037.00%2.9936.0728.19%3.3417.0229.68%
DG3L2.8995.9737.92%2.5544.9716.51%2.9275.9847.90%3.3907.2109.72%
PEMS-BAYDCRNN1.5923.7003.59%1.3122.7652.73%1.6523.7653.72%1.9704.6154.71%
STGCN1.6193.6913.67%1.3512.8292.87%1.6803.7773.81%1.9824.5484.70%
GWNet1.5983.7023.52%1.3062.7532.68%1.6563.7763.65%1.9924.6134.60%
STNorm1.5783.6533.49%1.3292.8262.76%1.6493.7823.64%1.9134.4424.45%
MTGNN1.5963.6653.51%1.3272.7922.77%1.6543.7473.65%1.9714.5414.49%
STWave1.5763.6093.53%1.3332.8402.73%1.6313.6993.65%1.9164.3614.48%
DGCRN1.5653.6193.54%1.3002.7392.71%1.6213.6943.67%1.9334.4894.60%
MegaCRN1.5583.6353.53%1.2922.7232.70%1.6163.7153.66%1.9164.5054.60%
D2STGNN1.5163.5333.43%1.2592.6602.63%1.5743.6263.57%1.8634.3484.42%
STAEformer1.5643.5833.55%1.3252.7942.82%1.6303.6663.72%1.8784.3064.41%
DG3L1.5393.5713.49%1.2902.7182.68%1.6013.6523.64%1.8714.3334.42%
Table 3. Ablation Study on PEMS03, METR-LA, and PEMS-BAY with best performances in bold.
Table 3. Ablation Study on PEMS03, METR-LA, and PEMS-BAY with best performances in bold.
DatasetsPEMS03METR-LAPEMS-BAY
MetricMAERMSEMAERMSEMAERMSE
w/o Dual-Gate14.64726.3292.9476.0941.5623.607
w/o Memory Graph-Bank15.49227.9363.1176.4661.5803.683
w/o SpatioTransformer-Encoder14.54326.1562.9136.0171.5413.579
Ours14.51025.6892.8995.9731.5393.571
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Zhang, Z.; Pi, S.; Zhang, H.; Pi, J. Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions. Entropy 2025, 27, 99. https://doi.org/10.3390/e27020099

AMA Style

Wang Y, Zhang Z, Pi S, Zhang H, Pi J. Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions. Entropy. 2025; 27(2):99. https://doi.org/10.3390/e27020099

Chicago/Turabian Style

Wang, Yuxuan, Zhouyuan Zhang, Shu Pi, Haishan Zhang, and Jiatian Pi. 2025. "Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions" Entropy 27, no. 2: 99. https://doi.org/10.3390/e27020099

APA Style

Wang, Y., Zhang, Z., Pi, S., Zhang, H., & Pi, J. (2025). Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions. Entropy, 27(2), 99. https://doi.org/10.3390/e27020099

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop